Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 3 additions & 1 deletion modules/about-using-gpu-operator.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -8,4 +8,6 @@

You can use the NVIDIA GPU Operator with {VirtProductName} to rapidly provision worker nodes for running GPU-enabled virtual machines (VMs). The NVIDIA GPU Operator manages NVIDIA GPU resources in an {product-title} cluster and automates tasks that are required when preparing nodes for GPU workloads.

Before you can deploy application workloads to a GPU resource, you must install components such as the NVIDIA drivers that enable the compute unified device architecture (CUDA), Kubernetes device plugin, container runtime, and other features, such as automatic node labeling and monitoring. By automating these tasks, you can quickly scale the GPU capacity of your infrastructure. The NVIDIA GPU Operator can especially facilitate provisioning complex artificial intelligence and machine learning (AI/ML) workloads.
Before you can deploy application workloads to a GPU resource, you must install components such as the NVIDIA drivers that enable the compute unified device architecture (CUDA), Kubernetes device plugin, container runtime, and other features, such as automatic node labeling and monitoring. By automating these tasks, you can quickly scale the GPU capacity of your infrastructure. The NVIDIA GPU Operator can especially facilitate provisioning complex artificial intelligence and machine learning (AI/ML) workloads.

To use the NVIDIA GPU Operator to configure mediated devices, see https://docs.nvidia.com/datacenter/cloud-native/openshift/latest/openshift-virtualization.html[NVIDIA GPU Operator with OpenShift Virtualization].
27 changes: 27 additions & 0 deletions modules/virt-label-nodes-with-mig-backed-profile.adoc
Original file line number Diff line number Diff line change
@@ -0,0 +1,27 @@
// Module included in the following assemblies:
//
// * virt/virtual_machines/advanced_vm_management/virt-configuring-virtual-gpus.adoc

:_mod-docs-content-type: REFERENCE
[id="virt-label-nodes-with-mig-backed-profile_{context}"]
= Labeling nodes with a MIG-backed vGPU profile

[role="_abstract"]
For GPUs that support MIG, you have the option to select MIG-backed vGPU instances instead of time-sliced vGPU instances.

.Prerequisites

* You have configured vGPU support. For more information see https://docs.nvidia.com/datacenter/cloud-native/openshift/latest/mig-ocp.html#[MIG Support in OpenShift Container Platform].
* You are using the AIE vGPU Manager image.

.Procedure

* Label the node with the name of the MIG-backed vGPU profile:
+
----
oc label node <node> --overwrite nvidia.com/vgpu.config=<profile>
----
+
* Replace <node> with the name FQDN of your compute node
* Replace <profile> with the profile name for your

82 changes: 0 additions & 82 deletions modules/virt-options-configuring-mdevs.adoc

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ The NVIDIA GPU Operator is supported only by NVIDIA. For more information, see l

include::modules/about-using-gpu-operator.adoc[leveloffset=+2]

include::modules/virt-options-configuring-mdevs.adoc[leveloffset=+2]
include::modules/virt-label-nodes-with-mig-backed-profile.adoc[leveloffset=+2]

[role="_additional-resources"]
.Additional resources
Expand Down