1
0
mirror of https://github.com/openshift/openshift-docs.git synced 2026-02-05 21:46:22 +01:00
Files
openshift-docs/modules/nvidia-gpu-sharing-methods.adoc
aireilly 8936b26f9e E810 hardware plugin doc updates
WPC T-GM + GNSS updates

General PTP docs reorg and clean up

Updates based on Aneesh's review comments

removing gnss-state-change from api/ocloudNotifications/v1/<resource_address>/CurrentState

Adding metric details

Adding new PTP image

Adding final PTP 4.14 image

Aneesh's comments

jack's comments

Adjust TOC

Peer review comments
2023-12-13 11:22:03 +00:00

28 lines
1.5 KiB
Plaintext

// Module included in the following assemblies:
//
// * architecture/nvidia-gpu-architecture-overview.adoc
:_mod-docs-content-type: CONCEPT
[id="nvidia-gpu-sharing-methods_{context}"]
= GPU sharing methods
Red{nbsp}Hat and NVIDIA have developed GPU concurrency and sharing mechanisms to simplify GPU-accelerated computing on an enterprise-level {product-title} cluster.
Applications typically have different compute requirements that can leave GPUs underutilized. Providing the right amount of compute resources for each workload is critical to reduce deployment cost and maximize GPU utilization.
Concurrency mechanisms for improving GPU utilization exist that range from programming model APIs to system software and hardware partitioning, including virtualization. The following list shows the GPU concurrency mechanisms:
* Compute Unified Device Architecture (CUDA) streams
* Time-slicing
* CUDA Multi-Process Service (MPS)
* Multi-instance GPU (MIG)
* Virtualization with vGPU
Consider the following GPU sharing suggestions when using the GPU concurrency mechanisms for different {product-title} scenarios:
Bare metal:: vGPU is not available. Consider using MIG-enabled cards.
VMs:: vGPU is the best choice.
Older NVIDIA cards with no MIG on bare metal:: Consider using time-slicing.
VMs with multiple GPUs and you want passthrough and vGPU:: Consider using separate VMs.
Bare metal with {VirtProductName} and multiple GPUs:: Consider using pass-through for hosted VMs and time-slicing for containers.