Use GPUs with clustered VMs

Applies to: Azure Local, version 22H2

You can include graphics processing units (GPUs) in your clusters to provide GPU acceleration to workloads running in clustered VMs. GPU acceleration can be provided via Discrete Device Assignment (DDA), which allows you to dedicate one or more physical GPUs to a VM, or through GPU Partitioning. Clustered VMs can take advantage of GPU acceleration, and clustering capabilities such as high availability via failover. Live migration of virtual machines (VMs) isn't currently supported, but VMs can be automatically restarted and placed where GPU resources are available if there's a failure.

In this article, you will learn how to use GPUs with clustered VMs to provide GPU acceleration to workloads using Discrete Device Assignment. This article guides you through preparing the cluster, assigning a GPU to a cluster VM, and failing over that VM using Windows Admin Center and PowerShell.

For information about how to manage GPUs in Azure Local, version 23H2, see Prepare GPUs for Azure Local.

Prerequisites

There are several requirements and things to consider before you begin to use GPUs with clustered VMs:

  • You need an Azure Local instance running Azure Stack HCI operating system, version 22H2 or later..
  • You need a Windows Server Failover cluster running Windows Server 2025 or later.
  • You must install the same make and model of the GPUs across all the servers in your cluster.

  • Review and follow the instructions from your GPU manufacturer to install the necessary drivers and software on each server in the cluster.

  • Depending on your hardware vendor, you might also need to configure any GPU licensing requirements.

  • You need a machine with Windows Admin Center installed. This machine could be one of your cluster nodes.

Note

Your system must be supported Azure Local solution with GPU support. To browse options, visit the Azure Local Catalog.

Prepare the cluster

When the prerequisites are complete, you can prepare the cluster to use GPUs with clustered VMs.

Preparing the cluster involves creating a resource pool that contains the GPUs that are available for assignment to VMs. The cluster uses this pool to determine VM placement for any started or moved VMs that are assigned to the GPU resource pool.

Using Windows Admin Center, follow these steps to prepare the cluster to use GPUs with clustered VMs.

To prepare the cluster and assign a VM to a GPU resource pool:

  1. Launch Windows Admin Center and make sure the GPUs extension is already installed.

  2. Select Cluster Manager from the top dropdown menu and connect to your cluster.

  3. From the Settings menu, select Extensions > GPUs.

  4. On the Tools menu, under Extensions, select GPUs to open the tool.

    Screenshot of the GPU tool in Windows Admin Center.

  5. On tool's main page, select the GPU pools tab, and then select Create GPU pool.

    Screenshot of the Create GPU pools page in Windows Admin Center.

  6. On the New GPU pool page, specify the following and then select Save:

    1. Server name
    2. GPU pool name
    3. GPUs that you want to add to the pool

    Screenshot of the New GPU pool page in Windows Admin Center to specify servers, pool name, and GPUs.

    After the process completes, you'll receive a success prompt that shows the name of the new GPU pool and the host server.

Assign a VM to a GPU resource pool

You can now assign a VM to a GPU resource pool. You can assign one or more VMs to a clustered GPU resource pool, and remove a VM from a clustered GPU resource pool.

Follow these steps to assign an existing VM to a GPU resource pool using Windows Admin Center.

Note

You also need to install drivers from your GPU manufacturer inside the VM so that apps in the VM can take advantage of the GPU assigned to them.

  1. On the Assign VM to GPU pool page, specify the following, then select Assign:

    1. Server name
    2. GPU pool name
    3. Virtual machine that you want to assign the GPU to from the GPU pool.

    You can also define advanced setting values for memory-mapped IO (MMIO) spaces to determine resource requirements for a single GPU.

    Screenshot of the Assign VM to GPU pool page in Windows Admin Center where you assign a VM to a GPU from the GPU pool.

    After the process completes, you'll receive a confirmation prompt that shows you successfully assigned the GPU from the GPU resource pool to the VM, which displays under Assigned VMs.

    Screenshot of success prompt showing GPU assigned to a VM and the VM displaying under Assigned VMs.

To unassign a VM from a GPU resource pool:

  1. On the GPU pools tab, select the GPU that you want to unassign, and then select Unassign VM.

  2. On the Unassign VM from GPU pool page, in the Virtual machines list box, specify the name of the VM, and then select Unassign.

    Screenshot of Unassign VM from GPU pool page showing VM to be unassigned.

    After the process completes, you receive a success prompt that the VM has been unassigned from the GPU pool, and under Assignment status the GPU shows Available (Not assigned).

When you start the VM, the cluster ensures that the VM is placed on a server with available GPU resources from this cluster-wide pool. The cluster also assigns the GPU to the VM through DDA, which allows the GPU to be accessed from workloads inside the VM.

Fail over a VM with an assigned GPU

To test the cluster’s ability to keep your GPU workload available, perform a drain operation on the server where the VM is running with an assigned GPU. To drain the server, follow the instructions in Failover cluster maintenance procedures. The cluster restarts the VM on another server in the cluster, as long as another server has sufficient available GPU resources in the pool that you created.

To test the cluster’s ability to keep your GPU workload available, perform a drain operation on the server where the VM is running with an assigned GPU. To drain the server, follow the instructions in Failover cluster maintenance procedures. The cluster restarts the VM on another server in the cluster, as long as another server has sufficient available GPU resources in the pool that you created.

For more information on using GPUs with your clustered VMs, see:

For more information on using GPUs with your VMs and GPU partitioning, see: