HBv3-series virtual machine overview

Applies to: ✔️ Linux VMs ✔️ Windows VMs ✔️ Flexible scale sets ✔️ Uniform scale sets

An HBv3-series server features 2 * 64-core EPYC 7V13 CPUs for a total of 128 physical "Zen3" cores. Simultaneous Multithreading (SMT) is disabled on HBv3. These 128 cores are divided into 16 sections (8 per socket), each section containing 8 processor cores with uniform access to a 32 MB L3 cache. Azure HBv3 servers also run the following AMD BIOS settings:

Nodes per Socket (NPS) = 2
L3 as NUMA = Disabled
NUMA domains within VM OS = 4
C-states = Enabled

As a result, the server boots with 4 NUMA domains (2 per socket) each 32-cores in size. Each NUMA has direct access to 4 channels of physical DRAM operating at 3200 MT/s.

To provide room for the Azure hypervisor to operate without interfering with the VM, we reserve 8 physical cores per server.

VM topology

The following diagram shows the topology of the server. We reserve these 8 hypervisor host cores (yellow) symmetrically across both CPU sockets, taking the first 2 cores from specific Core Complex Dies (CCDs) on each NUMA domain, with the remaining cores for the HBv3-series VM (green).

Topology of the HBv3-series server

Note that the CCD boundary is not equivalent to a NUMA boundary. On HBv3, a group of four consecutive (4) CCDs is configured as a NUMA domain, both at the host sever level and within a guest VM. Thus, all HBv3 VM sizes expose 4 NUMA domains that will appear to a OS and application as shown below, 4 uniform NUMA domains, each with different number of cores depending on the specific HBv3 VM size.

Topology of the HBv3-series VM

Each HBv3 VM size is similar in physical layout, features, and performance of a different CPU from the AMD EPYC 7003-series, as follows:

HBv3-series VM size NUMA domains Cores per NUMA domain Similarity with AMD EPYC
Standard_HB120rs_v3 4 30 Dual-socket EPYC 7713
Standard_HB120r-96s_v3 4 24 Dual-socket EPYC 7643
Standard_HB120r-64s_v3 4 16 Dual-socket EPYC 7543
Standard_HB120r-32s_v3 4 8 Dual-socket EPYC 7313
Standard_HB120r-16s_v3 4 4 Dual-socket EPYC 72F3


The constrained cores VM sizes only reduce the number of physical cores exposed to the VM. All global shared assets (RAM, memory bandwidth, L3 cache, GMI and xGMI connectivity, InfiniBand, Azure Ethernet network, local SSD) stay constant. This allows a customer to pick a VM size best tailored to a given set of workload or software licensing needs.

The virtual NUMA mapping of each HBv3 VM size is mapped to the underlying physical NUMA topology. There is no potentially misleading abstraction of the hardware topology.

The exact topology for the various HBv3 VM size appears as follows using the output of lstopo:

lstopo-no-graphics --no-io --no-legend --of txt

Click to view lstopo output for Standard_HB120rs_v3

lstopo output for HBv3-120 VM

Click to view lstopo output for Standard_HB120rs-96_v3

lstopo output for HBv3-96 VM

Click to view lstopo output for Standard_HB120rs-64_v3

lstopo output for HBv3-64 VM

Click to view lstopo output for Standard_HB120rs-32_v3

lstopo output for HBv3-32 VM

Click to view lstopo output for Standard_HB120rs-16_v3

lstopo output for HBv3-16 VM

InfiniBand networking

HBv3 VMs also feature Nvidia Mellanox HDR InfiniBand network adapters (ConnectX-6) operating at up to 200 Gigabits/sec. The NIC is passed through to the VM via SRIOV, enabling network traffic to bypass the hypervisor. As a result, customers load standard Mellanox OFED drivers on HBv3 VMs as they would a bare metal environment.

HBv3 VMs support Adaptive Routing, the Dynamic Connected Transport (DCT, in additional to standard RC and UD transports), and hardware-based offload of MPI collectives to the onboard processor of the ConnectX-6 adapter. These features enhance application performance, scalability, and consistency, and usage of them is strongly recommended.

Temporary storage

HBv3 VMs feature 3 physically local SSD devices. One device is preformatted to serve as a page file and will appear within your VM as a generic "SSD" device.

Two other, larger SSDs are provided as unformatted block NVMe devices via NVMeDirect. As the block NVMe device bypasses the hypervisor, it will have higher bandwidth, higher IOPS, and lower latency per IOP.

When paired in a striped array, the NVMe SSD provides up to 7 GB/s reads and 3 GB/s writes, and up to 186,000 IOPS (reads) and 201,000 IOPS (writes) for deep queue depths.

Hardware specifications

Hardware specifications HBv3-series VMs
Cores 120, 96, 64, 32, or 16 (SMT disabled)
CPU Frequency (non-AVX) 3.1 GHz (all cores), 3.675 GHz (up to 10 cores)
Memory 448 GB (RAM per core depends on VM size)
Local Disk 2 * 960 GB NVMe (block), 480 GB SSD (page file)
Infiniband 200 Gb/s Mellanox ConnectX-6 HDR InfiniBand
Network 50 Gb/s Ethernet (40 Gb/s usable) Azure second Gen SmartNIC

Software specifications

Software specifications HBv3-series VMs
Max MPI Job Size 36,000 cores (300 VMs in a single virtual machine scale set with singlePlacementGroup=true)
Additional Frameworks UCX, libfabric, PGAS
Azure Storage Support Standard and Premium Disks (maximum 32 disks)
OS Support for SRIOV RDMA CentOS/RHEL 7.6+, Ubuntu 18.04+, SLES 12 SP4+, WinServer 2016+
Recommended OS for Performance CentOS 8.1, Windows Server 2019+
Orchestrator Support Azure CycleCloud, Azure Batch, AKS; cluster configuration options


Windows Server 2012 R2 is not supported on HBv3 and other VMs with more than 64 (virtual or physical) cores. See Supported Windows guest operating systems for Hyper-V on Windows Server for more details.

Next steps