The ND-series virtual machines are a new addition to the GPU family designed for AI, and Deep Learning workloads. They offer excellent performance for training and inference. ND instances are powered by NVIDIA Tesla P40 GPUs and Intel Xeon E5-2690 v4 (Broadwell) CPUs. These instances provide excellent performance for single-precision floating point operations, for AI workloads utilizing Microsoft Cognitive Toolkit, TensorFlow, Caffe, and other frameworks. The ND-series also offers a much larger GPU memory size (24 GB), enabling to fit much larger neural net models. Like the NC-series, the ND-series offers a configuration with a secondary low-latency, high-throughput network through RDMA, and InfiniBand connectivity so you can run large-scale training jobs spanning many GPUs.
Premium Storage: Supported
Premium Storage caching: Supported
Ultra Disks: Supported (Learn more about availability, usage and performance)
Live Migration: Not Supported
Memory Preserving Updates: Not Supported
VM Generation Support: Generation 1 and 2
Accelerated Networking: Not Supported
Ephemeral OS Disks: Supported
Nvidia NVLink Interconnect: Not Supported
|Size||vCPU||Memory: GiB||Temp storage (SSD) GiB||GPU||GPU memory: GiB||Max data disks||Max uncached disk throughput: IOPS/MBps||Max NICs|
1 GPU = one P40 card.
Size table definitions
Storage capacity is shown in units of GiB or 1024^3 bytes. When you compare disks measured in GB (1000^3 bytes) to disks measured in GiB (1024^3) remember that capacity numbers given in GiB may appear smaller. For example, 1023 GiB = 1098.4 GB.
Disk throughput is measured in input/output operations per second (IOPS) and MBps where MBps = 10^6 bytes/sec.
Data disks can operate in cached or uncached modes. For cached data disk operation, the host cache mode is set to ReadOnly or ReadWrite. For uncached data disk operation, the host cache mode is set to None.
To learn how to get the best storage performance for your VMs, see Virtual machine and disk performance.
Expected network bandwidth is the maximum aggregated bandwidth allocated per VM type across all NICs, for all destinations. For more information, see Virtual machine network bandwidth.
Upper limits aren't guaranteed. Limits offer guidance for selecting the right VM type for the intended application. Actual network performance will depend on several factors including network congestion, application loads, and network settings. For information on optimizing network throughput, see Optimize network throughput for Azure virtual machines. To achieve the expected network performance on Linux or Windows, you may need to select a specific version or optimize your VM. For more information, see Bandwidth/Throughput testing (NTTTCP).
Supported operating systems and drivers
To take advantage of the GPU capabilities of Azure N-series VMs, NVIDIA GPU drivers must be installed.
The NVIDIA GPU Driver Extension installs appropriate NVIDIA CUDA or GRID drivers on an N-series VM. Install or manage the extension using the Azure portal or tools such as Azure PowerShell or Azure Resource Manager templates. See the NVIDIA GPU Driver Extension documentation for supported operating systems and deployment steps. For general information about VM extensions, see Azure virtual machine extensions and features.
If you choose to install NVIDIA GPU drivers manually, see N-series GPU driver setup for Windows or N-series GPU driver setup for Linux for supported operating systems, drivers, installation, and verification steps.
- General purpose
- Memory optimized
- Storage optimized
- GPU optimized
- High performance compute
- Previous generations
Learn more about how Azure compute units (ACU) can help you compare compute performance across Azure SKUs.