Key Features

AI Inference and Mainstream Compute for Every Enterprise

NVIDIA A30 Tensor Core GPU is the most versatile mainstream compute GPU for AI inference and mainstream enterprise workloads. Powered by NVIDIA Ampere architecture Tensor Core technology, it supports a broad range of math precisions, providing a single accelerator to speed up every workload. Built for AI inference at scale, the same compute resource can rapidly re-train AI models with TF32, as well as accelerate high-performance computing (HPC) applications using FP64 Tensor Cores. Multi-Instance GPU (MIG) and FP64 Tensor Cores combine with fast 933 gigabytes per second (GB/s) of memory bandwidth in a low 165W power envelope, all running on a PCIe card optimal for mainstream servers. The combination of third-generation Tensor Cores and MIG delivers secure quality of service across diverse workloads, all powered by a versatile GPU enabling an elastic data center. A30’s versatile compute capabilities across big and small workloads deliver maximum value for mainstream enterprises. A30 is part of the complete NVIDIA data center solution that incorporates building blocks across hardware, networking, software, libraries, and optimized AI models and applications from NGC™. Representing the most powerful end-to-end AI and HPC platform for data centers, it allows researchers to deliver real-world results and deploy solutions into production at scale.

The End-to-End Solution for Enterprises

NVIDIA A30 Tensor Core GPU— powered by the NVIDIA Ampere architecture, the heart of the modern data center—is an integral part of the NVIDIA data center platform. Built for deep learning, HPC, and data analytics, the platform accelerates over 2,000 applications, including every major deep learning framework. Additionally, NVIDIA AI Enterprise, an end-to-end, cloud-native suite of AI and data analytics software, is certified to run on A30 in hypervisor-based virtual infrastructure with VMware vSphere. This enables management and scaling of AI workloads in a hybrid cloud environment. The complete NVIDIA platform is available everywhere, from data center to edge, delivering both dramatic performance gains and cost-saving opportunities.

Categories

Powered by NVIDIA Ampere Architecture

NVIDIA Ampere Architecture-Based CUDA Cores

Accelerate graphics workflows with the latest CUDA® cores for up to 2.5X single-precision floating-point (FP32) performance compared to the previous generation.

Second-Generation RT Cores

Produce more visually accurate renders faster with hardware-accelerated motion blur and up to 2X faster ray-tracing performance than the previous generation.

Third-Generation Tensor Cores

Boost AI and data science model training with up to 10X faster training performance compared to the previous generation with hardware-support for structural sparsity.

Virtualization-Ready

Repurpose your personal workstation into multiple high-performance virtual workstations with support for NVIDIA RTX Virtual Workstation (vWS) software.

Third-Generation NVIDIA NVLink

Scale memory and performance across multiple GPUs with NVIDIA® NVLink™ to tackle larger datasets, models, and scenes.

PCI Express Gen 4

Improve data-transfer speeds from CPU memory for data-intensive tasks with support for PCI Express Gen 4.

Power Efficiency

Leverage a dual-slot, power efficient design that’s 2.5X more power efficient than the previous generation and crafted to fit a wide range of workstations.

Peak FP64 5.2TF
Peak FP64 Tensor Core 10.3 TF
Peak FP32 10.3 TF
TF32 Tensor Core 82 TF | 165 TF*
BFLOAT16 Tensor Core 165 TF | 330 TF*
Peak FP16 Tensor Core 165 TF | 330 TF*
Peak INT8 Tensor Core 330 TOPS | 661 TOPS*
Peak INT4 Tensor Core 661 TOPS | 1321 TOPS*
Media engines
1 optical flow accelerator (OFA)
1 JPEG decoder (NVJPEG)
4 Video decoders (NVDEC)
GPU Memory 24GB HBM2
GPU Memory Bandwidth 933GB/s
Interconnect
PCIe Gen4: 64GB/s
Third-gen NVIDIA® NVLINK® 200GB/s**
Form Factor 2-slot, full height, full length (FHFL)
Max thermal design power (TDP) 165W
Multi-Instance GPU (MIG)
4 MIGs @ 6GB each
2 MIGs @ 12GB each
1 MIGs @ 24GB
Virtual GPU (vGPU) software support
NVIDIA AI Enterprise
NVIDIA Virtual Compute Server
* Structural sparsity enabled
** NVLink Bridge for up to two GPUs.

Speak with an expert to learn more.