NVIDIA A30 TENSOR CORE GPU
NEWBring accelerated performance to every enterprise workload with NVIDIA® A30 Tensor Core GPUs. With NVIDIA Ampere architecture Tensor Cores and Multi-Instance GPU (MIG), it delivers speedups securely across diverse workloads, including AI inference at scale and high-performance computing (HPC) applications. By combining fast memory bandwidth and low-power consumption in a PCIe form factor—optimal for mainstream servers—A30 enables an elastic data centre and delivers maximum value for enterprises.
Get a Quote* With sparsity
** NVLink™ Bridge for up to two GPUs
FP64
5.2 teraFLOPS
FP64 Tensor Core
10.3 teraFLOPS
FP32
10.3 teraFLOPS
TF32 Tensor Core
82 teraFLOPS | 165 teraFLOPS*
BFLOAT16 Tensor Core
165 teraFLOPS | 330 teraFLOPS*
FP16 Tensor Core
165 teraFLOPS | 330 teraFLOPS*
INT8 Tensor Core
330 TOPS | 661 TOPS*
INT4 Tensor Core
661 TOPS | 1321 TOPS*
Media engines
- 1 optical flow accelerator (OFA)
- 1 JPEG decoder (NVJPEG)
- 4 video decoders (NVDEC)
GPU memory
24GB HBM2
GPU memory bandwidth
933GB/s
Interconnect
- PCIe Gen4: 64GB/s
- Third-gen NVLINK: 200GB/s**
Form factor
Dual-slot, full-height, full-length (FHFL)
Max thermal design power (TDP)
165W
Multi-Instance GPU (MIG)
- 4 GPU instances @ 6GB each
- 2 GPU instances @ 12GB each
- 1 GPU instance @ 24GB
Virtual GPU (vGPU) software support
- NVIDIA AI Enterprise for VMware
- NVIDIA Virtual Compute Server