NVIDIA H100 Tensor Core GPU
NEWAn Order-of-Magnitude Leap for Accelerated Computing
Tap into unprecedented performance, scalability, and security for every workload with the NVIDIA® H100 Tensor Core GPU. With NVIDIA NVLink® Switch System, up to 256 H100s can be connected to accelerate exascale workloads, along with a dedicated Transformer Engine to solve trillion-parameter language models. H100’s combined technology innovations can speed up large language models by an incredible 30X over the previous generation to deliver industry-leading conversational AI.
FP64
24 teraFLOPS
FP64 Tensor Core
48 teraFLOPS
FP32
48 teraFLOPS
TF32 Tensor Core
800 teraFLOPS* | 400 teraFLOPS
BFLOAT16 Tensor Core
1,600 teraFLOPS* | 800 teraFLOPS
FP16 Tensor Core
1,600 teraFLOPS* | 800 teraFLOPS
FP8 Tensor Core
3,200 teraFLOPS* | 1,600 teraFLOPS
INT8 Tensor Core
3,200 TOPS* | 1,600 TOPS
GPU memory
80GB
GPU memory bandwidth
2TB/s
Decoders
- 7 NVDEC
- 7 JPEG
Max thermal design power (TDP)
350W
Multi-Instance GPUs
Up to 7 MIGS @ 10GB each
Form factor
PCIe
Interconnect
- NVLINK: 600GB/s
- PCIe Gen5: 128GB/s
Server options
Partner and NVIDIA-Certified Systems with 1–8 GPUs