When you’re ready to scale, choose IPU-POD128 for production deployment in your enterprise datacenter, private or public cloud. Experience massive efficiency and productivity gains when large language training runs are completed in hours or minutes instead of months and weeks. IPU-POD128 delivers for AI at scale.
- Superior scaling & blazing fast performance
- Full systems integration support for datacenter installation.
- AI expert support to develop & deploy models at scale.
128x GC200 IPUs
8.3TB (includes 115.2GB In-Processor Memory and 8.2TB Streaming Memory)
- 32 petaFLOPS FP16.16
- 8 petaFLOPS FP32
100 GE RoCEv2
- TensorFlow, PyTorch, PyTorch Lightning, Keras, Paddle Paddle, Hugging Face, ONNX, HALO
- OpenBMC, Redfish DTMF, IPMI over LAN, Prometheus, and Grafana
- Slurm, Kubernetes
- OpenStack, VMware ESG
900kg + Host servers and switches
32U + Host servers and switches
Selection of approved host servers from Graphcore partners