SPECIFICATIONS
A100 40GB PCIe | A100 80GB PCIe | A100 40GB SXM | A100 80GB SXM | |
---|---|---|---|---|
FP64 | 9.7 TFLOPS | |||
FP64 Tensor Core | 19.5 TFLOPS | |||
FP32 | 19.5 TFLOPS | |||
Tensor Float 32 (TF32) | 156 TFLOPS | 312 TFLOPS* | |||
BFLOAT16 Tensor Core | 312 TFLOPS | 624 TFLOPS* | |||
FP16 Tensor Core | 312 TFLOPS | 624 TFLOPS* | |||
INT8 Tensor Core | 624 TOPS | 1248 TOPS* | |||
GPU Memory | 40GB HBM2 | 80GB HBM2e | 40GB HBM2 | 80GB HBM2e |
GPU Memory Bandwidth | 1,555GB/s | 1,935GB/s | 1,555GB/s | 2,039GB/s |
Max Thermal Design Power (TDP) | 250W | 300W | 400W | 400W |
Multi-Instance GPU | Up to 7 MIGs @ 5GB | Up to 7 MIGs @ 10GB | Up to 7 MIGs @ 5GB | Up to 7 MIGs @ 10GB |
Form Factor | PCIe | SXM | ||
Interconnect | NVIDIA® NVLink® Bridge for 2 GPUs: 600GB/s ** PCIe Gen4: 64GB/s | NVLink: 600GB/s PCIe Gen4: 64GB/s | ||
Server Options | Partner and NVIDIA-Certified Systems™ with 1-8 GPUs | NVIDIA HGX™ A100-Partner and NVIDIA-Certified Systems with 4,8, or 16 GPUs NVIDIA DGX™ A100 with 8 GPUs |
Customer reviews
Reviews
There are no reviews yet.
Write a customer review