Machines are high-performing computing for scaling AI applications.
Name | Generation | CUDA Cores | GPU Memory | Memory Bandwidth | Performance (TFLOPS) | Performance with Tensor Core (TFLOPS) | vCPUs | CPU RAM |
---|---|---|---|---|---|---|---|---|
GPU+ | Maxwell | 1,664 | 8 GB | 192 GB/s | 2.6 | - | 8 | 30 |
P4000 | Pascal | 1,792 | 8 GB | 243 GB/s | 5.3 | - | 8 | 30 |
P5000 | Pascal | 2,560 | 16 GB | 288 GB/s | 9.0 | - | 8 | 30 |
P6000 | Pascal | 3,840 | 24 GB | 432 GB/s | 12.0 | - | 8 | 30 |
V100 | Volta | 5,120 | 16 GB | 900 GB/s | 14.0 | 112 (FP16) | 8 | 30 |
V100x32 | Volta | 5,120 | 32 GB | 900 GB/s | 15.7 | 125 (FP16) | 8 | 30 |
RTX4000 | Turing | 2,304 | 8 GB | 416 GB/s | 7.1 | 57 (FP32) | 8 | 30 |
RTX5000 | Turing | 3,072 | 16 GB | 448 GB/s | 11.2 | 89 (FP32) | 8 | 30 |
A4000 | Ampere | 6,144 | 16 GB | 448 GB/s | 19.2 | 153 (FP16) | 8 | 45 |
A5000 | Ampere | 8,192 | 24 GB | 768 GB/s | 27.8 | 222 (FP16) | 8 | 45 |
A6000 | Ampere | 10,752 | 48 GB | 768 GB/s | 38.7 | 309 (FP16) | 8 | 45 |
A100 | Ampere | 6,912 | 40 GB HBM2 | 1,555 GB/s | 19.5 | 156 / 312 (FP32/16) | 12 | 90 |
A100-80G | Ampere | 6,912 | 80 GB HBM2 | 1,555 GB/s | 19.5 | 312 / 624 (FP32/16) | 12 | 90 |
H100 | Hopper | 16,896 | 80 GB HBM3 | 3,350 GB/s | 67.0 | 989 (TF32) 1979 (BFLOAT16/FP16) 3,958 (FP8/INT8) |
20 | 250 |