NVIDIA A100 Tensor Core GPU

NVIDIA A100 Tensor Core GPU

The NVIDIA® A100 Tensor Core GPU delivers unprecedented acceleration at every scale for AI, data analytics, and HPC to tackle the world’s toughest computing challenges. As the engine of the NVIDIA® data center platform, A100 can efficiently scale up to thousands of GPUs or, using new Multi-Instance GPU (MIG) technology, can be partitioned into seven isolated GPU instances to accelerate workloads of all sizes. A100’s third-generation Tensor Core technology now accelerates more levels of precision for diverse workloads, speeding time to insight as well as time to market.

A100PCIe 3QTR FrontRight 2000X1300

Specifications

Architecture Ampere
Process Size 7nm
Transistors TSMC
Die Size 54 Billion
CUDA Cores 826 mm2
Streaming Multiprocessors 6912
Tensor Cores 108
Multi-Instance GPU (MIG) Support Gen 3
FP64 432
FP64 Tensor Core Yes, up to seven instances per GPU
FP32 9.7 TFLOPS
TF32 Tensor Core 19.5 TFLOPS
BFLOAT16 Tensor Core 19.5 TFLOPS
FP16 Tensor Core 156 TFLOPS
INT8 Tensor Core 312 TFLOPS*
INT4 Tensor Core 312 TFLOPS
NVLink 624 TFLOPS*
NVLink Interconnect 312 TFLOPS
GPU Memory 624 TFLOPS*
Memory Interface 624 TOPS
Memory Bandwidth 1248 TOPS*
System Interface 1248 TOPS
Thermal Solution 2496 TOPS*
vGPU Support 2-Way Low Profile, 2-Slot
Secure and Measured Boot Hardware Root of Trust 600 GB/s Bidirectional
NEBS Ready 80GB HBM2e ECC on by Default
Power Connector 5120-bit
Maximum Power Consumption 1555 GB/s
PCIe 4.0 x16
Passive
NVIDIA® Virtual Compute Server with MIG support
CEC 1712
Level 3
8-pin CPU
300 W