NVIDIA A100 Tensor Core

NVIDIA A100 Tensor Core

Unprecedented acceleration with NVIDIA's A100 Tensor Core GPU

NVIDIA A100 Tensor Core

The NVIDIA A100 Tensor Core GPU delivers unprecedented acceleration for AI, data analytics, and high-performance computing (HPC) at every scale. Powering the world's highest-performing elastic data centers, the A100 offers up to 20x higher performance compared to the previous generation.

Key Features

  • Massive memory capacity - 80GB of HBM2e memory allows you to handle larger datasets and complex models, crucial for tackling demanding AI, machine learning, deep learning, and HPC tasks.
  • World-class memory bandwidth - Boasting over 2TB/s, this exceptional bandwidth speed significantly reduces processing time, accelerating your path to solutions and innovation.
  • Exceptional 3rd Gen Tensor Cores - These specialised cores deliver faster training and inference for deep learning workloads, enabling you to achieve breakthrough results in AI applications.

Specifications

Architecture Ampere
FP64 9.7 TFLOPS
FP64 Tensor Core 19.5 TFLOPs
FP32 19.5 TFLOPs
Tensor Flot 32 (TF32) 156 TFLOPs | 312 TFLOPS*
BFLOAT16 Tensor Core 312 TFLOPs | 624 TFLOPS*
FP16 Tensor Core 312 TFLOPs | 624 TFLOPS*
INT8 Tensor Core 624 TFLOPS | 1248 TFLOPS*
GPU Memory 80GB HBM2e
GPU Memory Bandwidth 1,935GB/s
Max Thermal Design Power (TDP) 300W
Multi-Instance GPU Up to 7 MIGs @ 10GB
Form Factor PCIe
Interconnect NVIDIA NVLink Bridge for up to 2 GPUs: 600GB/s
PCIe Gen4: 64GB/s
Server Options Partner and NVIDIA-certified Systems with 1-8 GPUs
*with sparsity