NVIDIA TESLA A100-80 GB Video Card Passive Ampere Graphics Card 80GB
NVIDIA TESLA A100
NVIDIA TESLA A100: Unprecedented Acceleration at Every Scale
The NVIDIA TESLA A100 Tensor Core GPU delivers unprecedented acceleration capabilities for AI, data analysis, and high performance computing (HPC) workflows to address the world's most complex computing challenges. As the engine of the NVIDIA data center platform, the A100 can help you interconnect thousands of GPUs or, with multi-instance GPU (MIG) technology, be partitioned into seven GPU instances to accelerate all types of workloads. The A100's third-generation NVIDIA Tensor cores now accelerate more levels of preThe third-generation NVIDIA Tensor cores in the A100 now accelerate more levels of precision for different workloads, reducing data access time and time to market.
ACCELERATE AI WORKFLOWS
- NVIDIA A100 Memory: 80 GB HBM2 ECC 5120 bits (Bandwidth: 1935 GB/s)
- CUDA cores: 6912
- FP64: 9.7 TFlops
- FP32: 19.5 TFlops
- TF32: 312 Tflops
- Tensor Float 32 (TF32): 156 TFlops
- BFLOAT16 Tensor Core: 312 TFlops
- FP16 Tensor Core: 1248 TOPs
- INT8 Tensor Core: 624 TOPs
- Up to 7 MIG instances at 10 GB
- Passive cooling
Groundbreaking Innovative Technology
NVIDIA AMPERE ARCHITECTUREA100 accelerates workloads big and small. Whether using MIG to partition an A100 GPU into smaller instances, or NVLink to connect multiple GPUs to accelerate large-scale workloads, A100 can readily handle different-sized acceleration needs, from the smallest job to the biggest multi-node workload. A100’s versatility means IT managers can maximize the utility of every GPU in their data center around the clock.
THIRD-GENERATION TENSOR CORESA100 delivers 312 teraFLOPS (TFLOPS) of deep learning performance. That’s 20X Tensor FLOPS for deep learning training and 20X Tensor TOPS for deep learning inference compared to NVIDIA Volta™ GPUs.
NEXT-GENERATION NVLINKNVIDIA NVLink in A100 delivers 2X higher throughput compared to the previous generation. When combined with NVIDIA NVSwitch™, up to 16 A100 GPUs can be interconnected at up to 600 gigabytes per second (GB/sec) to unleash the highest application performance possible on a single server. NVLink is available in A100 SXM GPUs via HGX A100 server boards and in PCIe GPUs via an NVLink Bridge for up to 2 GPUs.
MULTI-INSTANCE GPU (MIG)An A100 GPU can be partitioned into as many as seven GPU instances, fully isolated at the hardware level with their own high-bandwidth memory, cache, and compute cores. MIG gives developers access to breakthrough acceleration for all their applications, and IT administrators can offer right-sized GPU acceleration for every job, optimizing utilization and expanding access to every user and application.
HBM2With 40 gigabytes (GB) of high-bandwidth memory (HBM2), A100 delivers improved raw bandwidth of 1.6TB/sec, as well as higher dynamic random-access memory (DRAM) utilization efficiency at 95 percent. A100 delivers 1.7X higher memory bandwidth over the previous generation.
STRUCTURAL SPARSITYAI networks are big, having millions to billions of parameters. Not all of these parameters are needed for accurate predictions, and some can be converted to zeros to make the models “sparse” without compromising accuracy. Tensor Cores in A100 can provide up to 2X higher performance for sparse models. While the sparsity feature more readily benefits AI inference, it can also improve the performance of model training.
How Fast is the A100?
UP TO 7X HIGHER PERFORMANCE WITH MULTI INSTANCE GPU FOR AI INFERENCEThe NVIDIA A100 Tensor Core GPU is the flagship product of the NVIDIA data center platform for deep learning, HPC, and data analytics. The platform accelerates over 700 HPC applications and every major deep learning framework. It’s available everywhere, from desktops to servers to cloud services, delivering both dramatic performance gains and cost-saving opportunities.NVIDIA TESLA A100
NVIDIA TESLA A100-80 GB Video Card Passive Ampere Graphics Card 80GB Price List