NVIDIA A100 SXM 80GB GPU for HGX

Ultimate performance for all workloads

  • Released Nov 16th, 2020
  • NVLink network architecture
  • 6912 Cores
  • 300W TDP
  • 432 TMUS
  • 160 ROPS
  • 80GB Memory Size
  • HBM2e Memory Type
  • 5120 bit BUS Width

  • $USD $17,222.00

    *RRP Pricing*

    To View Channel Discounts Please Login


NVIDIA A100 SXM 80GB GPU for HGX

NVIDIA HGX AI Supercomputing Platform The A100 80GB GPU is a key element in NVIDIA HGX AI supercomputing platform, which brings together the full power of NVIDIA GPUs, NVIDIA NVLink, NVIDIA InfiniBand networking

NVIDIA HGX AI Supercomputing Platform

The A100 80GB GPU is a key element in NVIDIA HGX AI supercomputing platform, which brings together the full power of NVIDIA GPUs, NVIDIA NVLink, NVIDIA InfiniBand networking and a fully optimized NVIDIA AI and HPC software stack to provide the highest application performance. It enables researchers and scientists to combine HPC, data analytics and deep learning computing methods to advance scientific progress.


The A100 SXM4 80 GB is a professional graphics card by NVIDIA, launched on November 16th, 2020. Built on the 7 nm process, and based on the GA100 graphics processor, the card does not support DirectX. Since A100 SXM4 80 GB does not support DirectX 11 or DirectX 12, it might not be able to run all the latest games. The GA100 graphics processor is a large chip with a die area of 826 mm² and 54,200 million transistors. It features 6912 shading units, 432 texture mapping units, and 160 ROPs. Also included are 432 tensor cores which help improve the speed of machine learning applications. NVIDIA has paired 80 GB HBM2e memory with the A100 SXM4 80 GB, which are connected using a 5120-bit memory interface. The GPU is operating at a frequency of 1275 MHz, which can be boosted up to 1410 MHz, memory is running at 1593 MHz.
Its power draw is rated at 400 W maximum. This device has no display connectivity, as it is not designed to have monitors connected to it. A100 SXM4 80 GB is connected to the rest of the system using a PCI-Express 4.0 x16 interface.

Start configuring your GP-GPU Server now!


S5N | D43N-3U


HSGP1


GZ2 | G492-ZD2



The core of AI and HPC in the modern data center

Scientists, researchers, and engineers-the da Vincis and Einsteins of our time-are working to solve the world’s most important scientific, industrial, and big data challenges with AI and high-performance computing (HPC). Meanwhile businesses and even entire industries seek to harness the power of AI to extract new insights from massive data sets, both on-premises and in the cloud. The NVIDIA Ampere architecture, designed for the age of elastic computing, delivers the next giant leap by providing unmatched acceleration at every scale, enabling these innovators to do their life’s work.

Start configuring your GP-GPU Server now!



S5N | D43N-3U


HSGP1


GZ2 | G492-ZD2


A100 80GB PCIe A100 80GB SXM
FP64 9.7 TFLOPS
FP64 Tensor Core 19.5 TFLOPS
FP32 19.5 TFLOPS
Tensor Float 32 (TF32) 156 TFLOPS | 312 TFLOPS*
BFLOAT16 Tensor Core 312 TFLOPS | 624 TFLOPS*
FP16 Tensor Core 312 TFLOPS | 624 TFLOPS*
INT8 Tensor Core 624 TOPS | 1248 TOPS*
GPU Memory 80GB HBM2e 80GB HBM2e
GPU Memory Bandwidth 1,935 GB/s 2,039 GB/s
Max Thermal Design Power (TDP) 300W 400W ***
Multi-Instance GPU Up to 7 MIGs @ 10GB Up to 7 MIGs @ 10GB
Form Factor PCIe
Dual-slot air-cooled or single-slot liquid-cooled
SXM
Interconnect NVIDIA® NVLink® Bridge
for 2 GPUs: 600 GB/s **
PCIe Gen4: 64 GB/s
NVLink: 600 GB/s
PCIe Gen4: 64 GB/s
Server Options Partner and NVIDIA-Certified Systems™ with 1-8 GPUs NVIDIA HGX™ A100-Partner and NVIDIA-Certified Systems with 4,8, or 16 GPUs NVIDIA DGX™ A100 with 8 GPUs




Datasheet 1
Title Version Date Size
NVIADIA A100 DS 1 01-07-2021 480KB

Tags: NVIDIA, GP, GPU, A100, 80GB, SXM, Ampere, HPC, AI, Deep Learning, TESLA, 400W, HGX