# Line below added 29 Nov 2024

NVIDIA A100 PCIe 80GB GPU

Highest versatility for all workloads

  • Released June 28th, 2021
  • GA100 Graphic Processor
  • 6912 Cores
  • 300W TDP
  • 432 TMUS
  • 160 ROPS
  • 80GB Memory Size
  • HBM2e Memory Type
  • 5120 bit BUS Width

  • $USD $19,222.00

    *RRP Pricing*

    To View Channel Discounts Please Login


NVIDIA A100 PCIe 80GB GPU

Unprecedented Acceleration for World’s Highest-Performing Elastic Data Centers The NVIDIA A100 80GB Tensor Core GPU delivers unprecedented acceleration—at every scale—to power the world’s highest performing el

Unprecedented Acceleration for World’s Highest-Performing Elastic Data Centers

The NVIDIA A100 80GB Tensor Core GPU delivers unprecedented acceleration—at every scale—to power the world’s highest performing elastic data centers for AI, data analytics, and high-performance computing (HPC) applications. As the engine of the NVIDIA data center platform, A100 provides up to 20x higher performance over the prior NVIDIA Volta generation. A100 can efficiently scale up or be partitioned into seven isolated GPU instances, with Multi-Instance GPU (MIG) providing a unified platform that enables elastic data centers to dynamically adjust to shifting workload demands.

A100 is part of the complete NVIDIA data center solution that incorporates building blocks across hardware, networking, software, libraries, and optimized AI models and applications from NGC. Representing the most powerful end-to-end AI and HPC platform for data centers, it allows researchers to deliver real-world results and deploy solutions into production at scale, while allowing IT to optimize the utilization of every available A100 GPU.


The A100 PCIe 80 GB is a professional graphics card by NVIDIA, launched on June 28th, 2021. Built on the 7 nm process, and based on the GA100 graphics processor, the card does not support DirectX. Since A100 PCIe 80 GB does not support DirectX 11 or DirectX 12, it might not be able to run all the latest games. The GA100 graphics processor is a large chip with a die area of 826 mm² and 54,200 million transistors. It features 6912 shading units, 432 texture mapping units, and 160 ROPs. Also included are 432 tensor cores which help improve the speed of machine learning applications. NVIDIA has paired 80 GB HBM2e memory with the A100 PCIe 80 GB, which are connected using a 5120-bit memory interface. The GPU is operating at a frequency of 1065 MHz, which can be boosted up to 1410 MHz, memory is running at 1593 MHz.

Being a dual-slot card, the NVIDIA A100 PCIe 80 GB draws power from an 8-pin EPS power connector, with power draw rated at 250 W maximum. This device has no display connectivity, as it is not designed to have monitors connected to it. A100 PCIe 80 GB is connected to the rest of the system using a PCI-Express 4.0 x16 interface. The card measures 267 mm in length, and features a dual-slot cooling solution.

Start configuring your GP-GPU Server now!


S5N | D43N-3U


HSGP1


GZ2 | G492-ZD2



The core of AI and HPC in the modern data center

Scientists, researchers, and engineers-the da Vincis and Einsteins of our time-are working to solve the world’s most important scientific, industrial, and big data challenges with AI and high-performance computing (HPC). Meanwhile businesses and even entire industries seek to harness the power of AI to extract new insights from massive data sets, both on-premises and in the cloud. The NVIDIA Ampere architecture, designed for the age of elastic computing, delivers the next giant leap by providing unmatched acceleration at every scale, enabling these innovators to do their life’s work.

Start configuring your GP-GPU Server now!



S5N | D43N-3U


HSGP1


GZ2 | G492-ZD2


A100 80GB PCIeA100 80GB SXM
FP649.7 TFLOPS
FP64 Tensor Core19.5 TFLOPS
FP3219.5 TFLOPS
Tensor Float 32 (TF32)156 TFLOPS | 312 TFLOPS*
BFLOAT16 Tensor Core312 TFLOPS | 624 TFLOPS*
FP16 Tensor Core312 TFLOPS | 624 TFLOPS*
INT8 Tensor Core624 TOPS | 1248 TOPS*
GPU Memory80GB HBM2e80GB HBM2e
GPU Memory Bandwidth1,935 GB/s2,039 GB/s
Max Thermal Design Power (TDP)300W400W ***
Multi-Instance GPUUp to 7 MIGs @ 10GBUp to 7 MIGs @ 10GB
Form FactorPCIe
Dual-slot air-cooled or single-slot liquid-cooled
SXM
InterconnectNVIDIA® NVLink® Bridge
for 2 GPUs: 600 GB/s **
PCIe Gen4: 64 GB/s
NVLink: 600 GB/s
PCIe Gen4: 64 GB/s
Server OptionsPartner and NVIDIA-Certified Systems™ with 1-8 GPUsNVIDIA HGX™ A100-Partner and NVIDIA-Certified Systems with 4,8, or 16 GPUs NVIDIA DGX™ A100 with 8 GPUs




Datasheet 1
Title Version Date Size
A100 80GB for PCIe 1 01-07-2021 389KB

Tags: NVIDIA, GP, GPU, A100, 80GB, PCIe, Ampere, HPC, AI, Deep Learning, TESLA, 300W,