NVIDIA A800
The A800 PCIe 40 GB is a professional graphics card by NVIDIA, launched on November 8th, 2022. Built on the 7 nm process, and based on the GA100 graphics processor, the card does not support DirectX. Since A800 PCIe 40 GB does not support DirectX 11 or DirectX 12, it might not be able to run all the latest games. The GA100 graphics processor is a large chip with a die area of 826 mm² and 54,200 million transistors. It features 6912 shading units, 432 texture mapping units, and 160 ROPs. Also included are 432 tensor cores which help improve the speed of machine learning applications. NVIDIA has paired 40 GB HBM2e memory with the A800 PCIe 40 GB, which are connected using a 5120-bit memory interface. The GPU is operating at a frequency of 765 MHz, which can be boosted up to 1410 MHz, memory is running at 1215 MHz.
Being a dual-slot card, the NVIDIA A800 PCIe 40 GB draws power from an 8-pin EPS power connector, with power draw rated at 250 W maximum. This device has no display connectivity, as it is not designed to have monitors connected to it. A800 PCIe 40 GB is connected to the rest of the system using a PCI-Express 4.0 x16 interface. The card measures 267 mm in length, 111 mm in width, and features a dual-slot cooling solution.

NVIDIA A800 Specification
A800 40GB PCle | A800 80GB PCle | A800 80GB SXM | |
FP64 | 9.7 TFLOPS | ||
FP64 Tensor Core | 19.5TFLOPS | ||
FP32 | 19.5TFLOPS | ||
Tensor Float 32 (TF32) | 156 TFLOPS | 312 TFLOPS* | ||
BFLOAT 16 Tensor Core | 312 TFLOPS | 624 TFLOPS* | ||
INT8 Tensor Core | 624TOPSI 1248 TOPS* | ||
GPU Memory | 40GB HBM2 | 80GB HBM2e | 80GB HBM2e |
GPU Memory Bandwidth | 1,555GB/s | 1,935GB/s | 2,039GB/s |
Max Thermal Design Power (TDP) | 250W | 300W | 400W*** |
Multi-Instance GPU | Upto7MIGs@ 5GB | Upto7MIGs@ 10GB | Upto7MIGs@ 10GB |
Form Factor | PCle (dual-slot air cooled or single-slot liquid cooled) | SXM | |
Interconnect | NVIDIA NVLink" Bridge for 2 GPUs: 400GB/s PCle Gen4: 64GB/s | NVLink: 400GB/s PCle Gen4: 64GB/s | |
Server Options | Partner and NVIDIA-Certified Systems" with 1-8 GPUs | NVIDIA HGX™ A800-Partner and NVIDIA-Certified Systems with 4 or 8 GPUs |
NVIDIA A100 vs A800 : 80GB PCIe Version
Specifications | A100 | A800 |
FP64 | 9.7 TFLOPS | |
FP64 Tensor Core | 19.5 TFLOPS | |
FP32 | 19.5 TFLOPS | |
Tensor Float 32 | 156 TFLOPS | |
BFLOAT 16 Tensor Core | 312 TFLOPS | |
FP16 Tensor Core | 312 TFLOPS | |
INT8 Tensor Core | 624 TOPS | |
GPU Memory | 80 GB HBM2 | |
GPU Memory Bandwifth | 1,935 GB/s | |
TDP | 300 W | |
Multi-Instance GPU | Up to 7 MIGs @ 10 GB | |
Interconnect | NVLink : 600 GB/s | NVLink : 400 GB/s |
Server Options | 1-8 GPUs |
NVIDIA A100 vs A800 : 80GB SXM Version
Specifications | A100 | A800 |
FP64 | 9.7 TFLOPS | |
FP64 Tensor Core | 19.5 TFLOPS | |
FP32 | 19.5 TFLOPS | |
Tensor Float 32 | 156 TFLOPS | |
BFLOAT 16 Tensor Core | 312 TFLOPS | |
FP16 Tensor Core | 312 TFLOPS | |
INT8 Tensor Core | 624 TOPS | |
GPU Memory | 80 GB HBM2 | |
GPU Memory Bandwifth | 2,039 GB/s | |
TDP | 400 W | |
Multi-Instance GPU | Up to 7 MIGs @ 10 GB | |
Interconnect | NVLink : 600 GB/s | NVLink : 400 GB/s |
Server Options | 4/ 8 / 16 GPUs | 4 / 8 GPUs |
NVIDIA A100 vs A800 : 40GB PCIe Version
Specifications | A100 | A800 |
FP64 | 9.7 TFLOPS | |
FP64 Tensor Core | 19.5 TFLOPS | |
FP32 | 19.5 TFLOPS | |
Tensor Float 32 | 156 TFLOPS | |
BFLOAT 16 Tensor Core | 312 TFLOPS | |
FP16 Tensor Core | 312 TFLOPS | |
INT8 Tensor Core | 624 TOPS | |
GPU Memory | 40 GB HBM2 | |
GPU Memory Bandwifth | 1,555 GB/s | |
TDP | 250 W | |
Multi-Instance GPU | Up to 7 MIGs @ 10 GB | |
Interconnect | NVLink : 600 GB/s | NVLink : 400 GB/s |
Server Options | 1-8 GPUs |