Subscriptions

NVIDIA GPU Appliances

NVIDIA DGX A100

The world's fastest single node fully intergrated AI system

[DGX-A100]

More Image(s)

BSI and Nvidia are delighted to announce the launch of the World’s fastest single node fully integrated AI system, the DGX A100. The GPU is named AMPERE, after André-Marie Ampère, the French mathematician and physicist, considered to be the father of electrodynamics.
Availability: In stock

For details on how to purchase the NVIDIA DGX A100, please click the button below to send us your details and brief requirements. We can then quote you accordingly.

Details

Crafted with 54 billion transistors, NVIDIA Ampere is the largest 7 nanometer (nm) chip ever built and features six key ground-breaking innovations.

AMPERE - 3rd Generation Tensor Cores

First introduced in the NVIDIA Volta architecture, NVIDIA Tensor Core technology has brought dramatic speedups to AI, bringing down training times from weeks to hours and providing massive acceleration to inference.

The NVIDIA Ampere architecture builds upon these innovations by bringing new precisions -Tensor Float (TF32) and Floating Point 64 (FP64) - to accelerate and simplify AI adoption and extend the power of Tensor Cores to HPC.

TF32 works just like FP32 while delivering speedups of up to 20X for AI without requiring any code change. Using NVIDIA Automatic Mixed Precision, researchers can gain an additional 2X performance with automatic mixed precision and FP16 adding just a couple of lines of code. And with support for bfloat16, INT8, and INT4, Tensor Cores in NVIDIA A100 Tensor Core GPUs create an incredibly versatile accelerator for both AI training and inference. Bringing the power of Tensor Cores to HPC, A100 also enables matrix operations in full, IEEE-certified, FP64 precision.

The Universal System for AI Infrastructure

Deep neural networks are rapidly growing in size and complexity, in response to the most pressing challenges in business and research.

The computational capacity needed to support today’s modern AI workloads has outpaced traditional data center architectures. Modern techniques that exploit use of model parallelism are colliding with the limits of inter-GPU bandwidth, as developers build increasingly large accelerated computing clusters and push the limits of data center scale. A new approach is needed - one that delivers almost limitless AI computing scale in order to break through the barriers to achieving faster insights.

NVIDIA NVLink

GPU to GPU communication is achieved via Nvidia’s 3rd generation NVlink architecture using 8x internal NVswitches.

This doubles the GPU-to-GPU direct bandwidth to 600 gigabytes per second (GB/s), almost 10X higher than PCIe Gen 4, and a new NVIDIA NVSwitch that is 2X faster than the last generation. This unprecedented power delivers the fastest time-to-solution, allowing users to tackle challenges that weren't possible or practical before.

Data Center Scalability with Mellanox

Networking is via Nvidia’s recent Mellanox acquisition, and features 9x Mellanox ConnectX-6 200Gb/s network interfaces on a PCIe gen4 bus.

With the fastest I/O architecture of any DGX system, NVIDIA DGX A100 is the foundational building block for large AI clusters like NVIDIA DGX SuperPOD, the enterprise blueprint for scalable AI infrastructure. DGXA100 features eight single-port Mellanox ConnectX-6 VPI HDR InfiniBand adapters for clustering and 1 dualport ConnectX-6 VPI Ethernet adapter for storage and networking, all capable of 200Gb/s.

The combination of massive GPU-accelerated compute with state-of-the-art networking hardware and software optimizations means DGX A100 can scale to hundreds or thousands of nodes to meet the biggest challenges, such as conversational AI and large scale image classification.

Proven Infrastructure Solutions Built with Trusted Data Center Leaders

As an Nvidia Elite Partner, we offer a portfolio of infrastructure solutions that incorporates the Ampere architecture and the best of the NVIDIA DGX POD reference architecture.

Delivered as fully integrated, ready-to-deploy, these solutions make data center AI deployments simpler and faster for IT.

Part No. DGX-A100
Manufacturer nvidia
End of Life? No
Preceeded By DGX-2
Rack Units 6
System Weight 134 lbs
System Dimensions 866 D x 444 W x 131 H (mm)
Packing Dimensions 1,180 D x 730 W x 284 H (mm)
Operating Temperature Range 10–35 °C
EAN DGXA100
Specification AMPERE-DGXA100
NVIDIA CUDA Cores 81920
NVIDIA Tensor Cores 10240
NVIDIA NVLink 12
Performance 2 petaFLOPS
Compatible CPU(s) AMD 8168, 2.7 GHz, 24-cores
No. of GPUs 8x NVIDIA® Ampere A100
System Memory 1.5TB
GPU Memory 512 GB total system
Storage Capacity OS: 2X 960GB NVME SSDs Internal Storage: 30TB (8X 3.84TB) NVME SSDs
Supported OS Ubuntu Linux Host OS
Ports Dual 10 GbE
4 IB EDR
DGX-A100

    Please login to x.com to view...