NVIDIA GB200 NVL72

Faster Scale.Smarter Returns.

Scale without limits, innovate without compromise. NVIDIA GB200 NVL72 is here to maximize your AI returns. Deploy today and lead the change.

Supercharging Next-GenerationAI and Accelerated Computing

Performance Left
LLM Inference
30x
vs. NVIDIA H100 Tensor Core GPU
BREAKTHROUGH TRAINING & INFERENCE EFFICIENCY

BREAKTHROUGH TRAINING & INFERENCE EFFICIENCY

Up to 30× faster inference and significantly shorter training cycles compared to previous generations.

LLM Training
4x
vs. H100
High-Bandwidth, Low-Latency Interconnect

High-Bandwidth, Low-Latency Interconnect

With 130 TB/s NVLink bandwidth across GPUs and CPUs, it unlocks large-scale distributed training and multi-node workloads.

Energy Efficiency
25x
vs. H100
Sustainable, Liquid-Cooled Design

Sustainable, Liquid-Cooled Design

Advanced liquid cooling reduces energy and operational costs while supporting dense deployments at scale.

data processing
18x
vs. CPU
Massive Memory Capacity

Massive Memory Capacity

Powered by HBM3e, with memory measured in terabytes and extreme bandwidth to handle trillion-parameter models and long-context inference.

The Journey to Next-Gen Performance

Experience the unseen story of NVIDIA GB200 NVL72, where every step reflects the innovation, precision, and scale driving the future of AI infrastructure.

Redefining AI Infrastructure.
Enter the Blackwell Era.

Scaling Enterprise AI Models

Scaling Enterprise AI Models

Shorten time-to-market for foundation models and domain-specific LLMs. NVIDIA GB200 NVL72 accelerates training and fine-tuning, helping businesses deploy competitive AI products faster.

Scaling Enterprise AI Models

Powering AI Agents at Scale

Support enterprise-wide AI copilots and assistants with real-time inference and long-context reasoning. This enables employees and customers to access intelligent support instantly and reliably.

Scaling Enterprise AI Models

Driving Multimodal Applications

Run advanced applications that combine text, images, video, and audio. NVIDIA GB200 NVL72 provides the performance and bandwidth required for smarter search, compliance automation, and richer customer experiences.

Scaling Enterprise AI Models

Accelerating Data-Intensive Decisions

Process massive datasets in near real time for forecasting, risk analysis, and scenario planning. Enterprises can make faster, more confident decisions powered by AI-driven insights.

Use cases

Finance and Banking

Finance and Banking

Accelerate risk modeling/fraud detection, and algorithmic trading with ultra-low latency and high throughput for faster, more accurate decisions.

Finance and Banking

Healthcare and Life Sciences

Harness teraflyte-scale memory and FP64 performance to speed drug discovery, genomic analysis, and medical imaging, reducing research cycles and improving outcomes.

Finance and Banking

Manufacturing and Industrial Engineering

Leverage liquid-cooled, energy-efficient clusters to power digital twins, predictive maintenance, and generative design that enhance productivity and cut downtime.

Finance and Banking

Retail and Customer Experience

Deploy large-scale AI copilots, recommendation engines, and personalization systems that deliver real-time insights and seamless customer engagement.

NVIDIA GB200 NVL72 Technical Specs

GB200 NVL72

NVIDIA GB200 NVL72

GB200 Grace Blackwell Superchip

NVIDIA GB200 NVL72 Grace Blackwell Superchip

Configuration
36 Grace CPU : 72 Blackwell GPUs
1 Grace CPU : 2 Blackwell GPU
FP4 Tensor Core1
1,440 PFLOPS
40 PFLOPS
FP8/FP6 Tensor Core1
720 PFLOPS
20 PFLOPS
INT8 Tensor Core1
720 POPS
20 POPS
FP16/BF16 Tensor Core1
360 PFLOPS
10 PFLOPS
TF32 Tensor Core
180 PFLOPS
5 PFLOPS
FP32
5,760 TFLOPS
160 TFLOPS
FP64
2,880 TFLOPS
80 TFLOPS
FP64 Tensor Core
2,880 TFLOPS
80 TFLOPS
GPU Memory | Bandwidth
Up to 13.4 TB HBM3e | 576 TB/s
Up to 372GB HBM3e | 16 TB/s
NVLink Bandwidth
130TB/s
3.6TB/s
CPU Core Count
2,592 Arm® Neoverse V2 cores
72 Arm Neoverse V2 cores
CPU Memory | Bandwidth
Up to 17 TB LPDDR5X | Up to 18.4 TB/s
Up to 480GB LPDDR5X | Up to 512 GB/s
1 - With sparsity

Frequently Asked Questions

NVIDIA GB200 NVL72 clusters are now live and available on Bitdeer AI Cloud. Customers can start deploying and scaling AI workloads immediately.
A deposit may be required to secure your NVIDIA GB200 NVL72 reservation. Please check with our sales team for details and flexible arrangements.
NVIDIA GB200 NVL72 surpasses H100 and H200, packing 72 Blackwell GPUs into one rack for unmatched performance, bandwidth, and efficiency. It offers up to 30x faster training on GPT-scale trillion-parameter models and more than 4x higher energy efficiency versus H100, making it the leading choice for large-scale AI, HPC, and multi-agent workloads.
NVIDIA GB200 NVL72 excels in trillion-parameter training, real-time inference, multi-agent AI, HPC simulations, and large-scale data analytics.
For existing Bitdeer AI users, workloads and data scale smoothly to NVIDIA GB200 NVL72 with minimal changes. New users receive comprehensive support to ensure a faster and simpler migration.
Power and cooling are managed with rack-level liquid cooling and optimized power delivery, ensuring high efficiency, dense deployment, and stable performance at scale.

Power Your AI with NVIDIA GB200 NVL72

Get unmatched performance and scale

Turn AI performance into real impact today

GB200 NVL72GB200 NVL72
Bitdeer AI with Nasdaq LogoNVIDIA Preferred Partner Logo
Bitdeer Technologies Group (Straitdeer Pte. Ltd.) is a preferred NVIDIA Cloud Partner

Certified and trusted by industry standards

(ISO/IEC 27001:2022 and SOC2 Type I & Type II)

Bitdeer AI with Nasdaq LogoNVIDIA Preferred Partner Logo
Bitdeer Technologies Group (Straitdeer Pte. Ltd.) is a preferred NVIDIA Cloud Partner

Certified and trusted by industry standards

(ISO/IEC 27001:2022 and SOC2 Type I & Type II)