# GB200 NVL2 | NVIDIA

#  NVIDIA GB200 NVL2 

Bringing the new era of computing to every data center.

[ Read the Press Release ](https://nvidianews.nvidia.com/news/computer-industry-ai-factories-data-centers)

  * Introduction
  * Highlights
  * Features
  * GB200 NVL Platforms
  * Specs
  * Get Started






  * Introduction
  * Highlights
  * Features
  * GB200 NVL Platforms
  * Specs
  * Get Started



  * Introduction
  * Highlights
  * Features
  * GB200 NVL Platforms
  * Specs
  * Get Started



##  Unparalleled Single-Server Performance 

The NVIDIA GB200 NVL2 platform brings the new era of computing to every data center, delivering unparalleled performance for mainstream large language model (LLM) inference, vector database search, and data processing through 2 Blackwell GPUs and 2 Grace CPUs. With its scale-out, single node [NVIDIA MGX™ architecture](https://www.nvidia.com/en-us/data-center/products/mgx/), its design enables a wide variety of system designs and networking options to seamlessly integrate accelerated computing into existing data center infrastructure.

###  Computer Industry Joins NVIDIA to Build AI Factories and Data Centers for the Next Industrial Revolution 

At Computex 2024, the world’s top computer manufacturers joined NVIDIA to unveil the latest NVIDIA Blackwell-powered systems, including the GB200 NVL2, to lead the next industrial revolution.

[Press Release  ](https://nvidianews.nvidia.com/news/computer-industry-ai-factories-data-centers)

Highlights 

##  Turbocharging Accelerated Computing 

###  Llama 3 Inference 

5Xvs. NVIDIA H100 Tensor Core GPU

###  Vector Database Search 

9X vs. H100

###  Data Processing 

18X vs. CPU

Llama3 LLM inference: Token-to-token latency (TTL) = 50 milliseconds (ms) real time, first token latency (FTL) = 2s, input sequence length = 2.048, output sequence length = 128 output, 8x NVIDIA HGX™ H100 air-cooled vs. GB200 NVL2 air-cooled single node, per-GPU performance comparison  
Vector database search performance within RAG pipeline using memory shared by NVIDIA Grace CPU and Blackwell GPU. 1x x86, 1x H100 GPU, and 1x GPU from GB200 NVL2 node.  
Data processing: A database join and aggregation workload with Snappy/Deflate compression derived from TPC-H Q4 query. Custom query implementations for x86, H100 single GU, and single GPU from GB200 NVL2 node: GB200 vs. Intel Xeon 8480+  
Projected performance subject to change.

###  Real-Time Mainstream LLM Inference 

GB200 NVL2 introduces massive coherent memory up to 1.3 terabytes (TB) shared between two Grace CPUs and two Blackwell GPUs. This shared memory is coupled with fifth-generation NVIDIA® NVLink™ and high-speed, chip-to-chip (C2C) connections to deliver 5X faster real-time LLM inference performance for mainstream language models such as Llama 3 70B.

###  Vector Database Search 

GB200 NLV2 accelerates RAG vector search operation by up to 9X. The vector database of the Wikipedia dataset is over 200 gigabytes (GB), and access to the Grace CPU’s 960GB of memory and 900GB/s high-speed C2C link supercharges low-latency vector search.

###  Data Processing 

Databases play critical roles in handling, processing, and analyzing large volumes of data for enterprises. GB200 NVL2 takes advantage of high-bandwidth memory performance, [NVLink-C2C](https://www.nvidia.com/en-us/data-center/nvlink-c2c/), and dedicated decompression engines in the [NVIDIA Blackwell architecture](https://www.nvidia.com/en-us/data-center/technologies/blackwell-architecture/) to speed up key database queries by 18X compared to CPU.

Features 

##  Technological Breakthroughs 

###  Blackwell Architecture 

The NVIDIA Blackwell architecture delivers groundbreaking advancements in accelerated computing, powering a new era of computing with unparalleled performance, efficiency, and scale.

[ Learn More About Blackwell ](https://www.nvidia.com/en-us/data-center/technologies/blackwell-architecture/)

###  NVIDIA Grace CPU 

The NVIDIA Grace CPU is a breakthrough processor designed for modern data centers running AI, cloud, and high-performance computing (HPC) applications. It provides outstanding performance and memory bandwidth with 2X the energy efficiency of today’s leading server processors.

[ Learn More About the Grace CPU Superchip ](https://www.nvidia.com/en-us/data-center/grace-cpu-superchip/)

###  NVIDIA NVLink-C2C 

NVIDIA NVLink-C2C coherently interconnects each Grace CPU and Blackwell GPU at 900GB/s. The GB200 NVL2 uses both NVLink-C2C and the fifth-generation NVLink to deliver a 1.4 TB coherent memory model for accelerated AI.

[ Explore NVLink-C2C ](https://www.nvidia.com/en-us/data-center/nvlink-c2c/)

###  Key Value Caching 

Key Value (KV) Caching improves LLM response speeds by storing conversation context and history. GB200 NVL2 optimizes KV Caching through its fully coherent Grace GPU and Blackwell GPU memory connected by NVLink-C2C, 7X faster than PCIe, enabling LLMs to predict words faster than x86-based GPU implementations.

[ Learn More About KV Caching ](https://developer.nvidia.com/blog/nvidia-gh200-superchip-accelerates-inference-by-2x-in-multiturn-interactions-with-llama-models/)

###  Fifth-Generation NVIDIA NVLink 

Unlocking the full potential of exascale computing and trillion-parameter AI models requires swift, seamless communication between every GPU in a server cluster. Fifth-generation NVLink is a scale-up interconnect that unleashes accelerated performance for trillion- and multi-trillion-parameter AI models.

[ Learn About NVLink and NVLink Switch ](https://www.nvidia.com/en-us/data-center/nvlink/)

###  NVIDIA Networking 

The data center’s network plays a crucial role in driving AI advancements and performance, serving as the backbone for distributed AI model training and generative AI performance. [NVIDIA Quantum-X800 InfiniBand](https://nvdam.widen.net/s/hbp8zz7fvt/solution-overview-gtcspring24-quantum-x800-3175164), [NVIDIA Spectrum™-X800 Ethernet](https://nvdam.widen.net/s/xfmlcbklg5/ethernet-solution-overview-spectrum-x800-gtcspring24-3175614), and [NVIDIA BlueField®-3 DPUs](https://www.nvidia.com/en-us/networking/products/data-processing-unit/) enable efficient scalability across hundreds and thousands of Blackwell GPUs for optimal application performance.

[ Explore End-to-End Networking Solutions ](https://www.nvidia.com/en-us/networking/)

##  More GB200 Grace Blackwell Platforms 

###  NVIDIA GB200 Grace Blackwell NVL4 Superchip 

NVIDIA GB200 Grace Blackwell NVL4 Superchip unlocks the future of converged HPC and AI, delivering revolutionary performance through four NVIDIA NVLink™-connected Blackwell GPUs unified with two Grace CPUs over NVLink-C2C.

[Learn More About HPC Solutions  ](https://www.nvidia.com/en-us/high-performance-computing/)

###  NVIDIA GB200 NVL72 

The NVIDIA GB200 NVL72 connects 36 GB200 Superchips in a rack-scale design. The GB200 NVL72 is a liquid-cooled, rack-scale solution that boasts a 72-GPU NVLink domain that acts as a single, massive GPU.

[Learn More  ](https://www.nvidia.com/en-us/data-center/gb200-nvl72/)

Specifications¹ 

##  NVIDIA GB200 NVL2 

Configuration | 2x Grace CPUs, 2x Blackwell GPUs  
---|---  
FP4 Tensor Core² | 40 PFLOPS  
FP8/FP6 Tensor Core² | 20 PFLOPS  
INT8 Tensor Core² | 20 POPS  
FP16/BF16 Tensor Core² | 10 PFLOPS  
TF32 Tensor Core² | 5 PFLOPS  
FP32 | 180 TFLOPS  
FP64/FP64 Tensor Core | 90 TFLOPS  
GPU Memory | Bandwidth | Up to 384GB | 16TB/s  
CPU Core Count | 144 Arm® Neoverse V2 cores  
LPDDR5X Memory | Bandwidth | Up to 960GB | Up to 1,024GB/s  
Interconnect | NVLink: 1.8TB/s  
NVLink-C2C: 2x 900GB/s  
PCIe Gen6: 2x 256GB/s  
Server Options | Various NVIDIA GB200 NVL2 configuration options using NVIDIA MGX  
1 Preliminary specifications. May be subject to change.  
2 With sparsity.  
  
##  NVIDIA GB200 NVL72 

The NVIDIA GB200 NVL72 connects 36 GB200 Superchips in a rack-scale design. The GB200 NVL72 is a liquid-cooled, rack-scale solution that boasts a 72-GPU NVLink domain that acts as a single, massive GPU.

[ Learn More ](https://www.nvidia.com/en-us/data-center/gb200-nvl72/)

Get Started 

##  Stay Up to Date 

Sign up to hear when NVIDIA Blackwell becomes available.

Notify Me

Products

  * [Data Center GPUs](https://resources.nvidia.com/l/en-us-gpu)
  * [NVIDIA DGX Platform](https://www.nvidia.com/en-us/data-center/dgx-platform/)
  * [NVIDIA HGX Platform](https://www.nvidia.com/en-us/data-center/hgx/)
  * [Networking Products](https://www.nvidia.com/en-us/networking/products/)
  * [Virtual GPUs](https://www.nvidia.com/en-us/data-center/virtual-solutions/)



Technologies

  * [NVIDIA Blackwell Architecture](https://www.nvidia.com/en-us/data-center/technologies/blackwell-architecture/)
  * [NVIDIA Hopper Architecture](https://www.nvidia.com/en-us/data-center/technologies/hopper-architecture/)
  * [MGX](https://www.nvidia.com/en-us/data-center/products/mgx/)
  * [Confidential Computing](https://www.nvidia.com/en-us/data-center/solutions/confidential-computing/)
  * [Multi-Instance GPU](https://www.nvidia.com/en-us/technologies/multi-instance-gpu/)
  * [NVLink-C2C](https://www.nvidia.com/en-us/data-center/nvlink-c2c/)
  * [ NVLink/NVSwitch](https://www.nvidia.com/en-us/data-center/nvlink/)
  * [Tensor Cores](https://www.nvidia.com/en-us/data-center/tensor-cores/)



Resources

  * [Accelerated Apps Catalog](https://www.nvidia.com/en-us/accelerated-applications/)
  * [Blackwell Resources Center](https://resources.nvidia.com/l/en-us-blackwell-architecture)
  * [Data Center GPUs](https://www.nvidia.com/en-us/data-center/data-center-gpus/)
  * [Data Center GPU Line Card](https://docs.nvidia.com/data-center-gpu/line-card.pdf)
  * [Data Center GPUs Resource Center](https://www.nvidia.com/en-us/data-center/tesla-product-literature/)
  * [Data Center Product Performance](https://developer.nvidia.com/deep-learning-performance-training-inference)
  * [Deep Learning Institute](https://www.nvidia.com/en-us/training/)
  * [Energy Efficiency Calculator](https://www.nvidia.com/en-us/data-center/sustainable-computing/energy-efficiency-calculator)
  * [GPU Cloud Computing](https://www.nvidia.com/en-us/data-center/gpu-cloud-computing/)
  * [MLPerf Benchmarks](https://www.nvidia.com/en-us/data-center/resources/mlperf-benchmarks/)
  * [NGC Catalog](https://www.nvidia.com/en-us/gpu-cloud/)
  * [NVIDIA-Certified Systems](https://www.nvidia.com/en-us/data-center/products/certified-systems/)
  * [NVIDIA Data Center Corporate Blogs](https://blogs.nvidia.com/blog/category/enterprise/)
  * [NVIDIA Data Center Technical Blogs](https://developer.nvidia.com/blog/category/data-center-cloud/)
  * [Qualified System Catalog](https://www.nvidia.com/en-us/data-center/data-center-gpus/tesla-qualified-servers-catalog/)
  * [Where to Buy](https://www.nvidia.com/en-us/data-center/where-to-buy/)



Company Info

  * [About Us](https://www.nvidia.com/en-us/about-nvidia/)
  * [Company Overview](https://www.nvidia.com/en-us/about-nvidia/ai-computing/)
  * [Investors](https://investor.nvidia.com/home/default.aspx)
  * [Venture Capital (NVentures)](https://www.nventures.ai/)
  * [NVIDIA Foundation](https://www.nvidia.com/en-us/foundation/)
  * [Research](https://www.nvidia.com/en-us/research/)
  * [Social Responsibility](https://www.nvidia.com/en-us/sustainability/)
  * [Technologies](https://www.nvidia.com/en-us/technologies/)
  * [Careers](https://www.nvidia.com/en-us/about-nvidia/careers/)



Follow Data Center

[ __](http://www.facebook.com/NVIDIA "<util:I18n key="Follow GeForce on Facebook" />") [ __](https://www.linkedin.com/company/nvidia) [ __](http://twitter.com/nvidiadc "<util:I18n key="Follow GeForce on Twitter" />") [ __](http://www.youtube.com/user/nvidia)

NVIDIA

[ United States  ](https://www.nvidia.com/object/country-selector.html)

  * [Privacy Policy](https://www.nvidia.com/en-us/about-nvidia/privacy-policy/)
  * [Manage My Privacy](https://www.nvidia.com/en-us/about-nvidia/privacy-center/)
  * [Do Not Sell or Share My Data](https://www.nvidia.com/en-us/preferences/email-preferences/)
  * [Terms of Service](https://www.nvidia.com/en-us/about-nvidia/terms-of-service/)
  * [Accessibility](https://www.nvidia.com/en-us/about-nvidia/accessibility/)
  * [Corporate Policies](https://www.nvidia.com/en-us/about-nvidia/company-policies/)
  * [Product Security](https://www.nvidia.com/en-us/product-security/)
  * [Contact](https://www.nvidia.com/en-us/contact/)



Copyright © 2025 NVIDIA Corporation
