Powerful Server Platform For AI & HPC | NVIDIA HGX A100
Supercharging AI and high-performance computing for every data center.
Explore Systems- Overview
- Inference
- Training
- Networking
- Specifications
- Overview
- Inference
- Training
- Networking
- Specifications
- Overview
- Inference
- Training
- Networking
- Specifications
Overview
Supercharging AI and HPC for Every Data Center
The NVIDIA HGX™ platform brings together the full power of NVIDIA GPUs, NVIDIA NVLink™, NVIDIA networking, and fully optimized AI and high-performance computing (HPC) software stacks to provide the highest application performance and drive the fastest time to insights for every data center.
The NVIDIA HGX Rubin NVL8 integrates eight NVIDIA Rubin GPUs with sixth-generation high-speed NVLink interconnects, delivering 5.5x more NVFP4 FLOPS than HGX B200 to propel the data center into a new era of accelerated computing and generative AI.
NVIDIA Kicks Off the Next Generation of AI With Rubin — Six New Chips, One Incredible AI Supercomputer
The next-generation platform scales mainstream adoption, slashing cost per token with five breakthroughs for reasoning and agentic AI models.
Read the Press ReleaseAI Reasoning Performance and Versatility
AI, complex simulations, and massive datasets require multiple GPUs with extremely fast interconnections and a fully accelerated software stack. The NVIDIA HGX™ platform brings together the full power of NVIDIA GPUs, NVIDIA NVLink™, NVIDIA networking, and fully optimized AI and high-performance computing (HPC) software stacks to provide the highest application performance and drive the fastest time to insights for every data center.
Unmatched End-to-End Accelerated Computing Platform
The NVIDIA HGX B300 integrates NVIDIA Blackwell Ultra GPUs with high-speed interconnects to propel the data center into a new era of accelerated computing and generative AI. As a premier accelerated scale-up platform with up to 11x more inference performance than the previous generation, NVIDIA Blackwell-based HGX systems are designed for the most demanding generative AI, data analytics, and HPC workloads.
NVIDIA HGX includes advanced networking options—at speeds up to 800 gigabits per second (Gb/s)—using NVIDIA Quantum-X800 InfiniBand and Spectrum™-X Ethernet for the highest AI performance. HGX also includes NVIDIA BlueField®-3 data processing units (DPUs) to enable cloud networking, composable storage, zero-trust security, and GPU compute elasticity in hyperscale AI clouds.
AI Reasoning Performance and Versatility
Projected Performance subject to change. Token-to-token Latency (TTL) = 12ms, first token latency (FTL) = 2s, input sequence length =32,768, output sequence length = 8192, eight-way HGX B200 vs eight-way HGX Rubin NVL8, per GPU performance, served using disaggregated inference.
Real-Time Agentic AI Inference
HGX Rubin NVL8 delivers up to 3.5× higher inference performance compared to the previous NVIDIA Blackwell generation, powering advanced reasoning models like DeepSeek-R1. With Rubin’s Transformer Engine coupled with sixth-generation NVLink, Rubin accelerates NVFP4 inference to meet the demands of reasoning-based Agentic AI workloads
Scalable Training for Large AI Models
Next-Level Training Performance
The HGX B300 platform delivers up to 2.6x higher training performance for large language models such as DeepSeek-R1. With over 2 TB of high-speed memory and 14.4 TB/s of NVLink Switch bandwidth, it enables massive-scale model training and high-throughput inter-GPU communication.
Projected performance subject to change. Perf per GPU, FP8, 16K BS, 16K sequence length.
Accelerating HGX With NVIDIA Networking
AI factories and supercomputing centers span thousands of GPUs as a single distributed computing engine. To keep accelerators fully utilized, AI and scientific workloads demand deterministic latency, lossless throughput, stable iteration times, and the ability to scale not only within a data center but also across multiple sites.
NVIDIA networking delivers the full-stack fabric that makes this possible, combining NVIDIA NVLink scale-up, NVIDIA Quantum InfiniBand and Spectrum-X™ Ethernet scale-out, Spectrum-XGS Ethernet multi-data-center scale-across, NVIDIA® BlueField® DPU and DOCA™ for infrastructure services, and next-generation silicon-photonics platforms—enabling the world’s most demanding AI data centers.
NVIDIA HGX Specifications
NVIDIA HGX is available in a single baseboard with eight NVIDIA Rubin, NVIDIA Blackwell, or NVIDIA Blackwell Ultra SXMs. These powerful combinations of hardware and software lay the foundation for unprecedented AI supercomputing performance.
- NVIDIA Rubin NVL8
- NVIDIA Blackwell
| HGX Rubin NVL8* | |
|---|---|
| Form Factor | 8x NVIDIA Rubin SXM |
| NVFP4 Inference | 400 PFLOPS |
| NVFP4 Training | 280 PFLOPS |
| FP8/FP6 Training | 140 PF |
| INT8 Tensor Core<sup>1</sup> | 2 PFLOPS |
| FP16/BF16 Tensor Core<sup>1</sup> | 32 PFLOPS |
| TF32 Tensor Core<sup>1</sup> | 16 PFLOPS |
| FP32 | 1040 TFLOPS |
| FP64/FP64 Tensor Core | 264 TFLOPS |
| FP32 SGEMM | FP64 DGEMMCore<sup>2</sup> | 3200 TF | 1600 TF |
| Total Memory | 2.3 TB |
| NVIDIA NVLink | Sixth generation |
| NVIDIA NVLink Switch | NVLink 6 Switch |
| NVLink GPU-to-GPU Bandwidth | 3.6 TB/s |
| Total NVLink Switch Bandwidth | 28.8 TB/s |
| Networking Bandwidth | 1.6 TB/s |
* Preliminary specification, subject to change 1. Specification in Dense. 2. Peak performance using tensor core-based emulation algorithms.
| HGX Rubin NVL8 | |
|---|---|
| Form Factor | 8x NVIDIA Rubin SXM |
| FP4 Tensor Core<sup>1</sup> | 400 PFLOPS | 144 PFLOPS |
| FP8/FP6 Tensor Core<sup>2</sup> | 272 PFLOPS |
| INT8 Tensor Core<sup>2</sup> | 4 PFLOPS |
| FP16/BF16 Tensor Core<sup>2</sup> | 64 PFLOPS |
| TF32 Tensor Core<sup>2</sup> | 32 PFLOPS |
| FP32 | 1040 TFLOPS |
| FP64/FP64 Tensor Core | 264 TFLOPS |
| Total Memory | 2.3 TB |
| NVIDIA NVLink | Sixth generation |
| NVIDIA NVLink Switch™ | NVLink 6 Switch |
| NVLink GPU-to-GPU Bandwidth | 3.6 TB/s |
| Total NVLink Bandwidth | 28.8 TB/s |
| Networking Bandwidth | 1.6 TB/s |
| Attention Performance<sup>3</sup> | <awaiting results>2x |
| HGX B300 | HGX B200 | |
|---|---|---|
| Form Factor | 8x NVIDIA Blackwell Ultra SXM | 8x NVIDIA Blackwell SXM |
| FP4 Tensor Core<sup>1</sup> | 144 PFLOPS | 108 PFLOPS | 144 PFLOPS | 72 PFLOPS |
| FP8/FP6 Tensor Core<sup>2</sup> | 72 PFLOPS | 72 PFLOPS |
| INT8 Tensor Core<sup>2</sup> | 3 POPS | 72 POPS |
| FP16/BF16 Tensor Core<sup>2</sup> | 36 PFLOPS | 36 PFLOPS |
| TF32 Tensor Core<sup>2</sup> | 18 PFLOPS | 18 PFLOPS |
| FP32 | 600 TFLOPS | 600 TFLOPS |
| FP64/FP64 Tensor Core | 10 TFLOPS | 296 TFLOPS |
| Total Memory | 2.1 TB | 1.4 TB |
| NVIDIA NVLink | Fifth generation | Fifth generation |
| NVIDIA NVLink Switch™ | NVLink 5 Switch | NVLink 5 Switch |
| NVLink GPU-to-GPU Bandwidth | 1.8 TB/s | 1.8 TB/s |
| Total NVLink Bandwidth | 14.4 TB/s | 14.4 TB/s |
| Networking Bandwidth | 1.6 TB/s | 0.8 TB/s |
| Attention Performance<sup>3</sup> | 2x | 1x |
1. Specification in Sparse | Dense 2. Specification in Sparse. Dense is ½ sparse spec shown. 3. vs. NVIDIA Blackwell.
| HGX B300 | HGX B200 | |
|---|---|---|
| Form Factor | 8x NVIDIA Blackwell Ultra SXM | 8x NVIDIA Blackwell SXM |
| FP4 Tensor Core<sup>1</sup> | 144 PFLOPS | 108 PFLOPS | 144 PFLOPS | 72 PFLOPS |
| FP8/FP6 Tensor Core<sup>2</sup> | 72 PFLOPS | 72 PFLOPS |
| INT8 Tensor Core<sup>2</sup> | 3 POPS | 72 POPS |
| FP16/BF16 Tensor Core<sup>2</sup> | 36 PFLOPS | 36 PFLOPS |
| TF32 Tensor Core<sup>2</sup> | 18 PFLOPS | 18 PFLOPS |
| FP32 | 600 TFLOPS | 600 TFLOPS |
| FP64/FP64 Tensor Core | 10 TFLOPS | 296 TFLOPS |
| Total Memory | 2.1 TB | 1.4 TB |
| NVIDIA NVLink | Fifth generation | Fifth generation |
| NVIDIA NVLink Switch™ | NVLink 5 Switch | NVLink 5 Switch |
| NVLink GPU-to-GPU Bandwidth | 1.8 TB/s | 1.8 TB/s |
| Total NVLink Bandwidth | 14.4 TB/s | 14.4 TB/s |
| Networking Bandwidth | 1.6 TB/s | 0.8 TB/s |
| Attention Performance<sup>3</sup> | 2x | 1x |
Learn more about the NVIDIA Rubin Platform.
Learn More Products- Data Center GPUs
- NVIDIA DGX Platform
- NVIDIA HGX Platform
- Networking Products
- Virtual GPUs
- NVIDIA Blackwell Architecture
- NVIDIA Hopper Architecture
- MGX
- Confidential Computing
- Multi-Instance GPU
- NVLink-C2C
- NVLink/NVSwitch
- Tensor Cores
- Accelerated Apps Catalog
- Blackwell Resources Center
- Data Center GPUs
- Data Center GPU Line Card
- Data Center GPUs Resource Center
- Data Center Product Performance
- Deep Learning Institute
- Energy Efficiency Calculator
- GPU Cloud Computing
- MLPerf Benchmarks
- NGC Catalog
- NVIDIA-Certified Systems
- NVIDIA Data Center Corporate Blogs
- NVIDIA Data Center Technical Blogs
- Qualified System Catalog
- Where to Buy
- About Us
- Company Overview
- Investors
- Venture Capital (NVentures)
- NVIDIA Foundation
- Research
- Social Responsibility
- Technologies
- Careers
- Privacy Policy
- Your Privacy Choices
- Terms of Service
- Accessibility
- Corporate Policies
- Product Security
- Contact
Từ khóa » Hgx
-
Hg Exchange: SEA's First Member-Driven Private Exchange
-
HG Exchange - Fundnel
-
Hg Exchange - LinkedIn
-
Private Exchange HGX Begins Trading, Companies & Markets
-
Binance Acquires 18% Stake In Singapore-regulated Hg Exchange
-
Binance Asia Services Takes Strategic Stake In Singapore-based Hg ...
-
MAS Spins Digital Exchange HGX Out Of Its Sandbox - DigFin
-
Overview For HGX - Nasdaq Global Indexes
-
HGX PTE. LTD. - Financial Institutions Directory
-
HGX Pte Ltd - Company Profile And News - Bloomberg Markets
-
HGX PTE. LTD. - MyCareersFuture Singapore
-
Private Exchange HGX Expands Leadership Team And Products
-
HGX Pte. Ltd. - SG FinTech Directory