NVidia Blackwell Fleet
Blackwells are here.
NVIDIA B200, GB200, B300 and GB300s are now available for your most demanding workloads. Reserve yours today.
High-performing teams run on Blackwell clusters - operated by Voltage Park.
The B200
Faster training, efficient scalability
The second-generation Transformer Engine features FP8 and new precisions for 3X faster training on GPT MoE 1.8T.
The GB200
Real-time inference for LLM
Achieve up to 30X faster real-time trillion-parameter LLM inference compared to the NVIDIA H100 TensorCore GPU
Compare the entire Blackwell fleet
Instance
GPU
GPU Memory
VCPUS
Storage
Network Bandwidth
NVIDIA H100
GPU
HGX H100
GPU Memory
640 GB HBM3 total
VCPUs
2x Intel Xeon 8480C PCIe Gen5 CPUs
Storage
OS: 2× 1.92 TB NVMe M.2 SSDs (RAID 1); Data: 8× 3.84 TB NVMe U.2 SSDs (RAID 0)
Network Bandwidth
3.2 TB/s
NVIDIA B300
GPU
HGX B300
GPU Memory
Up to 2.3 TB total
VCPUs
2× Intel Xeon CPUs (exact model unspecified)
Storage
OS: 2× 1.9 TB NVMe M.2 SSDs; Internal: 8× 3.84 TB NVMe E1.S SSDs
Network Bandwidth
1.6 TB/s
NVIDIA GB300*
GPU
HGX GB300
GPU Memory
20.1 TB HBM3e total
VCPUs
72x NVIDIA Blackwell Ultra GPUs (2592 CPU cores)
Storage
Not Available
Network Bandwidth
1PB/s
*Liquid-cooled system
B200
For any stage of your develop-to-deploy pipeline.
An engine built to boost
Second-generation Blackwell Transformer Engine’s custom Tensor Core technology meets NVIDIA’s software stack. Lightning-fast inference and training for LLMs and MoE models.
Hardware-based security
The first TEE-I/O capable GPU in the industry with capable hosts and inline protection over NVLink.
NVLink and NVLink switch
Fifth-generation NVLink interconnect scales up to 576 GPUs. The NVLink Switch chip delivers 130 TB/s in a 72-GPU domain, plus 4x better bandwidth efficiency with SHARP FP8 support.
GB200
Architecture for real-time trillion-parameter inference and training
AI superchip
208 billion transistors in a single AI engine. All Blackwells feature two reticle-limited dies connected by a 10 TB/s chip-to-chip interconnect.
An engine built to boost
Second-generation Blackwell Transformer Engine’s custom Tensor Core technology meets NVIDIA’s software stack. Lightning-fast inference and training for LLMs and MoE models.
Massive Scale-Training
Second-generation Transformer Engine featuring FP8 precision gives 4x faster training for LLMs at scale.
B300
The building block of reasoning
Real-time inference and training
The same capabilities hyperscalers rely on with 144 petaFLOPS for inference and 72 petaFLOPS for training.
Automate AI operations
NVIDIA Mission Control and AI Enterprise helps automate infrastructure, cluster management, and model deployment to help teams scale efficiently.
Reasoning at scale
Blackwell Ultra Tensor Cores feature 2x faster attention-layer acceleration and 1.5x more AI compute FLOPS than standard Blackwell GPUs.
GB300
Efficiency for data center workloads, without performance compromise.
AI reasoning inference
NVIDIA Blackwell Ultra’s Tensor Cores boast 1.5x more AI compute FLOPS compared to Blackwell GPUs.
More memory
288 GB of HBM3e for larger batch sizing and boosted throughput for massive context lengths.
Superior RDMA
NVIDIA ConnectX-8 SuperNIC’s input/output (IO) module hosts two ConnectX-8 devices for 800 Gb/s of network connectivity per GPU.