Blog Update

YOUR B200 SXM6
QUICKSTART GUIDE

Follow our step-by-step guide to set up your B200 environment and
run an state-of-art Large Language Model with vLLM and SGLang.
NVIDIA B200 SXM6 QUICKSTART GUIDE
New Arrival

NVIDIA B200 SXM6
ALIVE ON CLOUD

Designed for the most demanding AI workloads.
Early and instant access to the Blackwell GPUs starting at $0.99/h.
NVIDIA B200 SXM6 ALIVE ON CLOUD
Blog Update

YOUR B200 SXM6
QUICKSTART GUIDE

Follow our step-by-step guide to set up your B200 environment and
run an state-of-art Large Language Model with vLLM and SGLang.
NVIDIA B200 SXM6 QUICKSTART GUIDE
New Arrival

NVIDIA B200 SXM6
ALIVE ON CLOUD

Designed for the most demanding AI workloads.
Early and instant access to the Blackwell GPUs starting at $0.99/h.
NVIDIA B200 SXM6 ALIVE ON CLOUD
3x
Faster Training Performance
15x
Faster real-time LLM inference
2.25x
The throughput of NVIDIA HGX H200

Compared to NVIDIA H200-SXM5, B200 excels in all technical benchmark tests. Learn More >

Performance

Enjoy Unmatched GPU Performance

NVIDIA B200 SXM6 GPU delivers breakthrough performance for AI training, fine-tuning, inference
tasks with massive memory capacity and ultra-high bandwidth.Start Now >

Computational Tasks

B200 offers nearly double the throughput, ideal for AI tasks that are heavily compute-bound.

Compute Performance: GEMM Operator Benchmark

H200-SXM-141GB
B200-SXM-180GB
600
800
1000
1200
1400
1600

Distributed Training

B200's improved communication performance contributes to better overall scaling, as seen in the PyTorch Distributed and Colossal-AI benchmarks.

Scaling Efficiency: All-Reduce Communication Benchmark

28163264128256size(MiB)085170255340425510595busbw(GB/s)H200-SXM-141GBB200-SXM-180GB

Real-world LLM training

B200 achieves significantly higher sample throughput and TFLOPS per GPU. This translates directly to faster training and inference times, making the B200 an excellent choice for production choices.

Real-time LLM inference throughput: Colossal-AI Benchmark

H200-SXM-141GB
B200-SXM-180GB
TFLOPS/GPU
900
800
700
600
500
400
300
200
100
0

7B model on 8 GPUs

TFLOPS/GPU
900
800
700
600
500
400
300
200
100
0

70B model on 16 GPUs

Technical Specs

NVIDIA B200 ON-DEMAND

CPU2 × Intel® Xeon® 6767P Processors
GPUNVIDIA® HGX™ 8 × B200-180 GB
Memory32 × 64 GB DDR5 6400 MHz
System disk2 × 960 GB NVMe SSD
Data disk8 × 3.84 TB NVMe SSD
Storage Network200 Gb/s
Compute Network8 × 400 Gb/s InfiniBand
NVIDIA DGX B200 SXM Blackwell 8 x 180GB Full Machine Cloud Instance

Flexible Pricing Options

Choose the payment model that best fits your workload requirements. Learn More >

On-Demand

$0.99 /hr
Per B200 SXM6 GPU
  • Instant deployment
  • Pay per second billing
  • No long-term commitment
  • Global availability
  • 99% uptime SLA

Reserved Cluster

Contact Sales
for customized requirements
  • Bare Metal or Cloud Access
  • Priority access
  • Custom configurations
  • Dedicated support
  • 99% uptime SLA

Dynamic Pricing

from $1.99 /hr
Per B200 SXM6 GPU
  • Up to 36% savings
  • Flexible workloads
  • Pay per second billing
  • No long-term commitment
  • Global availability

Ready to accelerate your AI projects?

Contact our team to discuss your team's specific requirements.

Rent NVIDIA B200 Servers | Unleash Blackwell AI Performance at HPC-AI Cloud