Updated minutes ago
B200 SXM
nvidia · blackwell · 180 GB HBM3e · 1000W TDP
VRAM
180 GB
BF16 TFLOPS
2250
Bandwidth
8000 GB/s
From
$4.49/hr
Spec Sheet
VRAM180 GB HBM3e
Memory Bandwidth8000 GB/s
BF16 TFLOPS2250
FP16 TFLOPS2250
FP8 TFLOPS4500
INT8 TOPS4500
TDP1000W
InterconnectNVLINK
NVLink Bandwidth1800 GB/s
Max per Node8
PCIe Gen6
CUDA Compute Capability10
Tensor CoresYes
Pricing by Provider
| Provider | On-Demand | Reserved | Spot | Badge |
|---|---|---|---|---|
| lambda | $5.99/hr | $4.49/hr | - | Cheapest |
| coreweave | $7.50/hr | $5.50/hr | - | |
| runpod | $7.20/hr | - | - |
Compatible Models (251)
Single GPU (226 models)
Mixtral 8x22B141B FP8DBRX Base132B FP8DBRX Instruct132B FP8Mistral Large 2411123B FP8Mistral Large 2123B FP8Llama 4 Scout109B FP8Command R+104B FP8Yi-Large102.6B FP8YaLM 100B100B FP8Llama 3.2 90B Vision90B FP8Llama 3.2 90B Vision Instruct88.8B FP8Qwen 2.5 72B72.7B FP8Qwen 2.5 Math 72B72.7B FP8Qwen 2.5 VL 72B72.7B FP8Dolphin 2.9 72B72B FP8DeepSeek R1 Distill 70B70.6B FP8Llama 3 70B 1M Context70.6B FP8Llama 3 70B70.6B FP8Llama 3.1 70B70.6B FP8Llama 3.3 70B70.6B FP8+206 more
Multi-GPU (25 models)
Training Capabilities
Estimated GPU count for full fine-tuning (AdamW, BF16) and QLoRA
| Model Size | Full Fine-Tune | QLoRA |
|---|---|---|
| 7B model | 1 GPU | 1 GPU |
| 13B model | 2 GPUs | 1 GPU |
| 70B model | 8 GPUs | 1 GPU |
Energy Efficiency
Estimated tokens/second per Watt for popular models
Mistral 7B
1.10 t/s/WFP8
Qwen 2.5 7B
1.05 t/s/WFP8
Llama 3.1 8B
1.00 t/s/WFP8
DeepSeek V3
0.22 t/s/WFP8
Llama 3.1 70B
0.11 t/s/WFP8
Qwen 2.5 72B
0.11 t/s/WFP8
Similar GPUs
| GPU | VRAM | BF16 TFLOPS | BW (GB/s) | From |
|---|---|---|---|---|
| B100 SXM | 192 GB | 1750 | 8000 | $4.50/hr |
| GB200 NVL72 (per GPU) | 192 GB | 2250 | 8000 | $6.50/hr |
| GB300 NVL72 (per GPU) | 192 GB | 2500 | 8000 | $7.50/hr |
| B300 | 288 GB | 2800 | 12000 | $0.00/hr |
| RTX 5090 | 32 GB | 210 | 1792 | $0.89/hr |