Updated minutes ago
H200 SXM
nvidia · hopper · 141 GB HBM3e · 700W TDP
VRAM
141 GB
BF16 TFLOPS
990
Bandwidth
4800 GB/s
From
$2.69/hr
Spec Sheet
VRAM141 GB HBM3e
Memory Bandwidth4800 GB/s
BF16 TFLOPS990
FP16 TFLOPS990
FP8 TFLOPS1979
INT8 TOPS1979
TDP700W
InterconnectNVLINK
NVLink Bandwidth900 GB/s
Max per Node8
PCIe Gen5
CUDA Compute Capability9
Tensor CoresYes
Pricing by Provider
| Provider | On-Demand | Reserved | Spot | Badge |
|---|---|---|---|---|
| lambda | $3.49/hr | $2.69/hr | - | Cheapest |
| tensordock | $3.80/hr | - | $2.90/hr | |
| coreweave | $4.25/hr | $3.19/hr | - | |
| runpod | $4.69/hr | - | - |
Pricing History
coreweave
$5.49/hr→ 0.0%
2024-09-012025-03-01
Low: $5.49High: $6.99
lambda
$3.99/hr→ 0.0%
2024-09-012025-03-01
Low: $3.99High: $5.49
Compatible Models (249)
Single GPU (216 models)
Llama 4 Scout109B FP8Command R+104B FP8Yi-Large102.6B FP8YaLM 100B100B FP8Llama 3.2 90B Vision90B FP8Llama 3.2 90B Vision Instruct88.8B FP8Qwen 2.5 72B72.7B FP8Qwen 2.5 Math 72B72.7B FP8Qwen 2.5 VL 72B72.7B FP8Dolphin 2.9 72B72B FP8DeepSeek R1 Distill 70B70.6B FP8Llama 3 70B 1M Context70.6B FP8Llama 3 70B70.6B FP8Llama 3.1 70B70.6B FP8Llama 3.3 70B70.6B FP8Hermes 3 70B70.6B FP8HelpSteer2 Llama 3.1 70B70.6B FP8Llama 3.1 Nemotron 70B Instruct70.6B FP8Llama 3.1 Nemotron 70B Reward70.6B FP8Nemotron 70B70.6B FP8+196 more
Multi-GPU (33 models)
Training Capabilities
Estimated GPU count for full fine-tuning (AdamW, BF16) and QLoRA
| Model Size | Full Fine-Tune | QLoRA |
|---|---|---|
| 7B model | 1 GPU | 1 GPU |
| 13B model | 2 GPUs | 1 GPU |
| 70B model | 10 GPUs | 1 GPU |
Energy Efficiency
Estimated tokens/second per Watt for popular models
Mistral 7B
0.94 t/s/WFP8
Qwen 2.5 7B
0.90 t/s/WFP8
Llama 3.1 8B
0.85 t/s/WFP8
DeepSeek V3
0.19 t/s/WFP8
Llama 3.1 70B
0.10 t/s/WFP8
Qwen 2.5 72B
0.09 t/s/WFP8
Similar GPUs
| GPU | VRAM | BF16 TFLOPS | BW (GB/s) | From |
|---|---|---|---|---|
| H20 | 96 GB | 148 | 4000 | $0.99/hr |
| GH200 | 96 GB | 990 | 4000 | $2.99/hr |
| H100 NVL | 94 GB | 835 | 3938 | $3.09/hr |
| H100 NVL 94GB (per GPU pair) | 188 GB | 1670 | 7876 | $5.49/hr |
| H100 SXM | 80 GB | 990 | 3350 | $1.89/hr |