Updated minutes ago
GH200
nvidia · hopper · 96 GB HBM3 · 900W TDP
VRAM
96 GB
BF16 TFLOPS
990
Bandwidth
4000 GB/s
From
$2.99/hr
Spec Sheet
VRAM96 GB HBM3
Memory Bandwidth4000 GB/s
BF16 TFLOPS990
FP16 TFLOPS990
FP8 TFLOPS1980
INT8 TOPS1980
TDP900W
InterconnectNVLINK
NVLink Bandwidth900 GB/s
Max per Node8
PCIe Gen5
CUDA Compute Capability9
Tensor CoresYes
Pricing by Provider
| Provider | On-Demand | Reserved | Spot | Badge |
|---|---|---|---|---|
| coreweave | $3.99/hr | $2.99/hr | - | Cheapest |
| lambda | $3.49/hr | - | - |
Compatible Models (249)
Single GPU (207 models)
Qwen 2.5 72B72.7B FP8Qwen 2.5 Math 72B72.7B FP8Qwen 2.5 VL 72B72.7B FP8Dolphin 2.9 72B72B FP8DeepSeek R1 Distill 70B70.6B FP8Llama 3 70B 1M Context70.6B FP8Llama 3 70B70.6B FP8Llama 3.1 70B70.6B FP8Llama 3.3 70B70.6B FP8Hermes 3 70B70.6B FP8HelpSteer2 Llama 3.1 70B70.6B FP8Llama 3.1 Nemotron 70B Instruct70.6B FP8Llama 3.1 Nemotron 70B Reward70.6B FP8Nemotron 70B70.6B FP8Llama 3.1 70B Turbo70.6B FP8Code Llama 70B70B FP8Llama 2 70B70B FP8WizardMath 70B70B FP8Meditron 70B70B FP8Japanese StableLM 70B70B FP8+187 more
Multi-GPU (42 models)
Mixtral 8x22Bx2 FP8DBRX Basex2 FP8DBRX Instructx2 FP8Mistral Large 2411x2 FP8Mistral Large 2x2 FP8Llama 4 Scoutx2 FP8Command R+x2 FP8Yi-Largex2 FP8YaLM 100Bx2 FP8Llama 3.2 90B Visionx2 FP8Llama 3.2 90B Vision Instructx2 FP8Claude Sonnet 4x2 BF16o1-minix2 BF16o3-minix2 BF16Claude 3 Sonnetx2 BF16+27 more
Training Capabilities
Estimated GPU count for full fine-tuning (AdamW, BF16) and QLoRA
| Model Size | Full Fine-Tune | QLoRA |
|---|---|---|
| 7B model | 2 GPUs | 1 GPU |
| 13B model | 3 GPUs | 1 GPU |
| 70B model | 14 GPUs | 1 GPU |
Energy Efficiency
Estimated tokens/second per Watt for popular models
Mistral 7B
0.61 t/s/WFP8
Qwen 2.5 7B
0.58 t/s/WFP8
Llama 3.1 8B
0.55 t/s/WFP8
Llama 3.1 70B
0.06 t/s/WFP8
Qwen 2.5 72B
0.06 t/s/WFP8