Skip to content
Updated minutes ago
nvidia

L4

nvidia · ada · 24 GB GDDR6 · 72W TDP

VRAM

24 GB

BF16 TFLOPS

121

Bandwidth

300 GB/s

From

$0.29/hr

Calculate ROI with this GPU →

Spec Sheet

VRAM24 GB GDDR6
Memory Bandwidth300 GB/s
BF16 TFLOPS121
FP16 TFLOPS121
FP8 TFLOPS242
INT8 TOPS242
TDP72W
InterconnectPCIE
Max per Node8
PCIe Gen4
CUDA Compute Capability8.9
Tensor CoresYes

Pricing by Provider

ProviderOn-DemandReservedSpotBadge
tensordock$0.39/hr-$0.29/hrCheapest
vast_ai$0.45/hr-$0.30/hr
runpod$0.69/hr-$0.49/hr
gcp$0.70/hr$0.49/hr-
aws$0.81/hr$0.52/hr-
lambda$0.59/hr--

Compatible Models (230)

Training Capabilities

Estimated GPU count for full fine-tuning (AdamW, BF16) and QLoRA

Model SizeFull Fine-TuneQLoRA
7B model6 GPUs1 GPU
13B model11 GPUs1 GPU
70B model55 GPUs2 GPUs

Energy Efficiency

Estimated tokens/second per Watt for popular models

Mistral 7B
0.57 t/s/WFP8
Qwen 2.5 7B
0.55 t/s/WFP8
Llama 3.1 8B
0.52 t/s/WFP8
Llama 3.1 70B
0.06 t/s/WFP8
Qwen 2.5 72B
0.06 t/s/WFP8

Similar GPUs

GPUVRAMBF16 TFLOPSBW (GB/s)From
RTX 409024 GB1651008$0.39/hr
RTX 408016 GB97717$0.32/hr
RTX 4060 Ti 16GB16 GB44288$0.30/hr
RTX 4070 Ti12 GB93504$0.25/hr
RTX 4070 Super12 GB55504$0.22/hr