Compare GPU Performance on AI Workloads
L40
![L40 image](/static/images/gpu/l40-gpu.webp)
![nvidia logo](/static/images/companies/nvidia-emblem.webp)
Vs.
L4
![L4 image](/static/images/gpu/l4-gpu.webp)
![nvidia logo](/static/images/companies/nvidia-emblem.webp)
LLM Benchmarks
Benchmarks were run on RunPod gpus using vllm. For more details on vllm, check out the vllm github repository.
Output Throughput (tok/s)
Select a model
128 input, 128 output
1
Output Throughput (tok/s)
Get started with RunPodÂ
today.
We handle millions of gpu requests a day. Scale your machine learning workloads while keeping costs low with RunPod.
Get Started