Cloud GPUs
Rent NVIDIA L40 GPUs from $0.99/hr
High-performance data center GPU with 48 GB GDDR6 memory and Ada Lovelace architecture, designed for AI inference, 3D rendering, and virtualization workloads with 300W power consumption in a dual-slot form factor.

Powering the next generation of AI & high-performance computing.

Engineered for large-scale AI training, deep learning, and high-performance workloads, delivering unprecedented compute power and efficiency.

NVIDIA Ada Lovelace Architecture

Revolutionary neural graphics architecture delivering unprecedented visual computing performance with advanced AI capabilities.

Fourth-Generation Tensor Cores

Enhanced AI acceleration delivering over 1 petaFLOP of inference performance for deep learning workloads.

48GB GDDR6 Memory

Industry-leading memory capacity enables handling of large AI models and complex 3D scenes.

Third-Generation RT Cores

Advanced ray tracing acceleration with 2X faster real-time performance for photorealistic rendering workflows.
Performance

Key specs at a glance.

Performance benchmarks that push AI, ML, and HPC workloads further.

Memory Bandwidth

864

GB/s

FP16 Tensor Performance

181

TFLOPS

PCIe Gen4 ×16 Bandwidth

63

GB/s
Use Cases

Popular use cases.

Designed for demanding workloads
—learn if this GPU fits your needs.
Technical Specs

Ready for your most
demanding workloads.

Essential technical specifications to help you choose the right GPU for your workload.

Specification

Details

Great for...

Memory Bandwidth
864
GB/s
Feeding massive multimodal and high-resolution image and LLM inference workloads without memory stalls.
Memory Bandwidth
864
GB/s
Feeding massive multimodal and high-resolution image and LLM inference workloads without memory stalls.
FP16 Tensor Performance
181
TFLOPS
Accelerating mixed-precision transformer and convolution operations in generative AI and graphics workloads.
FP16 Tensor Performance
181
TFLOPS
Accelerating mixed-precision transformer and convolution operations in generative AI and graphics workloads.
PCIe Gen4 ×16 Bandwidth
63
GB/s
Enabling high-speed GPU-to-GPU and host-to-device transfers when NVLink isn’t available, ensuring smooth multi-GPU scaling for training and inference.
PCIe Gen4 ×16 Bandwidth
63
GB/s
Enabling high-speed GPU-to-GPU and host-to-device transfers when NVLink isn’t available, ensuring smooth multi-GPU scaling for training and inference.
Comparison

Powerful GPUs. Globally available.
Reliability you can trust.

30+ GPUs, 31 regions, instant scale. Fine-tune or go full Skynet—we’ve got you.

Community Cloud

$
0.69
/hr
N/A

Secure Cloud

$
0.99
/hr
Unique GPU Models
25
19
Unique GPU Models
Secure Cloud
19
Community Cloud
25
Global Regions
17
14
Global Regions
Secure Cloud
14
Community Cloud
17
Network Storage
Network Storage
Secure Cloud
✔️
Community Cloud
✖️
Enterprise-Grade Reliability
Enterprise-Grade Reliability
Secure Cloud
✔️
Community Cloud
✖️
Savings Plans
Savings Plans
Secure Cloud
✔️
Community Cloud
✖️
24/7 Support
24/7 Support
Secure Cloud
✔️
Community Cloud
✔️
Delightful Dev Experience
Delightful Dev Experience
Secure Cloud
✔️
Community Cloud
✔️

7,035,265,000

Requests since launch & 400k developers worldwide

Build what’s next.

The most cost-effective platform for building, training, and scaling machine learning models—ready when you are.