Cloud GPUs
Rent NVIDIA L40 GPUs from $0.99/hr
High-performance data center GPU with 48 GB GDDR6 memory and Ada Lovelace architecture, designed for AI inference, 3D rendering, and virtualization workloads with 300W power consumption in a dual-slot form factor.

Powering the next generation of AI & high-performance computing.

Engineered for large-scale AI training, deep learning, and high-performance workloads, delivering unprecedented compute power and efficiency.

NVIDIA Ada Lovelace Architecture

Revolutionary neural graphics architecture delivering unprecedented visual computing performance with advanced AI capabilities.

Fourth-Generation Tensor Cores

Enhanced AI acceleration delivering over 1 petaFLOP of inference performance for deep learning workloads.

48GB GDDR6 Memory

Industry-leading memory capacity enables handling of large AI models and complex 3D scenes.

Third-Generation RT Cores

Advanced ray tracing acceleration with 2X faster real-time performance for photorealistic rendering workflows.
Performance

Key specs at a glance.

Performance benchmarks that push AI, ML, and HPC workloads further.

Memory Bandwidth

864

GB/s

FP16 Tensor Performance

181

TFLOPS

PCIe Gen4 ×16 Bandwidth

63

GB/s
Use Cases

Popular use cases.

Designed for demanding workloads
—learn if this GPU fits your needs.
Technical Specs

Ready for your most
demanding workloads.

Essential technical specifications to help you choose the right GPU for your workload.

Specification

Details

Great for...

Memory Bandwidth
864
GB/s
Feeding massive multimodal and high-resolution image and LLM inference workloads without memory stalls.
Memory Bandwidth
181
Feeding massive multimodal and high-resolution image and LLM inference workloads without memory stalls.
Feeding massive multimodal and high-resolution image and LLM inference workloads without memory stalls.
864
FP16 Tensor Performance
181
TFLOPS
Accelerating mixed-precision transformer and convolution operations in generative AI and graphics workloads.
FP16 Tensor Performance
181
Accelerating mixed-precision transformer and convolution operations in generative AI and graphics workloads.
Accelerating mixed-precision transformer and convolution operations in generative AI and graphics workloads.
181
PCIe Gen4 ×16 Bandwidth
63
GB/s
Enabling high-speed GPU-to-GPU and host-to-device transfers when NVLink isn’t available, ensuring smooth multi-GPU scaling for training and inference.
PCIe Gen4 ×16 Bandwidth
63
Enabling high-speed GPU-to-GPU and host-to-device transfers when NVLink isn’t available, ensuring smooth multi-GPU scaling for training and inference.
Enabling high-speed GPU-to-GPU and host-to-device transfers when NVLink isn’t available, ensuring smooth multi-GPU scaling for training and inference.
63
Comparison

Powerful GPUs. Globally available.
Reliability you can trust.

30+ GPUs, 31 regions, instant scale. Fine-tune or go full Skynet—we’ve got you.

Community Cloud

$
0.69
/hr
N/A

Secure Cloud

$
0.99
/hr
Unique GPU Models
25
19
Unique GPU Models
Lorem ipsum
19
Lorem ipsum
25
Global Regions
17
14
Global Regions
Lorem ipsum
14
Lorem ipsum
17
Network Storage
Network Storage
Lorem ipsum
Lorem ipsum
Lorem ipsum
Lorem ipsum
Enterprise-Grade Reliability
Network Storage
Lorem ipsum
Lorem ipsum
Lorem ipsum
Lorem ipsum
Savings Plans
Network Storage
Lorem ipsum
Lorem ipsum
Lorem ipsum
Lorem ipsum
24/7 Support
Network Storage
Lorem ipsum
Lorem ipsum
Lorem ipsum
Lorem ipsum
Delightful Dev Experience
Network Storage
Lorem ipsum
Lorem ipsum
Lorem ipsum
Lorem ipsum

7,035,265,000

Requests since launch & 400k developers worldwide

Build what’s next.

The most cost-effective platform for building, training, and scaling machine learning models—ready when you are.

12:22