Cloud GPUs
Rent NVIDIA B200 GPUs from $5.99/hr
Next-generation data center GPU based on Blackwell architecture that features 192GB of HBM3e memory with 8TB/s bandwidth, delivering up to 20 petaFLOPS of FP4 AI compute performance.

Powering the next generation of AI & high-performance computing.

Engineered for large-scale AI training, deep learning, and high-performance workloads, delivering unprecedented compute power and efficiency.

Built on NVIDIA Blackwell Architecture

Revolutionary dual-die design with 208 billion transistors delivering breakthrough performance for trillion-parameter AI models.

Second-Generation Transformer Engine

Custom Tensor Core technology accelerates LLM inference and training with new FP4 precision for up to 2.5X performance gains.

192GB HBM3e at 8TB/s

Massive memory capacity with ultra-high bandwidth enables training and inference on the largest AI models without memory constraints.

Fifth-Generation NVLink

1.8TB/s GPU-to-GPU interconnect allows seamless scaling across multiple GPUs for distributed AI workloads.
Performance

Key specs at a glance.

Performance benchmarks that push AI, ML, and HPC workloads further.

Memory Bandwidth

8

TB/s

FP16 Tensor Performance

2.2

PFLOPS

NVLink Bandwidth

1.8

TB/s
Use Cases

Popular use cases.

Designed for demanding workloads
—learn if this GPU fits your needs.
Technical Specs

Ready for your most
demanding workloads.

Essential technical specifications to help you choose the right GPU for your workload.

Specification

Details

Great for...

Memory Bandwidth
8
TB/s
Delivering large datasets to GPU memory without stalls for high-resolution simulations and LLM inference.
Memory Bandwidth
8
TB/s
Delivering large datasets to GPU memory without stalls for high-resolution simulations and LLM inference.
FP16 Tensor Performance
2.2
PFLOPS
Accelerating mixed-precision transformer training and inference, reducing fine-tuning times and boosting token throughput
FP16 Tensor Performance
2.2
PFLOPS
Accelerating mixed-precision transformer training and inference, reducing fine-tuning times and boosting token throughput
NVLink Bandwidth
1.8
TB/s
Enabling fast GPU-to-GPU communication in multi-card setups, avoiding PCIe bottlenecks for large model scaling.
NVLink Bandwidth
1.8
TB/s
Enabling fast GPU-to-GPU communication in multi-card setups, avoiding PCIe bottlenecks for large model scaling.
Comparison

Powerful GPUs. Globally available.
Reliability you can trust.

30+ GPUs, 31 regions, instant scale. Fine-tune or go full Skynet—we’ve got you.

Community Cloud

$
/hr
N/A

Secure Cloud

$
5.99
/hr
Unique GPU Models
25
19
Unique GPU Models
Secure Cloud
19
Community Cloud
25
Global Regions
17
14
Global Regions
Secure Cloud
14
Community Cloud
17
Network Storage
Network Storage
Secure Cloud
✔️
Community Cloud
✖️
Enterprise-Grade Reliability
Enterprise-Grade Reliability
Secure Cloud
✔️
Community Cloud
✖️
Savings Plans
Savings Plans
Secure Cloud
✔️
Community Cloud
✖️
24/7 Support
24/7 Support
Secure Cloud
✔️
Community Cloud
✔️
Delightful Dev Experience
Delightful Dev Experience
Secure Cloud
✔️
Community Cloud
✔️

7,035,265,000

Requests since launch & 400k developers worldwide

Build what’s next.

The most cost-effective platform for building, training, and scaling machine learning models—ready when you are.