Explore our credit programs for startups
You've unlocked a referral bonus! Sign up today and you'll get a random credit bonus between $5 and $500
You've unlocked a referral bonus!
Claim Your Bonus
Claim Bonus
Blog

Runpod Blog

Our team’s insights on building better and scaling smarter.
All
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
How Krnl Scaled to Millions—and Cut Infra Costs by 65%

How Krnl Scaled to Millions—and Cut Infra Costs by 65%

Discover how Krnl transitioned from AWS to Runpod’s Serverless GPUs to support millions of users—slashing idle cost and scaling more efficiently.
Read article
Mixture of Experts (MoE): A Scalable AI Training Architecture

Mixture of Experts (MoE): A Scalable AI Training Architecture

MoE models scale efficiently by activating only a subset of parameters. Learn how this architecture works, why it’s gaining traction, and how Runpod supports MoE training and inference.
Read article
AI Workloads
RunPod Global Networking Expands to 14 More Data Centers

RunPod Global Networking Expands to 14 More Data Centers

RunPod’s global networking feature is now available in 14 new data centers, improving latency and accessibility across North America, Europe, and Asia.
Read article
AI Infrastructure
How to Fine-Tune LLMs with Axolotl on RunPod

How to Fine-Tune LLMs with Axolotl on RunPod

Learn how to fine-tune large language models using Axolotl on RunPod. This guide covers LoRA, 8-bit quantization, DeepSpeed, and GPU infrastructure setup.
Read article
AI Workloads
RTX 5090 LLM Benchmarks: Is It the Best GPU for AI?

RTX 5090 LLM Benchmarks: Is It the Best GPU for AI?

See how the NVIDIA RTX 5090 stacks up in large language model benchmarks. We explore real-world performance and whether it’s the top GPU for AI workloads today.
Read article
Hardware & Trends
The RTX 5090 Is Here: Serve 65,000+ Tokens Per Second on RunPod

The RTX 5090 Is Here: Serve 65,000+ Tokens Per Second on RunPod

The new NVIDIA RTX 5090 is now live on RunPod. With blazing-fast inference speeds and large memory capacity, it’s ideal for real-time LLM workloads and AI scaling.
Read article
AI Workloads
Cost-Effective AI with Autoscaling on RunPod

Cost-Effective AI with Autoscaling on RunPod

Learn how RunPod autoscaling helps teams cut costs and improve performance for both training and inference. Includes best practices and real-world efficiency gains.
Read article
AI Workloads

Build what’s next.

The most cost-effective platform for building, training, and scaling machine learning models—ready when you are.

You’ve unlocked a
referral bonus!

Sign up today and you’ll get a random credit bonus between $5 and $500 when you spend your first $10 on Runpod.