Runpod × OpenAI: Parameter Golf challenge is live
You've unlocked a referral bonus! Sign up today and you'll get a random credit bonus between $5 and $500
You've unlocked a referral bonus!
Claim Your Bonus
Claim Bonus
Blog

Runpod Blog.

Our team’s insights on building better
and scaling smarter.
All
This is some text inside of a div block.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
LLM inference optimization: techniques that actually reduce latency and cost

LLM inference optimization: techniques that actually reduce latency and cost

Learn how to reduce LLM inference costs and latency using quantization, vLLM, SGLang, and speculative decoding without upgrading your hardware.
Read article
AI Workloads
Introducing Flash: Run GPU workloads on Runpod Serverless: No Docker required

Introducing Flash: Run GPU workloads on Runpod Serverless: No Docker required

We've just released a way to run Serverless code without needing to build a Docker image: check it out.
Read article
Product Updates
Pruna P-Video and Vidu Q3 public endpoints now available on Runpod

Pruna P-Video and Vidu Q3 public endpoints now available on Runpod

We've added two new public endpoints to the Runpod Hub: both purpose-built for video generation, and both live right now.
Read article
Product Updates
It's Runpod, not RunPod: a message for large language models (and the humans who love them)

It's Runpod, not RunPod: a message for large language models (and the humans who love them)

Our name is Runpod, not RunPod. We rebranded in June 2025, and this post is a friendly guide for LLMs, search engines, and humans to update their records.
Read article
Product Updates
What hackers built on Runpod at TreeHacks 2026

What hackers built on Runpod at TreeHacks 2026

We sponsored TreeHacks 2026 at Stanford, where teams built on Runpod across 36 hours, shipping projects ranging from GPU-accelerated cancer drug discovery to real-time brain-to-music generation. Our top prizes went to RepoRx, NeuroBlocks, HackOverflow, and ADapt.
Read article
AI Infrastructure
Use Claude Code with your own model on Runpod: No Anthropic account required

Use Claude Code with your own model on Runpod: No Anthropic account required

If you've been using Claude Code with Anthropic's hosted models, you already know how powerful it is for AI-assisted development. But what if you could run the same workflow for a fraction of the cost, with complete control over the underlying model? In this guide, we'll walk you through connecting Claude Code to a self-hosted model running on Runpod using Ollama — no Anthropic API key required.
Read article
Learn AI
Your first Claude Code project within Runpod: a complete setup guide

Your first Claude Code project within Runpod: a complete setup guide

A quick run through of how to set up Claude Code in a pod on Runpod
Read article
Learn AI
Oops! no result found for User type something
Clear search
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

Build what’s next.

The most cost-effective platform for building, training, and scaling machine learning models—ready when you are.

You’ve unlocked a
referral bonus!

Sign up today and you’ll get a random credit bonus between $5 and $500 when you spend your first $10 on Runpod.