We're officially SOC 2 Type II Compliant
You've unlocked a referral bonus! Sign up today and you'll get a random credit bonus between $5 and $500
You've unlocked a referral bonus!
Claim Your Bonus
Claim Bonus
Brendan McKeag

Brendan McKeag

Announcing Global Networking for Secure Pod-to-Pod Communication Across Data Centers

Runpod now supports secure internal communication between pods across data centers. With Global Networking enabled, your pods can talk to each other privately via .runpod.internal—no open ports required.
Read article
Product Updates

Scoped API Keys Now Live: Secure, Fine-Grained Access Control on Runpod

Runpod now supports scoped API keys with per-endpoint access, usage tracking, and on/off toggles. Create safer, more flexible keys that align with the principle of least privilege.
Read article
Product Updates

How to Build and Deploy an AI Chatbot from Scratch with Runpod: A Community Project Breakdown

Explore how Code in a Jiffy built a fully functional AI-powered coffee shop chatbot using Runpod. This community spotlight covers agentic chatbot structures, full-stack architecture, and how Runpod’s serverless infra simplifies deployment.
Read article
Learn AI

Stable Diffusion 3.5 Is Here — Better Quality, Easier Prompts, and Real Photorealism

Stable Diffusion 3.5 delivers a major quality leap, fixing past flaws while generating photorealistic images from minimal prompts. Learn what’s new, how to get started on Runpod, and what to expect next from the community.
Read article
Hardware & Trends

Why NVidia's Llama 3.1 Nemotron 70B Might Be the Most Reasonable LLM Yet

NVidia’s Llama 3.1 Nemotron 70B is outperforming larger and closed models on key reasoning tasks. In this post, Brendan tests it against a long-unsolved challenge: consistent, in-character roleplay with zero internal monologue or user coercion—and finds it finally up to the task.
Read article
AI Workloads

Why LLMs Can't Spell 'Strawberry' And Other Odd Use Cases

Large language models can write poetry and solve logic puzzles—but fail at tasks like counting letters or doing math. Here’s why, and what it tells us about their design.
Read article
Learn AI

Run GGUF Quantized Models Easily with KoboldCPP on Runpod

Lower VRAM usage and improve inference speed using GGUF quantized models in KoboldCPP with just a few environment variables.
Read article
AI Workloads

Build what’s next.

The most cost-effective platform for building, training, and scaling machine learning models—ready when you are.

You’ve unlocked a
referral bonus!

Sign up today and you’ll get a random credit bonus between $5 and $500 when you spend your first $10 on Runpod.