Blog

Runpod Blog

Our team’s insights on building better and scaling smarter.
All
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
A1111 Serverless API – Step-by-Step Video Tutorial

A1111 Serverless API – Step-by-Step Video Tutorial

This post features a video tutorial by generativelabs.co that walks users through deploying a Stable Diffusion A1111 API using Runpod Serverless. It covers setup, Dockerfile and handler edits, endpoint deployment, and testing via Postman—great for beginners and advanced users alike.
Read article
Learn AI
KoboldAI – The Other Roleplay Front End, And Why You May Want to Use It

KoboldAI – The Other Roleplay Front End, And Why You May Want to Use It

While Oobabooga is a popular choice for text-based AI roleplay, KoboldAI offers a powerful alternative with smart context handling, more flexible editing, and better long-term memory retention. This guide compares the two frontends and walks through deploying KoboldAI on Runpod for writers and roleplayers looking for a deeper, more persistent AI interaction experience.
Read article
Learn AI
Breaking Out of the 2048 Token Context Limit in Oobabooga

Breaking Out of the 2048 Token Context Limit in Oobabooga

Oobabooga now supports up to 8192 tokens of context, up from the previous 2048-token limit. Learn how to upgrade your install, download compatible models, and optimize your setup to take full advantage of expanded memory capacity in longform text generation.
Read article
Learn AI
Groundbreaking H100 NVidia GPUs Now Available On Runpod

Groundbreaking H100 NVidia GPUs Now Available On Runpod

Runpod now offers access to NVIDIA’s powerful H100 GPUs, designed for generative AI workloads at scale. These next-gen GPUs deliver 7–12x performance gains over the A100, making them ideal for training massive models like GPT-4 or deploying demanding inference tasks.
Read article
Hardware & Trends
Faster-Whisper: 3x Cheaper and 4x Faster Than Whisper for Speech Transcription

Faster-Whisper: 3x Cheaper and 4x Faster Than Whisper for Speech Transcription

Runpod's new Faster-Whisper endpoint delivers 2–4x faster transcription speeds than the original Whisper API—at a fraction of the cost. Perfect for podcasts, interviews, and multilingual speech recognition.
Read article
Product Updates
Creating a Vlad Diffusion Template for RunPod

Creating a Vlad Diffusion Template for RunPod

Want a custom spin on Stable Diffusion? This post shows you how to create and launch your own Vlad Diffusion template inside RunPod.
Read article
AI Workloads
How to Work With Long Term Memory In Oobabooga and Text Generation

How to Work With Long Term Memory In Oobabooga and Text Generation

Oobabooga has a 2048-token context limit, but with the Long Term Memory extension, you can store and retrieve relevant memories across conversations. This guide shows how to install the plugin, use the Character panel for persistent memory, and work around current context limitations.
Read article
Learn AI

Build what’s next.

The most cost-effective platform for building, training, and scaling machine learning models—ready when you are.