Our team’s insights on building better and scaling smarter.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Brendan McKeag
16 May 2025
Automated Image Captioning with Gemma 3 on Runpod Serverless
Learn how to deploy a lightweight Gemma 3 model to generate image captions using Runpod Serverless. This walkthrough includes setup, deployment, and sample outputs.
From OpenAI API to Self-Hosted Model: A Migration Guide
Tired of usage limits or API costs? This guide walks you through switching from OpenAI’s API to your own self-hosted LLM using open-source models on Runpod.
From Pods to Serverless: When to Switch and Why It Matters
Finished training your model in a Pod? This guide helps you decide when to switch to Serverless, what trade-offs to expect, and how to optimize for fast, cost-efficient inference.
How a Solo Dev Built an AI for Dads—No GPU, No Team, Just $5
No GPU. No team. Just $5. This is how one solo developer used Runpod Serverless to build and deploy a working AI product—"AI for Dads"—without writing any custom training code.
RunPod now integrates directly with AI IDEs like Cursor and Claude Desktop using MCP. Launch pods, deploy endpoints, and manage infrastructure—right from your editor.
Alibaba’s Qwen3 is here—with major performance improvements and a full range of models from 0.5B to 72B parameters. This post breaks down what’s new, how it compares to other open models, and what it means for developers.
GPU Clusters: Powering High-Performance AI (When You Need It)
Different stages of AI development call for different infrastructure. This post breaks down when GPU clusters shine—and how to scale up only when it counts.