
Wan 2.2 Releases With a Plethora Of New Features
Deploy Wan 2.2 on Runpod to unlock next-gen video generation with Mixture-of-Experts architecture, TI2V-5B support, and 83% more training data—run text-to-video and image-to-video models at scale using A100–H200 GPUs and customizable ComfyUI workflows.
AI Infrastructure

Deep Cogito Releases Suite of LLMs Trained with Iterative Policy Improvement
Deploy DeepCogito’s Cogito v2 models on Runpod to experience frontier-level reasoning at lower inference costs—choose from 70B to 671B parameter variants and leverage Runpod’s optimized templates and Instant Clusters for scalable, efficient AI deployment.
AI Infrastructure

Comparing the 5090 to the 4090 and B200: How Does It Stack Up?
Benchmark Qwen2.5-Coder-7B-Instruct across NVIDIA’s B200, RTX 5090, and 4090 to identify optimal GPUs for LLM inference—compare token throughput, cost per token, and memory efficiency to match your workload with the right performance tier.
Hardware & Trends

How to Run MoonshotAI’s Kimi-K2-Instruct on RunPod Instant Cluster
Run MoonshotAI’s Kimi-K2-Instruct on RunPod Instant Clusters using H200 SXM GPUs and a 2TB shared network volume for seamless multi-node training. This guide shows how to deploy with PyTorch templates, optimize Docker environments, and accelerate LLM inference with scalable, low-latency infrastructure.
AI Workloads

Iterative Refinement Chains with Small Language Models: Breaking the Monolithic Prompt Paradigm
As prompt complexity increases, large language models (LLMs) hit a “cognitive wall,” suffering up to 40% performance drops due to task interference and overload. By decomposing workflows into iterative refinement chains (e.g., the Self-Refine framework) and deploying each stage on serverless platforms like RunPod, you can maintain high accuracy, scalability, and cost efficiency.
AI Workloads

Running a 1-Trillion Parameter AI Model In a Single Pod: A Guide to MoonshotAI’s Kimi-K2 on Runpod
Moonshot AI’s Kimi-K2-Instruct is a trillion-parameter, mixture-of-experts open-source LLM optimized for autonomous agentic tasks—with 32 billion active parameters, Muon-trained performance rivaling proprietary models (89.5 % MMLU, 97.4 % MATH-500, 65.8 % pass@1), and the ability to run inference on as little as 1 TB of VRAM using 8-bit quantization.
AI Workloads

Streamline Your AI Workflows with RunPod’s New S3-Compatible API
RunPod’s new S3-compatible API lets you manage files on your network volumes without launching a Pod. With support for standard tools like the AWS CLI and Boto3, you can upload, sync, and automate data flows directly from your terminal — simplifying storage operations and saving on compute costs. Whether you’re prepping datasets or archiving model outputs, this update makes your AI workflows faster, cleaner, and more flexible.
Product Updates