
Shaamil Karim
Runpod and Defined.ai launch a pilot program to provide startups with access to high-quality training data and compute, enabling sector-specific fine-tuning and closing the data wealth gap.
Product Updates

Brendan McKeag
Runpod Serverless now supports multi-GPU workers, enabling full-precision deployment of large models like Llama-3 70B. With optimized VLLM support, flashboot, and network volumes, it's never been easier to run massive LLMs at scale.
Product Updates

Moritz Wallawitsch
Learn how vLLM achieves up to 24x higher throughput than Hugging Face Transformers by using PagedAttention to eliminate memory waste, boost inference performance, and enable efficient GPU usage.
AI Workloads

Moritz Wallawitsch
Learn how to run vLLM on Runpod’s serverless GPU platform. This guide walks you through fast, efficient LLM inference without complex setup.
AI Infrastructure

Brendan McKeag
Runpod introduces Serverless CPU: high-performance VM containers with customizable CPU options, ideal for cost-effective and versatile workloads not requiring GPUs.
Product Updates

Brendan McKeag
Our new Serverless CPU offering lets you launch high-performance containers without GPUs—perfect for lighter workloads, dev tasks, and automation.
Product Updates

River Snow
Learn how to securely access your Runpod Pod using SSH with a username and password by configuring the SSH daemon and setting a root password.
Learn AI
Oops! no result found for User type something