Explore our credit programs for startups

Emmett Fear

Emmett runs Growth at Runpod. He lives in Utah with his wife and dog, and loves to spend time hiking and paddleboarding. He has worked in many different facets of tech, from marketing, operations, product, and most recently, growth.

How to Serve Phi-2 on a Cloud GPU with vLLM and FastAPI

Provides step-by-step instructions to serve the Phi-2 language model on a cloud GPU using vLLM and FastAPI. Covers setting up vLLM for efficient inference and deploying a FastAPI server to expose the model via a REST API.
Guides

How to Run OpenChat on a Cloud GPU Using Docker

Offers a guide on running the OpenChat model on a cloud GPU using Docker. Explains how to configure the Docker environment for OpenChat and deploy it for inference, so you can interact with the model without local installation.
Guides

How to Run StarCoder2 as a REST API in the Cloud

Shows how to deploy StarCoder2 as a REST API on a cloud GPU. Walks through containerizing the code-generation model and setting up an API service, enabling you to query the model remotely with GPU-accelerated performance.
Guides

Train Any AI Model Fast with PyTorch 2.1 + CUDA 11.8 on Runpod: The Ultimate Guide

Demonstrates how to train any AI model quickly using PyTorch 2.1 with CUDA 11.8 on Runpod. Covers preparing the environment and using Runpod’s GPUs to accelerate training, with tips for optimizing training speed in the cloud.
Guides

Using Ollama to Serve Quantized Models from a GPU Container

Shows how to use Ollama to serve quantized AI models from a GPU-accelerated Docker container. Details how model quantization improves efficiency and how to set up Ollama in the container for faster, lighter-weight inference.
Guides

LLM Training with Runpod GPU Pods: Scale Performance, Reduce Overhead

Describes how to scale large language model (LLM) training using Runpod GPU pods. Highlights performance tuning and cost optimization strategies to maximize training efficiency and reduce overhead in cloud environments.
Guides

Instant Clusters for AI Research: Deploy and Scale in Minutes

Highlights how Runpod’s Instant Clusters can accelerate AI research. Discusses deploying GPU clusters within minutes and how this capability allows rapid scaling for experiments and collaborative projects without lengthy setup.
Guides

Automate AI Image Workflows with ComfyUI + Flux on Runpod: Ultimate Creative Stack

Shows how to automate AI image generation workflows by integrating ComfyUI with Flux on Runpod. Details setting up an automated pipeline using cloud GPUs and workflow tools to streamline the creation of AI-generated art.
Guides

Finding the Best Docker Image for vLLM Inference on CUDA 12.4 GPUs

Guides you in choosing the optimal Docker image for vLLM inference on CUDA 12.4–compatible GPUs. Compares available images and configurations to ensure you select one that maximizes performance for serving large language models.
Guides

Build what’s next.

The most cost-effective platform for building, training, and scaling machine learning models—ready when you are.

You’ve unlocked a
referral bonus!

Sign up today and you’ll get a random credit bonus between $5 and $500 when you spend your first $10 on Runpod.