We've cooked up a bunch of improvements designed to reduce friction and make the.



Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
Unordered list
Bold text
Emphasis
Superscript
Subscript
The way you build with AI is changing—and now, so is the way you interact with your infrastructure. We just shipped the official RunPod MCP server, unlocking first-class, chat-native access to your GPU fleet from any AI-first IDE.
Drop it into Cursor. Or Windsurf. Or Cline. Or Claude Desktop. If your editor speaks Model Context Protocol (MCP), it now speaks fluent RunPod.
No more context switching. No more curl commands. Just talk to your editor, and let your model do the heavy lifting—spin up pods, deploy endpoints, manage volumes, and more.
MCP (short for Model Context Protocol) is an open JSON‑RPC 2.0 standard that lets language model interfaces and tool providers speak the same language.
Instead of bolting together fragile glue code and one-off REST wrappers, you just define a few message types (initialize, request, result, etc.) and let the model decide what tools to invoke—and when. The MCP server handles the call. The client (Cursor, Claude Desktop, etc.) handles the UI. Your model handles the logic.
With the RunPod MCP server, that means full access to your RunPod account—directly from your editor, no context switch required.
Once you’ve plugged the MCP server into your setup, your LLM can access all of this (don’t worry, you can toggle each tool off and on at will, so what your LLM can do is completely within your control):
Pods
create-podlist-podsstart-podstop-poddelete-podget-podServerless Endpoints
create-endpointlist-endpointget-endpointdelete-endpointupdate-endpointTemplates
list-templateget-templatecreate-templateupdate-templatedelete-templateNetwork Volumes
list-network-volumesget-network-volumecreate-network-volumeupdate-network-volumedelete-network-volumeContainer Registry Auth
list-container-registry-authsget-container-registry-authcreate-container-registry-authdelete-container-registry-authUnder the hood, each tool wraps the same REST operations you’re already familiar with—just simplified. Your LLM handles the parameters, validation, and error handling for you.
Want to wire RunPod into Cursor? Just drop this JSON block into .cursor/mcp.json:
Swap in your RunPod personal API key from the console, restart your IDE, and you’re done.
Or just run the one-liner:
Now your model knows how to talk to RunPod.
Once set up, you can talk to your AI assistant like this:
“Create a serverless endpoint using my template called jacobs-comfyui.”
Behind the scenes, your IDE routes that natural-language request to your LLM, which selects the right MCP tool (create-endpoint), fills in the parameters, and fires it off to RunPod. A few seconds later, your endpoint is live.
No terminal. No docs. Just code and chat.
For Smithery-supported clients (like Claude Desktop):
Smithery handles install, build, and registration in one shot.
For Cursor, add this to .cursor/mcp.json:
Restart Cursor and ask something like:
“Spin up a 1×A100 pod using my template nightly-train.”
Your LLM will call create-pod and stream the pod ID back in seconds.
We’re big believers in developer flow. The future of infrastructure isn’t menus and dashboards—it’s context-aware, assistant-driven, and conversational. By building RunPod’s MCP server, we’re making it feel native to every AI-first editor you already use.
Less friction. Fewer tabs. More building.
Try it out. Clone the repo. Give it a star. And let us know what you want to see next.

RunPod now integrates directly with AI IDEs like Cursor and Claude Desktop using MCP. Launch pods, deploy endpoints, and manage infrastructure—right from your editor.

The way you build with AI is changing—and now, so is the way you interact with your infrastructure. We just shipped the official RunPod MCP server, unlocking first-class, chat-native access to your GPU fleet from any AI-first IDE.
Drop it into Cursor. Or Windsurf. Or Cline. Or Claude Desktop. If your editor speaks Model Context Protocol (MCP), it now speaks fluent RunPod.
No more context switching. No more curl commands. Just talk to your editor, and let your model do the heavy lifting—spin up pods, deploy endpoints, manage volumes, and more.
MCP (short for Model Context Protocol) is an open JSON‑RPC 2.0 standard that lets language model interfaces and tool providers speak the same language.
Instead of bolting together fragile glue code and one-off REST wrappers, you just define a few message types (initialize, request, result, etc.) and let the model decide what tools to invoke—and when. The MCP server handles the call. The client (Cursor, Claude Desktop, etc.) handles the UI. Your model handles the logic.
With the RunPod MCP server, that means full access to your RunPod account—directly from your editor, no context switch required.
Once you’ve plugged the MCP server into your setup, your LLM can access all of this (don’t worry, you can toggle each tool off and on at will, so what your LLM can do is completely within your control):
Pods
create-podlist-podsstart-podstop-poddelete-podget-podServerless Endpoints
create-endpointlist-endpointget-endpointdelete-endpointupdate-endpointTemplates
list-templateget-templatecreate-templateupdate-templatedelete-templateNetwork Volumes
list-network-volumesget-network-volumecreate-network-volumeupdate-network-volumedelete-network-volumeContainer Registry Auth
list-container-registry-authsget-container-registry-authcreate-container-registry-authdelete-container-registry-authUnder the hood, each tool wraps the same REST operations you’re already familiar with—just simplified. Your LLM handles the parameters, validation, and error handling for you.
Want to wire RunPod into Cursor? Just drop this JSON block into .cursor/mcp.json:
Swap in your RunPod personal API key from the console, restart your IDE, and you’re done.
Or just run the one-liner:
Now your model knows how to talk to RunPod.
Once set up, you can talk to your AI assistant like this:
“Create a serverless endpoint using my template called jacobs-comfyui.”
Behind the scenes, your IDE routes that natural-language request to your LLM, which selects the right MCP tool (create-endpoint), fills in the parameters, and fires it off to RunPod. A few seconds later, your endpoint is live.
No terminal. No docs. Just code and chat.
For Smithery-supported clients (like Claude Desktop):
Smithery handles install, build, and registration in one shot.
For Cursor, add this to .cursor/mcp.json:
Restart Cursor and ask something like:
“Spin up a 1×A100 pod using my template nightly-train.”
Your LLM will call create-pod and stream the pod ID back in seconds.
We’re big believers in developer flow. The future of infrastructure isn’t menus and dashboards—it’s context-aware, assistant-driven, and conversational. By building RunPod’s MCP server, we’re making it feel native to every AI-first editor you already use.
Less friction. Fewer tabs. More building.
Try it out. Clone the repo. Give it a star. And let us know what you want to see next.
The most cost-effective platform for building, training, and scaling machine learning models—ready when you are.