RightNow-AI/forge-mcp-server
Forge: Swarm Agents That Turn Slow PyTorch Into Fast CUDA/Triton Kernels
<p align="center"> </p>
<h1 align="center">Forge MCP Server</h1>
<p align="center"> <strong>Swarm agents that turn slow PyTorch into fast CUDA/Triton kernels, from any AI coding agent.</strong> </p>
<p align="center"> <a href="https://www.npmjs.com/package/@rightnow/forge-mcp-server"></a> <a href="https://www.npmjs.com/package/@rightnow/forge-mcp-server"></a> <a href="https://modelcontextprotocol.io"></a> <a href="https://github.com/RightNow-AI/forge-mcp-server/blob/main/LICENSE"></a> <a href="https://www.typescriptlang.org/"></a> </p>
<p align="center"> <a href="#installation">Installation</a> · <a href="#tools">Tools</a> · <a href="#resources">Resources</a> · <a href="#prompts">Prompts</a> · <a href="#security">Security</a> · <a href="#development">Development</a> </p>
---
Forge transforms PyTorch models into production-grade CUDA/Triton kernels through automated multi-agent optimization. Using 32 parallel AI agents with inference-time scaling, it achieves up to 14x faster inference than torch.compile(mode='max-autotune-no-cudagraphs') while maintaining 100% numerical correctness.
This MCP server connects any MCP-compatible AI coding agent to Forge. Your agent submits PyTorch code, Forge optimizes it with swarm agents on real datacenter GPUs, and returns the faste
Loading reviews...