RocketRide
RocketRide is a self-hosted, open-source AI pipeline platform (MIT) that exposes pipelines as MCP tools. Supports Claude, OpenAI, Gemini, and xAI. Pipelines connect via WebSocket and are made available to MCP clients like Claude Desktop, Cursor, and Windsurf.
README
<p align="center"> <img src="./images/header.jpeg" alt="Header"> </p>
<p align="center"> <a href="https://github.com/rocketride-org/rocketride-server/actions/workflows/ci.yml"><img src="https://github.com/rocketride-org/rocketride-server/actions/workflows/ci.yml/badge.svg" alt="CI"></a> <a href="https://opensource.org/licenses/MIT"><img src="https://img.shields.io/badge/License-MIT-yellow.svg" alt="License: MIT"></a> <a href="https://nodejs.org/"><img src="https://img.shields.io/badge/Node.js-18%2B-green.svg" alt="Node.js 18+"></a> <a href="https://discord.gg/9hr3tdZmEG"><img src="https://img.shields.io/badge/Discord-Join%20us-5865F2.svg" alt="Discord"></a> </p>
RocketRide is a high-performance data processing engine built on a C++ core with a Python-extensible node system. With 50+ pipeline nodes, native AI/ML support, and SDKs for TypeScript, Python, and MCP, it lets you process, transform, and analyze data at scale — entirely on your own infrastructure.
Key Capabilities
- Stay in your IDE — Build, debug, test, and scale heavy AI and data workloads with an intuitive visual builder in the environment you're used to. Stop using your browser.
- High-performance C++ engine — Native multithreading. No bottleneck. Purpose-built for throughput, not prototypes.
- Multi-agent workflows — Orchestrate and scale agents with built-in support for CrewAI and LangChain.
- 50+ pipeline nodes — Python-extensible, with 13 LLM providers, 8 vector databases, OCR, NER, PII anonymization, and more.
- TypeScript, Python & MCP SDKs — Integrate pipelines into native applications or expose them as tools for AI assistants.
- One-click deploy — Run on Docker, on-prem, or RocketRide Cloud (👀coming soon). Our architecture is made for production, not demos.
⚡ Quick Start
-
Install the extension for your IDE. Search for RocketRide in the extension marketplace:
<p align="center"> <img src="./images/install.png" alt="Install RocketRide extension"> </p>
<sub>Not seeing your IDE? Open an issue · Download directly</sub>
-
Click the RocketRide (🚀) extension in your IDE
-
Deploy a server — you'll be prompted on how you want to run the server. Choose the option that fits your setup:
- Local (Recommended) — This pulls the server directly into your IDE without any additional setup.
- On-Premises — Run the server on your own hardware for full control and data residency. Pull the image and deploy to Docker or clone this repo and build from source.
- RocketRide Cloud (👀coming soon) — Managed hosting with our proprietary model server. No infrastructure to maintain.
-
Create a
.pipefile and start building
🔧 Building your first pipe
-
All pipelines are recognized with the
*.pipeformat. Each pipeline and configuration is a JSON object - but the extension in your IDE will render within our visual builder canvas. -
All pipelines begin with source node: webhook, chat, or dropper. For specific usage, examples, and inspiration 💡 on how to build pipelines, check out our guides and documentation
-
Connect input lanes and output lanes by type to properly wire your pipeline. Some nodes like agents or LLMs can be invoked as tools for use by a parent node as shown below:
<p align="center"> <img src="./images/agent_pipeline.png" alt="Pipeline canvas example"> </p>
-
You can run a pipeline from the canvas by pressing the ▶️ button on the source node or from the
Connection Managerdirectly. -
View all available and running pipelines below the
Connection Manager. Selecting running pipelines allows for in depth analytics. Trace call trees, token usage, memory consumption, and more to optimize your pipelines before scaling and deploying. -
📦 Deploy your pipelines to RocketRide.ai cloud or run them on your own infrastructure.
-
Docker — Download the RocketRide server image and create a container. Requires Docker to be installed.
docker pull ghcr.io/rocketride-org/rocketride-engine:latest docker create --name rocketride-engine -p 5565:5565 ghcr.io/rocketride-org/rocketride-engine:latest -
RocketRide Cloud (👀coming soon) — Managed hosting with our proprietary model server and batched processing. The cheapest option to run AI workflows and pipelines at scale (seriously).
-
-
Run your pipelines as standalone processes or integrate them into your existing Python and TypeScript/JS applications utilizing our SDK.
-
Use it, commit it, ship it. 🚚
Useful Links
- 📚 Documentation
- 💬 Discord
- 🤝 Contributions
- 🔒 Security
- ⚖️ License
<p align="center">Made with ❤️ in 🌁 SF & 🇪🇺 EU</p>
推荐服务器
Baidu Map
百度地图核心API现已全面兼容MCP协议,是国内首家兼容MCP协议的地图服务商。
Playwright MCP Server
一个模型上下文协议服务器,它使大型语言模型能够通过结构化的可访问性快照与网页进行交互,而无需视觉模型或屏幕截图。
Magic Component Platform (MCP)
一个由人工智能驱动的工具,可以从自然语言描述生成现代化的用户界面组件,并与流行的集成开发环境(IDE)集成,从而简化用户界面开发流程。
Audiense Insights MCP Server
通过模型上下文协议启用与 Audiense Insights 账户的交互,从而促进营销洞察和受众数据的提取和分析,包括人口统计信息、行为和影响者互动。
VeyraX
一个单一的 MCP 工具,连接你所有喜爱的工具:Gmail、日历以及其他 40 多个工具。
graphlit-mcp-server
模型上下文协议 (MCP) 服务器实现了 MCP 客户端与 Graphlit 服务之间的集成。 除了网络爬取之外,还可以将任何内容(从 Slack 到 Gmail 再到播客订阅源)导入到 Graphlit 项目中,然后从 MCP 客户端检索相关内容。
Kagi MCP Server
一个 MCP 服务器,集成了 Kagi 搜索功能和 Claude AI,使 Claude 能够在回答需要最新信息的问题时执行实时网络搜索。
e2b-mcp-server
使用 MCP 通过 e2b 运行代码。
Neon MCP Server
用于与 Neon 管理 API 和数据库交互的 MCP 服务器
Exa MCP Server
模型上下文协议(MCP)服务器允许像 Claude 这样的 AI 助手使用 Exa AI 搜索 API 进行网络搜索。这种设置允许 AI 模型以安全和受控的方式获取实时的网络信息。