marm-mcp

marm-mcp

MARM MCP provides persistent memory and structured session context beneath any AI tool, so your agents learn, remember, and collaborate across all your workflows.

Category
访问服务器

README

<div align="center"> <picture> <img src="https://raw.githubusercontent.com/Lyellr88/MARM-Systems/MARM-main/media/marm-logo.svg" alt="MARM - The AI That Remembers Your Conversations" width="700" height="350"> </picture> <h1 align="center">MARM: The AI That Remembers Your Conversations</h1>

Memory Accurate Response Mode v2.2.5 - The intelligent persistent memory system for AI agents, stop fighting your memory and control it. Experience long-term recall, session continuity, and reliable conversation history, so your LLMs never lose track of what matters.

GitHub stars GitHub forks License Python FastAPI Docker Pulls PyPI Downloads

pip install MCP Registry

Official MARM

Note: This is the official MARM repository. All official versions and releases are managed here.

Forks may experiment, but official updates will always come from this repo.

</div>


Why MARM MCP: The Problem & Solution

Your AI forgets everything. MARM MCP doesn't.

Modern LLMs lose context over time, repeat prior ideas, and drift off requirements. MARM MCP solves this with a unified, persistent, MCP‑native memory layer that sits beneath any AI client you use. It blends semantic search, structured session logs, reusable notebooks, and smart summaries so your agents can remember, reference, and build on prior work—consistently, across sessions, and across tools.

MCP in One Sentence: MARM MCP provides persistent memory and structured session context beneath any AI tool, so your agents learn, remember, and collaborate across all your workflows.

The Problem → The MARM Solution

  • Problem: Conversations reset; decisions get lost; work scatters across multiple AI tools.
  • Solution: A universal, persistent memory layer that captures and classifies the important bits (decisions, configs, code, rationale), then recalls them by meaning—not keywords.

Before vs After

  • Without MARM: lost context, repeated suggestions, drifting scope, "start from scratch."
  • With MARM: session memory, cross-session continuity, concrete recall of decisions, and faster, more accurate delivery.

<br> <div align="center"> <picture> <img src="https://raw.githubusercontent.com/Lyellr88/MARM-Systems/MARM-main/media/google-overview.png" alt="MARM appears in Google AI Overview for AI memory protocol queries" width="900" height="550" /> </picture> </div> <p align="center"><i>Appears in Google AI Overview for AI memory protocol queries (as of Aug 2025)</i></p>

What MARM MCP Delivers

Memory Multi-AI Architecture
Semantic Search - Find by meaning using AI embeddings Unified Memory Layer - Works with Claude, Qwen, Gemini, MCP clients 18 Complete MCP Tools - Full Model Context Protocol coverage
Auto-Classification - Content categorized (code, project, book, general) Cross-Platform Intelligence - Different AIs learn from shared knowledge Database Optimization - SQLite with WAL mode and connection pooling
Persistent Cross-Session Memory - Memories survive across agent conversations User-Controlled Memory - "Bring Your Own History," granular control Rate Limiting - IP-based tiers for stability
Smart Recall - Vector similarity search with context-aware fallbacks MCP Compliance - Response size management for predictable performance
Docker Ready - Containerized deployment with health/readiness checks

Learn More


What Users Are Saying

“MARM successfully handles our industrial automation workflows in production. We've validated session management, persistent logging, and smart recall across container restarts in our Windows 11 + Docker environment. The system reliably tracks complex technical decisions and maintains data integrity through deployment cycles.”
@Ophy21, GitHub user (Industrial Automation Engineer)

“MARM proved exceptionally valuable for DevOps and complex Docker projects. It maintained 100% memory accuracy, preserved context on 46 services and network configurations, and enabled standards-compliant Python/Terraform work. Semantic search and automated session logs made solving async and infrastructure issues far easier. Value Rating: 9.5/10 - indispensable for enterprise-grade memory, technical standards, and long-session code management.”
@joe_nyc, Discord user (DevOps/Infrastructure Engineer)


🚀 Quick Start for MCP

<br> <div align="center"> <picture> <img src="https://raw.githubusercontent.com/Lyellr88/MARM-Systems/MARM-main/media/installation-flow.svg" width="850" height="500" </picture> </div> <br>

Docker (Fastest - 30 seconds):

docker pull lyellr88/marm-mcp-server:latest
docker run -d --name marm-mcp-server -p 8001:8001 -v marm_data:/app/data lyellr88/marm-mcp-server:latest
claude mcp add --transport http marm-memory http://localhost:8001/mcp

Quick Local Install:

pip install marm-mcp-server==2.2.5
marm-mcp-server
claude mcp add --transport http marm-memory http://localhost:8001/mcp

Key Information:

  • Server Endpoint: http://localhost:8001/mcp
  • API Documentation: http://localhost:8001/docs
  • Supported Clients: Claude Code, Qwen CLI, Gemini CLI, and any MCP-compatible LLM client or LLM platform

All Installation Options:

  • Docker (Fastest): One command, works everywhere
  • Automated Setup: One command with dependency validation
  • Manual Installation: Step-by-step with virtual environment
  • Quick Test: Zero-configuration trial run

Choose your installation method:

Installation Type Guide Best For
Docker INSTALL-DOCKER.md Cross-platform, production deployment
Windows INSTALL-WINDOWS.md Native Windows development
Linux INSTALL-LINUX.md Native Linux development
Platforms INSTALL-PLATFORM.md App & API integration

🛠️ MARM MCP Server Guide

Now that you understand the ecosystem, here's info and how to use the MCP server with your AI agents


<div align="center"> <picture> <img src="https://raw.githubusercontent.com/Lyellr88/MARM-Systems/MARM-main/media/feature-showcase.svg" height="550"
width="800"
</picture> </div>


🛠️ Complete MCP Tool Suite (18 Tools)

💡 Pro Tip: You don't need to manually call these tools! Just tell your AI agent what you want in natural language:

  • "Claude, log this session as 'Project Alpha' and add this conversation as 'database design discussion'"
  • "Remember this code snippet in your notebook for later"
  • "Search for what we discussed about authentication yesterday"

The AI agent will automatically use the appropriate tools. Manual tool access is available for power users who want direct control.

Category Tool Description
🧠 Memory Intelligence marm_smart_recall AI-powered semantic similarity search across all memories. Supports global search with search_all=True flag
marm_contextual_log Intelligent auto-classifying memory storage using vector embeddings
🚀 Session Management marm_start Activate MARM intelligent memory and response accuracy layers
marm_refresh Refresh AI agent session state and reaffirm protocol adherence
📚 Logging System marm_log_session Create or switch to named session container
marm_log_entry Add structured log entry with auto-date formatting
marm_log_show Display all entries and sessions (filterable)
marm_log_delete Delete specified session or individual entries
🔄 Reasoning & Workflow marm_summary Generate context-aware summaries with intelligent truncation for LLM conversations
marm_context_bridge Smart context bridging for seamless AI agent workflow transitions
📔 Notebook Management marm_notebook_add Add new notebook entry with semantic embeddings
marm_notebook_use Activate entries as instructions (comma-separated)
marm_notebook_show Display all saved keys and summaries
marm_notebook_delete Delete specific notebook entry
marm_notebook_clear Clear the active instruction list
marm_notebook_status Show current active instruction list
⚙️ System Utilities marm_current_context Background Tool - Automatically provides current date/time for log entries (AI agents use automatically)
marm_system_info Comprehensive system information, health status, and loaded docs
marm_reload_docs Reload documentation into memory system

🏗️ Architecture Overview

Core Technology Stack

FastAPI (0.115.4) + FastAPI-MCP (0.4.0)
├── SQLite with WAL Mode + Custom Connection Pooling  
├── Sentence Transformers (all-MiniLM-L6-v2) + Semantic Search
├── Structured Logging (structlog) + Memory Monitoring (psutil)
├── IP-Based Rate Limiting + Usage Analytics
├── MCP Response Size Compliance (1MB limit)
├── Event-Driven Automation System
├── Docker Containerized Deployment + Health Monitoring
└── Advanced Memory Intelligence + Auto-Classification

Database Schema (5 Tables)

memories - Core Memory Storage

CREATE TABLE memories (
    id TEXT PRIMARY KEY,
    session_name TEXT NOT NULL,
    content TEXT NOT NULL,
    embedding BLOB,              -- AI vector embeddings for semantic search
    timestamp TEXT NOT NULL,
    context_type TEXT DEFAULT 'general',  -- Auto-classified content type
    metadata TEXT DEFAULT '{}',
    created_at TEXT DEFAULT CURRENT_TIMESTAMP
);

sessions - Session Management

CREATE TABLE sessions (
    session_name TEXT PRIMARY KEY,
    marm_active BOOLEAN DEFAULT FALSE,
    created_at TEXT DEFAULT CURRENT_TIMESTAMP,
    last_accessed TEXT DEFAULT CURRENT_TIMESTAMP,
    metadata TEXT DEFAULT '{}'
);

Plus: log_entries, notebook_entries, user_settings


<div align="center"> <picture> <img src="https://raw.githubusercontent.com/Lyellr88/MARM-Systems/MARM-main/media/memory-intelligence.svg" width="900" height="625" </picture> </div>


📈 Performance & Scalability

Production Optimizations

  • Custom SQLite Connection Pool: Thread-safe with configurable limits (default: 5)
  • WAL Mode: Write-Ahead Logging for concurrent access performance
  • Lazy Loading: Semantic models loaded only when needed (resource efficient)
  • Intelligent Caching: Memory usage optimization with cleanup cycles
  • Response Size Management: MCP 1MB compliance with smart truncation

Rate Limiting Tiers

  • Default: 60 requests/minute, 5min cooldown
  • Memory Heavy: 20 requests/minute, 10min cooldown (semantic search)
  • Search Operations: 30 requests/minute, 5min cooldown

📚 Documentation for MCP

Guide Type Document Description
Docker Setup INSTALL-DOCKER.md Cross-platform, production deployment
Windows Setup INSTALL-WINDOWS.md Native Windows development
Linux Setup INSTALL-LINUX.md Native Linux development
Platform Integration INSTALL-PLATFORM.md App & API integration
MCP Handbook MCP-HANDBOOK.md Complete usage guide with all 18 MCP tools, cross-app memory strategies, pro tips, and FAQ

🆚 Competitive Advantage

vs. Basic MCP Implementations

Feature MARM v2.2.5 Basic MCP Servers
Memory Intelligence AI-powered semantic search with auto-classification Basic key-value storage
Tool Coverage 18 complete MCP protocol tools 3-5 basic wrappers
Scalability Database optimization + connection pooling Single connection
MCP Compliance 1MB response size management No size controls
Deployment Docker containerization + health monitoring Local development only
Analytics Usage tracking + business intelligence No tracking
Codebase Maturity 2,500+ lines professional code 200-800 lines

🤝 Contributing

Aren't you sick of explaining every project you're working on to every LLM you work with?

MARM is building the solution to this. Support now to join a growing ecosystem - this is just Phase 1 of a 3-part roadmap and our next build will complement MARM like peanut butter and jelly.

Join the repo that's working to give YOU control over what is remembered and how it's remembered.

Why Contribute Now?

  • Ground floor opportunity - Be part of the MCP memory revolution from the beginning
  • Real impact - Your contributions directly solve problems you face daily with AI agents
  • Growing ecosystem - Help build the infrastructure that will power tomorrow's AI workflows
  • Phase 1 complete - Proven foundation ready for the next breakthrough features

Development Priorities

  1. Load Testing: Validate deployment performance under real AI workloads
  2. Documentation: Expand API documentation and LLM integration guides
  3. Performance: AI model caching and memory optimization
  4. Features: Additional MCP protocol tools and multi-tenant capabilities

Join the MARM Community

Help build the future of AI memory - no coding required!

Connect: MARM Discord | GitHub Discussions

Easy Ways to Get Involved

  • Try the MCP server or Coming soon CLI and share your experience
  • Star the repo if MARM solves a problem for you
  • Share on social - help others discover memory-enhanced AI
  • Open issues with bugs, feature requests, or use cases
  • Join discussions about AI reliability and memory

For Developers

  • Build integrations - MCP tools, browser extensions, API wrappers
  • Enhance the memory system - improve semantic search and storage
  • Expand platform support - new deployment targets and integrations
  • Submit Pull Requests - Every PR helps MARM grow. Big or small, I review each with respect and openness to see how it can improve the project

⭐ Star the Project

If MARM helps with your AI memory needs, please star the repository to support development!


<div align="center">

Star History Chart </div>


License & Usage Notice

This project is licensed under the MIT License. Forks and derivative works are permitted.

However, use of the MARM name and version numbering is reserved for releases from the official MARM repository.

Derivatives should clearly indicate they are unofficial or experimental.


📁 Project Documentation

Usage Guides

  • MARM-HANDBOOK.md - Original MARM protocol handbook for chatbot usage
  • MCP-HANDBOOK.md - Complete MCP server usage guide with commands, workflows, and examples
  • PROTOCOL.md - Quick start commands and protocol reference
  • FAQ.md - Answers to common questions about using MARM

MCP Server Installation

Chatbot Installation

Project Information


mcp-name: io.github.Lyellr88/marm-mcp-server

Built with ❤️ by MARM Systems - Universal MCP memory intelligence

推荐服务器

Baidu Map

Baidu Map

百度地图核心API现已全面兼容MCP协议,是国内首家兼容MCP协议的地图服务商。

官方
精选
JavaScript
Playwright MCP Server

Playwright MCP Server

一个模型上下文协议服务器,它使大型语言模型能够通过结构化的可访问性快照与网页进行交互,而无需视觉模型或屏幕截图。

官方
精选
TypeScript
Audiense Insights MCP Server

Audiense Insights MCP Server

通过模型上下文协议启用与 Audiense Insights 账户的交互,从而促进营销洞察和受众数据的提取和分析,包括人口统计信息、行为和影响者互动。

官方
精选
本地
TypeScript
Magic Component Platform (MCP)

Magic Component Platform (MCP)

一个由人工智能驱动的工具,可以从自然语言描述生成现代化的用户界面组件,并与流行的集成开发环境(IDE)集成,从而简化用户界面开发流程。

官方
精选
本地
TypeScript
VeyraX

VeyraX

一个单一的 MCP 工具,连接你所有喜爱的工具:Gmail、日历以及其他 40 多个工具。

官方
精选
本地
Kagi MCP Server

Kagi MCP Server

一个 MCP 服务器,集成了 Kagi 搜索功能和 Claude AI,使 Claude 能够在回答需要最新信息的问题时执行实时网络搜索。

官方
精选
Python
graphlit-mcp-server

graphlit-mcp-server

模型上下文协议 (MCP) 服务器实现了 MCP 客户端与 Graphlit 服务之间的集成。 除了网络爬取之外,还可以将任何内容(从 Slack 到 Gmail 再到播客订阅源)导入到 Graphlit 项目中,然后从 MCP 客户端检索相关内容。

官方
精选
TypeScript
Neon MCP Server

Neon MCP Server

用于与 Neon 管理 API 和数据库交互的 MCP 服务器

官方
精选
Exa MCP Server

Exa MCP Server

模型上下文协议(MCP)服务器允许像 Claude 这样的 AI 助手使用 Exa AI 搜索 API 进行网络搜索。这种设置允许 AI 模型以安全和受控的方式获取实时的网络信息。

官方
精选
mcp-server-qdrant

mcp-server-qdrant

这个仓库展示了如何为向量搜索引擎 Qdrant 创建一个 MCP (Managed Control Plane) 服务器的示例。

官方
精选