GitHub Analytics MCP Server

GitHub Analytics MCP Server

Enables querying and analysis of public GitHub repositories for statistics, contributor data, and commit history. It provides both a RESTful API and an MCP interface for seamless integration with AI agents.

Category
访问服务器

README

GitHub Analytics MCP Server — Architecture Reference Project

Query, analyze, and visualize any public GitHub repository — from the command line, browser, or AI agent.

Python Docker Kubernetes Terraform CI/CD License


Overview

GitHub Analytics MCP Server is a production-ready microservice that turns the GitHub API into a simple, self-hosted analytics endpoint. Point it at any public repository and instantly get structured data on stars, forks, contributors, commit history, and language distribution.

It exposes two interfaces: a RESTful API (FastAPI with auto-generated Swagger docs) for direct HTTP access, and a Model Context Protocol (MCP) server that lets AI agents like Claude Desktop query GitHub data as a native tool.

The entire stack — API gateway, MCP server, container orchestration, infrastructure provisioning, and CI/CD — is included and deployable with a single command.

This project also serves as an architecture reference implementation: every layer is accompanied by design-decision documentation explaining why it is structured this way, not just what it does.

Features

  • 🔍 Query any public GitHub repository by owner/name
  • 📊 Repository statistics — stars, forks, issues, watchers
  • 👥 Contributor analysis — top contributors with commit counts
  • 📝 Commit history — recent commits with author and message details
  • 🌐 RESTful API with auto-generated OpenAPI/Swagger docs
  • 🤖 MCP Protocol support for AI agent integration (Claude Desktop, etc.)
  • 🐳 Production-ready with Docker multi-stage builds and Docker Compose
  • ☸️ Kubernetes deployment with Deployments, Services, Ingress, and HPA
  • 📈 Auto-scaling — Horizontal Pod Autoscaler (2–5 replicas, 70% CPU target)
  • 🔄 Full CI/CD pipeline — lint, test, build, and deploy via GitHub Actions
  • 🏗️ Infrastructure as Code — Terraform provisions the entire K8s stack

Why This Project?

Concern This Project Traditional Approach
Setup docker-compose up or make k8s-deploy Manual server provisioning
Scalability Auto-scaling with K8s HPA (2–5 replicas) Manual capacity planning
Infrastructure terraform apply — one command Multiple manual steps
High Availability Multi-replica with health checks Complex setup required
Monitoring Liveness & readiness probes built in Separate monitoring stack
Deployment Automated CI/CD on every push Manual release process
Portability Runs anywhere Docker/K8s runs Environment-dependent
API Docs Auto-generated OpenAPI (Swagger UI) Manual documentation

This is not just a tool — it is a reference implementation designed for studying architecture patterns. Every layer includes design-decision documentation explaining the reasoning behind its structure.

Architecture

graph TB
    subgraph "User Interface"
        A[Web Browser / CLI]
    end

    subgraph "API Layer"
        B[FastAPI Gateway<br/>Port 8080]
        C[MCP Server<br/>stdio mode]
    end

    subgraph "Container Orchestration"
        D[Kubernetes Cluster]
        E[Docker Containers]
        F[Auto-scaling HPA]
    end

    subgraph "External Services"
        G[GitHub API]
    end

    subgraph "Infrastructure"
        H[Terraform IaC]
        I[CI/CD Pipeline]
    end

    A -->|HTTP/REST| B
    A -->|MCP Protocol| C
    B -->|GitHub Token| G
    C -->|GitHub Token| G
    B -.->|Deployed in| D
    C -.->|Deployed in| D
    D -->|Manages| E
    D -->|Auto-scales| F
    H -.->|Provisions| D
    I -.->|Deploys to| D

Design Philosophy

One domain, two interfaces, shared core

GitHubClient is the single business-logic layer. The MCP Server and FastAPI Gateway are both thin adapters — they translate between their respective protocols and the shared core. Neither contains business logic, and neither duplicates the other.

Why two interfaces: MCP serves AI agents over stdio; REST serves humans and programs over HTTP. Two protocols, two adapters, zero duplicated logic.

Error handling strategy

Custom exception hierarchy (RepositoryNotFoundError, AuthenticationError, RateLimitError) translates GitHub HTTP status codes into semantic domain errors. The MCP server converts these into user-friendly text messages; the FastAPI gateway converts them into the corresponding HTTP status codes (404/401/429/502). Callers never need to know how the GitHub API works internally.

Infrastructure: three layers for three use cases

  • Docker Compose — local development. One command (docker-compose up) starts everything.
  • Kubernetes manifests (k8s/) — direct kubectl apply. Good for learning K8s and quick testing.
  • Terraform (terraform/) — state management, drift detection, multi-environment support. For production.

All three coexist intentionally. Each serves a different stage of the deployment lifecycle.

Why these numbers

  • HPA 2-5 replicas: 2 guarantees availability (one pod can fail without downtime); 5 is a cost ceiling.
  • 70% CPU threshold: leaves 30% buffer so existing pods absorb traffic spikes while new pods start (10-30s scheduling window).
  • Resource limits (100m/500m CPU, 128Mi/256Mi memory): FastAPI + uvicorn idles at ~30m CPU / ~50MB RAM. Limits prevent a runaway process from starving other pods.

Deliberate omissions

  • No database — this is a stateless proxy. Every request fetches fresh data from GitHub. Adding a DB would obscure the core architecture pattern.
  • Redis is optional — available via docker-compose --profile with-cache up to demonstrate Docker Compose profiles, but not wired into the application.
  • No auth middleware — authentication is orthogonal to the architecture being demonstrated. Including it would distract from the layered design.

Architecture Documentation

For deeper dives into specific decisions:

Quick Start

Option 1: Docker Compose (Fastest)

# 1. Clone and configure
git clone https://github.com/Pyroxyl/github-analytics-mcp.git
cd github-analytics-mcp
cp .env.example .env
# Edit .env and add your GITHUB_TOKEN

# 2. Start services
docker-compose up -d

# 3. Test the API
curl http://localhost:8080/health
curl http://localhost:8080/api/v1/repo/facebook/react/stats | jq

Option 2: Kubernetes (Production)

# 1. Build and deploy
make build
make k8s-deploy

# 2. Access the API (LoadBalancer on port 80)
curl http://localhost/health
curl http://localhost/api/v1/repo/facebook/react/stats | jq

Option 3: Terraform (Full IaC)

cd terraform
cp terraform.tfvars.example terraform.tfvars
# Edit terraform.tfvars

terraform init
terraform plan
terraform apply

Usage Examples

Repository Statistics

curl "http://localhost/api/v1/repo/facebook/react/stats" | jq
{
  "repository": "facebook/react",
  "stars": 242591,
  "forks": 50472,
  "open_issues": 1138,
  "watchers": 6690,
  "description": "The library for web and native user interfaces.",
  "language": "JavaScript"
}

Recent Commits

curl "http://localhost/api/v1/repo/anthropics/anthropic-sdk-python/commits?limit=3" | jq

Top Contributors

curl "http://localhost/api/v1/repo/kubernetes/kubernetes/contributors?top_n=5" | jq

Language Distribution

curl "http://localhost/api/v1/repo/microsoft/vscode/languages" | jq
{
  "repository": "microsoft/vscode",
  "languages": {
    "TypeScript": 95.54,
    "CSS": 1.49,
    "JavaScript": 1.0,
    "Rust": 0.61
  }
}

Compare Projects

# Compare stars across projects
curl -s "http://localhost/api/v1/repo/facebook/react/stats" | jq '.stars'
curl -s "http://localhost/api/v1/repo/vuejs/vue/stats" | jq '.stars'

Interactive API Documentation

🌐 Live API Docs: http://localhost/docs (or http://localhost:8080/docs for Docker Compose)

FastAPI auto-generates interactive Swagger UI where you can:

  • 📖 Browse all available endpoints
  • 🎮 Test APIs directly in your browser with "Try it out"
  • 📊 View request/response schemas
  • 💡 See example values for all parameters
  • ✨ Execute real API calls and see live responses

MCP Client Configuration

Add to your MCP client configuration (e.g., Claude Desktop):

{
  "mcpServers": {
    "github-analytics": {
      "command": "python",
      "args": ["-m", "src.server"],
      "cwd": "/path/to/github-analytics-mcp",
      "env": {
        "GITHUB_TOKEN": "your_token_here"
      }
    }
  }
}

Or using Docker:

{
  "mcpServers": {
    "github-analytics": {
      "command": "docker",
      "args": ["run", "--rm", "-i", "--env-file", ".env", "github-analytics-mcp"],
      "cwd": "/path/to/github-analytics-mcp"
    }
  }
}

Tech Stack

Layer Technology
Backend Python 3.11+, FastAPI, PyGithub
Protocol Model Context Protocol (MCP)
Containerization Docker (multi-stage builds), Docker Compose
Orchestration Kubernetes — Deployments, Services, HPA, Ingress
Infrastructure Terraform
CI/CD GitHub Actions (lint → test → build → deploy)

DevOps Highlights

  • Multi-stage Docker builds for minimal image size
  • Kubernetes auto-scaling (2–5 replicas based on CPU)
  • Liveness & readiness probes for self-healing
  • Rolling updates with zero downtime
  • Automated lint, test, build, and deploy pipeline

Project Structure

github-analytics-mcp/
├── src/                        # MCP Server
│   ├── server.py               # MCP protocol entry point
│   ├── github_client.py        # GitHub API client wrapper
│   └── tools/                  # MCP tool implementations
│       ├── repo_stats.py       #   get_repo_stats
│       ├── commits.py          #   list_recent_commits
│       ├── contributors.py     #   analyze_contributors
│       └── languages.py        #   get_language_breakdown
├── api/                        # FastAPI Gateway
│   ├── main.py                 # App entry point
│   ├── routes.py               # API route definitions
│   ├── models.py               # Pydantic models
│   └── dependencies.py         # Dependency injection
├── k8s/                        # Kubernetes manifests
│   ├── namespace.yaml
│   ├── configmap.yaml
│   ├── secret.yaml
│   ├── deployment-api.yaml     # API gateway (2 replicas)
│   ├── deployment-mcp.yaml     # MCP server
│   ├── service-api.yaml        # LoadBalancer service
│   ├── hpa-api.yaml            # Horizontal Pod Autoscaler
│   ├── ingress.yaml
│   └── deploy.sh               # Deployment script
├── terraform/                  # Infrastructure as Code
│   ├── main.tf
│   ├── kubernetes.tf
│   ├── providers.tf
│   ├── variables.tf
│   └── outputs.tf
├── .github/workflows/          # CI/CD pipelines
│   ├── ci.yml                  # Lint & test
│   ├── docker-build.yml        # Build & push image
│   └── cd.yml                  # Deploy to K8s
├── tests/                      # Unit tests
├── Dockerfile                  # Multi-stage container build
├── docker-compose.yml          # Local multi-service setup
├── Makefile                    # Convenience commands
├── requirements.txt
└── .env.example                # Environment template

Development

Prerequisites

  • Python 3.11+
  • Docker & Docker Compose
  • kubectl (for Kubernetes deployment)
  • Terraform (for IaC deployment)
  • GitHub Personal Access Token (create one here)

Local Development

python3 -m venv venv
source venv/bin/activate
pip install -r requirements.txt

# Run the MCP server
python -m src.server

# Run the API gateway
uvicorn api.main:app --reload --port 8080

# Run tests
pytest tests/

Make Commands

Command Description
make build Build Docker image
make run Start with Docker Compose
make stop Stop all containers
make logs View container logs
make k8s-deploy Deploy to Kubernetes
make k8s-status Check K8s pod/service status
make clean Remove containers and images
make help Show all available commands

CI/CD Pipeline

Push/PR → [CI] Lint + Test → [Docker Build] → ghcr.io → [CD] → Kubernetes
  1. CI — Runs ruff lint and pytest on every push/PR (Python 3.11 & 3.12)
  2. Docker Build — Builds and pushes images to GitHub Container Registry
  3. CD — Deploys to Kubernetes via Terraform after successful build

See .github/workflows/README.md for details.

Production Deployment

High Availability

  • 2+ API gateway replicas with rolling updates
  • Automatic pod restart on failure via liveness probes
  • Readiness probes prevent traffic to unhealthy pods

Auto-Scaling

  • HPA scales from 2 to 5 replicas
  • Target: 70% CPU utilization
  • Handles traffic spikes automatically

Security

  • GitHub tokens stored as Kubernetes Secrets
  • No credentials in source code or git history
  • Ingress-ready for TLS termination

Use Cases

  • 📊 Project Evaluation — Quickly assess GitHub projects before adopting them
  • 🔍 Trend Research — Analyze language trends across popular repositories
  • 🤖 AI Integration — Enable AI agents to access GitHub data via MCP
  • 📈 Metrics Dashboards — Build custom dashboards with real-time GitHub stats
  • 🔬 Open Source Research — Study contributor patterns and project health

Roadmap

  • [ ] Redis caching layer for API responses
  • [ ] Prometheus metrics & Grafana dashboards
  • [ ] Rate limiting & API key authentication
  • [ ] Additional endpoints (pull requests, releases, workflows)
  • [ ] Multi-cloud examples (AWS EKS, GCP GKE, Azure AKS)

Contributing

See CONTRIBUTING.md for development workflow and guidelines.

License

This project is licensed under the MIT License — see the LICENSE file for details.

Acknowledgments

Built with Model Context Protocol by Anthropic, FastAPI, and PyGithub.


⭐ If you find this project useful, please star it on GitHub!

Issues · Project Link

推荐服务器

Baidu Map

Baidu Map

百度地图核心API现已全面兼容MCP协议,是国内首家兼容MCP协议的地图服务商。

官方
精选
JavaScript
Playwright MCP Server

Playwright MCP Server

一个模型上下文协议服务器,它使大型语言模型能够通过结构化的可访问性快照与网页进行交互,而无需视觉模型或屏幕截图。

官方
精选
TypeScript
Magic Component Platform (MCP)

Magic Component Platform (MCP)

一个由人工智能驱动的工具,可以从自然语言描述生成现代化的用户界面组件,并与流行的集成开发环境(IDE)集成,从而简化用户界面开发流程。

官方
精选
本地
TypeScript
Audiense Insights MCP Server

Audiense Insights MCP Server

通过模型上下文协议启用与 Audiense Insights 账户的交互,从而促进营销洞察和受众数据的提取和分析,包括人口统计信息、行为和影响者互动。

官方
精选
本地
TypeScript
VeyraX

VeyraX

一个单一的 MCP 工具,连接你所有喜爱的工具:Gmail、日历以及其他 40 多个工具。

官方
精选
本地
graphlit-mcp-server

graphlit-mcp-server

模型上下文协议 (MCP) 服务器实现了 MCP 客户端与 Graphlit 服务之间的集成。 除了网络爬取之外,还可以将任何内容(从 Slack 到 Gmail 再到播客订阅源)导入到 Graphlit 项目中,然后从 MCP 客户端检索相关内容。

官方
精选
TypeScript
Kagi MCP Server

Kagi MCP Server

一个 MCP 服务器,集成了 Kagi 搜索功能和 Claude AI,使 Claude 能够在回答需要最新信息的问题时执行实时网络搜索。

官方
精选
Python
e2b-mcp-server

e2b-mcp-server

使用 MCP 通过 e2b 运行代码。

官方
精选
Neon MCP Server

Neon MCP Server

用于与 Neon 管理 API 和数据库交互的 MCP 服务器

官方
精选
Exa MCP Server

Exa MCP Server

模型上下文协议(MCP)服务器允许像 Claude 这样的 AI 助手使用 Exa AI 搜索 API 进行网络搜索。这种设置允许 AI 模型以安全和受控的方式获取实时的网络信息。

官方
精选