LinkedIn MCP Server

LinkedIn MCP Server

Enables searching and scraping of LinkedIn for structured data on people, companies, and job listings. It allows AI clients to retrieve detailed profiles, experience, and activity sections using browser automation.

Category
访问服务器

README

LinkedIn MCP Server

A Model Context Protocol (MCP) server for LinkedIn. Search people, companies, and jobs, scrape profiles, and retrieve structured JSON data from any MCP-compatible AI client.

Built with FastMCP, Patchright, and a clean hexagonal architecture.


Features

Category Tools
People get_person_profile · search_people
Companies get_company_profile · get_company_posts
Jobs get_job_details · search_jobs
Browser close_browser

Person Profile Sections

The get_person_profile tool supports granular section scraping. Request only the sections you need:

  • Main profile (always included) — name, headline, location, followers, connections, about, profile image
  • Experience — title, company, dates, duration, description, company logo
  • Education — school, degree, dates, description, school logo
  • Contact info — email, phone, websites, birthday, LinkedIn URL
  • Interests — people, companies, and groups followed
  • Honors and awards — title, issuer, description
  • Languages — language name and proficiency level
  • Posts — recent activity with reactions and timestamps
  • Recommendations — received and given, with author details

Company Profile Sections

  • About (always included) — overview, website, industry, size, headquarters, specialties, logo
  • Posts — recent feed posts with engagement metrics
  • Jobs — current open positions

Job Search Filters

The search_jobs tool supports the following filters:

Filter Values
date_posted past_hour, past_24_hours, past_week, past_month
job_type full_time, part_time, contract, temporary, internship, other
experience_level entry, associate, mid_senior, director, executive
work_type on_site, remote, hybrid
easy_apply true / false
sort_by date, relevance

Prerequisites

  • Python 3.12 or later
  • uv package manager
  • A LinkedIn account for authentication

Quick Start

1. Clone and install

git clone https://github.com/eliasbiondo/linkedin-mcp-server.git
cd linkedin-mcp-server
uv sync

2. Authenticate with LinkedIn

uv run linkedin-mcp-server --login

A browser window will open. Log in to LinkedIn and the session will be persisted locally at ~/.linkedin-mcp-server/browser-data.

3. Run the server

stdio transport (default — for Claude Desktop, Cursor, and similar clients):

uv run linkedin-mcp-server

HTTP transport (for remote clients, the MCP Inspector, etc.):

uv run linkedin-mcp-server --transport streamable-http --host 0.0.0.0 --port 8000

Client Integration

Claude Desktop / Cursor

Add to your MCP configuration file:

{
  "mcpServers": {
    "linkedin": {
      "command": "uv",
      "args": [
        "--directory", "/path/to/linkedin-mcp-server",
        "run", "linkedin-mcp-server"
      ]
    }
  }
}

MCP Inspector

npx @modelcontextprotocol/inspector

Then connect to http://localhost:8000/mcp if using HTTP transport.


Configuration

Configuration follows a strict precedence chain: CLI args > environment variables > .env file > defaults.

CLI Arguments

Argument Description Default
--transport stdio or streamable-http stdio
--host Host for HTTP transport 127.0.0.1
--port Port for HTTP transport 8000
--log-level DEBUG, INFO, WARNING, ERROR WARNING
--login Open browser for LinkedIn login
--logout Clear stored credentials
--status Check session status

Environment Variables

Create a .env file in the project root:

# Server
LINKEDIN_TRANSPORT=stdio
LINKEDIN_HOST=127.0.0.1
LINKEDIN_PORT=8000
LINKEDIN_LOG_LEVEL=WARNING

# Browser
LINKEDIN_HEADLESS=true
LINKEDIN_SLOW_MO=0
LINKEDIN_TIMEOUT=5000
LINKEDIN_VIEWPORT_WIDTH=1280
LINKEDIN_VIEWPORT_HEIGHT=720
LINKEDIN_CHROME_PATH=
LINKEDIN_USER_AGENT=
LINKEDIN_USER_DATA_DIR=~/.linkedin-mcp-server/browser-data

Architecture

The project follows a hexagonal (ports and adapters) architecture with strict layer separation:

src/linkedin_mcp_server/
├── domain/              # Core business logic — zero external dependencies
│   ├── models/          # Data models (Person, Company, Job, Search)
│   ├── parsers/         # HTML to structured data parsers
│   ├── exceptions.py    # Domain exceptions
│   └── value_objects.py # Immutable configuration and content objects
├── ports/               # Abstract interfaces
│   ├── auth.py          # Authentication port
│   ├── browser.py       # Browser automation port
│   └── config.py        # Configuration port
├── application/         # Use cases — orchestration layer
│   ├── scrape_person.py
│   ├── scrape_company.py
│   ├── scrape_job.py
│   ├── search_people.py
│   ├── search_jobs.py
│   └── manage_session.py
├── adapters/            # Concrete implementations
│   ├── driven/          # Infrastructure adapters (browser, auth, config)
│   └── driving/         # Interface adapters (CLI, MCP tools, serialization)
└── container.py         # Dependency injection composition root

Design Decisions

  • Ports and adapters — Domain logic is fully decoupled from infrastructure. The browser engine, MCP framework, and configuration source can all be swapped independently.
  • Dependency injection — A single Container class acts as the composition root and is the only place that imports concrete adapter classes.
  • Structured JSON output — LinkedIn HTML is parsed into typed Python dataclasses, then serialized to JSON for reliable LLM consumption.
  • Session persistence — Browser state is saved to disk, so authentication is required only once.

Development

Setup

uv sync --group dev
uv run pre-commit install

Running tests

uv run pytest

With coverage:

uv run pytest --cov=linkedin_mcp_server

Linting and formatting

This project uses Ruff for both linting and formatting. Pre-commit hooks will run these automatically on each commit.

# Lint
uv run ruff check .

# Lint and auto-fix
uv run ruff check . --fix

# Format
uv run ruff format .

License

This project is licensed under the MIT License. See the LICENSE file for details.


Contributing

Contributions are welcome. Please read the contributing guide for details on the development workflow and submission process.


Disclaimer

This tool is intended for personal and educational use. Scraping LinkedIn may violate their Terms of Service. Use responsibly and at your own risk. The authors are not responsible for any misuse or consequences arising from the use of this software.

推荐服务器

Baidu Map

Baidu Map

百度地图核心API现已全面兼容MCP协议,是国内首家兼容MCP协议的地图服务商。

官方
精选
JavaScript
Playwright MCP Server

Playwright MCP Server

一个模型上下文协议服务器,它使大型语言模型能够通过结构化的可访问性快照与网页进行交互,而无需视觉模型或屏幕截图。

官方
精选
TypeScript
Magic Component Platform (MCP)

Magic Component Platform (MCP)

一个由人工智能驱动的工具,可以从自然语言描述生成现代化的用户界面组件,并与流行的集成开发环境(IDE)集成,从而简化用户界面开发流程。

官方
精选
本地
TypeScript
Audiense Insights MCP Server

Audiense Insights MCP Server

通过模型上下文协议启用与 Audiense Insights 账户的交互,从而促进营销洞察和受众数据的提取和分析,包括人口统计信息、行为和影响者互动。

官方
精选
本地
TypeScript
VeyraX

VeyraX

一个单一的 MCP 工具,连接你所有喜爱的工具:Gmail、日历以及其他 40 多个工具。

官方
精选
本地
graphlit-mcp-server

graphlit-mcp-server

模型上下文协议 (MCP) 服务器实现了 MCP 客户端与 Graphlit 服务之间的集成。 除了网络爬取之外,还可以将任何内容(从 Slack 到 Gmail 再到播客订阅源)导入到 Graphlit 项目中,然后从 MCP 客户端检索相关内容。

官方
精选
TypeScript
Kagi MCP Server

Kagi MCP Server

一个 MCP 服务器,集成了 Kagi 搜索功能和 Claude AI,使 Claude 能够在回答需要最新信息的问题时执行实时网络搜索。

官方
精选
Python
e2b-mcp-server

e2b-mcp-server

使用 MCP 通过 e2b 运行代码。

官方
精选
Neon MCP Server

Neon MCP Server

用于与 Neon 管理 API 和数据库交互的 MCP 服务器

官方
精选
Exa MCP Server

Exa MCP Server

模型上下文协议(MCP)服务器允许像 Claude 这样的 AI 助手使用 Exa AI 搜索 API 进行网络搜索。这种设置允许 AI 模型以安全和受控的方式获取实时的网络信息。

官方
精选