Claimify
Extracts verifiable, decontextualized factual claims from text using a research-based four-stage pipeline (sentence splitting, selection, disambiguation, and decomposition). Integrates with MCP clients to enable claim extraction and verification workflows.
README
Claimify: Research-Based Claim Extraction via MCP
An implementation of the "Claimify" methodology for factual claim extraction, delivered as a local Model Context Protocol (MCP) server. This tool implements the multi-stage claim extraction approach detailed in the academic paper "Towards Effective Extraction and Evaluation of Factual Claims" by Metropolitansky & Larson (2025).
Promps from the paper have been modified for use with Structured Outputs. THIS IS NOT AN OFFICIAL IMPLEMENTATION.
Overview
Claimify extracts verifiable, decontextualized factual claims from text using a sophisticated four-stage pipeline:
- Sentence Splitting: Breaks text into individual sentences with surrounding context
- Selection: Filters for sentences containing verifiable propositions, excluding opinions and speculation
- Disambiguation: Resolves ambiguities or discards sentences that cannot be clarified
- Decomposition: Breaks down sentences into atomic, self-contained factual claims
The tool uses OpenAI's structured outputs feature exclusively for improved reliability and exposes its functionality through the Model Context Protocol, making it available to MCP-compatible clients like Cursor and Claude Desktop.
Features
- Research-based methodology: Implements the peer-reviewed Claimify approach
- Structured outputs: Uses OpenAI's structured outputs for reliable, type-safe responses
- MCP integration: Seamlessly integrates with development environments
- Robust parsing: Handles various text formats including lists and paragraphs
- Context-aware: Uses surrounding sentences to resolve ambiguities
- Multi-language support: Preserves original language while extracting claims
- Resource storage: Automatically stores extracted claims as MCP resources for easy retrieval
- Comprehensive logging: Detailed logging of all LLM calls, responses, and pipeline stages
- Production-ready: Includes error handling, monitoring, and configuration management
Requirements
- OpenAI API: Requires an OpenAI API key (if your MCP host does not support sampling (Github Copilot in vsCode does))
- Compatible Model: Must use a model that supports structured outputs:
gpt-4o(recommended)gpt-4o-mini(faster and cheaper)
- Python 3.10+: For proper type hints and Pydantic support
Quick Start
1. Installation
# Clone the repository
git clone <repository-url>
cd ClaimsMCP
# Create and activate a virtual environment
python -m venv claimify-env
source claimify-env/bin/activate # On Windows: claimify-env\Scripts\activate
# Install dependencies
pip install -r requirements.txt
# Download required NLTK data (done automatically on first run)
python -c "import nltk; nltk.download('punkt_tab')"
2. Configuration
Create a .env file in the project root:
# Copy the example file
cp env.example .env
Edit .env and add your API key:
# API Keys
OPENAI_API_KEY="your-openai-api-key-here"
# LLM Configuration
LLM_MODEL="gpt-4o-2024-08-06" # Model that supports structured outputs
# Logging Configuration
LOG_LLM_CALLS="true" # Set to "false" to disable logging
LOG_OUTPUT="stderr" # "stderr" or "file" - where to send logs
LOG_FILE="claimify_llm.log" # Used only if LOG_OUTPUT="file"
MCP Client Configuration
For Cursor
- Open Cursor and navigate to Settings > MCP
- Click "Add a New Global MCP Server"
- Add the following configuration to your MCP settings file (usually
~/.cursor/mcp.json):
{
"mcpServers": {
"claimify-local": {
"command": "/path/to/your/claimify-env/bin/python",
"args": [
"/path/to/your/project/claimify_server.py"
]
}
}
}
- Replace the paths with the absolute paths to your Python executable and server script
The "Claimify Extraction Server" should now appear as a connected tool in your MCP-enabled chat.
Usage Examples
Once configured, you can use the tool in your MCP client:
Using the Extract Claims Tool
Using the Prompts
The server exposes two prompts to help verify and document extracted claims:
1. Verify Single Claim (verify_claim)
Provides a pre-built prompt that instructs the LLM to verify a single factual claim against external sources.
Arguments:
claim_text(required): The decontextualized factual claim to check.
Behavior:
- LLM is instructed to search authoritative sources (scholarly publications, reputable news outlets, official organizations).
- Returns one of three statuses:
- VERIFIED: Claim is clearly supported by reliable sources (provides at least 3 references with URLs + justification)
- UNCERTAIN: Claim may be correct but lacks precision, has ambiguity, or has limited/conflicting evidence
- DISPUTED: Claim is demonstrably false or contradicted by reliable sources
- Sources must not be fabricated; preference for primary references.
Example retrieval (conceptual – actual call depends on client API):
get_prompt(name="verify_claim", arguments={"claim_text": "Python was first publicly released in 1991."})
Example expected LLM response format:
**Claim:** Python was first publicly released in 1991.
**Status:** IN_PROGRESS
[After research...]
**Claim:** Python was first publicly released in 1991.
**Status:** VERIFIED
**Evidence:**
- Source 1: [Python.org Release History](https://www.python.org/doc/versions/) - Official Python release history page confirms the initial public release year
- Source 2: [Computer History Museum](https://www.computerhistory.org/collections/catalog/102726710) - Archive referencing Python's early development
- Source 3: [Wikipedia - Python](https://en.wikipedia.org/wiki/Python_(programming_language)) - Encyclopedia entry citing original release year
If uncertain:
**Claim:** Stockholm has 800,000 inhabitants.
**Status:** UNCERTAIN
**Evidence:**
- Source 1: [Statistics Sweden](https://www.scb.se/en/) - Reports varying population figures depending on whether measuring city proper, municipality, or metropolitan area
**Analysis:**
The claim lacks specificity about which definition of "Stockholm" is being referenced (city proper ~975k, municipality ~975k, or urban area ~1.6M as of 2023). The figure of 800,000 may have been accurate for certain definitions at specific time periods, but without temporal and geographic context, full verification is not possible.
If disputed:
**Claim:** The Earth is flat.
**Status:** DISPUTED
**Analysis:**
This claim contradicts overwhelming scientific evidence. The Earth's spherical shape has been confirmed by satellite imagery, space missions, and centuries of astronomical observations. Reliable sources universally reject this claim.
2. Create Claims Report (create_claims_report)
Generates an initial CLAIMS.md file with all claims marked as TODO. Claims can then be verified incrementally, updating their status through the workflow: TODO → IN_PROGRESS → VERIFIED/UNCERTAIN/DISPUTED.
Workflow:
- Initial Creation: All claims start with status TODO
- During Verification: Update individual claims to IN_PROGRESS
- After Verification: Update to VERIFIED, UNCERTAIN, or DISPUTED with evidence
Arguments:
- None (attach the extraction resource to the context in VS Code)
Behavior:
- Parses all claims from the extraction resource attached in the context
- Creates CLAIMS.md with all claims marked as TODO
- Provides a template structure ready for incremental verification
Example usage: When viewing an extraction resource in VS Code, attach it to the prompt context. The prompt will generate an initial CLAIMS.md file ready for verification.
Initial CLAIMS.md structure:
# Claims Report
**Extraction ID:** extraction_1_1730678400
**Generated:** 2025-11-03
**Total Claims:** 5
**TODO:** 5
**In Progress:** 0
**Verified:** 0
**Uncertain:** 0
**Disputed:** 0
---
## Claims
### Claim 1
**Text:** Apple Inc. was founded in 1976.
**Status:** TODO
---
### Claim 2
**Text:** Steve Jobs co-founded Apple Inc.
**Status:** TODO
---
...
After verification updates:
### Claim 1
**Text:** Apple Inc. was founded in 1976.
**Status:** VERIFIED
**Evidence:**
- Source 1: [Wikipedia - Apple Inc.](https://en.wikipedia.org/wiki/Apple_Inc.) - States company was founded in 1976
- Source 2: [Apple Official](https://www.apple.com/about/) - Corporate history confirms 1976 founding
- Source 3: [Britannica](https://www.britannica.com/topic/Apple-Inc) - Encyclopedia entry validates founding year
---
### Claim 2
**Text:** Stockholm has 800,000 inhabitants.
**Status:** UNCERTAIN
**Evidence:**
- Source 1: [Statistics Sweden](https://www.scb.se/en/) - Reports varying figures depending on definition
**Analysis:**
The claim lacks specificity about which geographic definition and time period. Population varies significantly between city proper (~975k), municipality (~975k), and metropolitan area (~1.6M) as of 2023. The 800k figure may have been accurate historically for certain definitions.
---
### Claim 3
**Text:** The company invented smartphones.
**Status:** DISPUTED
**Analysis:**
While Apple popularized smartphones with the iPhone in 2007, they did not invent smartphones. Earlier devices like the IBM Simon (1994) and BlackBerry devices (early 2000s) preceded the iPhone. The claim conflates innovation/popularization with invention.
---
...
Note: The server only supplies the prompts; external searching depends on the client/model capabilities.
Example 1: Simple Factual Text
Input: "The American flag contains 50 stars and 13 stripes."
Output: [
"The American flag contains 50 stars [representing the 50 states] and 13 stripes [representing the original 13 colonies].",
"The American flag was designed in 1777",
"The American flag has been modified 27 times"
]
Example 2: Mixed Fact and Opinion
Input: "Apple Inc. was founded in 1976 by Steve Jobs, Steve Wozniak, and Ronald Wayne. The company is incredibly innovative and has the best products in the world."
Output: [
"Apple Inc. was founded in 1976 by Steve Jobs, Steve Wozniak, and Ronald Wayne."
]
(Note: The subjective content about being "incredibly innovative" and having "the best products" is filtered out)
Example 3: Multi-language Support
Input: "String-systemet är en prisbelönt ikon som kombinerar elegant och minimalistisk design med ett brett utbud av färger och storlekar. Nisse Strinning skapade första hyllan redan 1949."
Output: [
"String-systemet [ett hyllsystem] är en prisbelönt ikon [inom design]",
"String-systemet kombinerar elegant och minimalistisk design med ett brett utbud av färger och storlekar",
"Nisse Strinning skapade den första String-hyllan [String-systemet] 1949"
]
(Note: Content preserved in original Swedish, with contextual clarifications added in brackets)
Accessing Extracted Claims as Resources
Each extraction generates two kinds of resources:
- Aggregate Extraction Resource (
claim://extraction_<n>_<timestamp>)- Contains metadata (timestamp, preview, question) and the full list of claims
- Returns JSON format
- Individual Claim Resources (
claim://<slug>)- Each claim is accessible via a unique slug (URL-safe identifier derived from claim text)
- Returns plain text (the claim itself)
Aggregate Extraction JSON Example
{
"id": "extraction_1_1730678400",
"timestamp": "2025-11-03T14:30:00.123456",
"question": "What is the history of Apple?",
"text_preview": "Apple Inc. was founded in 1976 by Steve Jobs...",
"claims": [
"Apple Inc. was founded in 1976 by Steve Jobs, Steve Wozniak, and Ronald Wayne."
],
"claim_count": 1
}
URI: claim://apple-inc-was-founded-in-1976-by-steve-jobs
Individual Claim Resource Examples
URI pattern: claim://<slug>
Examples:
claim://apple-inc-was-founded-in-1976-by-steve-jobs-steve-wozniak-and-ronaldclaim://stockholm-is-the-capital-of-swedenclaim://python-was-first-publicly-released-in-1991
Content: Plain text of the claim (no JSON wrapper)
Benefits of per-claim resources:
- Direct access: Retrieve any claim by its slug
- Simple format: Plain text, no parsing needed
- Unique identifiers: Each claim has a stable, readable URI
- Easy citation: Link directly to individual claims
Project Structure
ClaimsMCP/
├── README.md # This file
├── requirements.txt # Python dependencies
├── env.example # Environment configuration template
├── claimify_server.py # Main MCP server script
├── llm_client.py # LLM client with structured outputs support
├── pipeline.py # Core claim extraction pipeline
├── structured_models.py # Pydantic models for structured outputs
├── structured_prompts.py # Optimized prompts for structured outputs
├── setup.py # Package setup configuration
├── test_claimify.py # Test suite for the claim extraction pipeline
└── LICENSE # Apache 2.0 license
Architecture
The system follows a modular architecture with structured outputs:
- MCP Server: Exposes the claim extraction as a tool via the Model Context Protocol
- ClaimifyPipeline: Orchestrates the multi-stage extraction process using structured outputs
- LLMClient: Handles communication with OpenAI API using structured outputs and Pydantic models
- Structured Models: Pydantic models that define the expected response format for each stage
- Stage Functions: Individual functions for Selection, Disambiguation, and Decomposition
- Prompt Management: Simplified prompts optimized for structured outputs
Structured Outputs Benefits
The implementation uses OpenAI's structured outputs feature, which provides:
- Type Safety: Responses are automatically validated against Pydantic models
- Reliability: No more regex parsing failures or malformed JSON
- Explicit Refusals: Safety-based refusals are programmatically detectable
- Consistency: Guaranteed adherence to the expected response schema
- Performance: Reduced need for retry logic and error handling
Configuration Options
| Environment Variable | Description | Default | Options |
|---|---|---|---|
LLM_MODEL |
Specific model to use | gpt-4o-2024-08-06 |
Models supporting structured outputs |
OPENAI_API_KEY |
OpenAI API key | None | Your API key |
LOG_LLM_CALLS |
Enable detailed logging of all LLM interactions | true |
true, false |
LOG_OUTPUT |
Where to send log output | stderr |
stderr, file |
LOG_FILE |
Log file name (used when LOG_OUTPUT=file) | claimify_llm.log |
Any filename |
Troubleshooting
Common Issues
-
"Model does not support structured outputs" error
- Ensure you're using a compatible model:
gpt-4o-2024-08-06,gpt-4o-mini, orgpt-4o - Update your
.envfile:LLM_MODEL=gpt-4o-2024-08-06
- Ensure you're using a compatible model:
-
"API key not set" error
- Ensure your
.envfile exists and contains the correct OpenAI API key - Check that the key starts with
sk-
- Ensure your
-
"NLTK punkt tokenizer not found"
- Run:
python -c "import nltk; nltk.download('punkt_tab')"orpython -c "import nltk; nltk.download('punkt')"
- Run:
-
MCP client can't connect
- Check that the paths in your MCP configuration are absolute and correct
- Ensure your Python virtual environment is activated
- Verify the server script is executable:
chmod +x claimify_server.py
-
No claims extracted
- Check the logs for detailed information about each pipeline stage
- Ensure the input text contains verifiable factual statements
- Try with simpler, more direct factual sentences first
Development
To extend or modify the system:
- Adding new response fields: Update the Pydantic models in
structured_models.py - Modifying prompts: Edit the prompts in
structured_prompts.py - Adding new stages: Create new functions in
pipeline.pyfollowing the existing pattern - Testing: Use the built-in logging to debug pipeline behavior
The structured outputs approach makes the system much more reliable and easier to debug compared to traditional text parsing methods.
License
This project is licensed under the Apache License 2.0 - see the LICENSE file for details.
References
Metropolitansky & Larson (2025). "Towards Effective Extraction and Evaluation of Factual Claims"
Support
For issues related to:
- Setup and configuration: Check this README and the troubleshooting section
- MCP integration: Refer to the Model Context Protocol documentation
- Research methodology: Consult the original Claimify paper
推荐服务器
Baidu Map
百度地图核心API现已全面兼容MCP协议,是国内首家兼容MCP协议的地图服务商。
Playwright MCP Server
一个模型上下文协议服务器,它使大型语言模型能够通过结构化的可访问性快照与网页进行交互,而无需视觉模型或屏幕截图。
Magic Component Platform (MCP)
一个由人工智能驱动的工具,可以从自然语言描述生成现代化的用户界面组件,并与流行的集成开发环境(IDE)集成,从而简化用户界面开发流程。
Audiense Insights MCP Server
通过模型上下文协议启用与 Audiense Insights 账户的交互,从而促进营销洞察和受众数据的提取和分析,包括人口统计信息、行为和影响者互动。
VeyraX
一个单一的 MCP 工具,连接你所有喜爱的工具:Gmail、日历以及其他 40 多个工具。
graphlit-mcp-server
模型上下文协议 (MCP) 服务器实现了 MCP 客户端与 Graphlit 服务之间的集成。 除了网络爬取之外,还可以将任何内容(从 Slack 到 Gmail 再到播客订阅源)导入到 Graphlit 项目中,然后从 MCP 客户端检索相关内容。
Kagi MCP Server
一个 MCP 服务器,集成了 Kagi 搜索功能和 Claude AI,使 Claude 能够在回答需要最新信息的问题时执行实时网络搜索。
e2b-mcp-server
使用 MCP 通过 e2b 运行代码。
Neon MCP Server
用于与 Neon 管理 API 和数据库交互的 MCP 服务器
Exa MCP Server
模型上下文协议(MCP)服务器允许像 Claude 这样的 AI 助手使用 Exa AI 搜索 API 进行网络搜索。这种设置允许 AI 模型以安全和受控的方式获取实时的网络信息。