Fire in da houseTop Tip:Most people pay way too much for Perplexity, MidJourney, Runway, ChatGPT - but you can actually get all that for $15 instead of $225+ per month. And free to test!Fire in da houseCheck it out

deep-research-mcp

MCP.Pizza Chef: Ozamatash

The deep-research-mcp server is an AI-driven research assistant that integrates search engines, web scraping, and AI to perform deep, iterative exploration of any topic. It generates detailed, comprehensive reports by continuously refining and expanding research queries. Available as an MCP server or standalone CLI, it enables seamless integration with MCP hosts like Claude Desktop for enhanced research workflows.

Use This MCP server To

Generate comprehensive research reports on complex topics Iteratively refine research queries using AI and web data Integrate with MCP hosts for real-time research context Automate data gathering from search engines and web scraping Produce structured summaries from diverse online sources Support academic, market, and technical research workflows

README

Open Deep Research MCP Server

An AI-powered research assistant that performs deep, iterative research on any topic. It combines search engines, web scraping, and AI to explore topics in depth and generate comprehensive reports. Available as a Model Context Protocol (MCP) tool or standalone CLI. Look at exampleout.md to see what a report might look like.

Quick Start

  1. Clone and install:
git clone https://github.com/Ozamatash/deep-research
cd deep-research
npm install
  1. Set up environment in .env.local:
# Copy the example environment file
cp .env.example .env.local
  1. Build:
# Build the server
npm run build
  1. Run the cli version:
npm run start "Your research query here"
  1. Test MCP Server with Claude Desktop:
    Follow the guide thats at the bottom of server quickstart to add the server to Claude Desktop:
    https://modelcontextprotocol.io/quickstart/server

Features

  • Performs deep, iterative research by generating targeted search queries
  • Controls research scope with depth (how deep) and breadth (how wide) parameters
  • Evaluates source reliability with detailed scoring (0-1) and reasoning
  • Prioritizes high-reliability sources (≥0.7) and verifies less reliable information
  • Generates follow-up questions to better understand research needs
  • Produces detailed markdown reports with findings, sources, and reliability assessments
  • Available as a Model Context Protocol (MCP) tool for AI agents
  • For now MCP version doesn't ask follow up questions

How It Works

flowchart TB
    subgraph Input
        Q[User Query]
        B[Breadth Parameter]
        D[Depth Parameter]
        FQ[Feedback Questions]
    end

    subgraph Research[Deep Research]
        direction TB
        SQ[Generate SERP Queries]
        SR[Search]
        RE[Source Reliability Evaluation]
        PR[Process Results]
    end

    subgraph Results[Research Output]
        direction TB
        L((Learnings with
        Reliability Scores))
        SM((Source Metadata))
        ND((Next Directions:
        Prior Goals,
        New Questions))
    end

    %% Main Flow
    Q & FQ --> CQ[Combined Query]
    CQ & B & D --> SQ
    SQ --> SR
    SR --> RE
    RE --> PR

    %% Results Flow
    PR --> L
    PR --> SM
    PR --> ND

    %% Depth Decision and Recursion
    L & ND --> DP{depth > 0?}
    DP -->|Yes| SQ
    
    %% Final Output
    DP -->|No| MR[Markdown Report]

    %% Styling
    classDef input fill:#7bed9f,stroke:#2ed573,color:black
    classDef process fill:#70a1ff,stroke:#1e90ff,color:black
    classDef output fill:#ff4757,stroke:#ff6b81,color:black
    classDef results fill:#a8e6cf,stroke:#3b7a57,color:black,width:150px,height:150px

    class Q,B,D,FQ input
    class SQ,SR,RE,PR process
    class MR output
    class L,SM,ND results
Loading

Advanced Setup

Using Local Firecrawl (Free Option)

Instead of using the Firecrawl API, you can run a local instance. You can use the official repo or my fork which uses searXNG as the search backend to avoid using a searchapi key:

  1. Set up local Firecrawl:
git clone https://github.com/Ozamatash/localfirecrawl
cd localfirecrawl
# Follow setup in localfirecrawl README
  1. Update .env.local:
FIRECRAWL_BASE_URL="http://localhost:3002"

Optional: Observability

Add observability to track research flows, queries, and results using Langfuse:

# Add to .env.local
LANGFUSE_PUBLIC_KEY="your_langfuse_public_key"
LANGFUSE_SECRET_KEY="your_langfuse_secret_key"

The app works normally without observability if no Langfuse keys are provided.

License

MIT License

deep-research-mcp FAQ

How do I install the deep-research-mcp server?
Clone the GitHub repo, install dependencies with npm, set up your .env.local file, build the server, and run the CLI or integrate with an MCP host.
Can I use deep-research-mcp without an MCP host?
Yes, it can run as a standalone CLI tool for research queries.
How does deep-research-mcp perform iterative research?
It uses AI to generate and refine search queries, combining web scraping and search engine results to deepen research over multiple iterations.
What kind of reports does deep-research-mcp generate?
It produces comprehensive, structured reports that summarize findings from multiple sources.
Is deep-research-mcp compatible with popular LLM providers?
Yes, it works with models like OpenAI, Anthropic Claude, and Google Gemini for AI-driven research.
How do I integrate deep-research-mcp with Claude Desktop?
Follow the guide at https://modelcontextprotocol.io/quickstart/server to add the server to Claude Desktop.
What environment setup is required?
You need to copy the example environment file to .env.local and configure any necessary API keys or settings before building and running the server.
Can deep-research-mcp handle multiple research topics simultaneously?
It is designed to handle iterative research on individual queries; concurrent queries depend on host capabilities.