Fire in da houseTop Tip:Paying $100+ per month for Perplexity, MidJourney, Runway, ChatGPT and other tools is crazy - get all your AI tools in one site starting at $15 per month with Galaxy AI Fire in da houseCheck it out free

mcp-typescribe

MCP.Pizza Chef: yWorks

mcp-typescribe is an MCP server that bridges the gap between LLMs and new or internal APIs by providing structured API information. It overcomes the limitation of LLMs only knowing pre-trained APIs by efficiently exposing API details without overwhelming context, enabling LLMs to interact with large and evolving codebases and SDKs effectively.

Use This MCP server To

Expose new or internal APIs to LLMs for real-time usage Enable LLMs to understand large SDKs without full documentation context Improve LLM code generation with up-to-date API knowledge Facilitate integration of proprietary APIs into AI workflows Support dynamic API discovery for AI-assisted development tools

README

npm version

MCP-Typescribe - an MCP Server providing LLMs API information

The Problem

Large Language Models (LLMs) have made incredible strides in code generation and developer productivity. However, they face a key limitation: they can only reliably use APIs and libraries they’ve seen during training. This creates a bottleneck for adopting new tools, SDKs, or internal APIs — LLMs simply don’t know how to use them effectively.

While tools can be given source code access (when interacting with APIs for which the sources are available) or access to documentation files (e.g. typescript type definition files), this doesn't scale well for large APIs. LLMs need a more efficient way to learn more about an API. Putting all the documentation into context for every request is inefficient, unfeasible, and leads to poor results.

As a result:

Larger new or internal APIs remain "invisible" to LLMs.

Developers must manually guide LLMs or provide example usage.

Innovation is slowed by the lag between an API’s release and its widespread understanding by AI tools.

The Idea

This project is an open-source implementation of the Model Context Protocol (MCP)—a protocol designed to provide LLMs with contextual, real-time access to information. In this case it's the API documentation, and particularly for now in this project TypeScript definitions.

Our goal is to:

Parse TypeScript (and other) definitions into a machine-readable format.

Serve this context dynamically to LLMs through tools like Claude, Cline, Cursor, or Windsurf and other custom interfaces.

Enable agentic behavior by letting LLMs query, plan, and adapt to unfamiliar APIs without retraining.

What This Enables

Plug-and-play API support for LLM-based coding assistants.

Faster onboarding for new or proprietary SDKs.

A step toward more autonomous, context-aware coding agents.

Project Overview

Image

This project provides a way for AI agents to efficiently explore and understand unknown TypeScript APIs. It loads TypeDoc-generated JSON documentation and exposes it through a set of query endpoints that allow agents to search for symbols, get detailed information about specific parts of the API, and understand relationships between different components.

Current Features

  • TypeDoc Integration: Loads and indexes TypeDoc JSON documentation for efficient querying
  • Comprehensive Query Capabilities: Provides a wide range of tools for exploring TypeScript APIs
  • MCP Protocol: Follows the Model Context Protocol for seamless integration with AI agents

Query Capabilities

The server provides the following tools for querying the API:

  • search_symbols: Find symbols by name with optional filtering by kind
  • get_symbol_details: Get detailed information about a specific symbol
  • list_members: List methods and properties of a class or interface
  • get_parameter_info: Get information about function parameters
  • find_implementations: Find implementations of interfaces or subclasses
  • search_by_return_type: Find functions returning a specific type
  • search_by_description: Search in JSDoc comments
  • get_type_hierarchy: Show inheritance relationships
  • find_usages: Find where a type/function is used

Getting Started

Prerequisites

  • Node.js
  • npm

Installation

  1. Clone the repository
  2. Install dependencies:
    npm install

Usage

  1. Generate TypeDoc JSON for your TypeScript API:

    npx typedoc --json docs/api.json --entryPointStrategy expand path/to/your/typescript/files

    If you (only) have an existing.d.ts file, you can create an api json file like so:

    Create a separate tsconfig.docs.json:

    {
      "extends": "./tsconfig.json",
      "files": ["existing.d.ts"],
      "typedocOptions": {
        "entryPoints": ["existing.d.ts"],
        "json": "docs/api.json",
        "pretty": false
      }
    }

    Then do

    npx typedoc --tsconfig tsconfig.docs.json
  2. Build the project:

    npm run build
  3. Explore the MCP server:

    npx @modelcontextprotocol/inspector node ./dist/mcp-server/cli.js run-server docs/api.json
  4. Connect an AI agent to the server to query the API

    E.g. with cline in VSCode, specify the following MCP server in cline_mcp_settings.json:

    {
      "mcpServers": {
        "typescribe": {
          "command": "npx",
          "args": [
            "-y",
            "mcp-typescribe@latest",
            "run-server",
            "<PATH_TO_API_DOT_JSON>"
          ],
          "env": {}
        }
      }
    }
  5. Enable the server and likely auto-approve the various tools. Tell the agent to use the "typescribe" tool to learn about your API.

Project Structure

  • src/sample-api/: A sample TypeScript API for testing - it uses a weird German-like dialect for the API names to test that the LLM does not hallucinate the API
  • src/mcp-server/: The MCP server implementation
    • utils/: Utility functions
    • schemas/: JSON schemas for the MCP tools
    • core/: Core functionality
    • server.ts: The MCP server implementation
    • index.ts: Entry point for the library exports
    • cli.ts: the entry point for the CLI/binary
  • tests/: Tests for the API functionality

Development

Running Tests

npm test

Building

npm run build

License

MIT

Copyright 2025 yWorks GmbH - https://www.yworks.com

mcp-typescribe FAQ

How does mcp-typescribe improve LLM interaction with APIs?
It provides structured API information that LLMs can access in real-time, enabling them to use new or internal APIs beyond their training data.
Can mcp-typescribe handle large APIs efficiently?
Yes, it avoids loading full documentation context by exposing only relevant API details, making it scalable for large codebases.
Is mcp-typescribe limited to public APIs?
No, it is designed to expose both public and internal APIs to LLMs securely and efficiently.
How does mcp-typescribe integrate with existing MCP clients?
It acts as a server that MCP clients can query to retrieve structured API information for LLM consumption.
Does mcp-typescribe support multiple programming languages?
While focused on TypeScript APIs, its design can be extended to support other languages and frameworks.
What benefits does mcp-typescribe offer over traditional documentation files?
It provides a more scalable and efficient way for LLMs to learn APIs without overwhelming context, improving accuracy and performance.
Can mcp-typescribe be used with different LLM providers?
Yes, it is provider-agnostic and works with OpenAI, Anthropic Claude, Google Gemini, and others.
How secure is the API information exposed by mcp-typescribe?
It follows MCP principles for scoped and secure model interaction, ensuring controlled access to sensitive API details.