Fire in da houseTop Tip:Paying $100+ per month for Perplexity, MidJourney, Runway, ChatGPT and other tools is crazy - get all your AI tools in one site starting at $15 per month with Galaxy AI Fire in da houseCheck it out free

vibe-eyes

MCP.Pizza Chef: monteslu

Vibe-Eyes is an MCP server that captures and vectorizes canvas content and debug information from browser-based games and applications. It uses a lightweight client to send canvas visuals and debug logs to a Node.js server, which converts them into compact SVG representations. These visualizations and debug data are then exposed to LLMs via the Model Context Protocol, enabling models to 'see' and understand real-time game/app states for enhanced interaction and analysis.

Use This MCP server To

Visualize browser game canvas content for LLM analysis Provide real-time debug logs and error info to LLMs Enable LLMs to interpret game state through vectorized SVG visuals Support multi-step reasoning on browser app visuals and debug data Integrate with LLMs for automated game testing and debugging Facilitate AI-driven game walkthroughs using live canvas data Convert canvas graphics into structured SVG for model consumption Monitor browser app exceptions and logs in real time via MCP

README

Vibe-Eyes

An MCP server that enables LLMs to "see" what's happening in browser-based games and applications through vectorized canvas visualization and debug information.

Vibe-Eyes Logo

Vibe-Eyes uses a client-server architecture where a lightweight browser client captures canvas content and debug information, sends it to a Node.js server via WebSockets, which then vectorizes the images into compact SVG representations and makes them available to LLMs through the Model Context Protocol (MCP).

flowchart LR
    A["Browser Game/App<br/>(Canvas + JavaScript)"] -->|"Captures"| B["Vibe-Eyes Client<br/>(vibe-eyes-client)"]
    B -->|"WebSocket<br/>(CORS-free)"| C["Socket.IO Server"]
    
    subgraph server["Vibe-Eyes Server (mcp.js)"]
        C -->|"Process"| D["Vectorization<br/>(vectorizer.js)"]
        C -->|"Store"| E["Debug Data<br/>(logs, errors, exceptions)"]
        D -->|"Rough SVG"| F["MCP Tool: getGameDebug()"]
        E --> F
    end
    
    F -->|"SVG + Debug Info"| G["Claude/LLM<br/>(MCP Client)"]
    G -->|"Debugging<br/>Assistance"| A
    
    classDef default color:#000,font-weight:bold
    classDef edgeLabel color:#333,font-size:12px
    
    style A fill:#c0e0ff,stroke:#000,stroke-width:2px
    style B fill:#ffe0a0,stroke:#000,stroke-width:2px
    style C fill:#a0d0ff,stroke:#000,stroke-width:2px
    style D fill:#b0e0a0,stroke:#000,stroke-width:2px
    style E fill:#ffb0b0,stroke:#000,stroke-width:2px
    style F fill:#d0b0ff,stroke:#000,stroke-width:2px
    style G fill:#ffb0d0,stroke:#000,stroke-width:2px
    style server fill:#f0f0f0,stroke:#666,stroke-width:1px,stroke-dasharray: 5 5,color:#000
Loading

Note: This project is experimental and designed to enhance "vibe coding" sessions with LLMs by providing visual context and rich debug information.

Video Explanation

Vibe-Eyes Video Explanation

Key Capabilities

  • Captures and vectorizes canvas elements from browser games
  • Collects console logs and errors in real-time
  • Catches unhandled exceptions with full stack traces
  • Makes the visual and debug information available to LLMs via MCP
  • Creates a seamless debugging experience for developers working with LLMs

How It Works

  1. A lightweight client runs in the browser game/application
  2. The client captures canvas snapshots, console logs/errors, and unhandled exceptions
  3. Data is sent to the Vibe-Eyes server via WebSocket (avoiding CORS issues)
  4. The server vectorizes the canvas images and stores them with the debug information
  5. LLMs connect via the Model Context Protocol to access the latest data
  6. LLMs can "see" what's happening and help debug issues with full context

Components

1. Vibe-Eyes MCP Server (mcp.js)

The core server that:

  • Receives canvas snapshots via Socket.IO
  • Vectorizes images to compact SVG representation (rough approximation)
  • Stores debug information (logs, errors, exceptions, timing)
  • Exposes the data via Model Context Protocol (MCP)
  • Provides HTTP endpoints for direct access
  • Processes images sequentially to manage resources

2. Browser Client

The browser client is available at vibe-eyes-client repository.

A lightweight browser integration that:

  • Finds canvas elements in the page
  • Captures canvas content as data URLs
  • Intercepts console logs and errors
  • Catches global unhandled exceptions with stack traces
  • Sends data to the Vibe-Eyes server via WebSockets
  • Minimizes performance impact on games
  • Supports explicit initialization to control when capturing starts

3. Vectorization Engine (vectorizer.js)

A high-quality SVG vectorization library that:

  • Converts raster images to vector SVGs
  • Optimizes SVGs for size and clarity
  • Preserves visual information while reducing data size

Getting Started

Installation

# Clone the repository
git clone https://github.com/monteslu/vibe-eyes.git
cd vibe-eyes

# Install dependencies
npm install

Using with LLM Agents

Register the MCP server with your AI agent:

# For Claude Code
claude mcp add

This enables Claude to use the Vibe-Eyes capabilities via MCP.

Integrating with Games/Applications

Add the client to your browser application by including the required scripts:

<!-- Include Socket.IO client -->
<script src="https://cdn.socket.io/4.7.4/socket.io.min.js"></script>

<!-- Include Vibe-Eyes client -->
<script src="https://cdn.jsdelivr.net/npm/vibe-eyes-client/dist/index.min.js"></script>

<!-- Initialize the client -->
<script>
  // Import the initialization function if using as module
  // import { initializeVibeEyes } from 'vibe-eyes-client';
  
  // Initialize with configuration
  const vibeEyes = initializeVibeEyes({
    // WebSocket URL to the Vibe-Eyes server
    serverUrl: 'ws://localhost:8869',
    // Capture interval in milliseconds
    captureDelay: 1000,
    // Start capturing automatically after connection
    autoCapture: true
  });
</script>

Using with Claude or Other LLMs

The MCP server exposes a tool for LLMs to access the latest visual and debug information via Model Context Protocol (MCP):

getGameDebug({ includeSvg: true/false })

The LLM will receive:

  • Recent console logs and errors from the application
  • Unhandled exceptions with full stack traces (if any occurred)
  • Vectorized SVG approximation of the canvas (if includeSvg is true)
  • Timing and correlation information to connect visual state with logs

This allows the LLM to "see" what's happening in the application and provide better assistance.

Example MCP Config (for Claude Code)

To access Vibe-Eyes from Claude:

{
  "name": "vibe-eyes",
  "url": "http://localhost:8869",
  "tools": [
    {
      "name": "getGameDebug",
      "description": "Retrieves the most recent canvas visualization and debug information from a browser game or application"
    }
  ]
}

How Vibe-Eyes Helps with "Vibe Coding"

Traditional "vibe coding" sessions require developers to manually take screenshots and describe what's happening in their application. Vibe-Eyes automates this process by:

  1. Providing Visual Context: LLMs can see the actual visual state of the game/app
  2. Correlating Visual and Code Issues: Console logs are paired with visual state
  3. Reducing Manual Work: No need to manually capture and upload screenshots
  4. Enabling Real-Time Debugging: LLMs can observe changes as they happen
  5. Optimizing Data Transfer: Vector representation is more compact than screenshots

Performance Considerations

  • The browser client is designed to minimize impact on application performance
  • Creating canvas data URLs can be CPU-intensive, so capturing frequency is configurable
  • WebSocket transport avoids CORS issues common in cross-domain setups
  • The server processes images sequentially to prevent overload
  • SVG vectorization balances visual accuracy with size optimization

Direct SVG Access

For applications that want to reuse the vectorized SVG output:

  1. WebSocket Response: The server includes the SVG directly in WebSocket responses:

    socket.on('debugCapture', (data, callback) => {
      // Capture and process...
      callback({
        success: true,
        id: "capture_123",
        svg: "<svg>...</svg>",  // Vectorized SVG 
        stats: { /* stats data */ }
      });
    });
  2. HTTP Endpoint: Access the latest capture via the /latest endpoint:

    fetch('http://localhost:8869/latest')
      .then(res => res.json())
      .then(data => {
        const svg = data.vectorized?.svg;
        // Use the SVG...
      });

API Reference

Browser Client

// Initialize the client
const vibeEyes = initializeVibeEyes({
  serverUrl: 'ws://localhost:8869',
  captureDelay: 1000, // ms between captures
  maxLogs: 10,        // Max console.log entries to store
  maxErrors: 10,      // Max console.error entries to store
  autoCapture: true   // Start capturing automatically
});

// Manual control
vibeEyes.startCaptureLoop(); // Start auto-capturing
vibeEyes.stopCaptureLoop();  // Stop auto-capturing
vibeEyes.captureAndSend();   // Trigger one capture immediately

// The server responds with:
// {
//   success: true,
//   id: "capture_1234567890",
//   processedAt: 1616161616161,
//   svg: "<svg>...</svg>",  // The vectorized SVG for direct use
//   stats: {
//     vectorizeTime: 120,
//     optimizeTime: 30,
//     originalSize: 50000,
//     finalSize: 15000,
//     sizeReduction: 70
//   }
// }

MCP Tool

// MCP tool available to LLMs
getGameDebug({
  includeSvg: true  // Whether to include SVG visualization
})

// Returns
{
  success: true,
  capture: {
    id: "capture_123456789",
    timestamp: 1616161616161,
    console_logs: [
      { timestamp: 1616161616000, data: ["Player position:", {x: 10, y: 20}] },
      // ...more logs
    ],
    console_errors: [
      // Any errors captured
    ],
    unhandled_exception: {
      timestamp: 1616161616100,
      message: "Uncaught SyntaxError: Unexpected token ';'",
      stack: "SyntaxError: Unexpected token ';'\n    at game.js:42:10\n...",
      type: "SyntaxError",
      source: "game.js",
      line: 42,
      column: 10
    },
    vectorized: {
      svg: "<svg>...</svg>",  // Only if includeSvg is true (rough approximation)
      imageType: "png",
      stats: {
        vectorizeTime: 120,
        optimizeTime: 30,
        originalSize: 50000,
        finalSize: 15000,
        sizeReduction: 70
      }
    }
  }
}

Standalone Vectorizer CLI

The project also includes a standalone CLI tool for vectorizing individual files:

# Install CLI globally
npm install -g vibe-eyes

# Use the CLI
vibe-eyes-vectorize input.png output.svg

# With options
vibe-eyes-vectorize photo.jpg --color-precision 10 --max-iterations 100

License

ISC

vibe-eyes FAQ

How does Vibe-Eyes capture browser game visuals?
It uses a lightweight browser client to capture canvas content and debug info, sending it via WebSocket to the server.
What format does Vibe-Eyes use to represent visuals for LLMs?
It vectorizes canvas images into compact SVG representations for efficient model consumption.
Can Vibe-Eyes handle real-time debug information?
Yes, it collects logs, errors, and exceptions alongside visuals and exposes them to LLMs.
Is Vibe-Eyes limited to specific browsers or games?
It works with any browser-based game or app using canvas and JavaScript, leveraging standard WebSocket communication.
How does Vibe-Eyes integrate with the Model Context Protocol?
It exposes vectorized visuals and debug data as MCP tools, enabling LLMs to query game state contextually.
What technology stack does Vibe-Eyes use?
It uses a Node.js server with Socket.IO for WebSocket communication and JavaScript for client-side capture.
Can Vibe-Eyes improve automated game testing?
Yes, by providing LLMs with visual and debug context, it enables AI-driven testing and error detection.
Does Vibe-Eyes support multiple LLM providers?
Yes, it is provider-agnostic and works with OpenAI, Anthropic Claude, Google Gemini, and others.