MCP cover image
See in Github
2025-04-04

用于OPIK的模型上下文协议(MCP)实现实现了无缝的IDE集成并统一访问提示,项目,跟踪和指标。

5

Github Watches

4

Github Forks

52

Github Stars

Comet Opik logo
Opik MCP Server
(Model Context Protocol)

A Model Context Protocol (MCP) implementation for the Opik platform with support for multiple transport mechanisms, enabling seamless integration with IDEs and providing a unified interface for Opik's capabilities.

License Node.js Version TypeScript <img/ alt="OPIK-MCP image">

WebsiteSlack communityTwitterDocumentation

Opik Server MCP server

⚠️ Notice: SSE (Server-Sent Events) transport support is currently experimental and untested. For production use, we recommend using the direct process execution approach shown in the IDE integration examples.

🚀 What is Opik MCP Server?

Opik MCP Server is an open-source implementation of the Model Context Protocol for the Opik platform. It provides a unified interface for interacting with Opik's capabilities, supporting multiple transport mechanisms for flexible integration into various environments.


You can use Opik MCP Server for:

  • IDE Integration:

    • Seamlessly integrate with Cursor and other compatible IDEs
    • Provide direct access to Opik's capabilities from your development environment
  • Unified API Access:

    • Access all Opik features through a standardized protocol
    • Leverage multiple transport options (stdio, SSE) for different integration scenarios
  • Platform Management:

    • Manage prompts, projects, traces, and metrics through a consistent interface
    • Organize and monitor your LLM applications efficiently

Features

  • Prompts Management: Create, list, update, and delete prompts
  • Projects/Workspaces Management: Organize and manage projects
  • Traces: Track and analyze trace data
  • Metrics: Gather and query metrics data

Quick Start

Installation

Manual Installation

# Clone the repository
git clone https://github.com/comet-ml/opik-mcp.git
cd opik-mcp

# Install dependencies and build
npm install
npm run build

Configuration

Create a .env file based on the example:

cp .env.example .env
# Edit .env with your specific configuration

Starting the Server

# Start with stdio transport (default)
npm run start:stdio

# Start with SSE transport for network access (experimental)
npm run start:sse

IDE Integration

Cursor Integration

To integrate with Cursor IDE, create a .cursor/mcp.json file in your project directory with the following configuration:

{
  "mcpServers": {
    "opik": {
      "command": "/path/to/node",
      "args": [
        "/path/to/opik-mcp/build/index.js",
        "--apiUrl",
        "https://www.comet.com/opik/api",
        "--apiKey",
        "YOUR_API_KEY",
        "--workspace",
        "default",
        "--debug",
        "true"
      ],
      "env": {
        "OPIK_API_BASE_URL": "https://www.comet.com/opik/api",
        "OPIK_API_KEY": "YOUR_API_KEY",
        "OPIK_WORKSPACE_NAME": "default",
      }
    }
  }
}

Replace /path/to/node with the path to your Node.js executable and /path/to/opik-mcp with the path to your opik-mcp installation. Also replace YOUR_API_KEY with your actual Opik API key.

Available Commands

The project includes a Makefile for common operations:

# Display all available commands
make help

# Run tests
make test

# Run transport-specific tests
make test-transport

# Start the server with SSE transport (experimental)
make start-sse

# Start the server with stdio transport
make start-stdio

Transport Options

Standard Input/Output

Ideal for local integration where the client and server run on the same machine.

make start-stdio

Server-Sent Events (SSE)

Enables remote access and multiple simultaneous clients over HTTP. Note that this transport option is experimental.

make start-sse

For detailed information about the SSE transport, see docs/sse-transport.md.

Development

Testing

# Run all tests
npm test

# Run specific test suite
npm test -- tests/transports/sse-transport.test.ts

Pre-commit Hooks

This project uses pre-commit hooks to ensure code quality:

# Run pre-commit checks manually
make precommit

Documentation

License

Apache 2.0

相关推荐

  • https://suefel.com
  • Latest advice and best practices for custom GPT development.

  • Yusuf Emre Yeşilyurt
  • I find academic articles and books for research and literature reviews.

  • https://maiplestudio.com
  • Find Exhibitors, Speakers and more

  • Carlos Ferrin
  • Encuentra películas y series en plataformas de streaming.

  • Joshua Armstrong
  • Confidential guide on numerology and astrology, based of GG33 Public information

  • Emmet Halm
  • Converts Figma frames into front-end code for various mobile frameworks.

  • Elijah Ng Shi Yi
  • Advanced software engineer GPT that excels through nailing the basics.

  • Alexandru Strujac
  • Efficient thumbnail creator for YouTube videos

  • lumpenspace
  • Take an adjectivised noun, and create images making it progressively more adjective!

  • Lists Tailwind CSS classes in monospaced font

  • https://appia.in
  • Siri Shortcut Finder – your go-to place for discovering amazing Siri Shortcuts with ease

  • apappascs
  • 发现市场上最全面,最新的MCP服务器集合。该存储库充当集中式枢纽,提供了广泛的开源和专有MCP服务器目录,并提供功能,文档链接和贡献者。

  • ShrimpingIt
  • MCP系列GPIO Expander的基于Micropython I2C的操作,源自ADAFRUIT_MCP230XX

  • jae-jae
  • MCP服务器使用剧作《无头浏览器》获取网页内容。

  • Mintplex-Labs
  • 带有内置抹布,AI代理,无代理构建器,MCP兼容性等的多合一桌面和Docker AI应用程序。

  • ravitemer
  • 一个功能强大的Neovim插件,用于管理MCP(模型上下文协议)服务器

  • patruff
  • Ollama和MCP服务器之间的桥梁,使本地LLMS可以使用模型上下文协议工具

  • pontusab
  • 光标与风浪冲浪社区,查找规则和MCP

  • WangRongsheng
  • 🧑‍🚀 llm 资料总结(数据处理、模型训练、模型部署、 o1 模型、mcp 、小语言模型、视觉语言模型)|摘要世界上最好的LLM资源。

  • n8n-io
  • 具有本机AI功能的公平代码工作流程自动化平台。将视觉构建与自定义代码,自宿主或云相结合,400+集成。

    Reviews

    3 (1)
    Avatar
    user_pK7lLev4
    2025-04-17

    I've been using opik-mcp by comet-ml for several months and it's an absolute game changer. Its seamless integration and robust features make machine learning project management so much easier. The detailed documentation and active community support further enhance its utility. Highly recommended for anyone looking to streamline their ML workflows!