MCP cover image
See in Github
2025-03-27

MCP (Model context protocol) server with LLMling backend

1

Github Watches

2

Github Forks

5

Github Stars

mcp-server-llmling

PyPI License Package status Daily downloads Weekly downloads Monthly downloads Distribution format Wheel availability Python version Implementation Releases Github Contributors Github Discussions Github Forks Github Issues Github Issues Github Watchers Github Stars Github Repository size Github last commit Github release date Github language count Github commits this week Github commits this month Github commits this year Package status Code style: black PyUp

Read the documentation!

LLMling Server Manual

Overview

mcp-server-llmling is a server for the Machine Chat Protocol (MCP) that provides a YAML-based configuration system for LLM applications.

LLMLing, the backend, provides a YAML-based configuration system for LLM applications. It allows to set up custom MCP servers serving content defined in YAML files.

  • Static Declaration: Define your LLM's environment in YAML - no code required
  • MCP Protocol: Built on the Machine Chat Protocol (MCP) for standardized LLM interaction
  • Component Types:
    • Resources: Content providers (files, text, CLI output, etc.)
    • Prompts: Message templates with arguments
    • Tools: Python functions callable by the LLM

The YAML configuration creates a complete environment that provides the LLM with:

  • Access to content via resources
  • Structured prompts for consistent interaction
  • Tools for extending capabilities

Key Features

1. Resource Management

  • Load and manage different types of resources:
    • Text files (PathResource)
    • Raw text content (TextResource)
    • CLI command output (CLIResource)
    • Python source code (SourceResource)
    • Python callable results (CallableResource)
    • Images (ImageResource)
  • Support for resource watching/hot-reload
  • Resource processing pipelines
  • URI-based resource access

2. Tool System

  • Register and execute Python functions as LLM tools
  • Support for OpenAPI-based tools
  • Entry point-based tool discovery
  • Tool validation and parameter checking
  • Structured tool responses

3. Prompt Management

  • Static prompts with template support
  • Dynamic prompts from Python functions
  • File-based prompts
  • Prompt argument validation
  • Completion suggestions for prompt arguments

4. Multiple Transport Options

  • Stdio-based communication (default)
  • Server-Sent Events (SSE) for web clients
  • Support for custom transport implementations

Usage

With Zed Editor

Add LLMLing as a context server in your settings.json:

{
  "context_servers": {
    "llmling": {
      "command": {
        "env": {},
        "label": "llmling",
        "path": "uvx",
        "args": [
          "mcp-server-llmling",
          "start",
          "path/to/your/config.yml"
        ]
      },
      "settings": {}
    }
  }
}

With Claude Desktop

Configure LLMLing in your claude_desktop_config.json:

{
  "mcpServers": {
    "llmling": {
      "command": "uvx",
      "args": [
        "mcp-server-llmling",
        "start",
        "path/to/your/config.yml"
      ],
      "env": {}
    }
  }
}

Manual Server Start

Start the server directly from command line:

# Latest version
uvx mcp-server-llmling@latest

1. Programmatic usage

from llmling import RuntimeConfig
from mcp_server_llmling import LLMLingServer

async def main() -> None:
    async with RuntimeConfig.open(config) as runtime:
        server = LLMLingServer(runtime, enable_injection=True)
        await server.start()

asyncio.run(main())

2. Using Custom Transport

from llmling import RuntimeConfig
from mcp_server_llmling import LLMLingServer

async def main() -> None:
    async with RuntimeConfig.open(config) as runtime:
        server = LLMLingServer(
            config,
            transport="sse",
            transport_options={
                "host": "localhost",
                "port": 8000,
                "cors_origins": ["http://localhost:3000"]
            }
        )
        await server.start()

asyncio.run(main())

3. Resource Configuration

resources:
  python_code:
    type: path
    path: "./src/**/*.py"
    watch:
      enabled: true
      patterns:
        - "*.py"
        - "!**/__pycache__/**"

  api_docs:
    type: text
    content: |
      API Documentation
      ================
      ...

4. Tool Configuration

tools:
  analyze_code:
    import_path: "mymodule.tools.analyze_code"
    description: "Analyze Python code structure"

toolsets:
  api:
    type: openapi
    spec: "https://api.example.com/openapi.json"
    namespace: "api"

Server Configuration

The server is configured through a YAML file with the following sections:

global_settings:
  timeout: 30
  max_retries: 3
  log_level: "INFO"
  requirements: []
  pip_index_url: null
  extra_paths: []

resources:
  # Resource definitions...

tools:
  # Tool definitions...

toolsets:
  # Toolset definitions...

prompts:
  # Prompt definitions...

MCP Protocol

The server implements the MCP protocol which supports:

  1. Resource Operations

    • List available resources
    • Read resource content
    • Watch for resource changes
  2. Tool Operations

    • List available tools
    • Execute tools with parameters
    • Get tool schemas
  3. Prompt Operations

    • List available prompts
    • Get formatted prompts
    • Get completions for prompt arguments
  4. Notifications

    • Resource changes
    • Tool/prompt list updates
    • Progress updates
    • Log messages

相关推荐

  • NiKole Maxwell
  • I craft unique cereal names, stories, and ridiculously cute Cereal Baby images.

  • https://suefel.com
  • Latest advice and best practices for custom GPT development.

  • Yusuf Emre Yeşilyurt
  • I find academic articles and books for research and literature reviews.

  • https://duotify.com
  • Professional code reviewer for a unique language, friendly and helpful.

  • https://maiplestudio.com
  • Find Exhibitors, Speakers and more

  • https://hubeiqiao.com
  • IT problem solver with clear, step-by-step guidance.

  • Carlos Ferrin
  • Encuentra películas y series en plataformas de streaming.

  • Bora Yalcin
  • Evaluator for marketplace product descriptions, checks for relevancy and keyword stuffing.

  • Andris Teikmanis
  • Latvian GPT assistant for developing GPT applications

  • Joshua Armstrong
  • Confidential guide on numerology and astrology, based of GG33 Public information

  • Contraband Interactive
  • Emulating Dr. Jordan B. Peterson's style in providing life advice and insights.

  • https://cantaspinar.com
  • Summarizes videos and answers related questions.

  • apappascs
  • Discover the most comprehensive and up-to-date collection of MCP servers in the market. This repository serves as a centralized hub, offering an extensive catalog of open-source and proprietary MCP servers, complete with features, documentation links, and contributors.

  • Mintplex-Labs
  • The all-in-one Desktop & Docker AI application with built-in RAG, AI agents, No-code agent builder, MCP compatibility, and more.

  • modelcontextprotocol
  • Model Context Protocol Servers

  • ShrimpingIt
  • Micropython I2C-based manipulation of the MCP series GPIO expander, derived from Adafruit_MCP230xx

  • OffchainLabs
  • Go implementation of Ethereum proof of stake

  • n8n-io
  • Fair-code workflow automation platform with native AI capabilities. Combine visual building with custom code, self-host or cloud, 400+ integrations.

  • WangRongsheng
  • 🧑‍🚀 全世界最好的LLM资料总结(Agent框架、辅助编程、数据处理、模型训练、模型推理、o1 模型、MCP、小语言模型、视觉语言模型) | Summary of the world's best LLM resources.

    Reviews

    4 (1)
    Avatar
    user_5OruX6PG
    2025-04-16

    I've been using the mcp-server-llmling by phil65 for a while now, and it's truly impressive. The seamless integration and robust features have greatly enhanced my workflow. It's reliable, efficient, and the community support is fantastic. Highly recommend checking it out on GitHub!