Cover image
Try Now
2025-01-26

模型上下文协议服务器通过AI驱动的DeepSeek R1指导功能,为LLM代理提供第二意见,包括代码审查,设计批评,写作反馈和通过DeepSeek API进行思想进行集思广益。

3 years

Works with Finder

2

Github Watches

9

Github Forks

21

Github Stars

mentor-mcp-server

TypeScript Model Context Protocol Version License Status GitHub

A Model Context Protocol server providing LLM Agents a second opinion via AI-powered Deepseek-Reasoning (R1) mentorship capabilities, including code review, design critique, writing feedback, and idea brainstorming through the Deepseek API. Set your LLM Agent up for success with expert second opinions and actionable insights.

Model Context Protocol

The Model Context Protocol (MCP) enables communication between:

  • Clients: Claude Desktop, IDEs, and other MCP-compatible clients
  • Servers: Tools and resources for task management and automation
  • LLM Agents: AI models that leverage the server's capabilities

Table of Contents

Features

Code Analysis

  • Comprehensive code reviews
  • Bug detection and prevention
  • Style and best practices evaluation
  • Performance optimization suggestions
  • Security vulnerability assessment

Design & Architecture

  • UI/UX design critiques
  • Architectural diagram analysis
  • Design pattern recommendations
  • Accessibility evaluation
  • Consistency checks

Content Enhancement

  • Writing feedback and improvement
  • Grammar and style analysis
  • Documentation review
  • Content clarity assessment
  • Structural recommendations

Strategic Planning

  • Feature enhancement brainstorming
  • Second opinions on approaches
  • Innovation suggestions
  • Feasibility analysis
  • User value assessment

Installation

# Clone the repository
git clone git@github.com:cyanheads/mentor-mcp-server.git
cd mentor-mcp-server

# Install dependencies
npm install

# Build the project
npm run build

Configuration

Add to your MCP client settings:

{
  "mcpServers": {
    "mentor": {
      "command": "node",
      "args": ["build/index.js"],
      "env": {
        "DEEPSEEK_API_KEY": "your_api_key",
        "DEEPSEEK_MODEL": "deepseek-reasoner",
        "DEEPSEEK_MAX_TOKENS": "8192",
        "DEEPSEEK_MAX_RETRIES": "3",
        "DEEPSEEK_TIMEOUT": "30000"
      }
    }
  }
}

Environment Variables

Variable Required Default Description
DEEPSEEK_API_KEY Yes - Your Deepseek API key
DEEPSEEK_MODEL Yes deepseek-reasoner Deepseek model name
DEEPSEEK_MAX_TOKENS No 8192 Maximum tokens per request
DEEPSEEK_MAX_RETRIES No 3 Number of retry attempts
DEEPSEEK_TIMEOUT No 30000 Request timeout (ms)

Tools

Code Review

<use_mcp_tool>
<server_name>mentor-mcp-server</server_name>
<tool_name>code_review</tool_name>
<arguments>
{
  "file_path": "src/app.ts",
  "language": "typescript"
}
</arguments>
</use_mcp_tool>

Design Critique

<use_mcp_tool>
<server_name>mentor-mcp-server</server_name>
<tool_name>design_critique</tool_name>
<arguments>
{
  "design_document": "path/to/design.fig",
  "design_type": "web UI"
}
</arguments>
</use_mcp_tool>

Writing Feedback

<use_mcp_tool>
<server_name>mentor-mcp-server</server_name>
<tool_name>writing_feedback</tool_name>
<arguments>
{
  "text": "Documentation content...",
  "writing_type": "documentation"
}
</arguments>
</use_mcp_tool>

Feature Enhancement

<use_mcp_tool>
<server_name>mentor-mcp-server</server_name>
<tool_name>brainstorm_enhancements</tool_name>
<arguments>
{
  "concept": "User authentication system"
}
</arguments>
</use_mcp_tool>

Examples

Detailed examples of each tool's usage and output can be found in the examples directory:

Each example includes the request format and sample response, demonstrating the tool's capabilities and output structure.

Development

# Build TypeScript code
npm run build

# Start the server
npm run start

# Development with watch mode
npm run dev

# Clean build artifacts
npm run clean

Project Structure

src/
├── api/         # API integration modules
├── tools/       # Tool implementations
│   ├── second-opinion/
│   ├── code-review/
│   ├── design-critique/
│   ├── writing-feedback/
│   └── brainstorm-enhancements/
├── types/       # TypeScript type definitions
├── utils/       # Utility functions
├── config.ts    # Server configuration
├── index.ts     # Entry point
└── server.ts    # Main server implementation

License

Apache License 2.0. See LICENSE for more information.


Built with the Model Context Protocol

相关推荐

  • Joshua Armstrong
  • Confidential guide on numerology and astrology, based of GG33 Public information

  • https://suefel.com
  • Latest advice and best practices for custom GPT development.

  • Alexandru Strujac
  • Efficient thumbnail creator for YouTube videos

  • Emmet Halm
  • Converts Figma frames into front-end code for various mobile frameworks.

  • Elijah Ng Shi Yi
  • Advanced software engineer GPT that excels through nailing the basics.

  • lumpenspace
  • Take an adjectivised noun, and create images making it progressively more adjective!

  • Lists Tailwind CSS classes in monospaced font

  • https://maiplestudio.com
  • Find Exhibitors, Speakers and more

  • https://appia.in
  • Siri Shortcut Finder – your go-to place for discovering amazing Siri Shortcuts with ease

  • Carlos Ferrin
  • Encuentra películas y series en plataformas de streaming.

  • Yusuf Emre Yeşilyurt
  • I find academic articles and books for research and literature reviews.

  • tomoyoshi hirata
  • Sony α7IIIマニュアルアシスタント

  • apappascs
  • 发现市场上最全面,最新的MCP服务器集合。该存储库充当集中式枢纽,提供了广泛的开源和专有MCP服务器目录,并提供功能,文档链接和贡献者。

  • ShrimpingIt
  • MCP系列GPIO Expander的基于Micropython I2C的操作,源自ADAFRUIT_MCP230XX

  • jae-jae
  • MCP服务器使用剧作《无头浏览器》获取网页内容。

  • ravitemer
  • 一个功能强大的Neovim插件,用于管理MCP(模型上下文协议)服务器

  • patruff
  • Ollama和MCP服务器之间的桥梁,使本地LLMS可以使用模型上下文协议工具

  • pontusab
  • 光标与风浪冲浪社区,查找规则和MCP

  • av
  • 毫不费力地使用一个命令运行LLM后端,API,前端和服务。

  • WangRongsheng
  • 🧑‍🚀 llm 资料总结(数据处理、模型训练、模型部署、 o1 模型、mcp 、小语言模型、视觉语言模型)|摘要世界上最好的LLM资源。

  • Mintplex-Labs
  • 带有内置抹布,AI代理,无代理构建器,MCP兼容性等的多合一桌面和Docker AI应用程序。

  • appcypher
  • 很棒的MCP服务器 - 模型上下文协议服务器的策划列表

    Reviews

    2 (1)
    Avatar
    user_hPEwvnlI
    2025-04-18

    I have been using mentor-mcp-server and it has significantly streamlined my server management tasks. The simplicity and efficiency offered by cyanheads' creation make it invaluable. The open-source nature of the project ensures constant updates and community support. Highly recommend it for anyone looking to optimize their server operations!