
Bucear
Dive es una aplicación de escritorio de host MCP de código abierto que se integra perfectamente con cualquier LLMS que respalde las capacidades de llamadas de funciones. ✨
16
Github Watches
72
Github Forks
841
Github Stars
Dive AI Agent 🤿 🤖
Dive is an open-source MCP Host Desktop Application that seamlessly integrates with any LLMs supporting function calling capabilities. ✨
Features 🎯
- 🌐 Universal LLM Support: Compatible with ChatGPT, Anthropic, Ollama and OpenAI-compatible models
- 💻 Cross-Platform: Available for Windows, MacOS, and Linux
- 🔄 Model Context Protocol: Enabling seamless MCP AI agent integration on both stdio and SSE mode
- 🌍 Multi-Language Support: Traditional Chinese, Simplified Chinese, English, Spanish with more coming soon
- ⚙️ Advanced API Management: Multiple API keys and model switching support
- 💡 Custom Instructions: Personalized system prompts for tailored AI behavior
- 🔄 Auto-Update Mechanism: Automatically checks for and installs the latest application updates
Recent updates(2025/3/14)
- 🌍 Spanish Translation: Added Spanish language support
- 🤖 Extended Model Support: Added Google Gemini and Mistral AI models integration
Download and Install ⬇️
Get the latest version of Dive:
For Windows users: 🪟
- Download the .exe version
- Python and Node.js environments are pre-installed
For MacOS users: 🍎
- Download the .dmg version
- You need to install Python and Node.js (with npx uvx) environments yourself
- Follow the installation prompts to complete setup
For Linux users: 🐧
- Download the .AppImage version
- You need to install Python and Node.js (with npx uvx) environments yourself
- For Ubuntu/Debian users:
- You may need to add
--no-sandbox
parameter - Or modify system settings to allow sandbox
- Run
chmod +x
to make the AppImage executable
- You may need to add
MCP Tips
While the system comes with a default echo MCP Server, your LLM can access more powerful tools through MCP. Here's how to get started with two beginner-friendly tools: Fetch and Youtube-dl.
Quick Setup
Add this JSON configuration to your Dive MCP settings to enable both tools:
"mcpServers":{
"fetch": {
"command": "uvx",
"args": [
"mcp-server-fetch",
"--ignore-robots-txt"
],
"enabled": true
},
"filesystem": {
"command": "npx",
"args": [
"-y",
"@modelcontextprotocol/server-filesystem",
"/path/to/allowed/files"
],
"enabled": true
},
"youtubedl": {
"command": "npx",
"args": [
"@kevinwatt/yt-dlp-mcp"
],
"enabled": true
}
}
Using SSE Server for MCP
You can also connect to an external MCP server via SSE (Server-Sent Events). Add this configuration to your Dive MCP settings:
{
"mcpServers": {
"MCP_SERVER_NAME": {
"enabled": true,
"transport": "sse",
"url": "YOUR_SSE_SERVER_URL"
}
}
}
Additional Setup for yt-dlp-mcp
yt-dlp-mcp requires the yt-dlp package. Install it based on your operating system:
Windows
winget install yt-dlp
MacOS
brew install yt-dlp
Linux
pip install yt-dlp
Build 🛠️
See BUILD.md for more details.
Connect With Us 🌐
- 💬 Join our Discord
- 🐦 Follow us on Twitter/X
- ⭐ Star us on GitHub
- 🐛 Report issues on our Issue Tracker
相关推荐
I find academic articles and books for research and literature reviews.
Confidential guide on numerology and astrology, based of GG33 Public information
Emulating Dr. Jordan B. Peterson's style in providing life advice and insights.
Your go-to expert in the Rust ecosystem, specializing in precise code interpretation, up-to-date crate version checking, and in-depth source code analysis. I offer accurate, context-aware insights for all your Rust programming questions.
Converts Figma frames into front-end code for various mobile frameworks.
Advanced software engineer GPT that excels through nailing the basics.
Take an adjectivised noun, and create images making it progressively more adjective!
Descubra la colección más completa y actualizada de servidores MCP en el mercado. Este repositorio sirve como un centro centralizado, que ofrece un extenso catálogo de servidores MCP de código abierto y propietarios, completos con características, enlaces de documentación y colaboradores.
Manipulación basada en Micrypthon I2C del expansor GPIO de la serie MCP, derivada de AdaFruit_MCP230xx
La aplicación AI de escritorio todo en uno y Docker con trapo incorporado, agentes de IA, creador de agentes sin código, compatibilidad de MCP y más.
Plataforma de automatización de flujo de trabajo de código justo con capacidades de IA nativas. Combine el edificio visual con código personalizado, auto-anfitrión o nube, más de 400 integraciones.
Un poderoso complemento Neovim para administrar servidores MCP (protocolo de contexto del modelo)
🧑🚀 全世界最好的 llM 资料总结(数据处理、模型训练、模型部署、 O1 模型、 MCP 、小语言模型、视觉语言模型) | Resumen de los mejores recursos del mundo.
Servidor MCP para obtener contenido de la página web con el navegador sin cabeza de dramaturgo.
Puente entre los servidores Ollama y MCP, lo que permite a LLM locales utilizar herramientas de protocolo de contexto del modelo
Una lista curada de servidores de protocolo de contexto del modelo (MCP)
Reviews

user_GbLtFHGt
Dive by OpenAgentPlatform is an exceptional tool for developers and enthusiasts! Its seamless integration and user-friendly interface make complex tasks manageable. The documentation and support provided are top-notch, enhancing the overall experience. Highly recommended for anyone looking to streamline their workflow. Check it out at https://github.com/OpenAgentPlatform/Dive!