
Kontext7
Context7 MCP Server-aktuelle Dokumentation für LLM- und KI-Code-Editoren
3 years
Works with Finder
2.1k
Github Watches
126
Github Forks
2.1k
Github Stars
Context7 MCP - Up-to-date Docs For Any Prompt
❌ Without Context7
LLMs rely on outdated or generic information about the libraries you use. You get:
- ❌ Code examples are outdated and based on year-old training data
- ❌ Hallucinated APIs don't even exist
- ❌ Generic answers for old package versions
✅ With Context7
Context7 MCP pulls up-to-date, version-specific documentation and code examples straight from the source — and places them directly into your prompt.
Add use context7
to your prompt in Cursor:
Create a basic Next.js project with app router. use context7
Create a script to delete the rows where the city is "" given PostgreSQL credentials. use context7
Context7 fetches up-to-date code examples and documentation right into your LLM's context.
- 1️⃣ Write your prompt naturally
- 2️⃣ Tell the LLM to
use context7
- 3️⃣ Get working code answers
No tab-switching, no hallucinated APIs that don't exist, no outdated code generations.
🛠️ Getting Started
Requirements
- Node.js >= v18.0.0
- Cursor, Windsurf, Claude Desktop or another MCP Client
Installing via Smithery
To install Context7 MCP Server for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install @upstash/context7-mcp --client claude
Install in Cursor
Go to: Settings
-> Cursor Settings
-> MCP
-> Add new global MCP server
Pasting the following configuration into your Cursor ~/.cursor/mcp.json
file is the recommended approach. See Cursor MCP docs for more info.
{
"mcpServers": {
"context7": {
"command": "npx",
"args": ["-y", "@upstash/context7-mcp@latest"]
}
}
}
Alternative: Use Bun
{
"mcpServers": {
"context7": {
"command": "bunx",
"args": ["-y", "@upstash/context7-mcp@latest"]
}
}
}
Alternative: Use Deno
{
"mcpServers": {
"context7": {
"command": "deno",
"args": ["run", "--allow-net", "npm:@upstash/context7-mcp"]
}
}
}
Install in Windsurf
Add this to your Windsurf MCP config file. See Windsurf MCP docs for more info.
{
"mcpServers": {
"context7": {
"command": "npx",
"args": ["-y", "@upstash/context7-mcp@latest"]
}
}
}
Install in VS Code
Add this to your VS Code MCP config file. See VS Code MCP docs for more info.
{
"servers": {
"Context7": {
"type": "stdio",
"command": "npx",
"args": ["-y", "@upstash/context7-mcp@latest"]
}
}
}
Install in Claude Code
Run this command. See Claude Code MCP docs for more info.
claude mcp add context7 -- npx -y @upstash/context7-mcp@latest
Install in Claude Desktop
Add this to your Claude Desktop claude_desktop_config.json
file. See Claude Desktop MCP docs for more info.
{
"mcpServers": {
"Context7": {
"command": "npx",
"args": ["-y", "@upstash/context7-mcp@latest"]
}
}
}
Available Tools
-
resolve-library-id
: Resolves a general library name into a Context7-compatible library ID.-
libraryName
(required)
-
-
get-library-docs
: Fetches documentation for a library using a Context7-compatible library ID.-
context7CompatibleLibraryID
(required) -
topic
(optional): Focus the docs on a specific topic (e.g., "routing", "hooks") -
tokens
(optional, default 5000): Max number of tokens to return. Values less than 5000 are automatically increased to 5000.
-
Development
Clone the project and install dependencies:
bun i
Build:
bun run build
Local Configuration Example
{
"mcpServers": {
"context7": {
"command": "npx",
"args": ["tsx", "/path/to/folder/context7-mcp/src/index.ts"]
}
}
}
Testing with MCP Inspector
npx -y @modelcontextprotocol/inspector npx @upstash/context7-mcp@latest
Troubleshooting
ERR_MODULE_NOT_FOUND
If you see this error, try using bunx
instead of npx
.
{
"mcpServers": {
"context7": {
"command": "bunx",
"args": ["-y", "@upstash/context7-mcp@latest"]
}
}
}
This often resolves module resolution issues, especially in environments where npx
does not properly install or resolve packages.
MCP Client Errors
-
Try removing
@latest
from the package name. -
Try using
bunx
as an alternative. -
Try using
deno
as an alternative.
Context7 In Media
- Better Stack: "Free Tool Makes Cursor 10x Smarter"
- Cole Medin: "This is Hands Down the BEST MCP Server for AI Coding Assistants"
- Income stream surfers: "Context7 + SequentialThinking MCPs: Is This AGI?"
- Julian Goldie SEO: "Context7: New MCP AI Agent Update"
- JeredBlu: "Context 7 MCP: Get Documentation Instantly + VS Code Setup"
- Income stream surfers: "Context7: The New MCP Server That Will CHANGE AI Coding"
Star History
License
MIT
相关推荐
🔥 1Panel bietet eine intuitive Weboberfläche und einen MCP -Server, um Websites, Dateien, Container, Datenbanken und LLMs auf einem Linux -Server zu verwalten.
🧑🚀 全世界最好的 llm 资料总结(数据处理、模型训练、模型部署、 O1 模型、 MCP 、小语言模型、视觉语言模型) | Zusammenfassung der weltbesten LLM -Ressourcen.
PDF wissenschaftliche Papierübersetzung mit erhaltenen Formaten - 基于 ai 完整保留排版的 pdf 文档全文双语翻译 , 支持 支持 支持 支持 google/deeptl/ollama/openai 等服务 提供 cli/gui/mcp/docker/zotero
⛓️Rugele ist ein leichter, leistungsstarker, leistungsstarker, eingebetteter Komponenten-Orchestrierungsregel-Motor-Rahmen für GO.
Erstellen Sie einfach LLM -Tools und -Argarten mit einfachen Bash/JavaScript/Python -Funktionen.
😎简单易用、🧩丰富生态 - 大模型原生即时通信机器人平台 | 适配 qq / 微信(企业微信、个人微信) / 飞书 / 钉钉 / diskord / telegram / slack 等平台 | 支持 Chatgpt 、 Deepseek 、 Diffy 、 Claude 、 Gemini 、 xai 、 ppio 、 、 ulama 、 lm Studio 、阿里云百炼、火山方舟、 siliconflow 、 qwen 、 mondshot 、 chatglm 、 sillytraven 、 mcp 等 llm 的机器人 / agent | LLM-basierte Instant Messaging Bots-Plattform, unterstützt Zwietracht, Telegramm, Wechat, Lark, Dingtalk, QQ, Slack
Reviews

user_r5yyYSlq
As a dedicated user of MCP applications, I must say Context7 by Upstash has significantly improved my workflow. The intuitive design and user-friendly interface make it incredibly efficient to manage multiple contexts seamlessly. Highly recommend it to anyone looking to boost their productivity!

user_ADWcFbqh
I recently started using Context7 by Upstash and it has completely enhanced my workflow. The seamless integration and intuitive design make it an indispensable tool for any developer. It provides an efficient way to manage contexts effortlessly. Highly recommend!