MCP cover image
See in Github
2025-04-02

1

Github Watches

0

Github Forks

0

Github Stars

YouTube to LinkedIn MCP Server

A Model Context Protocol (MCP) server that automates generating LinkedIn post drafts from YouTube videos. This server provides high-quality, editable content drafts based on YouTube video transcripts.

Features

  • YouTube Transcript Extraction: Extract transcripts from YouTube videos using video URLs
  • Transcript Summarization: Generate concise summaries of video content using OpenAI GPT
  • LinkedIn Post Generation: Create professional LinkedIn post drafts with customizable tone and style
  • Modular API Design: Clean FastAPI implementation with well-defined endpoints
  • Containerized Deployment: Ready for deployment on Smithery

Setup Instructions

Prerequisites

  • Python 3.8+
  • Docker (for containerized deployment)
  • OpenAI API Key
  • YouTube Data API Key (optional, but recommended for better metadata)

Local Development

  1. Clone the repository:

    git clone <repository-url>
    cd yt-to-linkedin
    
  2. Create a virtual environment and install dependencies:

    python -m venv venv
    source venv/bin/activate  # On Windows: venv\Scripts\activate
    pip install -r requirements.txt
    
  3. Create a .env file in the project root with your API keys:

    OPENAI_API_KEY=your_openai_api_key
    YOUTUBE_API_KEY=your_youtube_api_key
    
  4. Run the application:

    uvicorn app.main:app --reload
    
  5. Access the API documentation at http://localhost:8000/docs

Docker Deployment

  1. Build the Docker image:

    docker build -t yt-to-linkedin-mcp .
    
  2. Run the container:

    docker run -p 8000:8000 --env-file .env yt-to-linkedin-mcp
    

Smithery Deployment

  1. Ensure you have the Smithery CLI installed and configured.

  2. Deploy to Smithery:

    smithery deploy
    

API Endpoints

1. Transcript Extraction

Endpoint: /api/v1/transcript
Method: POST
Description: Extract transcript from a YouTube video

Request Body:

{
  "youtube_url": "https://www.youtube.com/watch?v=VIDEO_ID",
  "language": "en",
  "youtube_api_key": "your_youtube_api_key"  // Optional, provide your own YouTube API key
}

Response:

{
  "video_id": "VIDEO_ID",
  "video_title": "Video Title",
  "transcript": "Full transcript text...",
  "language": "en",
  "duration_seconds": 600,
  "channel_name": "Channel Name",
  "error": null
}

2. Transcript Summarization

Endpoint: /api/v1/summarize
Method: POST
Description: Generate a summary from a video transcript

Request Body:

{
  "transcript": "Video transcript text...",
  "video_title": "Video Title",
  "tone": "professional",
  "audience": "general",
  "max_length": 250,
  "min_length": 150,
  "openai_api_key": "your_openai_api_key"  // Optional, provide your own OpenAI API key
}

Response:

{
  "summary": "Generated summary text...",
  "word_count": 200,
  "key_points": [
    "Key point 1",
    "Key point 2",
    "Key point 3"
  ]
}

3. LinkedIn Post Generation

Endpoint: /api/v1/generate-post
Method: POST
Description: Generate a LinkedIn post from a video summary

Request Body:

{
  "summary": "Video summary text...",
  "video_title": "Video Title",
  "video_url": "https://www.youtube.com/watch?v=VIDEO_ID",
  "speaker_name": "Speaker Name",
  "hashtags": ["ai", "machinelearning"],
  "tone": "professional",
  "voice": "first_person",
  "audience": "technical",
  "include_call_to_action": true,
  "max_length": 1200,
  "openai_api_key": "your_openai_api_key"  // Optional, provide your own OpenAI API key
}

Response:

{
  "post_content": "Generated LinkedIn post content...",
  "character_count": 800,
  "estimated_read_time": "About 1 minute",
  "hashtags_used": ["#ai", "#machinelearning"]
}

4. Output Formatting

Endpoint: /api/v1/output
Method: POST
Description: Format the LinkedIn post for output

Request Body:

{
  "post_content": "LinkedIn post content...",
  "format": "json"
}

Response:

{
  "content": {
    "post_content": "LinkedIn post content...",
    "character_count": 800
  },
  "format": "json"
}

Environment Variables

Variable Description Required
OPENAI_API_KEY OpenAI API key for summarization and post generation No (can be provided in requests)
YOUTUBE_API_KEY YouTube Data API key for fetching video metadata No (can be provided in requests)
PORT Port to run the server on (default: 8000) No

Note: While environment variables for API keys are optional (as they can be provided in each request), it's recommended to set them for local development and testing. When deploying to Smithery, users will need to provide their own API keys in the requests.

License

MIT

相关推荐

  • NiKole Maxwell
  • I craft unique cereal names, stories, and ridiculously cute Cereal Baby images.

  • https://suefel.com
  • Latest advice and best practices for custom GPT development.

  • Yusuf Emre Yeşilyurt
  • I find academic articles and books for research and literature reviews.

  • https://maiplestudio.com
  • Find Exhibitors, Speakers and more

  • Bora Yalcin
  • Evaluator for marketplace product descriptions, checks for relevancy and keyword stuffing.

  • Carlos Ferrin
  • Encuentra películas y series en plataformas de streaming.

  • Joshua Armstrong
  • Confidential guide on numerology and astrology, based of GG33 Public information

  • Contraband Interactive
  • Emulating Dr. Jordan B. Peterson's style in providing life advice and insights.

  • rustassistant.com
  • Your go-to expert in the Rust ecosystem, specializing in precise code interpretation, up-to-date crate version checking, and in-depth source code analysis. I offer accurate, context-aware insights for all your Rust programming questions.

  • Elijah Ng Shi Yi
  • Advanced software engineer GPT that excels through nailing the basics.

  • Emmet Halm
  • Converts Figma frames into front-end code for various mobile frameworks.

  • Alexandru Strujac
  • Efficient thumbnail creator for YouTube videos

  • apappascs
  • 发现市场上最全面,最新的MCP服务器集合。该存储库充当集中式枢纽,提供了广泛的开源和专有MCP服务器目录,并提供功能,文档链接和贡献者。

  • modelcontextprotocol
  • 模型上下文协议服务器

  • Mintplex-Labs
  • 带有内置抹布,AI代理,无代理构建器,MCP兼容性等的多合一桌面和Docker AI应用程序。

  • ShrimpingIt
  • MCP系列GPIO Expander的基于Micropython I2C的操作,源自ADAFRUIT_MCP230XX

  • n8n-io
  • 具有本机AI功能的公平代码工作流程自动化平台。将视觉构建与自定义代码,自宿主或云相结合,400+集成。

  • open-webui
  • 用户友好的AI接口(支持Ollama,OpenAi API,...)

  • WangRongsheng
  • 🧑‍🚀 llm 资料总结(数据处理、模型训练、模型部署、 o1 模型、mcp 、小语言模型、视觉语言模型)|摘要世界上最好的LLM资源。

    Reviews

    4 (1)
    Avatar
    user_m7udtNQ1
    2025-04-15

    I've been using LOTUS-MCP by blue-lotus-org and I couldn't be more impressed. It’s highly reliable, efficient, and truly game-changing. The interface is intuitive, which makes navigation easy even for beginners. For any professional seeking a robust MCP application, LOTUS-MCP is undoubtedly a top choice. Highly recommend checking it out at https://mcp.so/server/MCP/blue-lotus-org!