Confidential guide on numerology and astrology, based of GG33 Public information

mcp-teams-server
An MCP (Model Context Protocol) server implementation for Microsoft Teams integration, providing capabilities to read messages, create messages, reply to messages, mention members.
3 years
Works with Finder
1
Github Watches
5
Github Forks
48
Github Stars
MCP Teams Server
An MCP (Model Context Protocol) server implementation for Microsoft Teams integration, providing capabilities to read messages, create messages, reply to messages, mention members.
Features
https://github.com/user-attachments/assets/548a9768-1119-4a2d-bd5c-6b41069fc522
- Start thread in channel with title and contents, mentioning users
- Update existing threads with message replies, mentioning users
- Read thread replies
- List channel team members
- Read channel messages
Prerequisites
- uv package manager
- Python 3.10
- Microsoft Teams account with proper set-up
Installation
- Clone the repository:
git clone [repository-url]
cd mcp-teams-server
- Create a virtual environment and install dependencies:
uv venv
uv sync --frozen --all-extras --dev
Teams configuration
Please read this document to help you to configure Microsoft Teams and required Azure resources. It is not a step-by-step guide but can help you figure out what you will need.
Usage
Set up the following environment variables in your shell or in an .env file. You can use sample file as a template:
Key | Description |
---|---|
TEAMS_APP_ID | UUID for your MS Entra ID application ID |
TEAMS_APP_PASSWORD | Client secret |
TEAMS_APP_TYPE | SingleTenant or MultiTenant |
TEAMS_APP_TENANT_ID | Tenant uuid in case of SingleTenant |
TEAM_ID | MS Teams Group Id or Team Id |
TEAMS_CHANNEL_ID | MS Teams Channel ID with url escaped chars |
Start the server:
uv run mcp-teams-server
Development
Integration tests require the set-up the following environment variables:
Key | Description |
---|---|
TEST_THREAD_ID | timestamp of the thread id |
TEST_MESSAGE_ID | timestamp of the message id |
TEST_USER_NAME | test user name |
uv run pytest -m integration
Pre-built docker image
There is a pre-built image hosted in ghcr.io. You can install this image by running the following command
docker pull ghcr.io/inditextech/mcp-teams-server:latest
Build docker image
A docker image is available to run MCP server. You can build it with the following command:
docker build . -t inditextech/mcp-teams-server
Run docker image
Basic run configuration:
docker run -it inditextech/mcp-teams-server
Run with environment variables from .env file:
docker run --env-file .env -it inditextech/mcp-teams-server
Setup LLM to use MCP Teams Server
Please follow instructions on the following document
Changelog
See CHANGELOG.md for a list of changes and version history.
Contributing
Please read CONTRIBUTING.md for details on our code of conduct and the process for submitting pull requests.
Security
For security concerns, please see our Security Policy.
License
This project is licensed under the Apache-2.0 file for details.
© 2025 INDUSTRIA DE DISEÑO TEXTIL S.A. (INDITEX S.A.)
相关推荐
Converts Figma frames into front-end code for various mobile frameworks.
Advanced software engineer GPT that excels through nailing the basics.
Take an adjectivised noun, and create images making it progressively more adjective!
Siri Shortcut Finder – your go-to place for discovering amazing Siri Shortcuts with ease
I find academic articles and books for research and literature reviews.
Discover the most comprehensive and up-to-date collection of MCP servers in the market. This repository serves as a centralized hub, offering an extensive catalog of open-source and proprietary MCP servers, complete with features, documentation links, and contributors.
Micropython I2C-based manipulation of the MCP series GPIO expander, derived from Adafruit_MCP230xx
Mirror ofhttps://github.com/agentience/practices_mcp_server
Mirror ofhttps://github.com/bitrefill/bitrefill-mcp-server
An AI chat bot for small and medium-sized teams, supporting models such as Deepseek, Open AI, Claude, and Gemini. 专为中小团队设计的 AI 聊天应用,支持 Deepseek、Open AI、Claude、Gemini 等模型。
Bridge between Ollama and MCP servers, enabling local LLMs to use Model Context Protocol tools
Reviews

user_vNJAoQkW
I have been thoroughly impressed with mcp-teams-server by InditexTech. It's a robust and efficient solution for managing team communications. The setup was smooth, and the performance has been consistently reliable. This tool has significantly streamlined our team's workflow, and the support from the developers has been top-notch. Highly recommend for any organization looking to enhance team collaboration! https://github.com/InditexTech/mcp-teams-server