whisper-telegram-mcp
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@whisper-telegram-mcptranscribe the audio file at /Users/me/Downloads/meeting.mp3"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
whisper-telegram-mcp
Transcribe Telegram voice messages with Whisper -- as an MCP tool for Claude
An MCP server that transcribes audio files and Telegram voice messages using OpenAI's Whisper speech recognition. Works with Claude Desktop, Claude Code, and any MCP-compatible client.
What It Does
Transcribe local audio files -- OGG, WAV, MP3, FLAC, and more
Transcribe Telegram voice messages -- pass a
file_id, get text backTwo backends -- local inference with faster-whisper (free, private) or OpenAI Whisper API (cloud)
Auto mode -- tries local first, falls back to OpenAI if it fails
Language detection -- automatic or specify an ISO-639-1 code
Word-level timestamps -- optional fine-grained timing
Quick Start
One command with uvx
uvx whisper-telegram-mcpNo installation needed -- uvx handles everything.
Or install with pip
pip install whisper-telegram-mcp
whisper-telegram-mcpIntegration
Claude Desktop
Add to your Claude Desktop config (~/Library/Application Support/Claude/claude_desktop_config.json):
{
"mcpServers": {
"whisper-telegram-mcp": {
"command": "uvx",
"args": ["whisper-telegram-mcp"],
"env": {
"WHISPER_MODEL": "base",
"WHISPER_BACKEND": "auto"
}
}
}
}Claude Code
Add to your project's .mcp.json:
{
"mcpServers": {
"whisper-telegram-mcp": {
"command": "uvx",
"args": ["whisper-telegram-mcp"],
"env": {
"WHISPER_MODEL": "base",
"WHISPER_BACKEND": "auto"
}
}
}
}Tools
Tool | Description |
| Transcribe a local audio file (OGG, WAV, MP3, etc.) to text |
| Download and transcribe a Telegram voice message by |
| List available Whisper model sizes with speed/accuracy info |
| Check which backends (local/OpenAI) are available and configured |
transcribe_audio
file_path: str # Absolute path to audio file
language: str | None # ISO-639-1 code (e.g. "en"), None = auto-detect
word_timestamps: bool # Include word-level timestamps (default: false)transcribe_telegram_voice
file_id: str # Telegram voice message file_id
bot_token: str | None # Bot token (falls back to TELEGRAM_BOT_TOKEN env var)
language: str | None # ISO-639-1 code, None = auto-detect
word_timestamps: bool # Include word-level timestamps (default: false)Response Format
All transcription tools return:
{
"text": "Hello, this is a voice message.",
"language": "en",
"language_probability": 0.98,
"duration": 3.5,
"segments": [
{"start": 0.0, "end": 3.5, "text": "Hello, this is a voice message."}
],
"backend": "local",
"success": true,
"error": null
}Configuration
All configuration is via environment variables:
Variable | Default | Description |
|
|
|
|
| Whisper model size (see below) |
| -- | Required for |
| -- | Required for |
| auto-detect | ISO-639-1 language code |
How It Works
MCP Client (Claude)
|
[MCP stdio]
|
whisper-telegram-mcp
/ \
/ \
transcribe_audio transcribe_telegram_voice
| |
| [Download via Bot API]
| |
+--------+-----------+
|
auto_transcribe()
/ \
LocalBackend OpenAIBackend
(faster-whisper) (Whisper API)Claude sends a tool call via MCP (stdio transport)
For Telegram voice messages, the file is downloaded via Bot API
auto_transcribe()picks the best available backendTranscription result is returned as structured JSON
Local vs OpenAI
Local (faster-whisper) | OpenAI API | |
Cost | Free | $0.006/min |
Privacy | All data stays on device | Audio sent to OpenAI |
Speed | ~1-10s depending on model | ~1-3s |
Setup | Automatic (downloads model on first use) | Requires |
Accuracy | Excellent with | Excellent |
Offline | Yes | No |
Model Sizes
Model | Parameters | Speed | Accuracy | VRAM |
| 39M | Fastest | Lowest | ~1GB |
| 74M | Fast | Good | ~1GB |
| 244M | Moderate | Better | ~2GB |
| 769M | Slow | High | ~5GB |
| 1550M | Slowest | Highest | ~10GB |
| ~800M | Fast | High | ~6GB |
English-only variants (tiny.en, base.en, small.en, medium.en) are slightly more accurate for English.
Development
git clone https://github.com/abid-mahdi/whisper-telegram-mcp.git
cd whisper-telegram-mcp
python3.12 -m venv .venv
source .venv/bin/activate
pip install -e ".[dev]"
# Run unit tests
pytest tests/ -v -m "not integration"
# Run integration tests (downloads ~150MB model on first run)
pytest tests/ -m integration -v
# Run with coverage
pytest tests/ --cov=src/whisper_telegram_mcp --cov-report=term-missingMCP Inspector
uvx mcp dev src/whisper_telegram_mcp/server.pyContributing
Fork the repository
Create a feature branch (
git checkout -b feat/amazing-feature)Run tests (
pytest tests/ -v -m "not integration")Commit with conventional commits (
feat:,fix:,docs:, etc.)Open a pull request
License
This server cannot be installed
Resources
Unclaimed servers have limited discoverability.
Looking for Admin?
If you are the server author, to access and configure the admin panel.
Latest Blog Posts
MCP directory API
We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/abid-mahdi/whisper-telegram-mcp'
If you have feedback or need assistance with the MCP directory API, please join our Discord server