Transcriptor MCP is an MCP server that extracts video transcripts, subtitles, and metadata from 11 major platforms using yt-dlp, with Whisper fallback for audio transcription when subtitles are unavailable.
Get cleaned transcripts (
get_transcript): Fetch plain text transcripts from YouTube, Twitter/X, Instagram, TikTok, Twitch, Vimeo, Facebook, Bilibili, VK, Dailymotion, and Reddit — with auto-discovery of language and subtitle type.Get raw subtitles (
get_raw_subtitles): Retrieve raw SRT/VTT/ASS/LRC subtitle files with full pagination support, specifying language, type (official or auto-generated), and chunk size.List subtitle languages (
get_available_subtitles): Discover available official and auto-generated subtitle languages for any supported video.Fetch video metadata (
get_video_info): Get extended info including title, description, channel, duration, upload date, view/like/comment counts, tags, categories, thumbnails, and live status.Get chapter markers (
get_video_chapters): Extract chapter start/end times and titles from a video.Batch playlist transcripts (
get_playlist_transcripts): Fetch cleaned transcripts for multiple videos from a playlist in one call, with item selection and limits.Search YouTube videos (
search_videos): Search via yt-dlp with filters for result count, pagination offset, upload date, and output in JSON or Markdown format.Whisper fallback: Automatically transcribes audio via local Whisper or the OpenAI Whisper API when native subtitles are unavailable.
Pagination: Large transcripts and subtitle files are paginated using
response_limitandnext_cursorparameters.Production features: Optional Redis caching, Prometheus metrics, authentication support, and Docker containerization.
Flexible deployment: Available over stdio or HTTP/SSE, with direct integration for Cursor, Claude Code, n8n, Smithery, and Glama.
Enables the extraction of video transcripts and metadata from the Bilibili platform.
Allows for retrieving video transcripts and metadata from Dailymotion videos.
Supports fetching transcripts and metadata for videos hosted on Facebook.
Provides capabilities to extract transcripts and metadata from Instagram video content.
Integrates with OpenAI's Whisper API to provide audio transcription fallback when native subtitles are unavailable.
Enables fetching of transcripts and metadata from TikTok videos.
Supports extracting transcripts and metadata from Twitch videos and streams.
Allows for retrieving video transcripts and metadata from Vimeo.
Provides tools to fetch video transcripts, raw subtitles, and detailed metadata including chapters, tags, and thumbnails.
Transcriptor MCP
An MCP server (stdio + HTTP/SSE) that fetches video transcripts/subtitles via yt-dlp, with pagination for large responses. Supports YouTube, Twitter/X, Instagram, TikTok, Twitch, Vimeo, Facebook, Bilibili, VK, Dailymotion, Reddit. Whisper fallback — transcribes audio when subtitles are unavailable (local or OpenAI API). Works with Cursor and other MCP hosts.
Overview
This repository primarily ships an MCP server:
stdio: for local usage (e.g., Cursor running a local command).
HTTP/SSE: for remote usage (e.g., VPS + Tailscale).
It also includes an optional REST API (Fastify), but MCP is the primary focus.
Supported platforms
Unlike YouTube-only tools, Transcriptor MCP works across 11 major video platforms:
YouTube · Twitter/X · Instagram · TikTok · Twitch · Vimeo · Facebook · Bilibili · VK · Dailymotion · Reddit
All URL-based tools (get_transcript, get_raw_subtitles, get_available_subtitles, get_video_info, get_video_chapters, get_playlist_transcripts) accept video URLs from any supported platform. The search_videos tool is YouTube-specific (yt-dlp ytsearch).
When to use Transcriptor MCP
Transcriptor MCP is the best choice when you need transcripts and metadata for AI, summarization, or content analysis — without downloading video or audio files:
Transcripts and subtitles — cleaned text or raw SRT/VTT; multi-language; Whisper fallback when subtitles are unavailable (local or OpenAI).
Multi-platform — YouTube, Twitter/X, Instagram, TikTok, Twitch, Vimeo, Facebook, Bilibili, VK, Dailymotion, Reddit.
Remote and production — MCP over HTTP/SSE, optional auth, Redis cache, Prometheus metrics; connect in one click via Smithery with no local install.
No media downloads — we focus on text and metadata only. For downloading videos or audio.
See docs for step-by-step use cases: summarize video, search and transcript, IDE/Cursor/Claude, n8n automation, researchers and batch, and self-hosted/enterprise.
How to connect
Method | Links | Use case |
Smithery | smithery.ai/servers/samson-art/transcriptor-mcp, URL | No install, instant connection |
Glama | MCP directory, one-click install | |
Docker (stdio) | Image | Local Cursor, no Node |
Smithery — Add the MCP server by URL in your client (Cursor, Claude Code, etc.). No config needed for the public instance. Use tools like get_transcript or get_video_info right away.
For one-click install in VS Code: Install in VS Code Install in VS Code Insiders
Docker — See MCP quick start below.
Local Node — Build and run node dist/mcp.js; see MCP Server (stdio) below.
Features
Multi-platform — YouTube, Reddit, Twitter/X, Instagram, TikTok, Twitch, Vimeo, Facebook, Bilibili, VK, Dailymotion.
Connect by URL (Smithery, Glama) — use the server without installing Docker or Node; Smithery, Glama.
Transcripts + raw subtitles: cleaned text or raw SRT/VTT.
Language support: official subtitles with auto-generated fallback.
Video metadata: extended info (title, channel, tags, thumbnails, etc.) and chapter markers.
Pagination: safe for large transcripts.
Whisper fallback: when subtitles are unavailable, transcribes video audio via Whisper (local self-hosted or OpenAI API); configurable (see Configuration).
Optional Redis cache: cache subtitles and metadata to reduce yt-dlp calls; configurable (see Caching).
Docker-first: ready for local + remote deployment.
Production-friendly HTTP: optional auth + allowlists (see
CHANGELOG.md).Prometheus: metrics for API and MCP, list of failed subtitle extractions (see Monitoring).
Self-configurable: Whisper & caching
You can enable these features independently; both are off by default.
Whisper fallback — When native subtitles are unavailable, transcribe video audio via Whisper (local self-hosted or OpenAI API). Configure via
WHISPER_MODE,WHISPER_BASE_URL,WHISPER_API_KEY, etc. See Configuration.Redis caching — Cache subtitles and metadata to reduce yt-dlp calls. Configure via
CACHE_MODE=redisandCACHE_REDIS_URL. See Caching.
Example usage (screenshot)
Below is a real-world example of the same “summarize YouTube video” task without MCP vs with MCP:
MCP quick start (Docker and self-hosted)
For one-click connection without installing anything, use Smithery or Glama above. The sections below are for Docker or your own server.
Docker Hub (stdio)
Image:
artsamsonov/transcriptor-mcp:latest
Run locally (stdio mode):
docker run --rm -i artsamsonov/transcriptor-mcp:latestCursor MCP configuration (Docker)
Add to Cursor MCP settings (or create .cursor/mcp.json):
{
"mcpServers": {
"transcriptor": {
"command": "docker",
"args": ["run", "--rm", "-i", "artsamsonov/transcriptor-mcp:latest"]
}
}
}Remote MCP over HTTP/SSE (VPS + Tailscale)
Run the HTTP/SSE MCP server on your VPS (default port 4200) using docker-compose:
cp docker-compose.example.yml docker-compose.yml
docker compose up -d transcriptor-mcpClaude Code (HTTP / streamable HTTP):
claude mcp add --transport http transcriptor http://<tailscale-host>:4200/mcpCursor (SSE):
Add a new MCP server of type SSE with URL
http://<tailscale-host>:4200/sse
If you set MCP_AUTH_TOKEN, add Authorization: Bearer <token> in the client headers.
For more MCP configuration examples, see [docs/quick-start.mcp.md](docs/quick-start.mcp.md).
n8n MCP Client (streamable HTTP):
Use the MCP Server URL
http://<host>:4200/mcp(streamable HTTP transport).If n8n runs behind a reverse proxy that sets
X-Forwarded-For, setN8N_PROXY_HOPSto the number of proxy hops (commonly1) to avoidERR_ERL_UNEXPECTED_X_FORWARDED_FOR.
MCP tools
Tool | Purpose |
| Cleaned plain text (first chunk) |
| Raw SRT/VTT, paginated |
| List official/auto languages |
| Extended metadata |
| Chapter markers |
| Batch transcripts from playlist |
| YouTube search |
MCP tool reference
All URL-based tools share the same base input:
url(string, required) – Video URL from a supported platform or YouTube video ID. Supported: YouTube, Twitter/X, Instagram, TikTok, Twitch, Vimeo, Facebook, Bilibili, VK, Dailymotion, Reddit.
get_raw_subtitles supports pagination; get_transcript returns the first chunk only (no pagination input). Pagination parameters for get_raw_subtitles:
response_limit(number, optional) – max characters per response, default50000, min1000, max200000.next_cursor(string, optional) – opaque offset returned from the previous page; pass it to fetch the next chunk.
Each tool returns:
content– human-readable text (for MCP chat UIs).structuredContent– strongly typed JSON payload you can consume from automations or code.
get_transcript
Purpose: Fetch cleaned subtitles as plain text (no timestamps, HTML, or speaker metadata).
Input: Only url (video URL or ID). Type and language are auto-discovered; the tool returns the first chunk with default size (no pagination parameters).
Structured response:
videoId– resolved YouTube ID.type,lang– effective subtitle type and language.text– current text chunk.is_truncated–trueif more text is available.total_length– total length of the full transcript.start_offset,end_offset– character offsets of this chunk.next_cursor– present in response when truncated (omitted on the last page). Not accepted as input for this tool.
get_raw_subtitles
Purpose: Fetch raw subtitle file content (SRT or VTT) with pagination support.
Extra input fields:
type–"official"or"auto", optional.lang– subtitle language code, optional.response_limit,next_cursor– pagination (optional).
Structured response:
videoId,type,lang– same semantics as above.format–"srt"or"vtt"(auto-detected from content).content– raw subtitle text for this page.is_truncated,total_length,start_offset,end_offset,next_cursor– same pagination fields asget_transcript.
get_available_subtitles
Purpose: Inspect which languages are available for a video, split into official vs auto-generated tracks.
Input:
url– YouTube URL or video ID.
Structured response:
videoId– resolved YouTube ID.official– sorted list of language codes with official subtitles.auto– sorted list of language codes with auto-generated subtitles.
This is useful to first discover languages and then pick type/lang for get_raw_subtitles (or other tools).
get_video_info
Purpose: Fetch extended metadata about a video (based on yt-dlp JSON output).
Input:
url– YouTube URL or video ID.
Structured response (key fields):
videoId– resolved YouTube ID.title,description.uploader,uploaderId.channel,channelId,channelUrl.duration– in seconds.uploadDate–YYYYMMDDstring if available.webpageUrl.viewCount,likeCount,commentCount.tags,categories.liveStatus,isLive,wasLive,availability.thumbnail– primary thumbnail URL.thumbnails– list of thumbnail variants{ url, width?, height?, id? }.
See src/mcp-core.ts and src/youtube.ts for the full JSON schema used by the MCP SDK.
get_video_chapters
Purpose: Get chapter markers extracted by yt-dlp.
Input:
url– YouTube URL or video ID.
Structured response:
videoId– resolved YouTube ID.chapters– array of{ startTime: number; endTime: number; title: string }.
If the video has no chapters, chapters is an empty array; if yt-dlp cannot fetch chapter data at all, the tool returns an MCP error instead of structured chapters.
get_playlist_transcripts
Purpose: Fetch cleaned transcripts for multiple videos from a playlist in one call.
Input:
url(string, required) – Playlist URL or watch URL withlist=(e.g.https://www.youtube.com/playlist?list=XXX).type–"official"or"auto", optional.lang– Subtitle language code, optional.format– Subtitle format (srt,vtt,ass,lrc), optional.playlistItems– yt-dlp-Ispec (e.g.1:5,1,3,7,-1), optional.maxItems– Max videos to process, optional.
Structured response:
results– array of{ videoId, text }for each video in the playlist.
search_videos
Purpose: Search videos on YouTube via yt-dlp (ytsearch). Returns a list of videos with metadata.
Input:
query(string, required) – Search query.limit(number, optional) – Max results (default 10, max 50).offset(number, optional) – Skip first N results (pagination).uploadDateFilter(string, optional) – Filter by upload date:hour,today,week,month, oryear.response_format(string, optional) – Human-readable format:json(default) ormarkdown.
Structured response:
results– array of{ videoId, title, url, duration, uploader, viewCount, thumbnail }.
Requirements
Docker (recommended for production)
Node.js >= 20.0.0 (for local development)
yt-dlp (included in Docker image)
REST API (optional)
The repository also ships an HTTP API (Fastify).
Quick Docker usage
Build the image:
docker build -t transcriptor-mcp-api -f Dockerfile --target api .Run on the default port:
docker run -p 3000:3000 transcriptor-mcp-api
For a more complete REST quick start (including docker-compose and local Node.js),
see [docs/quick-start.rest.md](docs/quick-start.rest.md).
Swagger / OpenAPI
Once the REST API is running, interactive API docs are available at:
http://localhost:3000/docsIf you change PORT / HOST, adjust the URL accordingly, e.g. http://<HOST>:<PORT>/docs.
Troubleshooting: restricted / sign-in required videos
If yt-dlp is blocked by age gate, sign-in, or region restrictions, you will likely need
an authenticated cookies.txt file and the COOKIES_FILE_PATH environment variable.
The root of this repository includes a sample [cookies.example.txt](cookies.example.txt)
showing the expected Netscape cookies format. For a full guide on:
exporting real cookies
wiring them into Docker / docker-compose / local Node.js
and keeping them secure
see [docs/cookies.md](docs/cookies.md).
Run in background
docker run -d -p 3000:3000 --name transcriptor transcriptor-mcp-apiE2E smoke tests (REST API + MCP, Docker)
Before publishing Docker images, you can run a small e2e smoke test that:
Starts a REST API container and checks Swagger +
POST /subtitleswith a stable YouTube videoOptionally starts an MCP container and checks MCP stdio (initialize over stdin/stdout), streamable HTTP (
POST /mcpwith initialize), and SSE (GET /sse)
Run the smoke test (requires built images):
npm run build
docker build -t artsamsonov/transcriptor-mcp-api:latest -f Dockerfile --target api .
docker build -t artsamsonov/transcriptor-mcp:latest -f Dockerfile --target mcp .
npm run test:e2e:apiEnvironment variables:
Variable | Default | Description |
| — | Full API image reference (overrides name/tag). |
|
| API image name and tag. |
|
| API base URL and port. |
|
| Video used for |
| — | Set to |
| — | Full MCP image reference (overrides name/tag). |
|
| MCP image name and tag. |
|
| MCP base URL and port. |
| — | If set, passed to MCP container as |
Example: skip MCP and use a custom video:
SMOKE_SKIP_MCP=1 SMOKE_VIDEO_URL="https://www.youtube.com/watch?v=YOUR_ID" npm run test:e2e:apiView logs
docker logs -f transcriptorStop the container
docker stop transcriptor
docker rm transcriptorAPI Documentation
For detailed REST API endpoint documentation (request/response schemas, examples, etc.), use the built-in Swagger UI at:
http://localhost:3000/docsor see [docs/quick-start.rest.md](docs/quick-start.rest.md).
MCP Server (stdio)
This project also ships an MCP server over stdio. It reuses the same yt-dlp based extraction and can return full transcript text or raw subtitles. Cursor configuration examples are provided below, but it should work with any MCP host that supports stdio.
Pagination
Tools that return large text accept:
response_limit(default50000, min1000, max200000)next_cursor(string offset from a previous response)
If the response is truncated, the tool returns next_cursor so you can fetch the next chunk.
Local setup
npm install
npm run build
npm run start:mcpHTTP setup (remote)
npm run build
MCP_PORT=4200 MCP_HOST=0.0.0.0 npm run start:mcp:httpCursor MCP configuration (local)
Create .cursor/mcp.json (or add to your global Cursor MCP settings):
{
"mcpServers": {
"transcriptor": {
"command": "node",
"args": ["dist/mcp.js"]
}
}
}Docker setup
Build and run the MCP server in a container (stdio mode):
docker build -f Dockerfile --target mcp -t transcriptor-mcp .
docker run --rm -i transcriptor-mcpBuild and run the MCP server in a container (HTTP mode):
docker build -f Dockerfile --target mcp -t transcriptor-mcp .
docker run -p 4200:4200 -e MCP_PORT=4200 -e MCP_HOST=0.0.0.0 transcriptor-mcp npm run start:mcp:httpCursor MCP config for Docker:
{
"mcpServers": {
"transcriptor": {
"command": "docker",
"args": ["run", "--rm", "-i", "artsamsonov/transcriptor-mcp:latest"]
}
}
}How It Works
The API receives a video URL (YouTube or other supported platform) and parameters (subtitle type and language) from the client
Extracts the video ID from the URL
Uses
yt-dlpto download subtitles with the specified parameters:
Single
yt-dlpcommand call with explicit type (--write-subsor--write-auto-subs) and language (--sub-lang)
Parses the subtitle file (SRT/VTT) and removes:
Timestamps
Subtitle numbers
HTML tags
Formatting
Returns clean plain text (for
/subtitles) or raw content (for/subtitles/raw)
Development
Prerequisites
Node.js >= 20.0.0
npm or yarn
yt-dlp installed and available in PATH
Versioning
The app version is read from package.json at runtime ([src/version.ts](src/version.ts)). When cutting a release, update the version field in package.json, then create a git tag (e.g. v0.4.7). Changelog entries under [Unreleased] should be moved to the new version before tagging.
Scripts
npm run build- Build the TypeScript projectnpm start- Run the compiled applicationnpm run dev- Run with hot reload using ts-node-devnpm run start:mcp- Run the MCP server (stdio)npm run start:mcp:http- Run the MCP server (HTTP/SSE)npm run dev:mcp- Run the MCP server with hot reloadnpm test- Run testsnpm run test:watch- Run tests in watch modenpm run test:coverage- Run tests with coverage reportnpm run lint- Lint the codenpm run lint:fix- Fix linting errorsnpm run type-check- Type check without buildingnpm run format- Format code with Prettiernpm run format:check- Check code formatting
Project Structure
├── src/
│ ├── index.ts # Main application entry point
│ ├── mcp.ts # MCP server entry point (stdio)
│ ├── mcp-core.ts # MCP tools registration (shared)
│ ├── mcp-http.ts # MCP server entry point (HTTP/SSE)
│ ├── validation.ts # Request validation logic
│ └── youtube.ts # YouTube subtitle downloading and parsing
├── dist/ # Compiled JavaScript (generated)
├── Dockerfile # Docker image (API and MCP via --target api/mcp)
├── logo.webp # Project logo used in README
├── example-usage.webp # Example usage screenshot used in README
├── package.json
├── tsconfig.json
└── README.mdTechnologies
TypeScript - Type-safe JavaScript
Node.js - Runtime environment
Fastify - Fast and low overhead web framework
yt-dlp - YouTube content downloader
Docker - Containerization
Jest - Testing framework
ESLint - Code linting
Prettier - Code formatting
Security
Data and keys: Video URLs are sent to yt-dlp for subtitle extraction. Keys and tokens are stored only in your environment; we never log or share them.
Do not commit or log sensitive values. Use environment variables or a secret manager (e.g. vault, cloud secrets) for:
**WHISPER_API_KEY** – required when using Whisper API; never log or expose in client responses.**CACHE_REDIS_URL**– Redis connection string whenCACHE_MODE=redis; may contain credentials.**MCP_AUTH_TOKEN**– Bearer token for MCP HTTP; keep it secret.**COOKIES_FILE_PATH**– path to cookies; ensure the file is not committed and has restricted permissions.
See docs/cookies.md for safe handling of cookies.
Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
Fork the repository
Create your feature branch (
git checkout -b feature/amazing-feature)Commit your changes (
git commit -m 'Add some amazing feature')Push to the branch (
git push origin feature/amazing-feature)Open a Pull Request
Please make sure your code passes all tests and linting checks before submitting.
License
MIT License
Copyright (c) 2025 samson-art
See LICENSE file for details.
Support
Bug reports: GitHub Issues
Feature requests: GitHub Issues
Contact: GitHub Profile