MCP Docling Server
An MCP server that provides document processing capabilities using the Docling library.
Installation
You can install the package using pip:
Usage
Start the server using either stdio (default) or SSE transport:
If you're using uv, you can run the server directly without installing:
Available Tools
The server exposes the following tools:
- convert_document: Convert a document from a URL or local path to markdown format
source
: URL or local file path to the document (required)enable_ocr
: Whether to enable OCR for scanned documents (optional, default: false)ocr_language
: List of language codes for OCR, e.g. ["en", "fr"] (optional)
- convert_document_with_images: Convert a document and extract embedded images
source
: URL or local file path to the document (required)enable_ocr
: Whether to enable OCR for scanned documents (optional, default: false)ocr_language
: List of language codes for OCR (optional)
- extract_tables: Extract tables from a document as structured data
source
: URL or local file path to the document (required)
- convert_batch: Process multiple documents in batch mode
sources
: List of URLs or file paths to documents (required)enable_ocr
: Whether to enable OCR for scanned documents (optional, default: false)ocr_language
: List of language codes for OCR (optional)
- qna_from_document: Create a Q&A document from a URL or local path to YAML format
source
: URL or local file path to the document (required)no_of_qnas
: Number of expected Q&As (optional, default: 5)- Note: This tool requires IBM Watson X credentials to be set as environment variables:
WATSONX_PROJECT_ID
: Your Watson X project IDWATSONX_APIKEY
: Your IBM Cloud API keyWATSONX_URL
: The Watson X API URL (default: https://us-south.ml.cloud.ibm.com)
- get_system_info: Get information about system configuration and acceleration status
Example with Llama Stack
https://github.com/user-attachments/assets/8ad34e50-cbf7-4ec8-aedd-71c42a5de0a1
You can use this server with Llama Stack to provide document processing capabilities to your LLM applications. Make sure you have a running Llama Stack server, then configure your INFERENCE_MODEL
Caching
The server caches processed documents in ~/.cache/mcp-docling/
to improve performance for repeated requests.
This server cannot be installed
hybrid server
The server is able to function both locally and remotely, depending on the configuration or use case.
모델 컨텍스트 프로토콜을 사용하여 문서 처리 기능을 제공하는 서버로, 문서를 마크다운으로 변환하고, 표를 추출하고, 문서 이미지를 처리할 수 있습니다.
Related Resources
Related MCP Servers
- AsecurityAlicenseAqualityA Model Context Protocol server that provides web content fetching and conversion capabilities.Last updated -45802MIT License
- AsecurityAlicenseAqualityA Model Context Protocol server that converts various file formats (PDF, PowerPoint, Word, Excel, Images, etc.) to Markdown to make them accessible to LLMs.Last updated -1MIT License
- AsecurityFlicenseAqualityA Model Context Protocol server that intelligently fetches and processes web content, transforming websites and documentation into clean, structured markdown with nested URL crawling capabilities.Last updated -25223
- AsecurityFlicenseAqualityA comprehensive Model Context Protocol server that processes Microsoft Word documents with full formatting support, enabling text extraction, HTML/Markdown conversion, structure analysis, and image extraction.Last updated -51