Skip to main content
Glama

PDF Knowledgebase MCP Server

by juanqui

Server Configuration

Describes the environment variables required to run the server.

NameRequiredDescriptionDefault
LLM_DPINoDPI setting for LLM parser150
CACHE_DIRNoCache directory for processing./.cache
LLM_MODELNoModel to use with LLM parseranthropic/claude-3.5-sonnet
LOG_LEVELNoLogging levelINFO
CHUNK_SIZENoTarget chunk size for LangChain chunker1000
PDF_PARSERNoParser to use: marker, pymupdf4llm, mineru, docling, llmmarker
MINERU_LANGNoLanguage setting for MinerU parseren
MINERU_VRAMNoVRAM setting for MinerU parser
PDF_CHUNKERNoChunking strategy to useunstructured
CHUNK_OVERLAPNoLangChain chunk overlap200
MINERU_METHODNoMethod setting for MinerU parserauto
DOCLING_DO_OCRNoEnable OCR for Docling parserfalse
OPENAI_API_KEYYesOpenAI API key for embeddings (required)
EMBEDDING_MODELNoOpenAI embedding modeltext-embedding-3-large
LLM_CONCURRENCYNoConcurrency setting for LLM parser5
VECTOR_SEARCH_KNoDefault search results5
DOCLING_TABLE_MODENoTable extraction mode for Docling parserACCURATE
FILE_SCAN_INTERVALNoFile monitoring interval in seconds60
KNOWLEDGEBASE_PATHNoDirectory containing PDF files./pdfs
OPENROUTER_API_KEYNoOpenRouter API key, required for LLM parser
EMBEDDING_BATCH_SIZENoEmbedding batch size100
DOCLING_OCR_LANGUAGESNoLanguages for OCR with Docling parseren
DOCLING_DO_TABLE_STRUCTURENoEnable table structure extraction for Docling parsertrue

Schema

Prompts

Interactive templates invoked by user choice

NameDescription

No prompts

Resources

Contextual data attached and managed by the client

NameDescription

No resources

Tools

Functions exposed to the LLM to take actions

NameDescription

No tools

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/juanqui/pdfkb-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server