Skip to main content
Glama

RAG Document Server

by jaimeferj

Server Configuration

Describes the environment variables required to run the server.

NameRequiredDescriptionDefault
LLM_MODELNoGoogle LLM modelgemini-1.5-flash
CHUNK_SIZENoSize of text chunks in characters1000
QDRANT_PATHNoPath to Qdrant storage./qdrant_storage
FASTAPI_HOSTNoFastAPI server host0.0.0.0
FASTAPI_PORTNoFastAPI server port8000
CHUNK_OVERLAPNoOverlap between chunks200
TOP_K_RESULTSNoNumber of chunks to retrieve5
GOOGLE_API_KEYYesGoogle AI Studio API key (required)
EMBEDDING_MODELNoGoogle embedding modeltext-embedding-004
QDRANT_COLLECTION_NAMENoQdrant collection namedocuments

Schema

Prompts

Interactive templates invoked by user choice

NameDescription

No prompts

Resources

Contextual data attached and managed by the client

NameDescription

No resources

Tools

Functions exposed to the LLM to take actions

NameDescription

No tools

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/jaimeferj/mcp-rag-docs'

If you have feedback or need assistance with the MCP directory API, please join our Discord server