Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@Universal MCP Serverfetch the latest news headlines from BBC"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
MCP Server for LLMs
This is a basic implementation of an MCP (Model Context Protocol) server that can be used with LLMs, including GGUF models.
Features
Resource management (create, read, update, delete)
Tool management (create, read, delete)
Prompt handling with LLM integration
RESTful API
Automatic GGUF model detection and loading
Model selection via configuration
Built-in and dynamic tool support
Web data fetching and file downloading
Secure command execution
File read/write operations
System information retrieval
Related MCP server: MCP Documentation Server
Setup
Create a virtual environment:
python3 -m venv venvActivate the virtual environment:
source venv/bin/activateInstall dependencies:
pip install -r requirements.txt
Configuration
The server can be configured using the config.json file:
server: Server host and port settingslogging: Logging levelllm: LLM configuration including:model_path: Directory to search for GGUF modelsmodel_file: Specific model file to use (optional, will use first found if not specified)n_ctx: Context window sizen_threads: Number of CPU threads to use
cors: CORS settings for web integration
By default, the server looks for GGUF models in /home/kali/.lmstudio/models/lmstudio-community/. You can change this in the config file.
To see what models are available, run:
python check_models.pyRunning the Server
You can start the server in several ways:
Using the run script (recommended):
./run.shDirectly with Python:
python main.pyIn the background with logging:
nohup python3 main.py > server.log 2>&1 &
The server will start on http://localhost:3000.
Stopping the Server
To stop the server, you can use the stop script:
./stop.shOr if running in the foreground, use Ctrl+C to stop it.
If you started it manually in the background, you can stop it with:
pkill -f "python3 main.py"API Endpoints
GET /- Server informationGET /resources- List all resourcesGET /resources/{uri}- Get a specific resourcePOST /resources- Create a new resourcePUT /resources/{uri}- Update a resourceDELETE /resources/{uri}- Delete a resourceGET /tools- List all toolsGET /tools/{name}- Get a specific toolPOST /tools- Create a new toolDELETE /tools/{name}- Delete a toolPOST /tools/{name}/execute- Execute a toolPOST /prompts- Handle a prompt request
Tool Support
The server includes several built-in tools:
get_current_time- Get the current date and timecalculate- Perform basic arithmetic calculationssearch_resources- Search for resources containing specific textget_system_info- Get comprehensive system information including OS, memory, CPU, and disk usage
Additionally, the server includes dynamically loaded tools from the tools/ directory:
fetch_web_data- Fetch data from a web URLdownload_file- Download a file from a URLexecute_command- Execute a safe shell command (limited to safe commands only)read_file- Read content from a file (restricted to temporary directories)write_file- Write content to a file (restricted to temporary directories)
You can create custom tool plugins in the tools/ directory. See tools/example_tool.py for an example.
Security Considerations
The server implements several security measures:
File operations are restricted to temporary directories only
Command execution is limited to a whitelist of safe commands
Web requests include appropriate headers to avoid blocking by servers
Testing
You can test the server with the provided test clients:
Basic test client:
python test_client.pyEnhanced tool test client:
python test_tool_client.pyComprehensive tools test client:
python test_comprehensive_tools.pyLLM test client:
python llm_client.pyMCP LLM test client:
python mcp_llm_client.pyModel checking script:
python check_models.py
Make sure the server is running before running the test clients.
Integration with LLMs
To use this MCP server with LLMs:
Start the server
Configure your LLM application to connect to
http://localhost:3000Use the MCP API to manage resources and tools that your LLM can access
The server will automatically use GGUF models from LM Studio or Ollama
The server automatically detects and loads the specified GGUF model from the configured model directory. It uses llama-cpp-python for inference.
Extending the Server
To add more functionality:
Add new endpoints in
main.pyImplement additional business logic
Add new Pydantic models for request/response validation
Create custom tool plugins in the
tools/directory
This server cannot be installed
Resources
Unclaimed servers have limited discoverability.
Looking for Admin?
If you are the server author, to access and configure the admin panel.