Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@MCP LLM Integration Serversummarize this article about climate change"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
MCP LLM Integration Server
This is a Model Context Protocol (MCP) server that allows you to integrate local LLM capabilities with MCP-compatible clients.
Features
llm_predict: Process text prompts through a local LLM
echo: Echo back text for testing purposes
Setup
Install dependencies:
source .venv/bin/activate uv pip install mcpTest the server:
python -c " import asyncio from main import server, list_tools, call_tool async def test(): tools = await list_tools() print(f'Available tools: {[t.name for t in tools]}') result = await call_tool('echo', {'text': 'Hello!'}) print(f'Result: {result[0].text}') asyncio.run(test()) "
Integration with LLM Clients
For Claude Desktop
Add this to your Claude Desktop configuration (~/.config/claude-desktop/claude_desktop_config.json):
For Continue.dev
Add this to your Continue configuration (~/.continue/config.json):
For Cline
Add this to your Cline MCP settings:
Customizing the LLM Integration
To integrate your own local LLM, modify the perform_llm_inference function in main.py:
Testing
Run the server directly to test JSON-RPC communication:
Then send JSON-RPC requests via stdin: