Skip to main content
Glama

LibreModel MCP Server ๐Ÿค–

A Model Context Protocol (MCP) server that bridges Claude Desktop with your local LLM instance running via llama-server.

Features

  • ๐Ÿ’ฌ Full conversation support with Local Model through Claude Desktop

  • ๐ŸŽ›๏ธ Complete parameter control (temperature, max_tokens, top_p, top_k)

  • โœ… Health monitoring and server status checks

  • ๐Ÿงช Built-in testing tools for different capabilities

  • ๐Ÿ“Š Performance metrics and token usage tracking

  • ๐Ÿ”ง Easy configuration via environment variables

Related MCP server: Claude-LMStudio Bridge

Quick Start

npm install @openconstruct/llama-mcp-server

A Model Context Protocol (MCP) server that bridges Claude Desktop with your local LLM instance running via llama-server.

Features

  • ๐Ÿ’ฌ Full conversation support with LibreModel through Claude Desktop

  • ๐ŸŽ›๏ธ Complete parameter control (temperature, max_tokens, top_p, top_k)

  • โœ… Health monitoring and server status checks

  • ๐Ÿงช Built-in testing tools for different capabilities

  • ๐Ÿ“Š Performance metrics and token usage tracking

  • ๐Ÿ”ง Easy configuration via environment variables

Quick Start

1. Install Dependencies

cd llama-mcp npm install

2. Build the Server

npm run build

3. Start Your LibreModel

Make sure llama-server is running with your model:

./llama-server -m lm37.gguf -c 2048 --port 8080

4. Configure Claude Desktop

Add this to your Claude Desktop configuration (~/.config/claude/claude_desktop_config.json):

{ "mcpServers": { "libremodel": { "command": "node", "args": ["/home/jerr/llama-mcp/dist/index.js"] } } }

5. Restart Claude Desktop

Claude will now have access to LibreModel through MCP!

Usage

Once configured, you can use these tools in Claude Desktop:

๐Ÿ’ฌ chat - Main conversation tool

Use the chat tool to ask LibreModel: "What is your name and what can you do?"

๐Ÿงช quick_test - Test LibreModel capabilities

Run a quick_test with type "creative" to see if LibreModel can write poetry

๐Ÿฅ health_check - Monitor server status

Use health_check to see if LibreModel is running properly

Configuration

Set environment variables to customize behavior:

export LLAMA_SERVER_URL="http://localhost:8080" # Default llama-server URL

Available Tools

Tool

Description

Parameters

chat

Converse with MOdel

message

,

temperature

,

max_tokens

,

top_p

,

top_k

,

system_prompt

quick_test

Run predefined capability tests

test_type

(hello/math/creative/knowledge)

health_check

Check server health and status

None

Resources

  • Configuration: View current server settings

  • Instructions: Detailed usage guide and setup instructions

Development

# Install dependencies npm install # LibreModel MCP Server ๐Ÿค– A Model Context Protocol (MCP) server that bridges Claude Desktop with your local LLM instance running via llama-server. ## Features - ๐Ÿ’ฌ **Full conversation support** with LibreModel through Claude Desktop - ๐ŸŽ›๏ธ **Complete parameter control** (temperature, max_tokens, top_p, top_k) - โœ… **Health monitoring** and server status checks - ๐Ÿงช **Built-in testing tools** for different capabilities - ๐Ÿ“Š **Performance metrics** and token usage tracking - ๐Ÿ”ง **Easy configuration** via environment variables ## Quick Start ### 1. Install Dependencies ```bash cd llama-mcp npm install

2. Build the Server

npm run build

3. Start Your LibreModel

Make sure llama-server is running with your model:

./llama-server -m lm37.gguf -c 2048 --port 8080

4. Configure Claude Desktop

Add this to your Claude Desktop configuration (~/.config/claude/claude_desktop_config.json):

{ "mcpServers": { "libremodel": { "command": "node", "args": ["/home/jerr/llama-mcp/dist/index.js"] } } }

5. Restart Claude Desktop

Claude will now have access to LibreModel through MCP!

Usage

Once configured, you can use these tools in Claude Desktop:

๐Ÿ’ฌ chat - Main conversation tool

Use the chat tool to ask LibreModel: "What is your name and what can you do?"

๐Ÿงช quick_test - Test LibreModel capabilities

Run a quick_test with type "creative" to see if LibreModel can write poetry

๐Ÿฅ health_check - Monitor server status

Use health_check to see if LibreModel is running properly

Configuration

Set environment variables to customize behavior:

export LLAMA_SERVER_URL="http://localhost:8080" # Default llama-server URL

Available Tools

Tool

Description

Parameters

chat

Converse with MOdel

message

,

temperature

,

max_tokens

,

top_p

,

top_k

,

system_prompt

quick_test

Run predefined capability tests

test_type

(hello/math/creative/knowledge)

health_check

Check server health and status

None

Resources

  • Configuration: View current server settings

  • Instructions: Detailed usage guide and setup instructions

Development

# Install dependencies npm install openconstruct/llama-mcp-server # Development mode (auto-rebuild) npm run dev # Build for production npm run build # Start the server directly npm start

Architecture

Claude Desktop โ†โ†’ LLama MCP Server โ†โ†’ llama-server API โ†โ†’ Local Model

The MCP server acts as a bridge, translating MCP protocol messages into llama-server API calls and formatting responses for Claude Desktop.

Troubleshooting

"Cannot reach LLama server"

  • Ensure llama-server is running on the configured port

  • Check that the model is loaded and responding

  • Verify firewall/network settings

"Tool not found in Claude Desktop"

  • Restart Claude Desktop after configuration changes

  • Check that the path to index.js is correct and absolute

  • Verify the MCP server builds without errors

Poor response quality

  • Adjust temperature and sampling parameters

  • Try different system prompts

License

CC0-1.0 - Public Domain. Use freely!


Built with โค๏ธ for open-source AI and the LibreModel project. by Claude Sonnet4

Development mode (auto-rebuild)

npm run dev

Build for production

npm run build

Start the server directly

npm start

## Architecture

Claude Desktop โ†โ†’ LLama MCP Server โ†โ†’ llama-server API โ†โ†’ Local Model

The MCP server acts as a bridge, translating MCP protocol messages into llama-server API calls and formatting responses for Claude Desktop. ## Troubleshooting **"Cannot reach LLama server"** - Ensure llama-server is running on the configured port - Check that the model is loaded and responding - Verify firewall/network settings **"Tool not found in Claude Desktop"** - Restart Claude Desktop after configuration changes - Check that the path to `index.js` is correct and absolute - Verify the MCP server builds without errors **Poor response quality** - Adjust temperature and sampling parameters - Try different system prompts ## License CC0-1.0 - Public Domain. Use freely! --- Built with โค๏ธ for open-source AI and the LibreModel project. by Claude Sonnet4 ### 1. Install Dependencies ```bash cd llama-mcp npm install

2. Build the Server

npm run build

3. Start Your LibreModel

Make sure llama-server is running with your model:

./llama-server -m lm37.gguf -c 2048 --port 8080

4. Configure Claude Desktop

Add this to your Claude Desktop configuration (~/.config/claude/claude_desktop_config.json):

{ "mcpServers": { "libremodel": { "command": "node", "args": ["/home/jerr/llama-mcp/dist/index.js"] } } }

5. Restart Claude Desktop

Claude will now have access to LibreModel through MCP!

Usage

Once configured, you can use these tools in Claude Desktop:

๐Ÿ’ฌ chat - Main conversation tool

Use the chat tool to ask LibreModel: "What is your name and what can you do?"

๐Ÿงช quick_test - Test LibreModel capabilities

Run a quick_test with type "creative" to see if LibreModel can write poetry

๐Ÿฅ health_check - Monitor server status

Use health_check to see if LibreModel is running properly

Configuration

Set environment variables to customize behavior:

export LLAMA_SERVER_URL="http://localhost:8080" # Default llama-server URL

Available Tools

Tool

Description

Parameters

chat

Converse with MOdel

message

,

temperature

,

max_tokens

,

top_p

,

top_k

,

system_prompt

quick_test

Run predefined capability tests

test_type

(hello/math/creative/knowledge)

health_check

Check server health and status

None

Resources

  • Configuration: View current server settings

  • Instructions: Detailed usage guide and setup instructions

Development

# Install dependencies npm install # Development mode (auto-rebuild) npm run dev # Build for production npm run build # Start the server directly npm start

Architecture

Claude Desktop โ†โ†’ LLama MCP Server โ†โ†’ llama-server API โ†โ†’ Local Model

The MCP server acts as a bridge, translating MCP protocol messages into llama-server API calls and formatting responses for Claude Desktop.

Troubleshooting

"Cannot reach LLama server"

  • Ensure llama-server is running on the configured port

  • Check that the model is loaded and responding

  • Verify firewall/network settings

"Tool not found in Claude Desktop"

  • Restart Claude Desktop after configuration changes

  • Check that the path to index.js is correct and absolute

  • Verify the MCP server builds without errors

Poor response quality

  • Adjust temperature and sampling parameters

  • Try different system prompts

License

CC0-1.0 - Public Domain. Use freely!


Built with โค๏ธ for open-source AI and the LibreModel project. by Claude Sonnet4

Deploy Server
A
security โ€“ no known vulnerabilities
F
license - not found
A
quality - confirmed to work

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/openconstruct/llama-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server