Skip to main content
Glama

LibreModel MCP Server πŸ€–

A Model Context Protocol (MCP) server that bridges Claude Desktop with your local LLM instance running via llama-server.

Features

  • πŸ’¬ Full conversation support with Local Model through Claude Desktop

  • πŸŽ›οΈ Complete parameter control (temperature, max_tokens, top_p, top_k)

  • βœ… Health monitoring and server status checks

  • πŸ§ͺ Built-in testing tools for different capabilities

  • πŸ“Š Performance metrics and token usage tracking

  • πŸ”§ Easy configuration via environment variables

Related MCP server: Claude-LMStudio Bridge

Quick Start

npm install @openconstruct/llama-mcp-server

A Model Context Protocol (MCP) server that bridges Claude Desktop with your local LLM instance running via llama-server.

Features

  • πŸ’¬ Full conversation support with LibreModel through Claude Desktop

  • πŸŽ›οΈ Complete parameter control (temperature, max_tokens, top_p, top_k)

  • βœ… Health monitoring and server status checks

  • πŸ§ͺ Built-in testing tools for different capabilities

  • πŸ“Š Performance metrics and token usage tracking

  • πŸ”§ Easy configuration via environment variables

Quick Start

1. Install Dependencies

cd llama-mcp
npm install

2. Build the Server

npm run build

3. Start Your LibreModel

Make sure llama-server is running with your model:

./llama-server -m lm37.gguf -c 2048 --port 8080

4. Configure Claude Desktop

Add this to your Claude Desktop configuration (~/.config/claude/claude_desktop_config.json):

{
  "mcpServers": {
    "libremodel": {
      "command": "node",
      "args": ["/home/jerr/llama-mcp/dist/index.js"]
    }
  }
}

5. Restart Claude Desktop

Claude will now have access to LibreModel through MCP!

Usage

Once configured, you can use these tools in Claude Desktop:

πŸ’¬ chat - Main conversation tool

Use the chat tool to ask LibreModel: "What is your name and what can you do?"

πŸ§ͺ quick_test - Test LibreModel capabilities

Run a quick_test with type "creative" to see if LibreModel can write poetry

πŸ₯ health_check - Monitor server status

Use health_check to see if LibreModel is running properly

Configuration

Set environment variables to customize behavior:

export LLAMA_SERVER_URL="http://localhost:8080"  # Default llama-server URL

Available Tools

Tool

Description

Parameters

chat

Converse with MOdel

message, temperature, max_tokens, top_p, top_k, system_prompt

quick_test

Run predefined capability tests

test_type (hello/math/creative/knowledge)

health_check

Check server health and status

None

Resources

  • Configuration: View current server settings

  • Instructions: Detailed usage guide and setup instructions

Development

# Install dependencies
npm install # LibreModel MCP Server πŸ€–

A Model Context Protocol (MCP) server that bridges Claude Desktop with your local LLM instance running via llama-server.

## Features

- πŸ’¬ **Full conversation support** with LibreModel through Claude Desktop
- πŸŽ›οΈ **Complete parameter control** (temperature, max_tokens, top_p, top_k)
- βœ… **Health monitoring** and server status checks
- πŸ§ͺ **Built-in testing tools** for different capabilities
- πŸ“Š **Performance metrics** and token usage tracking
- πŸ”§ **Easy configuration** via environment variables

## Quick Start

### 1. Install Dependencies

```bash
cd llama-mcp
npm install

2. Build the Server

npm run build

3. Start Your LibreModel

Make sure llama-server is running with your model:

./llama-server -m lm37.gguf -c 2048 --port 8080

4. Configure Claude Desktop

Add this to your Claude Desktop configuration (~/.config/claude/claude_desktop_config.json):

{
  "mcpServers": {
    "libremodel": {
      "command": "node",
      "args": ["/home/jerr/llama-mcp/dist/index.js"]
    }
  }
}

5. Restart Claude Desktop

Claude will now have access to LibreModel through MCP!

Usage

Once configured, you can use these tools in Claude Desktop:

πŸ’¬ chat - Main conversation tool

Use the chat tool to ask LibreModel: "What is your name and what can you do?"

πŸ§ͺ quick_test - Test LibreModel capabilities

Run a quick_test with type "creative" to see if LibreModel can write poetry

πŸ₯ health_check - Monitor server status

Use health_check to see if LibreModel is running properly

Configuration

Set environment variables to customize behavior:

export LLAMA_SERVER_URL="http://localhost:8080"  # Default llama-server URL

Available Tools

Tool

Description

Parameters

chat

Converse with MOdel

message, temperature, max_tokens, top_p, top_k, system_prompt

quick_test

Run predefined capability tests

test_type (hello/math/creative/knowledge)

health_check

Check server health and status

None

Resources

  • Configuration: View current server settings

  • Instructions: Detailed usage guide and setup instructions

Development

# Install dependencies
npm install openconstruct/llama-mcp-server


# Development mode (auto-rebuild)
npm run dev

# Build for production
npm run build

# Start the server directly
npm start

Architecture

Claude Desktop ←→ LLama MCP Server ←→ llama-server API ←→ Local Model

The MCP server acts as a bridge, translating MCP protocol messages into llama-server API calls and formatting responses for Claude Desktop.

Troubleshooting

"Cannot reach LLama server"

  • Ensure llama-server is running on the configured port

  • Check that the model is loaded and responding

  • Verify firewall/network settings

"Tool not found in Claude Desktop"

  • Restart Claude Desktop after configuration changes

  • Check that the path to index.js is correct and absolute

  • Verify the MCP server builds without errors

Poor response quality

  • Adjust temperature and sampling parameters

  • Try different system prompts

License

CC0-1.0 - Public Domain. Use freely!


Built with ❀️ for open-source AI and the LibreModel project. by Claude Sonnet4

Development mode (auto-rebuild)

npm run dev

Build for production

npm run build

Start the server directly

npm start


## Architecture

Claude Desktop ←→ LLama MCP Server ←→ llama-server API ←→ Local Model


The MCP server acts as a bridge, translating MCP protocol messages into llama-server API calls and formatting responses for Claude Desktop.

## Troubleshooting

**"Cannot reach LLama server"**
- Ensure llama-server is running on the configured port
- Check that the model is loaded and responding
- Verify firewall/network settings

**"Tool not found in Claude Desktop"**
- Restart Claude Desktop after configuration changes
- Check that the path to `index.js` is correct and absolute
- Verify the MCP server builds without errors

**Poor response quality**
- Adjust temperature and sampling parameters
- Try different system prompts

## License

CC0-1.0 - Public Domain. Use freely!

---

Built with ❀️ for open-source AI and the LibreModel project. by Claude Sonnet4

### 1. Install Dependencies

```bash
cd llama-mcp
npm install

2. Build the Server

npm run build

3. Start Your LibreModel

Make sure llama-server is running with your model:

./llama-server -m lm37.gguf -c 2048 --port 8080

4. Configure Claude Desktop

Add this to your Claude Desktop configuration (~/.config/claude/claude_desktop_config.json):

{
  "mcpServers": {
    "libremodel": {
      "command": "node",
      "args": ["/home/jerr/llama-mcp/dist/index.js"]
    }
  }
}

5. Restart Claude Desktop

Claude will now have access to LibreModel through MCP!

Usage

Once configured, you can use these tools in Claude Desktop:

πŸ’¬ chat - Main conversation tool

Use the chat tool to ask LibreModel: "What is your name and what can you do?"

πŸ§ͺ quick_test - Test LibreModel capabilities

Run a quick_test with type "creative" to see if LibreModel can write poetry

πŸ₯ health_check - Monitor server status

Use health_check to see if LibreModel is running properly

Configuration

Set environment variables to customize behavior:

export LLAMA_SERVER_URL="http://localhost:8080"  # Default llama-server URL

Available Tools

Tool

Description

Parameters

chat

Converse with MOdel

message, temperature, max_tokens, top_p, top_k, system_prompt

quick_test

Run predefined capability tests

test_type (hello/math/creative/knowledge)

health_check

Check server health and status

None

Resources

  • Configuration: View current server settings

  • Instructions: Detailed usage guide and setup instructions

Development

# Install dependencies
npm install

# Development mode (auto-rebuild)
npm run dev

# Build for production
npm run build

# Start the server directly
npm start

Architecture

Claude Desktop ←→ LLama MCP Server ←→ llama-server API ←→ Local Model

The MCP server acts as a bridge, translating MCP protocol messages into llama-server API calls and formatting responses for Claude Desktop.

Troubleshooting

"Cannot reach LLama server"

  • Ensure llama-server is running on the configured port

  • Check that the model is loaded and responding

  • Verify firewall/network settings

"Tool not found in Claude Desktop"

  • Restart Claude Desktop after configuration changes

  • Check that the path to index.js is correct and absolute

  • Verify the MCP server builds without errors

Poor response quality

  • Adjust temperature and sampling parameters

  • Try different system prompts

License

CC0-1.0 - Public Domain. Use freely!


Built with ❀️ for open-source AI and the LibreModel project. by Claude Sonnet4

Install Server
A
security – no known vulnerabilities
F
license - not found
A
quality - confirmed to work

Resources

Unclaimed servers have limited discoverability.

Looking for Admin?

If you are the server author, to access and configure the admin panel.

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/openconstruct/llama-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server