Skip to content

openconstruct/llama-mcp-server

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

4 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

LibreModel MCP Server πŸ€–

A Model Context Protocol (MCP) server that bridges Claude Desktop with your local LLM instance running via llama-server.

Features

  • πŸ’¬ Full conversation support with Local Model through Claude Desktop
  • πŸŽ›οΈ Complete parameter control (temperature, max_tokens, top_p, top_k)
  • βœ… Health monitoring and server status checks
  • πŸ§ͺ Built-in testing tools for different capabilities
  • πŸ“Š Performance metrics and token usage tracking
  • πŸ”§ Easy configuration via environment variables

Quick Start

npm install @openconstruct/llama-mcp-server

A Model Context Protocol (MCP) server that bridges Claude Desktop with your local LLM instance running via llama-server.

Features

  • πŸ’¬ Full conversation support with LibreModel through Claude Desktop
  • πŸŽ›οΈ Complete parameter control (temperature, max_tokens, top_p, top_k)
  • βœ… Health monitoring and server status checks
  • πŸ§ͺ Built-in testing tools for different capabilities
  • πŸ“Š Performance metrics and token usage tracking
  • πŸ”§ Easy configuration via environment variables

Quick Start

1. Install Dependencies

cd llama-mcp
npm install

2. Build the Server

npm run build

3. Start Your LibreModel

Make sure llama-server is running with your model:

./llama-server -m lm37.gguf -c 2048 --port 8080

4. Configure Claude Desktop

Add this to your Claude Desktop configuration (~/.config/claude/claude_desktop_config.json):

{
  "mcpServers": {
    "libremodel": {
      "command": "node",
      "args": ["/home/jerr/llama-mcp/dist/index.js"]
    }
  }
}

5. Restart Claude Desktop

Claude will now have access to LibreModel through MCP!

Usage

Once configured, you can use these tools in Claude Desktop:

πŸ’¬ chat - Main conversation tool

Use the chat tool to ask LibreModel: "What is your name and what can you do?"

πŸ§ͺ quick_test - Test LibreModel capabilities

Run a quick_test with type "creative" to see if LibreModel can write poetry

πŸ₯ health_check - Monitor server status

Use health_check to see if LibreModel is running properly

Configuration

Set environment variables to customize behavior:

export LLAMA_SERVER_URL="http://localhost:8080"  # Default llama-server URL

Available Tools

Tool Description Parameters
chat Converse with MOdel message, temperature, max_tokens, top_p, top_k, system_prompt
quick_test Run predefined capability tests test_type (hello/math/creative/knowledge)
health_check Check server health and status None

Resources

  • Configuration: View current server settings
  • Instructions: Detailed usage guide and setup instructions

Development

# Install dependencies
npm install # LibreModel MCP Server πŸ€–

A Model Context Protocol (MCP) server that bridges Claude Desktop with your local LLM instance running via llama-server.

## Features

- πŸ’¬ **Full conversation support** with LibreModel through Claude Desktop
- πŸŽ›οΈ **Complete parameter control** (temperature, max_tokens, top_p, top_k)
- βœ… **Health monitoring** and server status checks
- πŸ§ͺ **Built-in testing tools** for different capabilities
- πŸ“Š **Performance metrics** and token usage tracking
- πŸ”§ **Easy configuration** via environment variables

## Quick Start

### 1. Install Dependencies

```bash
cd llama-mcp
npm install

2. Build the Server

npm run build

3. Start Your LibreModel

Make sure llama-server is running with your model:

./llama-server -m lm37.gguf -c 2048 --port 8080

4. Configure Claude Desktop

Add this to your Claude Desktop configuration (~/.config/claude/claude_desktop_config.json):

{
  "mcpServers": {
    "libremodel": {
      "command": "node",
      "args": ["/home/jerr/llama-mcp/dist/index.js"]
    }
  }
}

5. Restart Claude Desktop

Claude will now have access to LibreModel through MCP!

Usage

Once configured, you can use these tools in Claude Desktop:

πŸ’¬ chat - Main conversation tool

Use the chat tool to ask LibreModel: "What is your name and what can you do?"

πŸ§ͺ quick_test - Test LibreModel capabilities

Run a quick_test with type "creative" to see if LibreModel can write poetry

πŸ₯ health_check - Monitor server status

Use health_check to see if LibreModel is running properly

Configuration

Set environment variables to customize behavior:

export LLAMA_SERVER_URL="http://localhost:8080"  # Default llama-server URL

Available Tools

Tool Description Parameters
chat Converse with MOdel message, temperature, max_tokens, top_p, top_k, system_prompt
quick_test Run predefined capability tests test_type (hello/math/creative/knowledge)
health_check Check server health and status None

Resources

  • Configuration: View current server settings
  • Instructions: Detailed usage guide and setup instructions

Development

# Install dependencies
npm install openconstruct/llama-mcp-server


# Development mode (auto-rebuild)
npm run dev

# Build for production
npm run build

# Start the server directly
npm start

Architecture

Claude Desktop ←→ LLama MCP Server ←→ llama-server API ←→ Local Model

The MCP server acts as a bridge, translating MCP protocol messages into llama-server API calls and formatting responses for Claude Desktop.

Troubleshooting

"Cannot reach LLama server"

  • Ensure llama-server is running on the configured port
  • Check that the model is loaded and responding
  • Verify firewall/network settings

"Tool not found in Claude Desktop"

  • Restart Claude Desktop after configuration changes
  • Check that the path to index.js is correct and absolute
  • Verify the MCP server builds without errors

Poor response quality

  • Adjust temperature and sampling parameters
  • Try different system prompts

License

CC0-1.0 - Public Domain. Use freely!


Built with ❀️ for open-source AI and the LibreModel project. by Claude Sonnet4

Development mode (auto-rebuild)

npm run dev

Build for production

npm run build

Start the server directly

npm start


## Architecture

Claude Desktop ←→ LLama MCP Server ←→ llama-server API ←→ Local Model


The MCP server acts as a bridge, translating MCP protocol messages into llama-server API calls and formatting responses for Claude Desktop.

## Troubleshooting

**"Cannot reach LLama server"**
- Ensure llama-server is running on the configured port
- Check that the model is loaded and responding
- Verify firewall/network settings

**"Tool not found in Claude Desktop"**
- Restart Claude Desktop after configuration changes
- Check that the path to `index.js` is correct and absolute
- Verify the MCP server builds without errors

**Poor response quality**
- Adjust temperature and sampling parameters
- Try different system prompts

## License

CC0-1.0 - Public Domain. Use freely!

---

Built with ❀️ for open-source AI and the LibreModel project. by Claude Sonnet4

### 1. Install Dependencies

```bash
cd llama-mcp
npm install

2. Build the Server

npm run build

3. Start Your LibreModel

Make sure llama-server is running with your model:

./llama-server -m lm37.gguf -c 2048 --port 8080

4. Configure Claude Desktop

Add this to your Claude Desktop configuration (~/.config/claude/claude_desktop_config.json):

{
  "mcpServers": {
    "libremodel": {
      "command": "node",
      "args": ["/home/jerr/llama-mcp/dist/index.js"]
    }
  }
}

5. Restart Claude Desktop

Claude will now have access to LibreModel through MCP!

Usage

Once configured, you can use these tools in Claude Desktop:

πŸ’¬ chat - Main conversation tool

Use the chat tool to ask LibreModel: "What is your name and what can you do?"

πŸ§ͺ quick_test - Test LibreModel capabilities

Run a quick_test with type "creative" to see if LibreModel can write poetry

πŸ₯ health_check - Monitor server status

Use health_check to see if LibreModel is running properly

Configuration

Set environment variables to customize behavior:

export LLAMA_SERVER_URL="http://localhost:8080"  # Default llama-server URL

Available Tools

Tool Description Parameters
chat Converse with MOdel message, temperature, max_tokens, top_p, top_k, system_prompt
quick_test Run predefined capability tests test_type (hello/math/creative/knowledge)
health_check Check server health and status None

Resources

  • Configuration: View current server settings
  • Instructions: Detailed usage guide and setup instructions

Development

# Install dependencies
npm install

# Development mode (auto-rebuild)
npm run dev

# Build for production
npm run build

# Start the server directly
npm start

Architecture

Claude Desktop ←→ LLama MCP Server ←→ llama-server API ←→ Local Model

The MCP server acts as a bridge, translating MCP protocol messages into llama-server API calls and formatting responses for Claude Desktop.

Troubleshooting

"Cannot reach LLama server"

  • Ensure llama-server is running on the configured port
  • Check that the model is loaded and responding
  • Verify firewall/network settings

"Tool not found in Claude Desktop"

  • Restart Claude Desktop after configuration changes
  • Check that the path to index.js is correct and absolute
  • Verify the MCP server builds without errors

Poor response quality

  • Adjust temperature and sampling parameters
  • Try different system prompts

License

CC0-1.0 - Public Domain. Use freely!


Built with ❀️ for open-source AI and the LibreModel project. by Claude Sonnet4

About

MCP cerver to let claude use llamacpp. Uses API.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published