A web-based management interface for llama.cpp. Easily configure and run llama-server instances with a simple web UI.
- Web-based dashboard for managing llama-server
- Model selection and upload
- Interactive parameter configuration
- Chat interface with real-time responses
- Real-time server metrics and logging
- Integrated llama.cpp server management
- Debian/Ubuntu (or compatible)
- Python 3.10+
- 2GB+ RAM
- Internet connection for initial setup
Run the interactive installer:
sudo bash install.shThe installer will:
- Install system dependencies
- Set up Python virtual environment
- Configure llama.cpp server
- Create systemd service
- Set up logging
sudo systemctl start llama-managerOpen your browser to:
http://localhost:8484
sudo journalctl -u llama-manager -fConfiguration is stored in /opt/llama-manager/config/config.json.
Key settings:
port: Web interface port (default: 8484)llama_server_port: llama-server port (default: 8080)models_dir: Directory for downloaded modelslog_level: Logging level (debug, info, warning, error)
- Backend: Python Flask with SocketIO for real-time updates
- Frontend: Vanilla JavaScript with real-time metrics display
- Server: gevent async server with background threading
- Integration: Manages llama.cpp subprocess
If the service fails to start:
# Check logs
sudo journalctl -u llama-manager -n 50
# Check if port is in use
sudo lsof -i :8484
# Verify installation
ls -la /opt/llama-managerSee LICENSE file in repository.