KovalskyAI is a fully self-hosted, blazing-fast AI chat interface built for privacy, speed, and simplicity.
No external APIs. No cloud. No monthly fees.
Just you and your model — on your own terms.
Visit the project landing page hosted on GitHub Pages:
🔗 https://maxadov.github.io/KovalskyAI/
- ⚡ Real-time streaming responses (like ChatGPT)
- 🎨 Minimalist, dark-themed interface with elegant animations (GSAP-powered)
- 🧱 Self-hosted backend — works offline, no dependencies on OpenAI or others
- 🖥️ Cross-platform: Runs smoothly on both laptops and desktops
- 🔁 CPU & GPU support — CUDA-accelerated via
koboldcpp
- 💬 Memory-aware chat history (retains previous conversation context)
- 🛡️ CORS enabled — ready for frontend/backend interaction
- 🌐 Clean separation between backend (Flask + Kobold API) and frontend (HTML/CSS/JS)
- You run your own model locally with KoboldCpp
- The backend (
app.py
) connects to it and forwards your input - Streaming tokens are parsed and sent instantly to the frontend
Here's what we're working on next:
- Text-to-Speech (TTS): The model can speak responses aloud using voice synthesis
- Search the web from within chat (
Search on Web
button) - Attach documents: Feed
.txt
,.pdf
,.docx
, etc. to the model - Save/load chat sessions to local storage
- Multi-language interface support
KovalskyAI uses:
- Backend: Python
Flask
,requests
, and a KoboldCpp-compatible API - Frontend: Pure
HTML
,CSS
,JavaScript
+GSAP
for animations
-
📥 Download KoboldCpp Download the latest
KoboldCpp
build from Releases Use the version withcu12
in the name if you have anNVIDIA GPU with CUDA 12
. -
🤖 Download the Model Get the model file
qwen1_5-7b-chat-q4_k_m.gguf
from Hugging Face. Chooseq4_k_m
format for a good balance of performance and quality. Place the.gguf
model file in the same folder askoboldcpp_cu12.exe.
-
🚀 Launch KoboldCpp Start the local API with the following command:
koboldcpp_cu12.exe --model ./YOUR_MODEL_NAME.gguf --usecublas --gpulayers 33 --threads 8 --port 5001
Adjust --gpulayers
and --threads
based on your hardware.
Make sure you are in the same directory as koboldcpp_cu12.exe
git clone https://github.com/maxadov/KovalskyAI.git
cd KovalskyAI
Create a virtual environment
python -m venv venv
Activate it Linux/macOS:
source venv/bin/activate
Windows:
venv\Scripts\activate
Install dependencies
pip install -r requirements.txt
Create and setup your .env
configuration at the root folder
KOBOLD_API_URL=http://localhost:5001/v1/chat/completions
SYSTEM_PROMPT=You are a smart, concise assistant. Never say more than needed. Answer clearly and directly, like a human who values brevity. Avoid fluff and unnecessary details.
FLASK_SECRET_KEY=YOUR_DIFFICULT_KEY
FLASK_PORT=8000
Start the server
python app.py
✅ Frontend will be available at: http://localhost:8000
Contributions are warmly welcome!
🔧If you’d like to suggest a feature, fix a bug, or improve anything:
-
Create an issue with details
-
Fork the repo
-
Open a pull request when ready
Made with ❤️ and local compute.