Context
No documentation on how to deploy Intuno to production. Self-hosting is one of Intuno's key differentiators (Ollama embeddings, no OpenAI requirement), but without deployment docs, nobody will actually self-host it.
Proposed: docs/DEPLOYMENT.md
Sections
- Prerequisites — Docker, PostgreSQL 15+, Redis 7+, Qdrant 1.7+
- Docker Compose production setup — production-ready compose file with proper resource limits, restart policies, health checks
- Environment variables checklist — every required/optional env var with descriptions and example values
- Database setup — PostgreSQL with pgvector extension, connection pooling recommendations
- Vector database — Qdrant deployment options (Docker, Qdrant Cloud), collection sizing
- Redis — caching and rate limiting configuration
- Reverse proxy — nginx/Caddy configuration with SSL termination, WebSocket support for
/ws
- Health check monitoring — using the
/health endpoint, recommended alerting
- Scaling — horizontal scaling guidance (stateless app servers, shared DB/Redis/Qdrant)
- Embedding provider setup — OpenAI vs Ollama configuration for air-gapped deployments
Context
No documentation on how to deploy Intuno to production. Self-hosting is one of Intuno's key differentiators (Ollama embeddings, no OpenAI requirement), but without deployment docs, nobody will actually self-host it.
Proposed:
docs/DEPLOYMENT.mdSections
/ws/healthendpoint, recommended alerting