DocGPT is a high-performance Retrieval-Augmented Generation (RAG) system designed to transform static documentation into interactive, intelligent chat experiences. By leveraging a modern AI stack, DocGPT autonomously crawls documentation, generates high-dimensional vector embeddings, and provides contextually aware responses using local and cloud-based LLMs.
The system follows a modular pipeline architecture, separating ingestion, storage, and retrieval layers to ensure scalability and low-latency performance.
- Node.js: The robust foundation for the backend architecture.
- Express.js: Streamlined server-side routing and API management.
- PostgreSQL: Metadata storage and relational data management.
- React.js: Front-end framework for a dynamic, responsive user interface.
- Ollama: Local LLM execution for enhanced privacy and offline intelligence.
- Hugging Face: Powering state-of-the-art embedding and inference models.
- Qdrant DB: High-performance vector storage for fast, semantic retrieval at scale.
- 🔍 Intelligent Ingestion: Automated recursive crawling of web-based documentation with smart rate-limiting.
- 🧠 Multi-Model RAG: Seamlessly switch between local (Ollama) and cloud-based (Groq/OpenAI) LLMs.
- ⚡ Vector Search: Leverages Qdrant for semantic retrieval, ensuring the AI only answers based on your private data.
- 📅 Smart Scheduling: Built-in cron jobs for periodic documentation updates and indexing.
- 💬 Modern UI: Markdown support, syntax highlighting, and fluid animations for a premium chat experience.
- Node.js (v18+)
- PostgreSQL
- Ollama (For local LLM support)
- Docker (Optional, for Qdrant)
-
Clone the Repository
git clone https://github.com/your-username/DocGPT.git cd DocGPT -
Server Setup
cd server npm install # Configure your .env file based on .env.example npm run dev
-
Client Setup
cd client npm install npm run dev
To start indexing documentation, run the following command in the server directory:
npm run ingestThis project is licensed under the MIT License - see the LICENSE file for details.
