Skip to content

harishkotra/DocsAI

Repository files navigation

⚡️ DocsAI

Turn any llms.txt or documentation base into a production-ready, lightning-fast RAG assistant in 60 seconds. Built with Agno, Next.js, and FastAPI for developers who want private, reliable AI sidekicks.

DocsAI eliminates hallucinations by rigidly adhering to your documentation. It ships with a beautiful web dashboard for agent management and an embeddable widget that can be pasted effortlessly into any external website.

Video Demo

screencapture-localhost-3000-2026-03-03-21_56_03 screencapture-localhost-3000-dashboard-2026-03-03-21_56_17 screencapture-localhost-3000-setup-2026-03-03-21_56_11

🏗 Architecture

DocsAI features a secure hybrid-monorepo design. NextJS handles the front-facing dashboard and widgets. FastAPI orchestrates incoming chats with the vector store through Agno, securely fetching responses against the internal LLM.

graph LR
    A[Visitor] -->|Widget Chat| B(Next.js Web App)
    C[Admin] -->|Dashboard / Setup| B
    B -->|API Requests| D(FastAPI Backend)
    D -->|RAG Framework| E(Agno)
    E -->|Vector Store| F[(Chroma DB)]
    E -->|Inference Node| G[Ollama, OpenAI, Anthropic...]
Loading

✨ What can I build?

You can drop the embed script generated by DocsAI onto:

  • Developer Documentation portals (Nextra, Docusaurus, Mintlify)
  • Internal Support Helplines for Employee FAQs
  • Knowledge Base Sidekicks for internal HR / Finance / DevOps policies.
  • Customer support widgets.

🚀 Getting Started

DocsAI strongly favors running within Docker Compose. We ship Ollama natively as the default provider over host.docker.internal for a completely free, local, and keyless experience! (OpenAI, Anthropic, OpenRouter, and Aisa.one are also supported during agent setup).

🐳 Run using Docker Compose

  1. Clone the repository
  2. Setup your environment:
    cp apps/web/.env.example apps/web/.env.local
    # -> Insert your Privy App ID here!
  3. Spin up the stack:
    docker-compose up --build

Ports:

(Note: Ensure your local Ollama is running (ollama serve) and is accessible from docker if you chose the local inference mode!).


☁️ Deploy Your Own (Cloud Run & Serverless)

If you'd like to put DocsAI on the open web, you easily can by leveraging Docker deployments on platforms like GCP Cloud Run, Railway, Render, or ECS.

Broad Deployment Theory:

  1. Backend (FastAPI): Deploy the apps/api directory using its Dockerfile. It requires no persistent state assuming you are ok with volatile in-memory indexing via Chroma (though a mounted volume is highly recommended for production). Make sure to select an external model provider like OpenAI and pass in the keys when using Serverless, as local Ollama instances generally do not adapt to zero-scale networks.
  2. Frontend (Next.js): Deploy the apps/web directory (via Dockerfile or native Vercel). You must pass two build/run time environment variables: NEXT_PUBLIC_PRIVY_APP_ID (for authentication) and NEXT_PUBLIC_API_URL (pointing to your newly minted backend URL).

🎨 Tech Stack

  • Frontend Application: Next.js 14, React.js, Tailwind CSS V4, Lucide Icons, ShadCN (Vanilla), Privy (Auth)
  • Backend Application: Python 3.11+, FastAPI, Uvicorn (uvloop)
  • AI Core Framework: Agno (Knowledge Base / Semantic RAG)
  • Vector Database: ChromaDB (Embeddings automatically generated via sentence-transformers/all-MiniLM-L6-v2)

🔐 Security & Privacy

Privacy is baked into the foundation.

  • Tenancy: Every project requested has a customized ID.
  • BYO Model: We manage absolutely zero configurations. If you enter an API key for your assistant, the backend only uses it at runtime routing.
  • Strict Guidelines: System prompts are inherently designed to refuse questions outside of indexed material.

About

Turn any `llms.txt` or documentation base into a production-ready, lightning-fast RAG assistant in 60 seconds. Built with Agno for developers who want private, reliable AI sidekicks.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors