A lightweight and interactive chatbot powered by Metaβs LLaMA 3 and enhanced through Retrieval-Augmented Generation (RAG) for domain-specific answers. Comes with a clean Gradio UI for easy access and testing.
- π RAG architecture for better factual accuracy
- π€ Powered by Metaβs LLaMA 3 model
- ποΈ Easy-to-use Gradio interface
- π Plug in your own data (PDFs, text files, etc.)
- π Local or cloud-hosted deployment
Case Study: We provide specific details in plain text of an object and ask questions regarding its properties. The LLM must reply only based on the provided context.
rag-llama3-gradio/
βββ main.py # Entry point: runs the Gradio app
βββ loader.py # Loads and chunks documents for processing
βββ embedder.py # Handles embeddings and indexing using ChromaDB
βββ rag_engine.py # Retrieval + Generation logic using LLaMA 3
βββ requirements.txt # Python dependencies
βββ README.md # Project documentation
βββ data/ # Source documents for retrieval
β βββ demo.txt # Sample file for testing
βββ .gitignore # Files and folders to exclude from Git
git clone https://github.com/raf-init/rag-llama3-gradio.git
cd rag-llama3-gradio
pip install -r requirements.txt