Skip to content

coreyb42/querysaurus

Repository files navigation

QuerySaurus

Welcome! This project is intended to help people get started thinking about and using LLMs in their programming projects, in particular using local LLMs and vector stores to implement the Retrieval Augmented Generation (RAG) architecture that forms the backbone of many current generative AI applications. It is coded in Python, and uses Jupyter Notebooks for coding.

LLaMA 2

We're using LLaMA-based LLMs here, specifically LLMs which are based on the LLaMA 2 foundational model. That model is usable for commercial applications, and is highly capable at things like text summaries.

Langchain

LangChain is the premier framework for building LLM-based applications. It has a vibrant community and is constantly getting better and better.

Getting Started

Prerequisites

You need some way to run a local LLM, or you'll need to use a third party LLM such as ChatGPT. The examples in this project use Ollama, which only works on Mac OS at the moment (8/31/2023).

If you are not on Mac OS, LangChain supports other methods of running local LLMs, such as Llama.cpp, and other LLM classes can be readily substituted in the examples. ChatGPT is quite affordable for small scale projects.

Installation

  • Clone the project using git
  • Install Python 3.11+
  • Create a virtual environment for this project
  • Install the requirements
  • Run jupyter lab
  • Open a browser to localhost:8888

The Notebooks

If you're new to this stuff, the suggested order is:

  • LLMs
  • Prompting
  • Embeddings
  • Vector Databases
  • Retrieval Augmented Generation (RAG)

Further Reading and Links

About

A starter kit for building applications with Large Language Models (LLMs) implementing the Retrieval Augmented Generation (RAG) architecture, using LangChain and LLaMA 2.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors