Skip to content

Latest commit

 

History

History

02-open-source

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 

2. Open-Source LLMs

In the previous module, we used Gemini 1.5 Flash via Google API. It's a very convenient way to use an LLM, but you have to pay for the usage, and you don't have control over the model you get to use.

In this module, we'll look at using open-source LLMs instead.

2.1 Introduction

YouTube Class: 2.1 - Introduction to Open-Source

  • Open-Source LLMs
  • Replacing the LLM box in the RAG flow

2.2 Using a GPU in Saturn Cloud

YouTube Class: 2.2 - Using SaturnCloud for GPU Notebooks

  • Registering in Saturn Cloud
  • Configuring secrets and git
  • Creating an instance with a GPU

Bonus: Using Google Colab for GPU Notebooks

This is my personal choice!

2.3 Model: Google FLAN-T5

YouTube Class: 2.3 - HuggingFace and Google FLAN T5

2.4 More models

2.5 Ollama - Running LLMs on a CPU

YouTube Class: 2.7 - Running LLMs Locally without a GPU with Ollama

Install and start Ollama:

curl -fsSL https://ollama.com/install.sh | sh
ollama start

Pull (only once) and run locally a model:

ollama pull phi3
ollama run phi3

then a chat with the model will be opened from the command line interface.