Skip to content

mangiucugna/local_multimodal_ai

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Install

1. Clone this repository

git clone https://github.com/mangiucugna/local_multimodal_ai

2. Clone llama.cpp

git clone https://github.com/ggerganov/llama.cpp

3. Build llama.cpp

cd llama.cpp
mkdir build
cd build
cmake ..
cmake --build . --config Release

4. Download and install the AI Model

Download the following bakllava model files to the llama.cpp/models folder

and copy them in llama.cpp/models/ggml-bakllava-1/

5. Install requirements

Create a venv and install requirements

python -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt

6. Install prerequisites

Install FFMPEG: https://ffmpeg.org/download.html

For Mac (using brew)

brew install ffmpeg

7. Launch the llama.cpp server

First start the llama.cpp server:

Windows

cd llama.cpp\build\bin
Release\server.exe -m ..\..\models\ggml-bakllava-1\ggml-model-q4_k.gguf --mmproj ..\..\models\ggml-bakllava-1\mmproj-model-f16.gguf -ngl 1

Mac & Linux

cd llama.cpp/build/bin
./server -m ../../models/ggml-bakllava-1/ggml-model-q4_k.gguf --mmproj ../../models/ggml-bakllava-1/mmproj-model-f16.gguf -ngl 1

8. Launch the web UI

Open another terminal window

source .venv/bin/activate
python app.py

Credits

  1. Forked from https://github.com/cocktailpeanut/mirror/
  2. Llama.cpp
  3. Bakllava
  4. Built with gradio.

About

A demo to run audio/video summarization on a local machine

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages