Just some simple tooling help out experimenting with llms using llama.cpp and others.
This repository also includes an ambitious idea
Prepare the repo:
cd ~
git clone --recurse-submodules https://github.com/tty-pt/llm
cd llm
./init.sh💡 Add . $HOME/llm/source.sh to .bashrc to use the binaries from verywhere.
Dowload a model:
llm-hug download bartowski/Mistral-7B-Instruct-v0.3-GGUF Mistral-7B-Instruct-v0.3-Q2_K.ggufRun the service and start chatting with the model:
llm-askd `llm-path *v0.3-Q2*`
llm-chatLists available models in the default huggingpath download cache
Zero out models you don't need to free up space but still be able to finish a full gguf repo download
Chat with a model
Get a model's path from a few characters
Evoke huggingface-cli. Example:
Run a daemon service to query the model
Ask questions and get answers. Example: