Skip to content

agronholm/llama-cpp-binaries

 
 

Repository files navigation

llama-cpp-binaries

llama.cpp server in a Python wheel.

Installation

git clone --recurse-submodules https://github.com/oobabooga/llama-cpp-binaries
cd llama-cpp-binaries
CMAKE_ARGS="-DGGML_CUDA=ON -DGGML_NATIVE=off -DCMAKE_CUDA_ARCHITECTURES=all" pip install -v .

Usage

import subprocess

from llama_cpp_binaries import get_binary_path

server_binary_path = get_binary_path()

# start with subprocess.Popen(...)

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages

  • Python 100.0%