So, for anyone reading this who wants a freshly built llama.dll in windows with or without CUDA support,
stop struggling and do what I did:
- git clone https://github.com/abetlen/llama-cpp-python.git --recurse-submodules
- cd llama-cpp-python
- if you want CUDA: set LLAMA_CUBLAS=1
- pip install .
- Find the dll in your python's lib/site-packages/llama_cpp
Probably works just as easily in Linux, but idk.
Worked in Win10, python 3.10.6, CUDA 12.2.