Auto Complete anything using a gguf model
-
Updated
Dec 4, 2023 - Python
Auto Complete anything using a gguf model
Load larger models by offloading model layers to both GPU and CPU
Ask LLaMa about image in your clipboard
A Genshin Impact Question Answer Project supported by Qwen1.5-14B-Chat
Static builds of llama.cpp (Currently only amd64 server builds are available)
Presentation on Artificial Intelligence for the Free Drawing and Print Graphics class of the Muthesius Academy of Art.
Repo to download, save and run quantised LLM models using Llama.cpp and benchmark the results (private use)
Lightweight implementation of the OpenAI open API on top of local models
Llama-2 on apple mac using gpu
AgentX is an Open-source library that help people use LLMs on their own computers or help them to serve LLMs as easy as possible that support multi-backends like PyTorch, llama.cpp, Ollama and EasyDeL
UnOfficial Gradio Repo for ICML 2024 paper "Executable Code Actions Elicit Better LLM Agents" by Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhang, Yunzhu Li, Hao Peng, Heng Ji.
A custom framework for easy use of LLMs, VLMs, etc. supporting various modes and settings via web-ui
A chatbot with the ability to vocally respond (TTS) using llama
Add a description, image, and links to the llama-cpp topic page so that developers can more easily learn about it.
To associate your repository with the llama-cpp topic, visit your repo's landing page and select "manage topics."