#
gguf
Here are 6 public repositories matching this topic...
LLM inference server implementation based on llama.cpp.
-
Updated
Nov 18, 2024 - C++
C++ inference wrappers for running blazing fast embedding services on your favourite serverless like AWS Lambda. By Prithivi Da, PRs welcome.
-
Updated
Mar 4, 2024 - C++
llama.cpp 🦙 LLM inference in TypeScript
-
Updated
Sep 26, 2024 - C++
Improve this page
Add a description, image, and links to the gguf topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the gguf topic, visit your repo's landing page and select "manage topics."