llama.cpp 🦙 LLM inference in TypeScript
-
Updated
Sep 26, 2024 - C++
llama.cpp 🦙 LLM inference in TypeScript
C++ inference wrappers for running blazing fast embedding services on your favourite serverless like AWS Lambda. By Prithivi Da, PRs welcome.
LLM inference server implementation based on llama.cpp.
Add a description, image, and links to the gguf topic page so that developers can more easily learn about it.
To associate your repository with the gguf topic, visit your repo's landing page and select "manage topics."