llama
Here are 33 public repositories matching this topic...
LLM InferenceNet is a C++ project designed to facilitate fast and efficient inference from Large Language Models (LLMs) using a client-server architecture. It enables optimized interactions with pre-trained language models, making deployment on edge devices easier.
-
Updated
Jul 28, 2023 - C++
Code. Music. Live.
-
Updated
Aug 8, 2023 - C++
LLaVA server (llama.cpp).
-
Updated
Oct 20, 2023 - C++
Runs LLaMA with Extremely HIGH speed
-
Updated
Nov 21, 2023 - C++
Lightweight chat terminal-interface for llama.cpp server compilable for windows and linux.
-
Updated
Mar 1, 2024 - C++
Super easy to use library for doing LLaMA/GPT-J stuff! - Mirror of: https://gitlab.com/niansa/libjustlm
-
Updated
Mar 25, 2024 - C++
Multi-Model and multi-tasking llama Discord Bot - Mirror of: https://gitlab.com/niansa/discord_llama
-
Updated
Mar 27, 2024 - C++
LLaMA (Language Learning for Machine Translation) adalah proyek riset yang diprakarsai oleh Facebook AI Research (FAIR) yang bertujuan untuk meningkatkan kualitas terjemahan mesin menggunakan pendekatan yang lebih alami dan berfokus pada bahasa asal.
-
Updated
Apr 13, 2024 - C++
Modern, Header-only C++ bindings for the Ollama API.
-
Updated
Jun 11, 2024 - C++
Lightweight inference library for ONNX files, written in C++. It can run SDXL on a RPI Zero 2 but also Mistral 7B on desktops and servers.
-
Updated
Jun 19, 2024 - C++
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
-
Updated
Jul 1, 2024 - C++
Fast Multimodal LLM on Mobile Devices
-
Updated
Jul 4, 2024 - C++
Improve this page
Add a description, image, and links to the llama topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the llama topic, visit your repo's landing page and select "manage topics."