llama
Here are 33 public repositories matching this topic...
🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
-
Updated
Jul 11, 2024 - C++
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
-
Updated
Jul 1, 2024 - C++
Lightweight inference library for ONNX files, written in C++. It can run SDXL on a RPI Zero 2 but also Mistral 7B on desktops and servers.
-
Updated
Jun 19, 2024 - C++
Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM, ONNX). Powers 👋 Jan
-
Updated
Jul 11, 2024 - C++
RTP-LLM: Alibaba's high-performance LLM inference engine for diverse applications.
-
Updated
Jul 11, 2024 - C++
LLaMA (Language Learning for Machine Translation) adalah proyek riset yang diprakarsai oleh Facebook AI Research (FAIR) yang bertujuan untuk meningkatkan kualitas terjemahan mesin menggunakan pendekatan yang lebih alami dan berfokus pada bahasa asal.
-
Updated
Apr 13, 2024 - C++
A high-performance inference system for large language models, designed for production environments.
-
Updated
Jul 11, 2024 - C++
🤘 TT-NN operator library, and TT-Metalium low level kernel programming model.
-
Updated
Jul 11, 2024 - C++
-
Updated
Jul 10, 2024 - C++
WebAssembly binding for llama.cpp - Enabling in-browser LLM inference
-
Updated
Jul 10, 2024 - C++
Fast Multimodal LLM on Mobile Devices
-
Updated
Jul 4, 2024 - C++
LLaVA server (llama.cpp).
-
Updated
Oct 20, 2023 - C++
Improve this page
Add a description, image, and links to the llama topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the llama topic, visit your repo's landing page and select "manage topics."