Skip to content

Latest commit

 

History

History
55 lines (47 loc) · 7.97 KB

compatibility-table.md

File metadata and controls

55 lines (47 loc) · 7.97 KB

+++ disableToc = false title = "Model compatibility table" weight = 24 url = "/model-compatibility/" +++

Besides llama based models, LocalAI is compatible also with other architectures. The table below lists all the compatible models families and the associated binding repository.

{{% alert note %}}

LocalAI will attempt to automatically load models which are not explicitly configured for a specific backend. You can specify the backend to use by configuring a model with a YAML file. See [the advanced section]({{%relref "docs/advanced" %}}) for more details.

{{% /alert %}}

Backend and Bindings Compatible models Completion/Chat endpoint Capability Embeddings support Token stream support Acceleration
[llama.cpp]({{%relref "docs/features/text-generation#llama.cpp" %}}) Vicuna, Alpaca, LLaMa, Falcon, Starcoder, GPT-2, and many others yes GPT and Functions yes** yes CUDA, openCL, cuBLAS, Metal
gpt4all-llama Vicuna, Alpaca, LLaMa yes GPT no yes N/A
gpt4all-mpt MPT yes GPT no yes N/A
gpt4all-j GPT4ALL-J yes GPT no yes N/A
falcon-ggml (binding) Falcon (*) yes GPT no no N/A
dolly (binding) Dolly yes GPT no no N/A
gptj (binding) GPTJ yes GPT no no N/A
mpt (binding) MPT yes GPT no no N/A
replit (binding) Replit yes GPT no no N/A
gptneox (binding) GPT NeoX, RedPajama, StableLM yes GPT no no N/A
bloomz (binding) Bloom yes GPT no no N/A
rwkv (binding) rwkv yes GPT no yes N/A
bert (binding) bert no Embeddings only yes no N/A
whisper whisper no Audio no no N/A
stablediffusion (binding) stablediffusion no Image no no N/A
langchain-huggingface Any text generators available on HuggingFace through API yes GPT no no N/A
piper (binding) Any piper onnx model no Text to voice no no N/A
sentencetransformers BERT no Embeddings only yes no N/A
bark bark no Audio generation no no yes
autogptq GPTQ yes GPT yes no N/A
exllama GPTQ yes GPT only no no N/A
diffusers SD,... no Image generation no no N/A
vall-e-x Vall-E no Audio generation and Voice cloning no no CPU/CUDA
vllm Various GPTs and quantization formats yes GPT no no CPU/CUDA
exllama2 GPTQ yes GPT only no no N/A
transformers-musicgen no Audio generation no no N/A
tinydream stablediffusion no Image no no N/A
coqui Coqui no Audio generation and Voice cloning no no CPU/CUDA
petals Various GPTs and quantization formats yes GPT no no CPU/CUDA
transformers Various GPTs and quantization formats yes GPT, embeddings yes yes**** CPU/CUDA/XPU

Note: any backend name listed above can be used in the backend field of the model configuration file (See [the advanced section]({{%relref "docs/advanced" %}})).