#
qwen
Here are 6 public repositories matching this topic...
-
Updated
Aug 2, 2024 - C++
Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).
bloom
falcon
moe
gemma
mistral
mixture-of-experts
model-quantization
multi-gpu-inference
m2m100
llamacpp
llm-inference
internlm
llama2
qwen
baichuan2
mixtral
phi-2
deepseek
minicpm
-
Updated
Mar 15, 2024 - C++
Explore LLM model deployment based on AXera's AI chips
-
Updated
Aug 7, 2024 - C++
Improve this page
Add a description, image, and links to the qwen topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the qwen topic, visit your repo's landing page and select "manage topics."