LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
-
Updated
Jun 8, 2024 - Python
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
Openai style api for open large language models, using LLMs just as chatgpt! Support for LLaMA, LLaMA-2, BLOOM, Falcon, Baichuan, Qwen, Xverse, SqlCoder, CodeLLaMA, ChatGLM, ChatGLM2, ChatGLM3 etc. 开源大模型的统一后端接口
This is the official implementation of "LLM-QBench: A Benchmark Towards the Best Practice for Post-training Quantization of Large Language Models", and it is also an efficient LLM compression tool with various advanced compression methods, supporting multiple inference backends.
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
Firefly: 大模型训练工具,支持训练Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
🚀🚀🚀A collection of some awesome public projects about Large Language Model, Vision Foundation Model and AI Generated Content.
🐋MindChat(漫谈)——心理大模型:漫谈人生路, 笑对风霜途
Famous Vision Language Models and Their Architectures
[ACL 2024] Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation
Collection of Basic Prompt Templates for Various Chat LLMs (Chat LLM 的基础提示模板集合)
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & more LLMs
internlm funetuning
一个Node.js后端,React.js前端,Python-Flask调用InternLM2-Chat-7B的项目,内容待上传
Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).
Add a description, image, and links to the internlm topic page so that developers can more easily learn about it.
To associate your repository with the internlm topic, visit your repo's landing page and select "manage topics."