Starred repositories
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
EasyR1: An Efficient, Scalable, Multi-Modality RL Training Framework based on veRL
Research and development (R&D) is crucial for the enhancement of industrial productivity, especially in the AI era, where the core aspects of R&D are mainly focused on data and models. We are commi…
Production-tested AI infrastructure tools for efficient AGI development and community-driven innovation
Fully open data curation for reasoning models
This is a replicate of DeepSeek-R1-Zero and DeepSeek-R1 training on small models with limited data
Fully open reproduction of DeepSeek-R1
Finetune Llama 3.3, DeepSeek-R1 & Reasoning LLMs 2x faster with 70% less memory! 🦥
Clean, minimal, accessible reproduction of DeepSeek R1-Zero
Train a 1B LLM with 1T tokens from scratch by personal
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
Official code implementation of General OCR Theory: Towards OCR-2.0 via a Unified End-to-end Model
Financial portfolio optimisation in python, including classical efficient frontier, Black-Litterman, Hierarchical Risk Parity
📈 目前最大的工业缺陷检测数据库及论文集 Constantly summarizing open source dataset and critical papers in the field of surface defect research which are of great importance.
Use PEFT or Full-parameter to finetune 450+ LLMs (Qwen2.5, InternLM3, GLM4, Llama3.3, Mistral, Yi1.5, Baichuan2, DeepSeek-R1, ...) and 150+ MLLMs (Qwen2.5-VL, Qwen2-Audio, Llama3.2-Vision, Llava, I…
Deep Learning Deployment Framework: Supports tf/torch/trt/trtllm/vllm and other NN frameworks. Support dynamic batching, and streaming modes. It is dual-language compatible with Python and C++, off…
TrustRAG:The RAG Framework within Reliable input,Trusted output
Building a quick conversation-based search demo with Lepton AI.
本项目旨在分享大模型相关技术原理以及实战经验(大模型工程化、大模型应用落地)
Play ChatGPT and other LLM with Xiaomi AI Speaker
Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
LLM API 管理 & 分发系统,支持 OpenAI、Azure、Anthropic Claude、Google Gemini、DeepSeek、字节豆包、ChatGLM、文心一言、讯飞星火、通义千问、360 智脑、腾讯混元等主流模型,统一 API 适配,可用于 key 管理与二次分发。单可执行文件,提供 Docker 镜像,一键部署,开箱即用。LLM API management & k…
The official codes for "Aurora: Activating chinese chat capability for Mixtral-8x7B sparse Mixture-of-Experts through Instruction-Tuning"
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models