Skip to content

Releases: hiyouga/LLaMA-Factory

v0.9.2: MiniCPM-o, SwanLab, APOLLO

11 Mar 13:47
Compare
Choose a tag to compare

This is the last version before LLaMA-Factory v1.0.0. We are working hard to improve the efficiency and availability.

We will attend the vLLM Beijing Meetup on Mar 16th! See you in Beijing πŸ‘‹

New features

New models

  • Base models
    • GPT2 (0.1B/0.4B/0.8B/1.5B) πŸ“„
    • Granite 3.0-3.1 (1B/2B/3B/8B) πŸ“„
    • PaliGemma2 (3B/10B/28B) πŸ“„πŸ–ΌοΈ
    • Moonlight (16B) πŸ“„
    • DeepSeek V2-V2.5 Base (236B) πŸ“„
    • DeepSeek V3 Base (671B) πŸ“„
  • Instruct/Chat models
    • Granite 3.0-3.1 (1B/2B/3B/8B) by @Tuyohai in #5922 πŸ“„πŸ€–
    • DeepSeek R1 (1.5B/7B/8B/14B/32B/70B/671B) by @Qwtdgh in #6767 πŸ“„πŸ€–
    • TeleChat2 (3B/7B/12B/35B/115B) @ge-xing in #6313 πŸ“„πŸ€–
    • Qwen2.5-VL (3B/7B/72B) by @hiyouga in #6779 πŸ“„πŸ€–πŸ–ΌοΈ
    • PaliGemma2-mix (3B/10B/28B) by @Kuangdd01 in #7060 πŸ“„πŸ€–πŸ–ΌοΈ
    • Qwen2 Audio (7B) by @BUAADreamer in #6701 πŸ“„πŸ€–πŸ”ˆ
    • MiniCPM-V/MiniCPM-o (8B) by @BUAADreamer in #6598 and #6631 πŸ“„πŸ€–πŸ–ΌοΈπŸ”ˆ
    • InternLM3-Instruct (8B) by @hhaAndroid in #6640 πŸ“„πŸ€–
    • Marco-o1 (8B) πŸ“„πŸ€–
    • Skywork-o1 (8B) πŸ“„πŸ€–
    • Phi-4 (14B) πŸ“„πŸ€–
    • Moonlight Instruct (16B) πŸ“„
    • Mistral Small (24B) πŸ“„πŸ€–
    • QwQ (32B) πŸ“„πŸ€–
    • Llama-3.3-Instruct (70B) πŸ“„πŸ€–
    • QvQ (72B) πŸ“„πŸ€–πŸ–ΌοΈ
    • DeepSeek V2-V2.5 (236B) πŸ“„πŸ€–
    • DeepSeek V3 (671B) πŸ“„πŸ€–

New datasets

  • Supervised fine-tuning datasets
    • OpenO1 (en) πŸ“„
    • Open Thoughts (en) πŸ“„
    • Open-R1-Math (en) πŸ“„
    • Chinese-DeepSeek-R1-Distill (zh) πŸ“„

Changes

Bug fix

Full Changelog: v0.9.1...v0.9.2

v0.9.1: Many Vision Models, Qwen2.5 Coder, Gradient Fix

24 Nov 17:17
Compare
Choose a tag to compare

New features

Note: now you can install transformers>=4.46.0,<=4.46.1 to make the gradient accumulation fix enabled.

New models

  • Base models
    • Qwen2.5 (0.5B/1.5B/3B/7B/14B/32B/72B) πŸ“„
    • Qwen2.5-Coder (0.5B/1.5B/3B/7B/14B/32B) πŸ“„πŸ–₯️
    • Llama-3.2 (1B/3B) πŸ“„
    • OpenCoder (1.5B/8B) πŸ“„πŸ–₯️
    • Index (1.9B) πŸ“„
  • Instruct/Chat models
    • Qwen2.5-Instruct (0.5B/1.5B/3B/7B/14B/32B/72B) πŸ“„πŸ€–
    • Qwen2.5-Coder-Instruct (0.5B/1.5B/3B/7B/14B/32B) πŸ“„πŸ€–πŸ–₯️
    • Llama-3.2-Instruct (1B/3B) πŸ“„πŸ€–
    • OpenCoder-Instruct (1.5B/8B) πŸ“„πŸ€–πŸ–₯️
    • Index-Chat (1.9B) πŸ“„πŸ€–
    • LLaVA-NeXT (7B/8B/13B/34B/72B/110B) πŸ“„πŸ€–πŸ–ΌοΈ
    • LLaVA-NeXT-Video (7B/34B) πŸ“„πŸ€–πŸ–ΌοΈ
    • Video-LLaVA (7B) πŸ“„πŸ€–πŸ–ΌοΈ
    • Pixtral (12B) πŸ“„πŸ€–πŸ–ΌοΈ
    • EXAONE-3.0-Instruct (8B) πŸ“„πŸ€–

Security fix

Bug fix

Full Changelog: v0.9.0...v0.9.1

v0.9.0: Qwen2-VL, Liger-Kernel, Adam-mini

08 Sep 17:14
Compare
Choose a tag to compare

Congratulations on 30,000 stars πŸŽ‰ Follow us at X (twitter)

New features

New models

  • Base models
    • Qwen2-Math (1.5B/7B/72B) πŸ“„πŸ”’
    • Yi-Coder (1.5B/9B) πŸ“„πŸ–₯️
    • InternLM2.5 (1.8B/7B/20B) πŸ“„
    • Gemma-2-2B πŸ“„
    • Meta-Llama-3.1 (8B/70B) πŸ“„
  • Instruct/Chat models
    • MiniCPM/MiniCPM3 (1B/2B/4B) by @LDLINGLINGLING in #4996 #5372 πŸ“„πŸ€–
    • Qwen2-Math-Instruct (1.5B/7B/72B) πŸ“„πŸ€–πŸ”’
    • Yi-Coder-Chat (1.5B/9B) πŸ“„πŸ€–πŸ–₯️
    • InternLM2.5-Chat (1.8B/7B/20B) πŸ“„πŸ€–
    • Qwen2-VL-Instruct (2B/7B) πŸ“„πŸ€–πŸ–ΌοΈ
    • Gemma-2-2B-it by @codemayq in #5037 πŸ“„πŸ€–
    • Meta-Llama-3.1-Instruct (8B/70B) πŸ“„πŸ€–
    • Mistral-Nemo-Instruct (12B) πŸ“„πŸ€–

New datasets

  • Supervised fine-tuning datasets
    • Magpie-ultra-v0.1 (en) πŸ“„
    • Pokemon-gpt4o-captions (en&zh) πŸ“„πŸ–ΌοΈ
  • Preference datasets
    • RLHF-V (en) πŸ“„πŸ–ΌοΈ
    • VLFeedback (en) πŸ“„πŸ–ΌοΈ

Changes

  • Due to compatibility consideration, fine-tuning vision language models (VLMs) requires transformers>=4.35.0.dev0, try pip install git+https://github.com/huggingface/transformers.git to install it.
  • visual_inputs has been deprecated, now you do not need to specify this argument.
  • LlamaFactory now adopts lazy loading for multimodal inputs, see #5346 for details. Please use preprocessing_batch_size to restrict the batch size in dataset pre-processing (supported by @naem1023 in #5323 ).
  • LlamaFactory now supports lmf (equivalent to llamafactory-cli) as a shortcut command.

Bug fix

v0.8.3: Neat Packing, Split Evaluation

18 Jul 18:00
Compare
Choose a tag to compare

New features

New models

  • Base models
    • InternLM2.5-7B πŸ“„
    • Gemma2 (9B/27B) πŸ“„
  • Instruct/Chat models
    • TeleChat-1B-Chat by @hzhaoy in #4651 πŸ“„πŸ€–
    • InternLM2.5-7B-Chat πŸ“„πŸ€–
    • CodeGeeX4-9B-Chat πŸ“„πŸ€–
    • Gemma2-it (9B/27B) πŸ“„πŸ€–

Changes

  • Fix DPO cutoff len and deprecate reserved_label_len argument
  • Improve loss function for reward modeling

Bug fix

v0.8.2: PiSSA, Parallel Functions

19 Jun 13:06
Compare
Choose a tag to compare

New features

New models

  • Base models
    • DeepSeek-Coder-V2 (16B MoE/236B MoE) πŸ“„
  • Instruct/Chat models
    • MiniCPM-2B πŸ“„πŸ€–
    • DeepSeek-Coder-V2-Instruct (16B MoE/236B MoE) πŸ“„πŸ€–

New datasets

Bug fix

v0.8.1: Patch release

10 Jun 16:50
Compare
Choose a tag to compare
  • Fix #2666: Unsloth+DoRA
  • Fix #4145: The PyTorch version of the docker image does not match the vLLM requirement
  • Fix #4160: The problem in LongLoRA implementation with the help of @f-q23
  • Fix #4167: The installation problem in the Windows system by @yzoaim

v0.8.0: GLM-4, Qwen2, PaliGemma, KTO, SimPO

07 Jun 22:26
Compare
Choose a tag to compare

Stronger LlamaBoard πŸ’ͺπŸ˜€

  • Support single-node distributed training in Web UI
  • Add dropdown menu for easily resuming from checkpoints and picking saved configurations by @hiyouga and @hzhaoy in #4053
  • Support selecting checkpoints of full/freeze tuning
  • Add throughput metrics to LlamaBoard by @injet-zhou in #4066
  • Faster UI loading

New features

  • Add KTO algorithm by @enji-zhou in #3785
  • Add SimPO algorithm by @hiyouga
  • Support passing max_lora_rank to the vLLM backend by @jue-jue-zi in #3794
  • Support preference datasets in sharegpt format and remove big files from git repo by @hiyouga in #3799
  • Support setting system messages in CLI inference by @ycjcl868 in #3812
  • Add num_samples option in dataset_info.json by @seanzhang-zhichen in #3829
  • Add NPU docker image by @dongdongqiang2018 in #3876
  • Improve NPU document by @MengqingCao in #3930
  • Support SFT packing with greedy knapsack algorithm by @AlongWY in #4009
  • Add llamafactory-cli env for bug report
  • Support image input in the API mode
  • Support random initialization via the train_from_scratch argument
  • Initialize CI

New models

  • Base models
    • Qwen2 (0.5B/1.5B/7B/72B/MoE) πŸ“„
    • PaliGemma-3B (pt/mix) πŸ“„πŸ–ΌοΈ
    • GLM-4-9B πŸ“„
    • Falcon-11B πŸ“„
    • DeepSeek-V2-Lite (16B) πŸ“„
  • Instruct/Chat models
    • Qwen2-Instruct (0.5B/1.5B/7B/72B/MoE) πŸ“„πŸ€–
    • Mistral-7B-Instruct-v0.3 πŸ“„πŸ€–
    • Phi-3-small-8k-instruct (7B) πŸ“„πŸ€–
    • Aya-23 (8B/35B) πŸ“„πŸ€–
    • OpenChat-3.6-8B πŸ“„πŸ€–
    • GLM-4-9B-Chat πŸ“„πŸ€–
    • TeleChat-12B-Chat by @hzhaoy in #3958 πŸ“„πŸ€–
    • Phi-3-medium-8k-instruct (14B) πŸ“„πŸ€–
    • DeepSeek-V2-Lite-Chat (16B) πŸ“„πŸ€–
    • Codestral-22B-v0.1 πŸ“„πŸ€–

New datasets

  • Pre-training datasets
    • FineWeb (en)
    • FineWeb-Edu (en)
  • Supervised fine-tuning datasets
    • Ruozhiba-GPT4 (zh)
    • STEM-Instruction (zh)
  • Preference datasets
    • Argilla-KTO-mix-15K (en)
    • UltraFeedback (en)

Bug fix

v0.7.1: Ascend NPU Support, Yi-VL Models

15 May 18:16
Compare
Choose a tag to compare

🚨🚨 Core refactor 🚨🚨

  • Add CLIs usage, now we recommend using llamafactory-cli to launch training and inference, the entry point is located at the cli.py
  • Rename files: train_bash.py -> train.py, train_web.py -> webui.py, api_demo.py -> api.py
  • Remove files: cli_demo.py, evaluate.py, export_model.py, web_demo.py, use llamafactory-cli chat/eval/export/webchat instead
  • Use YAML configs in examples instead of shell scripts for a pretty view
  • Remove the sha1 hash check when loading datasets
  • Rename arguments: num_layer_trainable -> freeze_trainable_layers, name_module_trainable -> freeze_trainable_modules

The above changes are made by @hiyouga in #3596

REMINDER: Now installation is mandatory to use LLaMA Factory

New features

  • Support training and inference on the Ascend NPU 910 devices by @zhou-wjjw and @statelesshz (docker images are also provided)
  • Support stop parameter in vLLM engine by @zhaonx in #3527
  • Support fine-tuning token embeddings in freeze tuning via the freeze_extra_modules argument
  • Add Llama3 quickstart to readme

New models

  • Base models
    • Yi-1.5 (6B/9B/34B) πŸ“„
    • DeepSeek-V2 (236B) πŸ“„
  • Instruct/Chat models
    • Yi-1.5-Chat (6B/9B/34B) πŸ“„πŸ€–
    • Yi-VL-Chat (6B/34B) by @BUAADreamer in #3748 πŸ“„πŸ–ΌοΈπŸ€–
    • Llama3-Chinese-Chat (8B/70B) πŸ“„πŸ€–
    • DeepSeek-V2-Chat (236B) πŸ“„πŸ€–

Bug fix

v0.7.0: LLaVA Multimodal LLM Support

27 Apr 20:24
Compare
Choose a tag to compare

Congratulations on 20k stars πŸŽ‰ We are the 1st of the GitHub Trending at Apr. 23rd πŸ”₯ Follow us at X

New features

  • Support SFT/PPO/DPO/ORPO for the LLaVA-1.5 model by @BUAADreamer in #3450
  • Support inferring the LLaVA-1.5 model with both native Transformers and vLLM by @hiyouga in #3454
  • Support vLLM+LoRA inference for partial models (see support list)
  • Support 2x faster generation of the QLoRA model based on UnslothAI's optimization
  • Support adding new special tokens to the tokenizer via the new_special_tokens argument
  • Support choosing the device to merge LoRA in LlamaBoard via the export_device argument
  • Add a Colab notebook for getting into fine-tuning the Llama-3 model on a free T4 GPU
  • Automatically enable SDPA attention and fast tokenizer for higher performance

New models

  • Base models
    • OLMo-1.7-7B
    • Jamba-v0.1-51B
    • Qwen1.5-110B
    • DBRX-132B-Base
  • Instruct/Chat models
    • Phi-3-mini-3.8B-instruct (4k/128k)
    • LLaVA-1.5-7B
    • LLaVA-1.5-13B
    • Qwen1.5-110B-Chat
    • DBRX-132B-Instruct

New datasets

  • Supervised fine-tuning datasets
  • Preference datasets

Bug fix

v0.6.3: Llama-3 and 3x Longer QLoRA

21 Apr 15:43
Compare
Choose a tag to compare

New features

  • Support Meta Llama-3 (8B/70B) models
  • Support UnslothAI's long-context QLoRA optimization (56,000 context length for Llama-2 7B in 24GB)
  • Support previewing local datasets in directories in LlamaBoard by @codemayq in #3291

New algorithms

New models

  • Base models
    • CodeGemma (2B/7B)
    • CodeQwen1.5-7B
    • Llama-3 (8B/70B)
    • Mixtral-8x22B-v0.1
  • Instruct/Chat models
    • CodeGemma-7B-it
    • CodeQwen1.5-7B-Chat
    • Llama-3-Instruct (8B/70B)
    • Command R (35B) by @marko1616 in #3254
    • Command R+ (104B) by @marko1616 in #3254
    • Mixtral-8x22B-Instruct-v0.1

Bug fix