All the adapters and merged models are available on Hugging Face: model collection. Here is a summary of all the experiments.
Base model | Axolotl config | Training log | QLoRA r/alpha | Target modules |
---|---|---|---|---|
Llama-2-7b | yml file | W&B logs | 1 (16/16) | gate, up, down* |
Llama-2-13b | yml file | W&B logs | 1 (16/16) | gate, up, down |
Llama-2-13b | yml file | W&B logs | 1 (64/64) | gate,up, down |
Mistral-7b-v0.1 | yml file | W&B logs | 2 (32/16) | all modules |
Mistral-7b-v0.1 | yml file | W&B logs | 2 (32/16) | gate, up, down |
Mistral-7b-v0.1 | yml file | W&B logs | 2 (64/32) | gate, up, down |
* Targeting gate_proj
, down_proj
, and up_proj
modules follows the finetuning recipes reported by He et al. 2022 and Lee and Hunter et al. 2023.
Inspired by the performance of Open-Platypus dataset, we have curated a similar dataset while limiting each contributing subset to datasets with permissive licences. Please see HF for additional information on Open-Otter dataset.