- This repo releases our implementation for the PACIT model.
- PACIT is a simple and effective in-context instruction tuning method inspired by the pedagogical concept of desirable difficulty. The PACIT method unlocks the power of examples by encouraging the model to actively learn to grasp the distinctions between the positive and negative examples instead of merely reading. The model is expected to first verify the correctness of the provided example according to the task description, which is then set as the condition for generating a better response to the task instance.
- It is built based on the pretrained T5 model and LLaMA, and finetuned on SuperNI dataset data.
Model checkpoints are published on Huggingface hub: T5-770M, T5-3B and LLaMA-2-7B
Our main experiments and analysis are conducted on the following environment:
- CUDA (12.2)
- Pytorch (2.1.2)
- Transformers (4.36.2)
pip install -r requirements.txtFor a quick start, you can directly run PACIT.
Our models are trained and evaluated on Super-NaturalInstructions, which can be cloned by running:
git clone git@github.com:allenai/natural-instructions.git dataPut the data under the path: "./Tk-Instruct-main/data"
We use TK-instruct-main to generate our base dataset and then use our python code to further process. You can run the shell directly
bash script/data_processing.shA sample script for training the baseline models. You can run it as follows:
bash script/SuperNI_trainA sample script for training PACIT models. You can run it as follows:
bash script/PACIT_trainA sample script for evaluating basline and PACIT models. You can run it as follows:
bash script/SuperNI_evaluation_generation bash script/PACIT_GT_evaluation_generation bash script/PACIT_Random_evaluation_generationHere are the performance numbers (in ROUGE-L) for our tested models:
@misc{xue2024pacit,
title={PACIT: Unlocking the Power of Examples for Better In-Context Instruction Tuning},
author={Tianci Xue and Ziqi Wang and Yixia Li and Yun Chen and Guanhua Chen},
year={2024},
eprint={2310.00901},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
