This repo is for EMNLP 2023 Findings paper:
Not All Demonstration Examples are Equally Beneficial:
Reweighting Demonstration Examples for In-Context Learning
Download this repo and enter working directory:
cd wicl
This repo need the following packages:
datasets==2.12.0
numpy==1.24.3
torch==1.12.1
tqdm==4.65.0
transformers==4.24.0
A suitable conda environment named wicl
can be created and activated with:
conda env create -f environment.yml
conda activate wicl
Set proper hyper-parameters in re_w.sh
, then:
bash re_w.sh
We utilize 5 different GPT-like causal language models released by fairseq (https://github.com/facebookresearch/fairseq/tree/main/examples/moe_lm), and the number of parameters of these models is 355M, 1.3B, 2.7B, 6.7B, 13B, respectively. For example, if you want to use GPT 6.7B, you can set model_size = "6.7B"
in re_w.sh
.
8 text classification are supported: 'sst2', 'mr', 'subj', 'agnews', 'cb', 'dbpedia', 'rte', 'boolq'
For example, if you want to test on 'sst2' and 'mr', you can set tasks = ('sst2' 'mr')
in re_w.sh
.