VL-ICL Bench: The Devil in the Details of Benchmarking Multimodal In-Context Learning.
We host the our dataset at HuggingFace here.
git lfs install
git clone https://huggingface.co/datasets/ys-zong/VL-ICL
cd VL-ICL
bash unzip.sh
cd ..
Different conda environments may be needed for different models.
conda create -n {env_name} python==3.10 -y
pip install -r requirements/{model.txt}
conda activate {env_name}
Replace {model.txt}
with corresponding file.
Most of the models can be automatically downloaded from Huggingface. For Text-to-image models (Emu1, Emu2, GILL, SEED-LLaMA), please see here for detailed instructions.
CUDA_VISIBLE_DEVICES=0 python I2T_inference.py --engine {model_name} --n_shot {shots} --dataset {dataset_name} --task_description detailed
For example,
CUDA_VISIBLE_DEVICES=0 python I2T_inference.py --engine qwen-vl-chat --n_shot 0 1 2 4 5 --task_description detailed --dataset open_mi
python I2T_evaluate.py --dataset {dataset_name} --engine {model_name} --n_shot {shots}
CUDA_VISIBLE_DEVICES=0 python T2I_inference.py --engine {model_name} --n_shot {shots} --dataset {dataset_name} --task_description detailed
For example,
CUDA_VISIBLE_DEVICES=0 python T2I_inference.py --engine emu1-gen --n_shot 0 1 2 4 5 --task_description detailed --dataset open_t2i_mi
CUDA_VISIBLE_DEVICES=0 python T2I_evaluate.py --dataset open_t2i_mi --engine seed-llama
@article{zong2024vlicl,
title={VL-ICL Bench: The Devil in the Details of Benchmarking Multimodal In-Context Learning},
author={Zong, Yongshuo and Bohdal, Ondrej and Hospedales, Timothy},
journal={arXiv preprint arXiv:2403.13164},
year={2024}
}