This is the repository for MAGBIG (Multilingual Assessment of Gender Bias in Image Generation) proposed in "Multilingual Text-to-Image Generation Magnifies Gender Stereotypes and Prompt Engineering May Not Help You".
You can find a Python package requirements file at translate_script\requirements.txt
and our set of prompts for all prompt types and languages in the folder prompt
.
You can generate images for these prompts using python generate_evaluate\generate_images.py
. After that, you can classify them with python generate_evaluate\classify_images.py
. For the evaluation, you can use python generate_evaluate\exp_max_unfairness.py
for bias and python generate_evaluate\CLIPscore.py
for text-to-image alignment with CLIP. These Python scripts also reproduce our results.
To reproduce our prompts, you can run your bash script translate_script\run.sh
or modify it to compute your own translations for new prompts and languages.
Please cite our work if you find it helpful.
You can also use our benchmark in the huggingface dataset library: https://huggingface.co/datasets/felfri/MAGBIG
If you like or use our work, please consider citing us.
@misc{friedrich2024multilingual,
title={Multilingual Text-to-Image Generation Magnifies Gender Stereotypes and Prompt Engineering May Not Help You},
author={Felix Friedrich and Katharina Hämmerl and Patrick Schramowski and Jindrich Libovicky and Kristian Kersting and Alexander Fraser},
year={2024},
eprint={2401.16092},
archivePrefix={arXiv},
primaryClass={cs.CL}
}