Skip to content

For our NeurIPS21 paper "Are My Deep Learning Systems Fair? An Empirical Study of Fixed-Seed Training" by Shangshu Qian, Hung Viet Pham, Thibaud Lutellier, Zeou Hu, Jungwon Kim, Lin Tan, Yaoliang Yu, Jiahao Chen, and Sameena Shah

License

Notifications You must be signed in to change notification settings

lin-tan/fairness-variance

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Are My Deep Learning Systems Fair? An Empirical Study of Fixed-Seed Training

This is the artifact repository for the NeurIPS 2021 paper: Are My Deep Learning Systems Fair? An Empirical Study of Fixed-Seed Training.

Citation

If you use this repository in your project, please cite the following publication:

@inproceedings{NEURIPS2021_fdda6e95,
 author = {Qian, Shangshu and Pham, Viet Hung and Lutellier, Thibaud and Hu, Zeou and Kim, Jungwon and Tan, Lin and Yu, Yaoliang and Chen, Jiahao and Shah, Sameena},
 booktitle = {Advances in Neural Information Processing Systems},
 editor = {M. Ranzato and A. Beygelzimer and Y. Dauphin and P.S. Liang and J. Wortman Vaughan},
 pages = {30211--30227},
 publisher = {Curran Associates, Inc.},
 title = {Are My Deep Learning Systems Fair? An Empirical Study of Fixed-Seed Training},
 url = {https://proceedings.neurips.cc/paper/2021/file/fdda6e957f1e5ee2f3b311fe4f145ae1-Paper.pdf},
 volume = {34},
 year = {2021}
}

Artifacts list

URL Description
raw_data.csv This spreadsheet is the overall and per-class fairness variance results for 189 experiments (27 techniques and seven bias metrics).
stat_tests.csv This spreadsheet contains the statistical test results for 154 bias mitigation experiments (22 mitigation techinques and seven bias metrics), including U-test, Levene's test, and Cohen's d number.
overall Naming convention: ./raw_bias_numbers/overall/<Technique>.yaml
This folder contains the raw overall bias numbers for each technique from 16 FIT runs.
Each yaml file is a dictionary. The key of the dictionary is the bias metric. The value of the dictionary is a list of 16 bias values for each run.
per-class Naming convention: ./raw_bias_numbers/per_class/<Technique>/run_<idx>.yaml
This folder contains the raw per-class bias numbers for each technique from 16 FIT runs.
Each yaml file is a dictionary, and contains the bias value for all the classes in one run. The key of the dictionary is the bias metric. The value of the dictionary is a list of bias values for each class.
Raw prediction results This release contains the raw prediction results of 27 techniques in the paper, while each technique has the result for 16 FIT runs.
Docker image This release contains the docker image to reproduce our study.
Weights This release contains the weights for all the trained models.
Dataset This release contains a backup of all the datasets used in our study.
Replication package This folder contains the code and scripts to reproduce our study.
Code README This file is the guide to use our replication package.
Bias metric calculation README This file is the details about the bias metric calculation part of the replication package.

About

For our NeurIPS21 paper "Are My Deep Learning Systems Fair? An Empirical Study of Fixed-Seed Training" by Shangshu Qian, Hung Viet Pham, Thibaud Lutellier, Zeou Hu, Jungwon Kim, Lin Tan, Yaoliang Yu, Jiahao Chen, and Sameena Shah

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published