Skip to content

[MICCAI2022] Estimating Model Performance under Domain Shifts with Class-Specific Confidence Scores.

License

Notifications You must be signed in to change notification settings

ZerojumpLine/ModelEvaluationUnderClassImbalance

Repository files navigation

Update

Jul-18-2023: We provide more details about our experiments on skin lesion classfications under domain shifts here.

Introduction

It is difficult for the practitioner to assess the model reliability on the target domain as labeled data for a new test domain is usually not available. It would be of great practical value to develop a tool to estimate the performance of a trained model on an unseen test domain without access to ground truth.

In this study we aim to estimate the model performance by only making use of unlabeled test data from the target domain.


We find existing methods struggle with data that present class imbalance, because the methods used to calibrate confidence do not account for bias induced by class imbalance, consequently failing to estimate class-wise accuracy. Here, we introduce class-wise calibration within the framework of performance estimation for imbalanced datasets.


Installation

For Conda users, you can create a new Conda environment using

conda create -n modeleval python=3.9

after activating the environment with

source activate modeleval

try to install all the dependencies with

pip install -r requirements.txt

at last, let us install the conda environment for the jupyter notebook kernel.

python -m ipykernel install --user --name=modeleval

Data

Please download the data from the link, and put them under '/data/'.

Model evaluation on classification task

Refer to juypter notebook:

ImbalanceCIFAR10.ipynb

Model evaluation on segmentation task

Refer to jupyter notebook:

Prostate.ipynb

Note that the optimization process takes longer as we take probabiltiy maps as dense predictions.

We just show a result of one condition for simplicity. Please contact us (zeju.li18@imperial.ac.uk) for raw data if you want to reproduce more results in this paper.

Citation

If you find our work has positively influenced your projects, please kindly consider citing our work:

@article{li2022estimating,
  title={Estimating Model Performance under Domain Shifts with Class-Specific Confidence Scores},
  author={Li, Zeju and Kamnitsas, Konstantinos and Islam, Mobarakol and Chen, Chen and Glocker, Ben},
  journal={arXiv preprint arXiv:2207.09957},
  year={2022}
}

About

[MICCAI2022] Estimating Model Performance under Domain Shifts with Class-Specific Confidence Scores.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published