Skip to content

Repository of paper Consistency-preserving Visual Question Answering in Medical Imaging (MICCAI2022)

License

Notifications You must be signed in to change notification settings

sergiotasconmorales/consistency_vqa

Repository files navigation

Consistency-preserving Visual Question Answering in Medical Imaging

This is the official repository of the paper "Consistency-preserving Visual Question Answering in Medical Imaging," published in the proceedings of the MICCAI2022.

Our method consists of a loss function and corresponding training method to improve the consistency. Evaluated on a medical dataset, we achieve improvements both in consistency and accuracy. For more details, please refer to our paper.

method

Data

You can download our DME dataset from here. You can place the zip file in any location and then unzip it. We'll refer to the path to the unzipped folder as <path_data>. A more detailed description of the dataset can be found here.

Installing Requirements

After cloning the repo, create a new environment with Python 3.9, activate it, and then install the required packages by running:

pip install -r requirements.txt

Configuration file

In the folder config/idrid_regions/single/ you can find different configuration files that correspond to different scenarios, as shown in Table 1 of our paper. More specifically, you can find the following configuration files:

Config file Consistency method
default_baseline.yaml None
default_squint.yaml SQuINT by Selvaraju et al.
default_consistency.yaml Ours

In order to use a configuration file to train, you must first change the fields path_img, path_qa and path_masks to match the path to the downloaded data <path_data>. Please note that with these configuration files you should obtain results that are similar to the ones reported in our paper. However, since we reported the average for 10 runs of each model, your results may deviate.

If you have a comet ml account, you can set the parameter comet_ml to True in the configuration file to monitor the training process in real time. This requires the proper configuration to access your account from Python (see this for more info).

Training

To train a model just run the following command:

train.py --path_config <path_config>

Example:

train.py --path_config config/idrid_regions/single/default_baseline.yaml

After training, the logs folder, as defined in the YAML file, will contain the results of the training. This includes the model weights for the best and last epoch, as well as the answers produced by the model for each epoch. Additionally, a JSON file named logbook will be generated, which contains the information from the config file and the values of the metrics (loss and performance) for each epoch.

Inference for test set

In order to do inference on the test set, use the following command:

inference.py --path_config <path_config>

The inference results are stored in the logs folder, as defined in the config file, in the sub-folder answers. In total 6 answer files are generated, as follows:

File name Meaning
answers_epoch_0.pt best model on test set
answers_epoch_2000.pt best model on val set
answers_epoch_1000.pt best model on train set
answers_epoch_1000.pt best model on train set
answers_epoch_2001.pt last model on val set
answers_epoch_1001.pt last model on train set

Each of these files contains a matrix with two columns, the first one representing the question ID, and the second one corresponding to the answer provided by the model. The answer is an integer. To convert from integer to the textual answer, a dictionary is given in <path_data>/processed/map_index_answer.pickle

Inference for a single sample

The following command allows you to do inference on a single sample using a previously trained model (as specified by the config file in <path_config>):

inference_single.py --path_config <path_config> --path_image <path_image> --path_mask <path_mask> --question <question>

Plotting metrics and learning curves

To plot learning curves and accuracy, use the following command after having trained and done inference:

plotter.py --path_config <path_config>

The resulting plots are stored in the logs folder.

Computing consistency

After running the inference script, you can compute the consistency using:

compute_consistency.py --path_config <path_config>

By default, this only computes the consistency C1 (see paper). To compute the consistency C2 as well, set the parameter q3_too to True when calling the function compute_consistencyin the script compute_consistency.py.



This work was carried out at the AIMI Lab of the ARTORG Center for Biomedical Engineering Research of the University of Bern. Please cite this work as:

@inproceedings{tascon2022consistency,
title={Consistency-preserving Visual Question Answering in Medical Imaging},
author={Tascon-Morales, Sergio and M{'a}rquez-Neila, Pablo and Sznitman, Raphael},
booktitle={International Conference on Medical Image Computing and Computer-Assisted Intervention},
pages={386--395},
year={2022},
organization={Springer}
}

About

Repository of paper Consistency-preserving Visual Question Answering in Medical Imaging (MICCAI2022)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages