Skip to content

MICCAI 2022 Code for the paper: D'ARTAGNAN: Counterfactual Video Generation

Notifications You must be signed in to change notification settings

HReynaud/dartagnan

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DARTAGNAN

This repository contains the code for the paper "D'ARTAGNAN: Counterfactual Video Generation"

Presentation

Our model can generate counterfactual (ie. fake) videos, based on an input video and a property we want to inject in that input video to generate the counterfactual video. The method is very general and we apply it to healthcare by using the Echonet-Dynamic dataset, which contains ~10.000 echocardiograms (cardiac ultrasounds of the heart).

To modify a property of the videos, we need a model to evaluate that property. We call it the expert model. In our use case, we want to change the ejection fraction of the hearts in the echocardiograms of our dataset. We thus train a regression model than can infer the ejection fraction from a video.

Once that model is trained, we train the main model, that takes a video and a new ejection fraction as input and generates a counterfactual video with a heart showing an ejection fraction close to the input ejection fraction. This lets us answer the question What would this echocardiogram look like if the patient had a different ejection fraction?.

Echocardiograms

Run the code

1) Environment

You will need to have conda installed on your machine. Then, create a new environment with the following command:

conda env create -f environment.yml

The versions of the packages are the ones we used for the experiments. If you want to use different versions, you can do so by editing the environment.yml file. Be warned that many packages may be incompatible with each other depending on the versions you choose.

Once the environment is created, activate it with the following command:

conda activate dartagnan

Then, install the dartagnan package with the following command:

pip install -e .

Finally, configure the utils/constants.py file with your wandb credentials and project details to track the experiments.

2) Data

Download the data

The dataset we use is the Echonet-Dynamic dataset, which can be downloaded from here. You will need to request access.

Once downloaded, you need to extract the files in the data folder of this repository. The folder structure should look like this:

data
├── EchoNet-Dynamic
│   ├── Videos
│   │   ├── 0X1A0A263B22CCD966.avi
│   │   ├── ...
│   ├── FileList.csv
│   ├── VolumeTracings.csv

Process the data

The data needs to be processed before it can be used. To do so, run the following command:

# Assuming you are in the root of the repository 
# and you have activated the dartagnan environment
python tools/build_echo_dataset.py -i data/EchoNet-Dynamic -o data/processed_echonet

After running this command, the folder structure should look like this:

data
├── EchoNet-Dynamic
│   ├── Videos
│   │   ├── 0X1A0A263B22CCD966.avi
│   │   ├── ...
│   ├── FileList.csv
│   ├── VolumeTracings.csv
├── processed_echonet
│   ├── videos
│   │   ├── 0X1A0A263B22CCD966.npy
│   │   ├── ...
│   ├── metainfo.csv

2) Train the models

Expert model

D'ARTAGNAN relies on an Ejection Fraction Regression model, refered to as the expert model which has to be pre-trained before we train the counterfactual model. To train the expert model, run the following command:

python train.py --model ExpertTrainer --root data/echonet --batch_size 4 --name expert --internal_dim 32

This will train the expert model and save the weights in the checkpoints/expert folder. You can monitor the training with wandb. Feel free to play with the hyperparameters to get the best results with the minimal memory footprint.

D'ARTAGNAN - Counterfactual video generation model

To train the counterfactual video generation model, run the following command:

python train.py --model VideoTrainer --root data/processed_echonet --batch_size 2 --name dartagnan --internal_dim 32 --trained_expert_path checkpoints/expert/best.pt

This will train the counterfactual video generation model and save the weights in the checkpoints/dartagnan folder. You can monitor the training with wandb. There are many hyperparameters that you can play with to adjust the memory footprint and the performance of the model.

3) Inference

To generate counterfactual videos using a D'ARTAGNAN trained model, run the following command:

WIP

python tools/inference.py --model_path checkpoints/dartagnan/best.pt --video_path data/EchoNet-Dynamic/Videos/0X1A0A263B22CCD966.avi --output_path results/0X1A0A263B22CCD966.avi

About

MICCAI 2022 Code for the paper: D'ARTAGNAN: Counterfactual Video Generation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages