Skip to content

A python (PyTorch) implementation of federated multi-encoding U-Net (Fed-MENU) method for federated learning-based multi-organ segmentation with inconsistent labels.

DIAL-RPI/Fed-MENU

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

22 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Federated Multi-encoding U-Net (Fed-MENU)

Federated Multi-organ Segmentation with Inconsistent Labels

This is a python (PyTorch) implementation of federated multi-encoding U-Net (Fed-MENU) method for federated learning-based multi-organ segmentation with inconsistent labels proposed in our IEEE Transactions on Medical Imaging journal paper "Federated Multi-organ Segmentation with Inconsistent Labels" (early access). A preprint version of this paper is also available on arXiv.

Citation

X. Xu, H. H. Deng, J. Gateno and P. Yan, "Federated Multi-organ Segmentation with Inconsistent Labels," in IEEE Transactions on Medical Imaging, vol. 42, no. 10, pp. 2948-2960, October 2023, doi: 10.1109/TMI.2023.3270140.

@article{Xu2023FedMENU,
  title={Federated Multi-organ Segmentation with Inconsistent Labels}, 
  author={Xu, Xuanang and Deng, Hannah H. and Gateno, Jaime and Yan, Pingkun},
  journal={IEEE Transactions on Medical Imaging}, 
  year={2023},
  volume={42},
  number={10},
  pages={2948-2960},
  publisher={IEEE},
  doi={10.1109/TMI.2023.3270140}
}

Update

  • Oct 30, 2023: Add scripts data_downsampling for data downsampling preprocessing. In our experiments, we preprocessed the CT images from the six datasets by downsampling them to a spacing (resolution) no lower than 0.8mm x 0.8mm x 1.5mm (W x H x D). For the raw CT images with a spacing lower than 0.8mm x 0.8mm x 1.5mm (W x H x D), we will downsample them to a minimum spacing of 0.8mm x 0.8mm x 1.5mm to shrink the image pixel size. The purpose of this downsampling preprocessing is to reduce the data size and speed up the data loading process.

Abstract

Federated learning is an emerging paradigm allowing large-scale decentralized learning without sharing data across different data owners, which helps address the concern of data privacy in medical image analysis. However, the requirement for label consistency across clients by the existing methods largely narrows its application scope. In practice, each clinical site may only annotate certain organs of interest with partial or no overlap with other sites. Incorporating such partially labeled data into a unified federation is an unexplored problem with clinical significance and urgency. This work tackles the challenge by using a novel federated multi-encoding U-Net (Fed-MENU) method for multi-organ segmentation. In our method, a multi-encoding U-Net (MENU-Net) is proposed to extract organ-specific features through different encoding sub-networks. Each sub-network can be seen as an expert of a specific organ and trained for that client. Moreover, to encourage the organ-specific features extracted by different sub-networks to be informative and distinctive, we regularize the training of the MENU-Net by designing an auxiliary generic decoder (AGD). Extensive experiments on six public abdominal CT datasets show that our Fed-MENU method can effectively obtain a federated learning model using the partially labeled datasets with superior performance to other models trained by either localized or centralized learning methods. Source code is publicly available at https://github.com/DIAL-RPI/Fed-MENU.

Method

Scheme of Fed-MENU

Data

In our paper, we conducted experiments using six public abdominal CT image datasets for five-organ segmentation, including Liver, Kidney, Pancreas, Spleen, and Gallbladder. The first five client datasets (Client #1~5) are randomly split into training/validation/testing sets with a fixed ratio of 60%:10%:30%, respectively. The experimental results on the five testing sets are used for in-federation evaluation, which indicates the model performance when the testing data follows the same distribution as the training and validation data. For the BTCV dataset, we reserve it as an out-of-federation testing set, which is completely unseen to the model during training and validation.

Clients Dataset # of images
(train/val/test)
Organ Labels
Liver | Kidney | Pancreas | Spleen | Gallbladder
In-federation Client #1 LiTS 131 (79/13/39) ✅ | ⛔ | ⛔ | ⛔ | ⛔
In-federation Client #2 KiTS 210 (126/21/63) ⛔ | ✅ | ⛔ | ⛔ | ⛔
In-federation Client #3 MSD-Task07 281 (169/28/84) ⛔ | ⛔ | ✅ | ⛔ | ⛔
In-federation Client #4 MSD-Task09 41 (24/5/12) ⛔ | ⛔ | ⛔ | ✅ | ⛔
In-federation Client #5 AMOS 200 (120/20/60) ✅ | ✅ | ✅ | ✅ | ✅
Out-of-federation Client BTCV 30 (0/0/30) ✅ | ✅ | ✅ | ✅ | ✅

Contact

You are welcome to contact us:

About

A python (PyTorch) implementation of federated multi-encoding U-Net (Fed-MENU) method for federated learning-based multi-organ segmentation with inconsistent labels.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages