Skip to content

This is the research project for the USC Viterbi CSCI-567 Machine Learning course.

Notifications You must be signed in to change notification settings

SARIHUST/SAMMed

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Leveraging SAM for Single-Source Domain Generalization in Medical Image Segmentation

Abstract

Domain Generalization (DG) aims to reduce domain shifts between domains to achieve promising performance on the unseen target domain, which has been widely practiced in medical image segmentation. Single-source domain generalization (SDG) is the most challenging setting that trains on only one source domain. Although existing methods have made considerable progress on SDG of medical image segmentation, the performances are still far from the applicable standards when faced with a relatively large domain shift. To address this problem, in this paper, we leverage the Segment Anything Model (SAM) to SDG to greatly improve the ability of generalization. Specifically, we introduce a parallel framework, the source images are sent into the SAM module and normal segmentation module respectively. To reduce the calculation resources, we apply a merging strategy before sending images to the SAM module. We extract the bounding boxes from the segmentation module and send the refined version as prompts to the SAM module. We evaluate our model on a classic DG dataset and achieve competitive results compared to other state-of-the-art DG methods. Furthermore, We conducted a series of ablation experiments to prove the effectiveness of the proposed method.

Dataset

Download the pre-processed Prostate dataset provided by SAML and put it in the data folder for your data.

Pretrained Model

We use the checkpoint of SAM in vit_b version. Don't forget to follow the installation part of SAM. (We apologize that we currently can not provide a clear requirements.txt file for you because our server is down. It might help to use the file provided by SAMUS. We might be able to provide the full mirror image if needed.)

Training process

step 1

Generate the coarse prediction masks from the segmentation network (Resnet).

python resnet_prostate.py --domain domain_id --batch_size batchsize --gpu gpuid --epoch epochs

step 2

Formulated the refined bounding boxes.

python save_resnet_bbox.py --domain domain_id --thresh threshold

step 3

Generate the final prediction masks results by fine-tuning SAM with the refined bounding boxes.

python sam_4_preprocessed_bbox.py --domain domain_id --batch_size batchsize --gpu gpuid --epoch epochs

About

This is the research project for the USC Viterbi CSCI-567 Machine Learning course.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published