Skip to content

Latest commit

 

History

History
97 lines (72 loc) · 4.9 KB

README.md

File metadata and controls

97 lines (72 loc) · 4.9 KB

Primitive Generation and Semantic-related Alignment for Universal Zero-Shot Segmentation

PyTorch Python

🏠[Project page]📄[arXiv]📄[PDF]

This repository contains code for CVPR2023 paper:

Primitive Generation and Semantic-related Alignment for Universal Zero-Shot Segmentation
Shuting He, Henghui Ding, Wei Jiang

framework

Installation:

The code is tested under CUDA 11.2, Pytorch 1.9.0 and Detectron2 0.6.

  1. Install Detectron2 following the manual
  2. Run sh make.sh under PADing/modeling/pixel_decoder/ops (Note: 1-2 steps you can also follow the installation process of Mask2Former)
  3. Install other required packages: pip install -r requirements.txt
  4. Prepare the dataset following datasets/README.md

Inference

CUDA_VISIBLE_DEVICES=0 python train_net.py \
    --config-file configs/panoptic-segmentation/PADing.yaml \
    --num-gpus 1 --eval-only \
    MODEL.WEIGHTS [path_to_weights] \
    OUTPUT_DIR [output_dir]

Training

Firstly, download the pretrained weights here or you can train vanilla mask2former backbone using seen classes and convert it using the following command:

python train_net_pretrain.py --config-file configs/panoptic-segmentation/pretrain.yaml --num-gpus 8

python tools/preprocess_pretrained_weight.py --task_name panoptic --input_file panoptic_pretrain/model_final.pth

Then train PADing and finetune the last class embedding layer of the trained mask2former model:

CUDA_VISIBLE_DEVICES=0 python train_net.py  --config-file configs/panoptic-segmentation/PADing.yaml --num-gpus 1 MODEL.WEIGHTS pretrained_weight_panoptic.pth

Trained Models and logs

Download pretrained weights here.

Download final trained PADing weights for inference here.

Semantic-segmentation Instance-segmentation Panoptic-segmentation
Seen | Unseen (IoU) Seen | Unseen (mAP) PQs, SQs, RQs | PQu,SQu,RQu
40.6 | 26.4 53.1 | 8.5 37.7/79.0/44.8 | 14.4/75.3/17.4
model | log model | log model | log

Note: We reorganize code and the performances are slightly different from the paper's.

Acknowledgement

This project is based on Zegformer, Mask2Former. Many thanks to the authors for their great works!

BibTeX

Please consider to cite PADing if it helps your research.

@inproceedings{PADing,
  title={Primitive Generation and Semantic-Related Alignment for Universal Zero-Shot Segmentation},
  author={He, Shuting and Ding, Henghui and Jiang, Wei},
  booktitle={CVPR},
  year={2023}
}