Skip to content

Implementation of CVPR 2022 paper "Learning Distinctive Margin toward Active Domain Adaptation”

License

Notifications You must be signed in to change notification settings

TencentYoutuResearch/ActiveLearning-SDM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Learning Distinctive Margin toward Active Domain Adaptation

  • A Pytorch implementation of our CVPR 2022 paper "Learning Distinctive Margin toward Active Domain Adaptation"
  • arXiv

Installation

  • Python 3.7
  • Pytorch 1.8.0
  • torchvision 0.9
  • Numpy 1.20

Run the code

Preliminaries

Training

  • Setting

Modify the configuration in SDM_code/config/ini.config

Arg:
[data]
name : dataset
path = dataset location
source = the initial of certain scenario 
target = the initial of certain scenario
class = number of categories
[sample]
strategy = certain sample strategy
[param]
epoch : we set it to 40 in our experiments
lr : learning rate
batch : batch size
sdm_lambda : default value is 0.01
sdm_margin : default value is 1.0
  • Usage

After modify setting, just run the code:

python3 run.py
  • Log

We also provide our experiment logs saved in SDM_code/log/{dataset}_{source}{target}.log. For example, officehome_AC.log

Acknowledgement

This codebase is built upon TQS.

Citation

If you find our work helps your research, please kindly consider citing our paper in your publications.

@article{xie2022sdm
	title={Learning Distinctive Margin toward Active Domain Adaptation},
    author={Xie, Ming and Li, Yuxi and Wang, Yabiao and Luo, Zekun and Gan, Zhenye and Sun, Zhongyi and Chi, Mingmin and Wang, Chengjie and Wang, Pei},
    booktitle={IEEE/CVF International Conference on Computer Vision and Pattern Recognition},
    year={2022}
}

About

Implementation of CVPR 2022 paper "Learning Distinctive Margin toward Active Domain Adaptation”

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages