Skip to content
master
Switch branches/tags
Code

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
DA
Apr 5, 2020
Mar 4, 2021
Mar 17, 2020
Mar 17, 2020

BNM

code release for "Towards Discriminability and Diversity: Batch Nuclear-norm Maximization under Label Insufficient Situations" ( CVPR2020 oral)

One-sentence description

We prove in the paper that Batch Nuclear-norm Maximization (BNM) could ensure the prediction discriminability and diversity, which is an effective method under label insufficient situations.

Application

One line code under Pytorch and Tensorflow

Assume X is the prediction matrix. We could calculate BNM loss in both Pytorch and Tensorflow, as follows:

-Pytorch

  1. Direct calculation (Since there remains direct approach for nuclear-norm)
L_BNM = -torch.norm(X,'nuc')
  1. Calculation by SVD
L_BNM = -torch.sum(torch.svd(X, compute_uv=False)[1])

-Tensorflow

L_BNM = -tf.reduce_sum(tf.svd(X, compute_uv=False))

Tasks

We apply BNM to domain adaptation (DA) in DA, unsupervised open domain recognition (UODR) in UODR and semi-supervised learning (SSL) in SSL.

Training instructions for DA, UODR and SSL are in the README.md in DA, UODR and SSL respectively.

Citation

If you use this code for your research, please consider citing:

@InProceedings{Cui_2020_CVPR,
author = {Cui, Shuhao and Wang, Shuhui and Zhuo, Junbao and Li, Liang and Huang, Qingming and Tian, Qi},
title = {Towards Discriminability and Diversity: Batch Nuclear-Norm Maximization Under Label Insufficient Situations},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2020}
}

Supplementary could be found in Google driver and baidu cloud (z7yt).

量子位

Contact

If you have any problem about our code, feel free to contact

or describe your problem in Issues.

About

code of Towards Discriminability and Diversity: Batch Nuclear-norm Maximization under Label Insufficient Situations (CVPR2020 oral)

Topics

Resources

License

Releases

No releases published

Packages

No packages published

Languages