Skip to content

Learning with Noisy Labels by adopting a peer prediction loss function (deep learning & multi-class version).

License

Notifications You must be signed in to change notification settings

UCSC-REAL/Multi-class-Peer-Loss-functions

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Peer Loss functions

This repository is the (Multi-Class & Deep Learning) Pytorch implementation of "Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates" accepted by ICML2020.

Required Packages & Environment

Supported OS: Windows, Linux, Mac OS X; Python: 3.6/3.7;

Deep Learning Library: PyTorch (GPU required)

Required Packages: Numpy, Pandas, random, sklearn, tqdm, csv, torch (Keras is required if you want to estimate the noise transition matrix).

Utilities

This repository includes:

📋 Multi-class implementation of Peer Loss functions;

📋 Peer Loss functions in Deep Learning;

📋 Dynamical tunning strategies of Peer Loss functions to further improve the performance.

Details of running (equationweighted) Peer Loss functions on MNIST, Fashion MNIST, CIFAR-10, CIFAR-100 with different noise setting are mentioned in the README.md file in each folder.

The workflow of equationweighted Peer Loss functions comes to:

Figure1

Decision boundary visualization

Given a 2D syntheric dataset, the decision boundaries returned by training with Cross-entropy loss become loose when the noise rate is high. However, the decision boundaries w.r.t. Peer Loss functions remain tight despite high presence of label noise.

Figure2

Citation

If you use our code, please cite the following paper:

@inproceedings{liu2020peer,
  title={Peer loss functions: Learning from noisy labels without knowing noise rates},
  author={Liu, Yang and Guo, Hongyi},
  booktitle={International Conference on Machine Learning},
  pages={6226--6236},
  year={2020},
  organization={PMLR}
}

Related Code

📋 Peer Loss functions and its experiments on UCI datasets is available at: https://github.com/gohsyi/PeerLoss

📋 The following work extends Peer Loss functions to a family of divergence-type robust loss functions and is available at: https://github.com/weijiaheng/Robust-f-divergence-measures The corresponding paper is "When Optimizing f-Divergence is Robust with Label noise" accepted by ICLR2021.

Thanks for watching!

About

Learning with Noisy Labels by adopting a peer prediction loss function (deep learning & multi-class version).

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%