Skip to content

AnmingGu/kmixup-cifar10

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

k-Mixup-CIFAR10

Introduction

K-mixup is a generic and straightforward data augmentation principle that extends the popular mixup regularization technique. In essence, k-mixup uses optimal transport to match k-batches of training points with other k-batches of traning points. K-mixup further improves generalization and robustness over standard mixup.

This repository contains the implementation used for the results in our paper (https://arxiv.org/abs/2106.02933).

Citation

If you use this method or this code in your paper, then please cite our paper:

@article{greenewald2021kmixup,
  title={k-Mixup regularization for deep learning via optimal transport},
  author={Greenewald, Kristjan and Gu, Anming and Yurochkin, Mikhail and Solomon, Justin and Chien, Edward},
  journal={arXiv preprint arXiv:2106.02933},
  year={2021}
}

Requirements and Installation

  • A computer running macOS or Linux
  • For training new models, you'll also need a NVIDIA GPU and NCCL
  • Python version 3.6
  • A PyTorch installation

Training

Use python train.py to train a new model. Here is an example setting:

$ python train.py --lr=0.1 --alpha=10.0 --mixupBatch=16

License

This project is provided under the MIT License, see LICENSE file.

Acknowledgement

The CIFAR-10 reimplementation of k-mixup is adapted from the mixup-cifar10 repository by facebookresearch.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages