Skip to content

gjlee0802/LeNet-5

Repository files navigation

LeNet-5 (MNIST) and Sharpness Aware Minimization

Introduce

We performed a task on LeNet-5 in SSU deep learning programming and hands-on classes.
The challenge topic was to think about how to improve the performance of LeNet-5.
I thought generalization-ability was also an important indicator to evaluate the performance of the model.
I've seen previous studies that the sharpness of loss landscape affects the performance of generalization,
and I studied how to increase the performance of generalization by applying Sharpness Aware Minimization on LeNet-5.

Performance

  • Basic LeNet-5 accuracy
    • TEST 1. Pure dataset : 98.75%
    • TEST 2. Gaussian Noise dataset(standard deviation : 0.4) : 91.62%
    • TEST 3. Gaussian Noise dataset(standard deviation : 0.6) : 89.04%
    • TEST 4. Gaussian Noise dataset(standard deviation : 0.8) : 86.43%
  • LeNet-5 with SAM(Sharpness-aware minimization) accuracy
    • TEST 1. Pure dataset : 99%
    • TEST 2. Gaussian Noise dataset(standard deviation : 0.4) : 98.60%
    • TEST 3. Gaussian Noise dataset(standard deviation : 0.6) : 98.11%
    • TEST 4. Gaussian Noise dataset(standard deviation : 0.8) : 97.23%

Random seeding was specified when applying Gaussian noise.
In the case of the base LeNet-5, we show severe performance degradation on the Noise-applied Test dataset.
On the other hand, the improved version LeNet-5 with Sharpness-aware minimization shows a slight performance degradation on the Noise-enabled Test dataset, which is even better in generalization performance.

Reference

Sharpness-aware minimization is a way to improve generalization performance. Check out the paper below.

You can find a summary of Sharpness-aware minimization below.

Citation

@inproceedings{foret2021sharpnessaware,
  title={Sharpness-aware Minimization for Efficiently Improving Generalization},
  author={Pierre Foret and Ariel Kleiner and Hossein Mobahi and Behnam Neyshabur},
  booktitle={International Conference on Learning Representations},
  year={2021},
  url={https://openreview.net/forum?id=6Tm1mposlrM}
}
@inproceesings{pmlr-v139-kwon21b,
  title={ASAM: Adaptive Sharpness-Aware Minimization for Scale-Invariant Learning of Deep Neural Networks},
  author={Kwon, Jungmin and Kim, Jeongseop and Park, Hyunseo and Choi, In Kwon},
  booktitle ={Proceedings of the 38th International Conference on Machine Learning},
  pages={5905--5914},
  year={2021},
  editor={Meila, Marina and Zhang, Tong},
  volume={139},
  series={Proceedings of Machine Learning Research},
  month={18--24 Jul},
  publisher ={PMLR},
  pdf={http://proceedings.mlr.press/v139/kwon21b/kwon21b.pdf},
  url={https://proceedings.mlr.press/v139/kwon21b.html},
  abstract={Recently, learning algorithms motivated from sharpness of loss surface as an effective measure of generalization gap have shown state-of-the-art performances. Nevertheless, sharpness defined in a rigid region with a fixed radius, has a drawback in sensitivity to parameter re-scaling which leaves the loss unaffected, leading to weakening of the connection between sharpness and generalization gap. In this paper, we introduce the concept of adaptive sharpness which is scale-invariant and propose the corresponding generalization bound. We suggest a novel learning method, adaptive sharpness-aware minimization (ASAM), utilizing the proposed generalization bound. Experimental results in various benchmark datasets show that ASAM contributes to significant improvement of model generalization performance.}
}