Skip to content

Demonstrates finetuning of a NLP model with novel 'AdaFamily' optimizer and 'mini-batch trimming'

License

Notifications You must be signed in to change notification settings

hfassold/nlp_finetuning_adafamily

Repository files navigation

Fine-tuning of 'DistillBert' NLP transformer model for text classification, with 'AdaFamily' optimizer and 'mini-batch trimming'

Demonstrates finetuning of a NLP model with novel 'AdaFamily' optimizer and 'mini-batch trimming'. Code is taken and adapted from https://github.com/hfwittmann/transformer_finetuning_lightning Uses pytorch lightning. Demonstrates also how to modify (via 'monkey-patching') a huggingface transformer model so that it employs a custom loss function.

Regarding 'mini-batch trimming' (curriculum learning method), see my arxiv preprint at https://arxiv.org/abs/2110.13058 (accepted for ASPAI 2021 conference)

Regarding 'AdaFamily' (a family of novel adaptive gradient methods), see my arxiv preprint at https://arxiv.org/abs/2203.01603 (accepted for ISPR 2022 conference). We use the AdaFamily variant with myu = 0.25

About

Demonstrates finetuning of a NLP model with novel 'AdaFamily' optimizer and 'mini-batch trimming'

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published