Skip to content

Conversation

@rwightman
Copy link
Collaborator

  • remove dud attention, involution + my swin attention adaptation don't seem worth keeping
  • add or update several new 26/50 layer ResNe(X)t variants that were used in experiments
  • remove models associated with dead-end or uninteresting experiment results
  • weights coming soon...

* remove dud attention, involution + my swin attention adaptation don't seem worth keeping
* add or update several new 26/50 layer ResNe(X)t variants that were used in experiments
* remove models associated with dead-end or uninteresting experiment results
* weights coming soon...
…p / fixes for byoanet. Rename resnet26ts to tfs to distinguish (extra fc).
* add polynomial decay 'poly'
* cleanup cycle specific args for cosine, poly, and tanh sched, t_mul -> cycle_mul, decay -> cycle_decay, default cycle_limit to 1 in each opt
* add k-decay for cosine and poly sched as per https://arxiv.org/abs/2004.05909
* change default tanh ub/lb to push inflection to later epochs
…g as an alternate to CE w/ smoothing. For training experiments.
…ll trying to improve reliability of sgd test.
@rwightman rwightman merged commit a6e8598 into master Sep 14, 2021
guoriyue pushed a commit to guoriyue/pytorch-image-models that referenced this pull request May 24, 2024
Update attention / self-attn based models from a series of experiments
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants