Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
25 commits
Select commit Hold shift + click to select a range
4bc103f
Fix CUDA crash w/ channels-last + CSP models. Remove use of chunk()
rwightman Feb 23, 2021
0e16d4e
Add benchmark.py script, and update optimizer factory to be more frie…
rwightman Feb 23, 2021
f0ffdf8
Add numerous experimental ViT Hybrid models w/ ResNetV2 base. Update …
rwightman Feb 23, 2021
de97be9
Spell out diff between my small and deit small vit models.
rwightman Feb 24, 2021
2db2d87
Add epoch-repeats arg to multiply the number of dataset passes per ep…
rwightman Feb 24, 2021
0706d05
Benchmark models listed in txt file. Add more hybrid vit variants for…
rwightman Mar 1, 2021
17cdee7
Fix C&P patch_size error, and order of op patch_size arg resolution b…
rwightman Mar 2, 2021
4445eaa
Add img_size to benchmark output
rwightman Mar 6, 2021
4de57cc
Add weight init scheme that's closer to JAX impl
rwightman Mar 18, 2021
cbcb76d
Should have included Conv2d layers in original weight init. Lets see …
rwightman Mar 19, 2021
f42f1df
Improve evenness of per-worker split for validation set with TFDS
rwightman Mar 19, 2021
cf5fec5
Cleanup experimental vit weight init a bit
rwightman Mar 20, 2021
e2e3290
Add '--experiment' to train args for fixed exp name if desired, 'trai…
rwightman Mar 20, 2021
0dfc5a6
Add PiT model from https://github.com/naver-ai/pit
rwightman Apr 1, 2021
a760a4c
Some ViT cleanup, merge distilled model with main, fixup torchscript …
rwightman Apr 1, 2021
7953e5d
Fix pos_embed scaling for ViT and num_classes != 1000 for pretrained …
rwightman Apr 1, 2021
a5310a3
Merge remote-tracking branch 'origin/benchmark-fixes-vit_hybrids' int…
rwightman Apr 1, 2021
ea9c955
Fully move ViT hybrids to their own file, including embedding module.…
rwightman Apr 1, 2021
2886827
Update benchmark script to add precision arg. Fix some downstream (De…
rwightman Apr 1, 2021
c468c47
Add regnety_160 weights from DeiT teacher model, update that and my r…
rwightman Apr 1, 2021
9071568
Add weights for SE NFNet-L0 model, rename nfnet_l0b -> nfnet_l0. 82.7…
rwightman Apr 2, 2021
acbd698
Update README.md with updates. Small tweak to head_dist handling.
rwightman Apr 2, 2021
bf2ca6b
Merge jax and original weight init
rwightman Apr 2, 2021
2bb65bd
Wrong default_cfg pool_size for L1
rwightman Apr 2, 2021
37c71a5
Some further create_optimizer_v2 tweaks, remove some redudnant code, …
rwightman Apr 2, 2021
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
18 changes: 18 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,22 @@ I'm fortunate to be able to dedicate significant time and money of my own suppor

## What's New

### April 1, 2021
* Add snazzy `benchmark.py` script for bulk `timm` model benchmarking of train and/or inference
* Add Pooling-based Vision Transformer (PiT) models (from https://github.com/naver-ai/pit)
* Merged distilled variant into main for torchscript compatibility
* Some `timm` cleanup/style tweaks and weights have hub download support
* Cleanup Vision Transformer (ViT) models
* Merge distilled (DeiT) model into main so that torchscript can work
* Support updated weight init (defaults to old still) that closer matches original JAX impl (possibly better training from scratch)
* Separate hybrid model defs into different file and add several new model defs to fiddle with, support patch_size != 1 for hybrids
* Fix fine-tuning num_class changes (PiT and ViT) and pos_embed resizing (Vit) with distilled variants
* nn.Sequential for block stack (does not break downstream compat)
* TnT (Transformer-in-Transformer) models contributed by author (from https://gitee.com/mindspore/mindspore/tree/master/model_zoo/research/cv/TNT)
* Add RegNetY-160 weights from DeiT teacher model
* Add new NFNet-L0 w/ SE attn (rename `nfnet_l0b`->`nfnet_l0`) weights 82.75 top-1 @ 288x288
* Some fixes/improvements for TFDS dataset wrapper

### March 17, 2021
* Add new ECA-NFNet-L0 (rename `nfnet_l0c`->`eca_nfnet_l0`) weights trained by myself.
* 82.6 top-1 @ 288x288, 82.8 @ 320x320, trained at 224x224
Expand Down Expand Up @@ -189,6 +205,7 @@ A full version of the list below with source links can be found in the [document
* NFNet-F - https://arxiv.org/abs/2102.06171
* NF-RegNet / NF-ResNet - https://arxiv.org/abs/2101.08692
* PNasNet - https://arxiv.org/abs/1712.00559
* Pooling-based Vision Transformer (PiT) - https://arxiv.org/abs/2103.16302
* RegNet - https://arxiv.org/abs/2003.13678
* RepVGG - https://arxiv.org/abs/2101.03697
* ResNet/ResNeXt
Expand All @@ -204,6 +221,7 @@ A full version of the list below with source links can be found in the [document
* ReXNet - https://arxiv.org/abs/2007.00992
* SelecSLS - https://arxiv.org/abs/1907.00837
* Selective Kernel Networks - https://arxiv.org/abs/1903.06586
* Transformer-iN-Transformer (TNT) - https://arxiv.org/abs/2103.00112
* TResNet - https://arxiv.org/abs/2003.13630
* Vision Transformer - https://arxiv.org/abs/2010.11929
* VovNet V2 and V1 - https://arxiv.org/abs/1911.06667
Expand Down
Loading