Code for GLiT: Neural Architecture Search for Global and Local Image Transformer accepted by ICCV2021
- torch==1.7.0
- torchvision==0.8.1
- timm==0.3.2
The requirements.txt file lists other Python libraries that this project depends on, and they will be installed using: pip3 install -r requirements.txt
method | FLOPs | Params | Top-1 | model |
---|---|---|---|---|
GLiT-Tiny | 1.33G | 7.3M | 76.4 | Glit_Tiny |
python -m torch.distributed.launch --nproc_per_node=4 --use_env main.py --model glit_tiny_patch16_224 --clip-grad 1.0 --batch-size 256 --data-path /path/to/imagenet --output_dir /path/to/save
Download the model Glit_Tiny
python main.py --eval --resume glit_tiny.pth.tar --data-path /path/to/imagenet
This implementation is based on DeiT. Please ref to their reposity for more details.
If you find that this project helps your research, please consider citing our paper:
@inproceedings{chen2021glit,
title={Glit: Neural architecture search for global and local image transformer},
author={Chen, Boyu and Li, Peixia and Li, Chuming and Li, Baopu and Bai, Lei and Lin, Chen and Sun, Ming and Yan, Junjie and Ouyang, Wanli},
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
pages={12--21},
year={2021}
}