Skip to content

explainingai-code/VIT-Pytorch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Vision Transformer (VIT) Implementation in pytorch on mnist images on textures

This repository implements Vision Transformer on a synthetic dataset of mnist colored numbers on textures/solid background .

Vision Transformer Videos

PatchEmbedding Attention Block Building Vision Transformer

Sample from dataset

Data preparation

For setting up the mnist dataset: Follow - https://github.com/explainingai-code/Pytorch-VAE#data-preparation

Download Quarter RGB resolution texture data from ALOT Homepage In case you want to train on higher resolution, you can download that as well and but you would have to create new imdb.json Rest of the code should work fine as long as you create valid json files.

Download imdb.json from Drive Verify the data directory has the following structure after textures download

VIT-Pytorch/data/textures/{texture_number}
	*.png
VIT-Pytorch/data/train/images/{0/1/.../9}
	*.png
VIT-Pytorch/data/test/images/{0/1/.../9}
	*.png
VIT-Pytorch/data/imdb.json

Quickstart

  • Create a new conda environment with python 3.8 then run below commands
  • git clone https://github.com/explainingai-code/VIT-Pytorch.git
  • cd VIT-Pytorch
  • pip install -r requirements.txt
  • python -m tools.train for training vit
  • python -m tools.inference for running inference, attention visualizations and positional embedding plots

Configuration

  • config/default.yaml - Allows you to play with different aspects of VIT

Output

Outputs will be saved according to the configuration present in yaml files.

For every run a folder of task_name key in config will be created

  • Best Model checkpoint in task_name directory

During inference the following output will be saved

  • Attention map visualization for sample of test set in task_name/output
  • Positional embedding similarity plots in task_name/output/position_plot.png

Sample Output for VIT

Following is a sample attention map that you should get

Here is a positional embedding similarity plot you should get

Citations

@misc{dosovitskiy2021image,
      title={An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale}, 
      author={Alexey Dosovitskiy and Lucas Beyer and Alexander Kolesnikov and Dirk Weissenborn and Xiaohua Zhai and Thomas Unterthiner and Mostafa Dehghani and Matthias Minderer and Georg Heigold and Sylvain Gelly and Jakob Uszkoreit and Neil Houlsby},
      year={2021},
      eprint={2010.11929},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

About

This repo implements and trains Vision Transformer (VIT) on a synthetically generated dataset which has colored mnist images on texture backgrounds

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages