Implementation of TransGanFormer, an all-attention GAN that combines the finding from the recent GansFormer and TransGan paper. It will also contain a bunch of tricks I have picked up building transformers and GANs for the last year or so, including efficient linear attention and pixel level attention.
$ pip install transganformer
$ transganformer --data ./path/to/data
@misc{jiang2021transgan,
title = {TransGAN: Two Transformers Can Make One Strong GAN},
author = {Yifan Jiang and Shiyu Chang and Zhangyang Wang},
year = {2021},
eprint = {2102.07074},
archivePrefix = {arXiv},
primaryClass = {cs.CV}
}
@misc{hudson2021generative,
title = {Generative Adversarial Transformers},
author = {Drew A. Hudson and C. Lawrence Zitnick},
year = {2021},
eprint = {2103.01209},
archivePrefix = {arXiv},
primaryClass = {cs.CV}
}