This repository contains implementation of the CVPR - 2017 Paper "Image-to-Image Translation with Conditional Adversarial Networks".
Part of Winter of GANs
The architecture is of U-Net kind, which takes input as a sketch images of size (256 X 256 X 3) and outputs a coloured image of size (256 X 256 X 3). Encoder layer consists of 8 layers which convert images into latent space of size (1 X 1 X 512). Decoder contains 7 layer which upsamples the image. L1 loss is used to the generator along with the cGan loss
It takes sketch as well as coloured images a input and stack one on another. It return the probability that given the sketch, does the coloured image belongs to it or not.
Gan Model contains the generator and the discriminator architecture. Train contains loss function and the code to train the model. The model was trained on the kaggle dataset Anime-sketch-colorization-pair which contains 14k images. Training was done on Google Colab for 2 session of 8 hours and 100 epochs.
Generator weights can be found at - https://drive.google.com/file/d/1cu-MbW4fxP0F9MKQh2HXPjHrACFUzXux/view?usp=sharing
To test the model, run the command,
python3 test.py
Some of the results
Thanks Bhuvan Aggarwal for helping the training procedure ❤
- From Omkar with love 💙 !!!