Skip to content

emedinac/StyleTransfer

Repository files navigation

Style Augmentation

This code is an implementation of the paper Style Augmentation: Data Augmentation via Style Randomization . My main contributions here are the following:

  • We used Xception and InceptionV4 as classifier networks.
  • We used Learning Linear Transformations for Fast Image and Video Style Transfer as stylizing network. We changed internal configurations to choose intermediate style stregth and generate randomization in each iteration.
  • We trained the styling net using the r31 and saving the embedded features in the same manner as the original style tranfer paper (however, pretrained methods are available for the stylizing network).

Personal Targets

  • Unterstand Why texture bias the training as stated in ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness
  • Understand the influence between stylization in augmentation and networks that perform this task (I tested and read several style transfer, but this one was the best).
  • Train networks and use a network as data augmenter embedded in torchvision.
  • Understand the incluence of data augmentation in each network and hyperparameter fitting
  • Analyze the noise in the latent domain (uniform and gaussian were tested).

Usage

All parameters are detailed in main.py and config_classification.py, just download the wikiArt (training) DB, the pretrained The model for stylization from original authors or you can follow the author's procedure to get the model, thereafter move the model into models folder. Finally, run the training code.:

python FeatureExtractor.py # To get the feature vectors from WikiArt DB.
python training.py

hyperparameters can be edited in the config_classification.py

Results

Results reported on Xception and InceptionV4 using STL-10 for classfication task are the following:

Netwrok Trad Style Accuracy
Xception (96) 67.91%
Xception X 83.65%
Xception X 69.21%
Xception X X 82.67%
InceptionV3 (299) 79.17%
InceptionV3 X 86.49%
InceptionV3 X 80.52%
InceptionV3 X X 87.18%
InceptionV4 (299) 74.08%
InceptionV4 X 85.34%
InceptionV4 X 77.58%
InceptionV4 X X 86.28%
Xception (256) 72.14%
Xception X 86.85%
Xception X 74.67%
Xception X X 86.85%
WideResNet (96) 77.28%
WideResNet X 87.26%
WideResNet X 83.58%
WideResNet X X 87.83%

Weights tuned without pre-training are available in the following: drive. Some other results and configurations will be available soon!. Due to hardware limitations, WideResNet cannot be tested on images with size of 256x256 pixels.

Some qualitative results for style transfer.

No Augmentation (original images) No Augmentation Trad+Style augmentation ( ColorJitter, RandomHorizontalFlip, RandomAffine, RandomRotation, RandomErasing) Trad+Style augmentation

Acknowledgement:

  • Thanks to Federal University of Rio de Janeiro (UFRJ) which helped me with the infrastructure and motivation to do this project.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages