Skip to content

RicoFio/disentangle_mlp

Repository files navigation

MLP Semester 2 - Group Coursework Group 079

We introduce a modern retake on the paper "Autoencoding beyond pixels using a learned similarity matrix" by Larsen et al. (paper). By introducing a beta-VAE architecture we improve on the latent space disentangling and present a new beta-VAE-GAN hybrid.

Abstract

Recent research in probabilistic generative models based on deep neural networks has led to image generation systems of a quality previously unseen. We re-explore an algorithm first introduced by that combines Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs). Its aim is to combine the strength of the two approaches by informing the VAE's loss with the GAN's discriminator, thus creating a feature-wise loss. This helps find disentangledlatent representations, which often better capture features of the data capable of generalizing beyond training samples. These result in improved sample generation. We explore the reproducibility challenges of the algorithm, compare it with VAE and GAN and augment it with $\beta$-VAE, an extension that has been shown to improve the disentanglement of latent representation in VAE. This choice involves the tuning of one key hyperparameter: we avoid the expensive heuristics proposed in the literature and show improved results on our baselines with a simple Bayesian optimization procedure with a 10 % decrease in Frechet Inception Distance score.

Setup

  • Install dependencies from requirements.txt
  • Download data from here

References

  • "Autoencoding beyond pixels using a learned similarity matrix" by Larsen et al. (paper)
  • " beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework." (paper)

Change Log

  • Imported boilerplate from here

About

Improved disentangling of VAE-GAN on CelebA

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages