Skip to content

yaxingwang/UDIT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

30 Commits
 
 
 
 
 
 

Repository files navigation

Controlling biases and diversity in diverse image-to-image translation

Abstract:

The task of unpaired image-to-image translation is highly challenging due to the lack of explicitcross-domain pairs of instances. We consider here diverse image translation (DIT), an evenmore challenging setting in which an image can have multiple plausible translations. This isnormally achieved by explicitly disentangling content and style in the latent representationand sampling different styles codes while maintaining the image content. Despite the successof current DIT models, they are prone to suffer from bias. In this paper, we study the problemof bias in image-to-image translation. Biased datasets may add undesired changes (e.g. changegender or race in face images) to the output translations as a consequence of the particularunderlying visual distribution in the target domain. In order to alleviate the effects of thisproblem we propose the use of semantic constraints that enforce the preservation of desiredimage properties. Our proposed model is a step towards unbiased diverse image-to-imagetranslation (UDIT), and results in less unwanted changes in the translated images while stillperforming the wanted transformation. Experiments on several heavily biased datasets showthe effectiveness of the proposed techniques in different domains such as faces, objects, andscenes.

Overview

Dependences

Installation

  • Install tensorflow
  • Opencv

Instructions

git clone git@github.com:yaxingwang/UDIT.git

References

Contact

If you run into any problems with this code, please submit a bug report on the Github site of the project. For another inquries pleace contact with me: yaxing@cvc.uab.es

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages