Skip to content
Implementation of Apple's Learning from Simulated and Unsupervised Images through Adversarial Training
Branch: master
Clone or download
Latest commit f4920bd Jun 13, 2017


Keras implementation of Apple's Learning from Simulated and Unsupervised Images through Adversarial Training


Install dlutils from

$ pip install -U git+


$ git clone
$ python install develop


In apple's paper they use Unity Eyes to generate ~1.2 million synthetic images. I am on mac though so I just used the easily available SynthesEyes Dataset. This is small (only around ~11,000 images) so it would be much better if someone could generate a larger dataset w/ Unity Eyes and share it on s3.

The dataset of real image's used in apple's paper is the MPIIGaze Dataset. They use the normalized images provided in this dataset which are stored in matlab files. It was a bit of a pain to get these in an easily usable form so I'm sharing the ready to go datasets on s3.

Ready to go datasets


Implementation of 3.1 Appearance-based Gaze Estimation on UnityEyes and MPIIGaze datasets as described in paper.

  • Currently only Python 3 support.
  • Tensorflow support and maybe PyTorch support in future.


This is meant to be a light-weight and clean implementation that is easy to understand - no deep shit. It can also be used as a resource to understand GANs in general and how they can be implemented.

Running Online

You can see a interactive Jupyter Notebook version of this script with training data on Kaggle or just the raw training set

About is a company whose vision is a world where medical conditions are addressed early on, in their infancy. This approach will shift the health-care industry from a constant fire-fight against symptoms to a preventative approach where root causes are addressed and fixed. Our first step to make realize this vision is easy, accurate and available diagnosis. Please get in contact with me if this resonates with you!

You can’t perform that action at this time.