Multiple imputation utilising denoising autoencoder for approximate Bayesian inference
Switch branches/tags
Nothing to show
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Failed to load latest commit information.
Examples Initial restructure for installer Sep 15, 2018
midas Update Sep 22, 2018
LICENSE.txt Create LICENSE.txt Jan 8, 2018 Update Sep 22, 2018 Update Jan 9, 2018 test Sep 18, 2018

MIDAS - Multiple Imputation with Denoising Autoencoders

MIDAS draws on recent advances in deep learning to deliver a fast, scalable, and high-performance solution for multiply imputing missing data. MIDAS employs a class of unsupervised neural networks known as denoising autoencoders, which are capable of producing complex yet robust reconstructions of partially corrupted inputs. To enhance their efficiency and accuracy while preserving their robustness, these networks are trained with the recently developed technique of Monte Carlo dropout, which is mathematically equivalent to approximate Bayesian inference in deep Gaussian processes. Preliminary tests indicate that, in addition to handling larger datasets than existing multiple imputation algorithms, MIDAS generates more accurate and precise imputed values in ordinary applications.


To install via pip, input the following command into the terminal:
pip install git+

MIDAS requires

  • Python (>=3.5, 2.X coming)
  • Numpy (>=1.5)
  • Pandas (>=0.19)
  • Tensorflow (>= 1.10)
  • Matplotlib

Tensorflow also has a number of requirements, particularly if GPU acceleration is desired. See for details.


Variational autoencoder enabled. More flexibility in model specification, although defaulting to a simple mirrored system. Deeper analysis tools within .overimpute() for checking fit on continuous values. Constructor code deconflicted. Individual output specification enabled for very large datasets.

Key added features

  • Variational autoencoder capacity added, including encoding to and sampling from latent space

Planned features:

  • Time dependence handling through recurrent cells
  • Improving the pipeline methods for very large datasets
  • Tensorboard integration
  • Dropout scaling
  • A modified constructor that can generate embeddings for better interpolation of features
  • R support

Wish list:

  • Smoothing for time series (LOESS?)
  • Informative priors?

Previous versions

Alpha 0.1:

Basic functionality feature-complete.

  • Support for mixed categorical and continuous data types
  • An "additional data" pipeline, allowing data that may be relevant to the imputation to be included (without being included in error generating statistics)
  • Simplified calibration for model complexity through the "overimputation" function, including visualization of reconstructed features
  • Basic large dataset functionality