Skip to content

zdebruine/MMVAE

Repository files navigation

MMVAE

This repository is a base for a project on Mixture-of-experts Multimodal Variational Autoencoding (MMVAE).

Objective

The objective is to engineer an AI architecture for diagonal integration of multiple modalities by learning a shared latent space that gives cohesive and realistic joint generation across modalities which naturally cannot be encountered jointly.

Architecture

This model is designed for concurrent training on information across diverse modalities.

Generative core model

At the core of this model is a variational autoencoder that generates realistic human single-cell transcriptomes:

Expert multi-channel models

For zero-shot cross-generation across multiple modalities (e.g. species), modality-specific encoders and decoders are stacked at either end of the core VAE:

Adversarially-assisted multimodal integration

To help channels communicate with each other and synergize, adversarial feedback will be added to the encoder. This will ensure the core VAE encoder is not biased for one channel or the other:

Adversarial discrimination

Outputs from each channel will also be evaluated with a generative adversarial discriminator network to encourage realistic-looking distributions in context:

Literature Background

MMVAE is inspired by works on multi-modal MMVAE from Shi et. al (2019), adversarial integration from Kopp et. al. (2022), recognizing challenges highlighted in this review.

About

Mixture-of-Experts Multimodal Variational Autoencoder

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published