[Survey] Masked Modeling for Self-supervised Representation Learning on Vision and Beyond (https://arxiv.org/abs/2401.00897)
-
Updated
Jun 29, 2024 - Python
[Survey] Masked Modeling for Self-supervised Representation Learning on Vision and Beyond (https://arxiv.org/abs/2401.00897)
A collection of literature after or concurrent with Masked Autoencoder (MAE) (Kaiming He el al.).
Official repo for Recursion's accepted spotlight paper at NeurIPS 2023 Generative AI & Biology workshop.
An optimized implementation of masked autoencoders (MAEs)
Codebase for the paper 'EncodecMAE: Leveraging neural codecs for universal audio representation learning'
Masked Spectrogram Modeling using Masked Autoencoders for Learning General-purpose Audio Representations
Official Pytorch implementation of EVEREST: Efficient Masked Video Autoencoder by Removing Redundant Spatiotemporal Tokens [ICML2024].
R-MAE: Pre-training LiDAR Perception with Masked Autoencoders for Ultra-Efficient 3D Sensing
Masked Modeling Duo: Towards a Universal Audio Pre-training Framework
Design a patches masked autoencoder by CNN
Video Foundation Models & Data for Multimodal Understanding
Re-implementation of the method proposed in ''DreamDiffusion: Generating High-Quality Images from Brain EEG Signals'' by Y. Bai, X. Wang et al. for Neural Network Course exam Topics
[NeurIPS 2023] Masked Image Residual Learning for Scaling Deeper Vision Transformers
Cross-Sensor Masked Autoencoder for Content Based Image Retrieval in Remote Sensing
Investigate possibilities for Vision Transformers with multiscale grids
Official Implementation of the CrossMAE paper: Rethinking Patch Dependence for Masked Autoencoders
SimpleClick: Interactive Image Segmentation with Simple Vision Transformers (ICCV 2023)
An optimized implementation of spatiotemporal masked autoencoders
Official implementation of Matrix Variational Masked Autoencoder (M-MAE) for ICML paper "Information Flow in Self-Supervised Learning" (https://arxiv.org/abs/2309.17281)
HSIMAE: A Unified Masked Autoencoder with large-scale pretraining for Hyperspectral Image Classification
Add a description, image, and links to the masked-autoencoder topic page so that developers can more easily learn about it.
To associate your repository with the masked-autoencoder topic, visit your repo's landing page and select "manage topics."