Latent diffusion models have established a new state-of-the-art in high-resolution visual generation. Integrating Vision Foundation Model priors improves generative efficiency, yet existing latent designs remain largely heuristic. These approaches often struggle to unify semantic discriminability, reconstruction fidelity, and latent compactness. In this paper, we propose Geometric Autoencoder (GAE), a principled framework that systematically addresses these challenges. By analyzing various alignment paradigms, GAE constructs an optimized low-dimensional semantic supervision target from VFMs to provide guidance for the autoencoder. Furthermore, we leverage latent normalization that replaces the restrictive KL-divergence of standard VAEs, enabling a more stable latent manifold specifically optimized for diffusion learning. To ensure robust reconstruction under high-intensity noise, GAE incorporates a dynamic noise sampling mechanism. Empirically, GAE achieves compelling performance on the ImageNet-1K
- [2026.03.12]: Core code released! Includes DiT training and inference based on GAE latent space.
- [2026.03.12]: Pre-trained weights for GAE-AE and DiT are available on Hugging Face.
Geometric Autoencoder (GAE) is a principled framework designed to systematically address the heuristic nature of latent space design in Latent Diffusion Models (LDMs). GAE significantly enhances semantic discriminability and latent compactness without compromising reconstruction fidelity through three core innovations:
- Latent Normalization: Replaces the restrictive KL-divergence of standard VAEs with RMSNorm regularization. By projecting features onto a unit hypersphere, GAE prevents training collapse and provides a stable, scalable latent manifold optimized for diffusion learning.
- Latent Alignment: Leverages Vision Foundation Models (VFMs, e.g., DINOv2) as semantic teachers. Through a carefully designed semantic downsampler, the low-dimensional latent vectors directly inherit strong discriminative semantic priors.
- Dynamic Noise Sampling: Specifically addresses the high-intensity noise typical in diffusion processes, ensuring robust reconstruction performance even under extreme noise levels.
GAE achieves state-of-the-art performance on the ImageNet-1K
Some pre-trained weights are hosted on Hugging Face.
| Model | Epochs | Latent Dim | gFID (w/o) | Weights |
|---|---|---|---|---|
| GAE-LightningDiT-XL | 80 | 32 | 1.82 | 🔗 HF Link |
| GAE-LightningDiT-XL | 800 | 32 | 1.31 | 🔗 HF Link |
| Model | Epochs | Latent Dim | Weights |
|---|---|---|---|
| GAE | 200 | 32 | 🔗 HF Link |
We use LightningDiT for the DiT implementation.
git clone https://github.com/sii-research/GAE.git
cd GAE
conda create -n gae python=3.10.12
conda activate gae
pip install -r requirements.txtDownload the pre-trained weights from Hugging Face and place them in the checkpoints/ folder. Ensure update the paths in the configs/ folder to match your local setup.
bash extract_gae.sh $DIT_CONFIG $VAE_CONFIGbash train_gae.sh $DIT_CONFIG $VAE_CONFIGFor class-uniform sampling:
bash inference_gae.sh $DIT_CONFIG $VAE_CONFIGFor class-random sampling:
Change "from inference_sample import" -> "from inference import" in inference_gae.pyOur project is built upon the excellent foundations of the following open-source projects:
-
LightningDiT: For the PyTorch Lightning based DiT implementation.
-
RAE: For the timeshift and class-uniform sampling implementation.
-
ADM: For the evaluation suite to score generated samples.
We express our sincere gratitude to the authors for their valuable contributions to the community.
If you find this work useful, please consider citing:
@misc{liu2026geometricautoencoderdiffusionmodels,
title={Geometric Autoencoder for Diffusion Models},
author={Hangyu Liu and Jianyong Wang and Yutao Sun},
year={2026},
eprint={2603.10365},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2603.10365},
}


