NeurIPS 2025 π₯π₯π₯π₯
A novel entropy-aware inference framework that enhances the denoising generation capabilities of diffusion models
If you find this work useful, please give us a star π.
Paper β’ OpenReview β’ arXiv β’ Quick Start β’ Citation
EVODiff is an efficient diffusion model inference framework grounded in entropy-aware information flow optimization. It systematically improves image quality and accelerates generation by optimizing conditional variance at each step, all without relying on reference trajectories.
We reveal that successful denoising generation of diffusion models fundamentally operates by reducing conditional entropy during reverse transitions, grounded in information-theoretic principles.
EVODiff stands out by being the first entropy-aware diffusion model inference framework for better generation by optimizing the denoising information flow of diffusion models.
- π‘οΈ Significant Theoretical Contribution: Provides the first rigorous mathematical proof that data-prediction parameterization is superior to noise-prediction for diffusion model inference, theoretically grounding previous empirical findings. π₯
- π Entropy-aware Denoising: Directly optimizes reconstruction error by systematically leveraging entropy-aware information flow and variance reduction.
- π Reference-free via On-the-fly Optimization: Achieves superior performance without relying on reference trajectories or costly optimization procedures (unlike methods that require optimization or distillation from
$\tilde{x}_0$ ).
Comparison of strategies employed for optimizing reconstruction error across different methods. EVODiff uniquely leverages entropy-aware denoising information flow for better inference with training-free optimization.
- [2025.9.19] π EVODiff has been accepted by NeurIPS 2025!
- [Coming Soon] π The official implementation code will be released soon!
EVODiff significantly improves generation quality, especially at low number of function evaluations (NFEs), by effectively reducing uncertainty and mitigating visual artifacts.
Qualitative comparisons on text-to-image generation using the simple prompt "Giant caterpillar riding a bicycle". EVODiff leverages entropy-aware information flow to reduce artifacts and enhance fidelity compared to SOTA solvers.
Extensive experiments demonstrate that EVODiff consistently outperforms SOTA gradient-based solvers in terms of both speed (NFE) and quality (FID).
Quantitative comparisons demonstrating EVODiff's consistent superior performance (lower FID) across diverse datasets (CIFAR-10 shown here) and varying NFEs.
# Clone the repository
git clone https://github.com/ShiguiLi/EVODiff.git
cd EVODiff
# Install dependencies
pip install -r requirements.txtTo use EVODiff, simply initialize the EVODiff_edm class and wrap the sampling process.
python sample.py \
--ckp_path="path/to/checkpoint.pkl" \
--sample_folder="my_output_folder" \
--method="evodiff" \
--steps=10 \
--order=2 \
--skip_type="logSNR" \
--denoise_to_zeroIf you find EVODiff useful for your research and applications, please cite our work:
π§© Conference Version (NeurIPS 2025)
@inproceedings{li2025evodiff,
title={{EVOD}iff: Entropy-aware Variance Optimized Diffusion Inference},
author={Shigui Li and Wei Chen and Delu Zeng},
booktitle={The Thirty-ninth Annual Conference on Neural Information Processing Systems},
year={2025},
url={https://openreview.net/forum?id=rKASv92Myl}
}π Preprint Version (arXiv)
@article{li2025evodiff,
title={{EVOD}iff: Entropy-aware Variance Optimized Diffusion Inference},
author={Li, Shigui and Chen, Wei and Zeng, Delu},
journal={arXiv preprint arXiv:2509.26096},
year={2025}
}


