Controllable Animation Video Generation with Large Models-based Multimodal Agents. The long-term aim of this project is to construct a long video generation agent to help everyone become a director, visualising their ideas.
If you appreciate our project, please consider giving us a star β on GitHub to stay updated with the latest developments.
2025.09.13 β¨ The AniMaker Homepage is now live!
2025.08.11 π AniMaker is conditionally accepted by Siggraph Asia 2025.
2025.06.10 π We release AniMaker, a multi-agent framework designed to efficiently generate coherent, long-form storytelling animations from text input.
2024.08.09 π We release Anim-Director, the first animation generation agent powered by LMM.
2024.07.30 π Anim-Director is conditionally accepted by Siggraph Asia 2024.
If you find this project useful in your research, please consider cite:
@inproceedings{li2024anim,
title={Anim-director: A large multimodal model powered agent for controllable animation video generation},
author={Li, Yunxin and Shi, Haoyuan and Hu, Baotian and Wang, Longyue and Zhu, Jiashun and Xu, Jinyi and Zhao, Zhen and Zhang, Min},
booktitle={SIGGRAPH Asia 2024 Conference Papers},
pages={1--11},
year={2024}
}
@inproceedings{shi2025animaker,
title={AniMaker: Multi-Agent Animated Storytelling with MCTS-Driven Clip Generation},
author={Shi, Haoyuan and Li, Yunxin and Chen, Xinyu and Wang, Longyue and Hu, Baotian and Zhang, Min},
booktitle={Proceedings of the SIGGRAPH Asia 2025 Conference Papers},
pages={1--11},
year={2025}
}