![]() Matrix-Game-2 |
![]() Hunyuan-GameCraft |
![]() Hunyuan-Worldplay |
![]() Lingbot-World |
![]() YUME-1.5 |
![]() FlashWorld |
![]() Wan-2.2-IT2V |
![]() WoW |
![]() Cosmos-Predict-2.5 |
![]() Pi3 |
![]() Libero |
![]() Ai2-THOR |
We define a world model as: A model or framework centered on perception, equipped with interaction and long-term memory capabilities, for understanding and predicting the complex world. Accordingly, π Multimodal Understanding, π€ Visual Action Prediction, and πΌοΈ Visual Generation are all sub-tasks that a world model needs to accomplish.
We warmly welcome researchers to share their views on this framework or thoughts on world models in the Issues section. We also hope that you can submit valuable world-model-related methods to our framework via Pull Requests, or document and submit them to [awesome_world_models]. Feel free to give our repo a star π to follow the latest progress of OpenWorldLib!
The following three documents are essential to this project (click to navigate):
- docs/planning.md: This document tracks the short-term optimization goals and future development plans for OpenWorldLib.
- docs/awesome_world_models.md: This document records cutting-edge research, related surveys, and open-source projects on world models.
- docs/installation.md: This document provides installation instructions for different methods in OpenWorldLib.
The main goals of OpenWorldLib include:
- Establishing a unified and standardized world model framework to make the invocation of existing world-model-related code more consistent and well-structured;
- Integrating open-source world model research outcomes and systematically curating related papers for researchers' reference and use.
OpenWorldLib covers the following research directions related to World Models, We sincerely thank all the excellent methods included in this framework for their significant contributions to world model:
| Task Category | Sub-task | Representative Methods/Models |
|---|---|---|
| Video Generation | Navigation Generation | lingbot, matrix-game, hunyuan-worldplay, genie3, etc. |
| Long Video Generation | sora-2, veo-3, wan, etc. | |
| 3D Scene Generation | 3D Scene Generation | flash-world, vggt, etc. |
| Reasoning | VQA (Visual Question Answering) | spatialVLM, omnivinci and other VLMs with world understanding |
| VLA (Vision-Language-Action) | pi-0, pi-0.5, giga-brain, etc. |
First, create a conda environment:
conda create -n "openworldlib" python=3.10 -y
conda activate "openworldlib"Then install using the provided script:
cd OpenWorldLib
bash scripts/setup/default_install.shSome methods have special installation requirements. All installation scripts are located in ./scripts/setup.
π For the full installation guide, please refer to docs/installation.md
After installing the base environment, you can test matrix-game-2 generation and multi-turn interaction with the following commands:
cd OpenWorldLib
bash scripts/test_inference/test_nav_video_gen.sh matrix-game-2
bash scripts/test_stream/test_nav_video_gen.sh matrix-game-2Scripts for other methods can be found under scripts/test_inference and scripts/test_stream. Currently, we are primarily using GPUs with 80GB and 141GB of VRAM for testing. In the future, we will test more models and provide updates in the ./docs/installation.md file.
To help developers and users better understand OpenWorldLib, we provide details about our codebase. The framework structure is as follows:
OpenWorldLib
ββ assets
ββ data # Test data
β ββ benchmarks
β β ββ reasoning
β ββ test_case
β ββ ...
ββ docs # Documentation
ββ examples # Benchmark examples
ββ scripts # All key test scripts
ββ src
β ββ openworldlib # Main source path
β ββ base_models # Base models
β β ββ diffusion_model
β β β ββ image
β β β ββ video
β β β ββ ...
β β ββ llm_mllm_core
β β β ββ llm
β β β ββ mllm
β β β ββ ...
β β ββ perception_core
β β β ββ detection
β β β ββ general_perception
β β β ββ ...
β β ββ three_dimensions
β β ββ depth
β β ββ general_3d
β β ββ ...
β ββ memories # Memory module
β β ββ reasoning
β β ββ visual_synthesis
β ββ operators # Input & interaction signal processing
β ββ pipelines # All runtime pipelines
β ββ reasoning # Reasoning module
β β ββ audio_reasoning
β β ββ general_reasoning
β β ββ spatial_reasoning
β ββ representations # Representation module
β β ββ point_clouds_generation
β β ββ simulation_environment
β ββ synthesis # Generation module
β ββ audio_generation
β ββ visual_generation
β ββ vla_generation
ββ submodules # Auxiliary installs (e.g., diff-gaussian-raster)
ββ test # All test code
ββ test_stream # All interactive test code
ββ tools # Utilities
ββ installing
ββ vibe_codeWhen using OpenWorldLib, users typically call the pipeline class directly, which handles weight loading, environment initialization, and other tasks. Users interact with the operator class, and leverage the synthesis, reasoning, and representation classes for generation. In multi-turn interactions, the memory class is used to maintain the running context.
- We document the latest cutting-edge world model research in docs/awesome_world_models.md, and welcome contributions of valuable research.
- We document our upcoming training and optimization plans in docs/planning.md.
We welcome all developers to contribute and help improve OpenWorldLib as a unified world model repository. We recommend using Vibe Coding for quick code contributions β related prompts can be found under tools/vibe_code/prompts. You are also encouraged to add high-quality world model works to docs/planning.md and docs/awesome_world_models.md. We look forward to your contributions!
Related documents: [Project Overview] | [Development Guide] | [Task Assignment Details] | [Code Submission Guidelines]
This project is an extension of DataFlow and DataFlow-MM for world model tasks. We are also actively collaborating with RayOrch, Paper2Any, and other projects.
If OpenWorldLib has been helpful to you, please consider giving our repo a star π and citing the related papers:
@misc{dataflow-team-openworldlib,
author = {{OpenDCAI}},
title = {OpenWorldLib: A Unified Codebase for Advanced World Models},
year = {2026},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/OpenDCAI/OpenWorldLib}}
}
@article{zeng2026research,
title={Research on World Models Is Not Merely Injecting World Knowledge into Specific Tasks},
author={Zeng, Bohan and Zhu, Kaixin and Hua, Daili and Li, Bozhou and Tong, Chengzhuo and Wang, Yuran and Huang, Xinyi and Dai, Yifan and Zhang, Zixiang and Yang, Yifan and others},
journal={arXiv preprint arXiv:2602.01630},
year={2026}
}To further elaborate on our framework's design philosophy and our understanding of world models, we will release a technical report for OpenWorldLib. We hope our work is helpful to you!












