Skip to content

The official implementation for "Diffusion Probabilistic Models for Scene-Scale 3D Categorical Data"

License

Notifications You must be signed in to change notification settings

zoomin-lee/scene-scale-diffusion

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

37 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Diffusion Probabilistic Models for Scene-Scale 3D Categorical Data

📌Paper

Comparison of object-scale and scene scale generation (ours). Our result includes multiple objects in a generated scene, while the object-scale generation crafts one object at a time. (a) is obtained by Point-E

Abstract

In this paper, we learn a diffusion model to generate 3D data on a scene-scale. Specifically, our model crafts a 3D scene consisting of multiple objects, while recent diffusion research has focused on a single object. To realize our goal, we represent a scene with discrete class labels, i.e., categorical distribution, to assign multiple objects into semantic categories. Thus, we extend discrete diffusion models to learn scene-scale categorical distributions. In addition, we validate that a latent diffusion model can reduce computation costs for training and deploying. To the best of our knowledge, our work is the first to apply discrete and latent diffusion for 3D categorical data on a scene-scale. We further propose to perform semantic scene completion (SSC) by learning a conditional distribution using our diffusion model, where the condition is a partial observation in a sparse point cloud. In experiments, we empirically show that our diffusion models not only generate reasonable scenes, but also perform the scene completion task better than a discriminative model.

Instructions

Dataset

: We use CarlaSC cartesian dataset.

Training

: There are some argparse in 'SSC_train.py'.

python SSC_train.py 
  • For multi-GPU : --distribution True
  • For Discrete Diffusion Model : --mode gen/con/vis
  • For Latent Diffusion Model : --mode l_vae/l_gen --l_size 882/16162/32322 --init_size 32 --l_attention True --vq_size 100

Example for training l_gen mode

python SSC_train.py --mode l_gen --vq_size 100 --l_size 32322 --init_size 32 --l_attention True --log_path ./result --vqvae_path ./lst_stage.tar

Visualization

: We save the result to a txt file using the utils/table.py/visulization function. If you use open3d, you will be able to easily visualize it.

Result

3D Scene Generation

image

Semantic Scene Completion

image

image

Acknowledgments

This project is based on the following codebase.

About

The official implementation for "Diffusion Probabilistic Models for Scene-Scale 3D Categorical Data"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages