We provide an available conda environment named ldm:
sh init/create_env.sh
conda activate ldm
You can find instructions to download the dataset in site
In Our Method, we use the sequences:
boreas-2021-05-06-13-19boreas-2021-04-08-12-44boreas-2020-12-01-13-26boreas-2021-01-26-11-22
-
Set
data_rootinconfigs/autoencoder/boreas/autoencoder.yamlandconfigs/ldm/boreas/ldm.yaml -
To train LDM, add
path/to/your/autoencoder/ckpttockpt_pathinconfigs/ldm/boreas/ldm.yaml
# train an autoencoder
python main.py -b configs/autoencoder/boreas/autoencoder.yaml -t --gpus 0,
# train LDM
python main.py -b configs/ldm/boreas/ldm.yaml -t --gpus 0,
- To resume your training from an existing log directory or an existing checkpoint file, use the flag
-r:
# using a log directory
python main.py -b path/to/your/config.yaml -t --gpus 0, -r path/to/your/log/
# or, using a checkpoint
python main.py -b path/to/your/config.yaml -t --gpus 0, -r path/to/your/ckpt
Pre-trained weights are avilable in site
- To generate augmentation with adverse weather conditions (including postprocessing):
python scripts/generate_snow_data -r path/to/your/ldm/ckpt
- To open the generated .bin file:
import numpy as np
path = path/to/your/generated/file
pcd = np.fromfile(os.path.join(path), dtype=np.float64).reshape((-1, 3))
- Statistical metric:
- Jensen-Shannon Divergence (JSD)
- Distance metric:
- Chamfer Distance (CD)
pip install open3d
python scripts/autoencoder_metrics.py -r path/to/your/autoencoder/ckpt
You can evaluate Our Method in 3D Object Detection using our benchmark in site
Our codebase builds on site
@inproceedings{matteazzi2025augmentation,
title={Augmentation of LiDAR Scenes with Adverse Weather Conditions Using Latent Diffusion Models},
author={Matteazzi, Andrea and Arnold, Michael and Tutsch, Dietmar},
booktitle={International Conference on Neural Information Processing},
pages={198--212},
year={2025},
organization={Springer}
}