Zihui Zhang, Bo Yang, Bing Wang, Bo Li
We propose the first unsupervised 3D semantic segmentation method, learning from growing superpoints in point clouds.
Our method demonstrates promising results on multiple datasets:
- S3DIS Dataset
- ScanNet Dataset
- SemanticKITTI Dataset
Setting up for this project involves installing dependencies.
To install all the dependencies, please run the following:
sudo apt install build-essential python3-dev libopenblas-dev
conda env create -f env.yml
conda activate growsp
pip install -U MinkowskiEngine --install-option="--blas=openblas" -v --no-deps
S3DIS dataset can be found here.
Download the files named "Stanford3dDataset_v1.2_Aligned_Version.zip". Uncompress the folder and move it to
${your_S3DIS}
.
And there is an error in line 180389
of file Area_5/hallway_6/Annotations/ceiling_1.txt
.
It need to be fixed manually.
- Preparing the dataset:
python data_prepare/data_prepare_S3DIS.py --data_path ${your_S3DIS}
This code will preprcocess S3DIS and put it under ./data/S3DIS/input
- Construct initial superpoints:
python data_prepare/initialSP_prepare_S3DIS.py
This code will construct superpoints on S3DIS and put it under ./data/S3DIS/initial_superpoints
- Training:
CUDA_VISIBLE_DEVICES=0, python train_S3DIS.py
The output model and log file will be saved in ./ckpt/S3DIS
by default.
Download the ScanNet dataset from the official website.
You need to sign the terms of use. Uncompress the folder and move it to
${your_ScanNet}
.
- Preparing the dataset:
python data_prepare/data_prepare_ScanNet.py --data_path ${your_ScanNet}
This code will preprcocess ScanNet and put it under ./data/ScanNet/processed
- Construct initial superpoints:
python data_prepare/initialSP_prepare_ScanNet.py
This code will construct superpoints on ScanNet and put it under ./data/ScanNet/initial_superpoints
- Training:
CUDA_VISIBLE_DEVICES=0, python train_ScanNet.py
The output model and log file will be saved in ./ckpt/ScanNet
by default.
Please first download the following iterms from SemanticKITTI:
Uncompressed and Merge the velodyne
and labels
of each sequence.
The organized dataset should be as follows:
your_SemanticKITTI
└── sequences
└── 00
│ ├── velodyne
│ ├── labels
└── 01
...
- Preparing the dataset:
python data_prepare/data_prepare_SemanticKITTI.py --data_path ${your_SemanticKITTI}/sequences
This code will preprcocess SemanticKITTI and put it under ./data/SemanticKITTI/dataset
- Construct initial superpoints:
python data_prepare/initialSP_prepare_SemanticKITTI.py
This code will construct superpoints on SemanticKITTI and put it under ./data/SemanticKITTI/initial_superpoints
- Training:
CUDA_VISIBLE_DEVICES=0, python train_SemanticKITTI.py
The output model and log file will be saved in ./ckpt/SemanticKITTI
by default.
The trained models for these three datasets can be found at here