Skip to content

aRI0U/RandLA-Net-pytorch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

49 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

RandLA-Net: Efficient Semantic Segmentation of Large-Scale Point Clouds

This repository contains a PyTorch implementation of RandLA-Net.

Preparation

  1. Clone this repository

    git clone https://github.com/aRI0U/RandLA-Net-pytorch.git
  2. Install all Python dependencies

  cd RandLA-Net-pytorch
  pip install -r requirements

Common issue: the setup file from torch-points-kernels package needs PyTorch to be previously installed. You may thus need to install PyTorch first and then torch-points-kernels.

  1. Download a dataset and prepare it. We conducted experiments with Semantic3D and S3DIS.

To setup Semantic3D:

cd RandLA-Net-pytorch/utils
./download_semantic3d.sh
python3 prepare_semantic3d.py # Very slow operation

To setup SDIS, register and then download the zip archive containing the files here. We used the archive which contains only the 3D point clouds with ground truth annotations.

Assuming that the archive is located in folder RandLA-Net-pytorch/datasets, then run:

cd RandLA-Net-pytorch/utils
python3 prepare_s3dis.py
  1. Finally, in order to subsample the point clouds using a grid subsampling, run:
cd RandLA-Net-pytorch/utils/cpp_wrappers
./compile_wrappers.sh   # you might need to chmod +x before
cd ..
python3 subsample_data.py

Usage

  • Train a model

    python3 train.py

    A lot of options can be configured through command-line arguments. Type python3 train.py --help for more details.

  • Evaluate a model

    python3 test.py

Visualization

One can visualize the evolution of the loss with Tensorboard.

On a separate terminal, launch:

tensorboard --logdir runs

Citation

This work implements the work presented in RandLA-Net: Efficient Semantic Segmentation of Large-Scale Point Clouds.

The original implementation (in TensorFlow 1) can be found here.

To cite the original paper:

@article{RandLA-Net,
  arxivId = {1911.11236},
  author = {Hu, Qingyong and Yang, Bo and Xie, Linhai and Rosa, Stefano and Guo, Yulan and Wang, Zhihua and Trigoni, Niki and Markham, Andrew},
  eprint = {1911.11236},
  title = {{RandLA-Net: Efficient Semantic Segmentation of Large-Scale Point Clouds}},
  url = {http://arxiv.org/abs/1911.11236},
  year = {2019}
}

Warning

This repository is still on update, and the segmentation results we reach with our implementation are for now not as good as the ones obtained by the original paper.