Skip to content

Processing pipeline designed to segment point clouds acquired in forests

Notifications You must be signed in to change notification settings

julesmorel/PointNet2

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Segmentation based on Pointnet++ (PyTorch)

Processing pipeline designed to segment point clouds acquired in forests


Overview

This method relies on both a geometric and a deep learning approach to:

  1. Identify the ground points on a complete TLS scan (Morel et al. 2020).
  2. Separate the wood points from the leaves points on scanned trees (Morel et al. 2020).

Setup

See the official code release "PointNet++" for model definitions and hyper-parameters. The custom operations used by Pointnet++ are ONLY supported using CUDA.

Requirements

  • Linux (tested on Ubuntu 20.04.4 LTS and 21.10)
  • PCL 1.11
  • PDAL 2.4.2
  • Python 3.9.7
  • PyTorch 1.10.2
  • CUDA Toolkit 11.3

Install

Install this library by running the following command:

./install.sh

Segmentation

Several pre-trained models are provided in this package, they are stored into the models folder.

.
├── ...
├── models                    	             # Pytorch models file
│   ├── terrain_segmentation  	             #  Terrain/Vegetation segmentation models
│   │   ├── model_terrain_01_128             #   Fine model (2D grid resolution:10cm, N neighbors PCA:128)
│   │   └── model_terrain_02_64              #   Coarse model (2D grid resolution:20cm, N neighbors PCA:64)
│   ├── wood_segmentation    	             #  Wood/leaves segmentation models 	
│   │   ├── model_seg_landes_pca5cm	     #   Model trained on vegetation from Landes (Radius PCA:5cm)
│   │   ├── model_seg_landes_pca10cm	     #   Model trained on vegetation from Landes (Radius PCA:10cm)
│   │   ├── model_seg_landes_low_pca5cm	     #   Model trained on smaller trees from Landes (Radius PCA:5cm)
|   │   ├── model_seg_landes_mixed_pca10cm   #   Model trained on both big and small trees from Landes (Radius PCA:10cm)
│   │   ├── model_seg_sologne_5cm	         #   Model trained on vegetation from Sologne (Radius PCA:5cm)
│   │   ├── model_seg_sologne_10cm	        #   Model trained on vegetation from Sologne (Radius PCA:10cm)  
│   │   └── model_seg_vosges                 #   Model trained on vegetation from Vosges (Radius PCA:5cm)					          
└── ...

For both terrain and wood points segmentation, the method follows the steps described in (Morel et al. 2020):

  1. The point cloud is subsampled.
  2. The point cloud is enriched by computing local geometric descriptors at each points.
  3. The cloud is divided into several overlapping batches whose size fits the input layer of the deep learning model.
  4. The model predicts a label for each point of each batch.
  5. As the batches are overlapping, each point of the input point cloud receives several prediction. The final label is then inferred by scrolling the set of batches and applying a voting process.

Note: The initial filtering on the input point cloud and by the computation of the geometric local descriptors differ for both segmentation problem:

  • Ground points segmentation: filtering of the input scan through a coarse 2D XY grid (~10cm), which tends to make the point density uniform so the local descriptors are computed with PCA amongst the K neighbors
  • Wood points segmentation: filtering of the input scan through a fine 3D grid (0.5cm). As the point density stays non uniform the local descriptors are computed with PCA considering the neighbors in a sphere of given radius.

The processing of complete LiDAR multi-scans have revealed different problematics in the terrain and in the wood points segmentation. Therefore, the implementation of the preparation and the data slicing differs. Both following subsections give the details of the two approaches.

Ground points segmentation

Considering a set of files, each file storing a LiDAR scan of the forest plot, the method first subsamples each point cloud and crops it to the desired extent. The resulting point clouds are then merged, then segmented.

screenshot

In order to segment the ground points from the vegetation points, first edit the parameters in segment_terrain.sh then call the script:

./segment_terrain.sh INPUT_FILE_1 ... INPUT_FILE_N MODEL_PATH

where INPUT_FILE_1 ... INPUT_FILE_N (N>=1) are the paths to the files (las/laz/ascii) containing the point clouds to segment and MODEL_PATH is the path to the model used for the inference.

Note: if the input files are in las/laz format, the offset related to the georeference translation is stored in offset.txt in the local folder.

Wood points segmentation

Besides the set LiDAR files, this approach takes also as input the digital terrain model in order to filter out the points close to the ground.

The implementation is made of the following steps:

  • For each LiDAR scan of the forest plot:
  1. the point cloud is cropped to the desired extent then subsampled
  2. The points below a given threshold Zmin are filtered out.
  3. The remaining points are tilled into fixed size tiles in order to limit memory consumption.
  4. For each tile, the segmentation occurs. Then, the results of each tile are merged back together.
  • The result of each scan are merged together.

screenshot

In order to segment the wood from the leaves points, first edit the parameters in segment_wood.sh then call the script:

./segment_wood.sh INPUT_FILE_1 ... INPUT_FILE_N DTM_FILE MODEL_PATH

where INPUT_FILE_1 ... INPUT_FILE_N (N>=1) are the paths to the files (las/laz/ascii) containing the point clouds to segment, DTM_FILE is a GDAL raster of the digital terrain model on the extent and MODEL_PATH is the path to the model used for the inference.

Note: if the input files are in las/laz format, the offset related to the georeference translation is stored in offset.txt in the local folder.


Additional steps

Two additional computation steps have been implemented to improve the terrain segmentation results:

  1. a filtering method based on Statistical Outlier Removal and Radius Outlier Removal to clean the input point cloud
  2. a clustering method which aims at cleaning the segmentation results

Filtering

As a preliminary step before the inference, filtering of the noise can be applied by a custom filter based on Statistical Outlier Removal and Radius Outlier Removal:

.outliersFilter/outliersFilter INPUT_FILE OUTPUT_FILE meanK stddevMulThresh radiusSearch minNeighborsInRadius

where:

  • INPUT_FILE and OUTPUT_FILE are the paths to the ascii file in input and output respectively.
  • meanK and stddevMulThresh are the number of neighbors to analyze for each point and the standard deviation multiplier.
  • radiusSearch is the sphere radius that is to be used for determining the k-nearest neighbors for filtering and minNeighborsInRadius is the minimum number of neighbors that a point needs to have in the given search radius in order to be considered an inlier.

We usually use:

.outliersFilter/outliersFilter INPUT_FILE OUTPUT_FILE 128 1.0 1. 50

Clustering

We observed that the usual segmentation errors result in small patch of points, clearly away from the main ground points cluster.

screenshot

To cope with this issue, we propose to retrieve the main cluster of points using the following script:

.clustering/clustering INPUT_FILE OUTPUT_FILE clusterTolerance minClusterSize maxClusterSize

where:

  • INPUT_FILE and OUTPUT_FILE are the paths to the ascii file in input and output respectively.
  • clusterTolerance is the minimum distance between 2 clusters
  • minClusterSize and maxClusterSize are the lower and upper limits of the clusters size

We usually use:

.clustering/clustering INPUT_FILE OUTPUT_FILE 0.2 10 10000000

Training of a custom model

The library provides also several scritps to train new pytorch models. The data should be labelized point clouds stored in ASCII files (4 columns: X Y Z label) and separated in 2 folders: training and validation.

Dataset preparation

Given TRAIN_DIR and VALIDATION_DIR, both directories containing the training data and the validation data respectively, format the input data by running:

./prepare_data_terrain.sh TRAIN_DIR
./prepare_data_terrain.sh VALIDATION_DIR

(replace prepare_data_terrain.sh with prepare_data_wood.sh if you are training a model to segment wood from leaves points)

Training

Edit train.sh to setup the parameters, then run:

./train.sh

Classification

Once the DTM has been computed and the trees isolated, the classification can be assigned to the input .laz files by running the following command:

./classify.sh LAZ_FILES DTM_TIF TREES_ASCII

Where:

  • LAZ_FILES is a glob that lists all the input .laz files
  • DTM_TIF is a raster file storing the DTM
  • TREES_ASCII is an ASCII file containing the point clouds of every isolated trees

We use for instance:

./classify.sh ~/Data/LiDAR/GEDI/GEDI009/211011_gedi_090* ~/Data/LiDAR/GEDI/GEDI009/dtm.tif ~/Data/LiDAR/GEDI/GEDI009/alltrees.xyz

Acknowledgement

About

Processing pipeline designed to segment point clouds acquired in forests

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published