A pipeline for online 3D reconstruction of objects only using visual data from RGB-D camera (Kinect V2).
A simple demo of 3D reconstruction of a sneaker 'Kyrie4 CNY' with this pipeline. More examples see 'Demo' part.
Kinect V2 as the RGB-D camera for my pipeline, and use ROS framework to allow communication between hardware and software, as well as different programs.
I used ROS package iai_kinect2 to drive the RGB-D camera, do registration of range image and color image, and generate raw point cloud as input data of this 3D reconstruction pipeline.
The pipeline is divided into three major parts. I used PCL to preprocessing raw point cloud and visualize registration process. For registration part, Open3D.
Open3D is a brand new library for 3D data processing, which was just released in Jan 2018. The reason I used this library for point cloud registration is that they implemented the state-of-the-art Colored ICP algorithm. Plus, the library has Python interface with complete documentation and examples, which is easy to use.
The image above shows how effective this algorithm is. The left only used ICP with normal info of each points, while the right image first used Normal ICP for rough registration, plus Colored ICP for refinement. For some objects with rich textures on surface, with color information synthesized, this algorithm would be super powerful.
This pipeline works in ROS framework.
- Input data is raw point cloud (from topic
/kinect2/qhd/points
generated by Kinect V2) - Preprocessing (PCL, C++)
- input: raw point cloud
- Crop ROI of raw point cloud
- Segment and get rid of the bottom turntable plane
- Remove outliers with statistical method
- output: preprocessed point cloud
- input: raw point cloud
- Registration (Open3D, Python)
- input: preprocessed point cloud
- Set the first point cloud as fixed base cloud
- For every coming point cloud, locally regitering it into last point cloud
- Normal ICP for rough registration
- Colored ICP for refinement
- Utilize some schemes to get rid of bad local registration
- When local registration is successful, regiter it into the base cloud
- Repetitive do step 2 and step 3 after a complete model is generated
- output: compelete point cloud rendered model of objects
- input: preprocessed point cloud
- Visualization (PCL, C++)
- input: compelete point cloud rendered model of objects
- Update the visualization window step by step, showing the process of point cloud registration
- input: compelete point cloud rendered model of objects
The image above shows the setup of the 3D reconstruction pipeline. With all the ROS nodes ready, when people rotates the turntable slowly, the 3D reconstruction process is shown on the computer screen.
Feel free to download this pipeline and play with it. Hope you enjoy it. There are some tips you may want to know for an easy start.
- Kinect V2 (Kinect for XBOX One)
- Turntable
- An Ubuntu computer with ROS installed
Tips:
- The turntable should not be reflective, or Kinect V2 will generate point cloud with distortion.
- The relative position of Kinect V2 and the turntable should be set up before running the software. For this part, I will provide with an interface for easy use.
- ROS
- catkin_tools (use
catkin build
to build this ROS package) - iai_kinect2 (ROS package as driver for Kinect V2)
- PCL1.7.2 (C++)
- Open3D (Python)
- numpy (Python)
Intsructions for use:
- In a ROS environment, command
$ roslaunch kinect_bridge kinect_bridge.launch
to connect Kinect V2 thtough iai_kinect2; - Command
$ roslaunch scanner pipeline.launch
to run the pipeline. - Intermediate PCD files are saved in ./data/ folder, and the final reconstruction PCD file is in ./data/result/ folder.