Baseline model for "GraspNet-1Billion: A Large-Scale Benchmark for General Object Grasping" (CVPR 2020).
- Python 3
- PyTorch 1.6
- Open3d >=0.8
- TensorBoard 2.3
- NumPy
- SciPy
- Pillow
- tqdm
Get the code.
git clone https://github.com/graspnet/graspnet-baseline.git
cd graspnet-baselineInstall packages via Pip.
pip install -r requirements.txtCompile and install pointnet2 operators (code adapted from votenet).
cd pointnet2
python setup.py installCompile and install knn operator (code adapted from pytorch_knn_cuda).
cd knn
python setup.py installInstall graspnetAPI for evaluation.
git clone https://github.com/graspnet/graspnetAPI.git
cd graspnetAPI
pip install .install realsene
pip install pyrealsense2Tolerance labels are not included in the original dataset, and need additional generation. Make sure you have downloaded the orginal dataset from GraspNet. The generation code is in dataset/generate_tolerance_label.py. You can simply generate tolerance label by running the script: (--dataset_root and --num_workers should be specified according to your settings)
cd dataset
sh command_generate_tolerance_label.shOr you can download the tolerance labels from Google Drive/Baidu Pan and run:
mv tolerance.tar dataset/
cd dataset
tar -xvf tolerance.tarTraining examples are shown in command_train.sh. --dataset_root, --camera and --log_dir should be specified according to your settings. You can use TensorBoard to visualize training process.
Testing examples are shown in command_test.sh, which contains inference and result evaluation. --dataset_root, --camera, --checkpoint_path and --dump_dir should be specified according to your settings. Set --collision_thresh to -1 for fast inference.
The pretrained weights can be downloaded from:
checkpoint-rs.tar[Google Drive] [Baidu Pan]checkpoint-kn.tar[Google Drive] [Baidu Pan]
checkpoint-rs.tar and checkpoint-kn.tar are trained using RealSense data and Kinect data respectively.
A demo program is provided for grasp detection and visualization using RGB-D images. You can refer to command_demo.sh to run the program. --checkpoint_path should be specified according to your settings (make sure you have downloaded the pretrained weights, we recommend the realsense model since it might transfer better). The output should be similar to the following example:
Try your own data by modifying get_and_process_data() in demo.py. Refer to doc/example_data/ for data preparation. RGB-D images and camera intrinsics are required for inference. factor_depth stands for the scale for depth value to be transformed into meters. You can also add a workspace mask for denser output.
Results "In repo" report the model performance with single-view collision detection as post-processing. In evaluation we set --collision_thresh to 0.01.
Evaluation results on RealSense camera:
| Seen | Similar | Novel | |||||||
|---|---|---|---|---|---|---|---|---|---|
| AP | AP0.8 | AP0.4 | AP | AP0.8 | AP0.4 | AP | AP0.8 | AP0.4 | |
| In paper | 27.56 | 33.43 | 16.95 | 26.11 | 34.18 | 14.23 | 10.55 | 11.25 | 3.98 |
| In repo | 47.47 | 55.90 | 41.33 | 42.27 | 51.01 | 35.40 | 16.61 | 20.84 | 8.30 |
Evaluation results on Kinect camera:
| Seen | Similar | Novel | |||||||
|---|---|---|---|---|---|---|---|---|---|
| AP | AP0.8 | AP0.4 | AP | AP0.8 | AP0.4 | AP | AP0.8 | AP0.4 | |
| In paper | 29.88 | 36.19 | 19.31 | 27.84 | 33.19 | 16.62 | 11.51 | 12.92 | 3.56 |
| In repo | 42.02 | 49.91 | 35.34 | 37.35 | 44.82 | 30.40 | 12.17 | 15.17 | 5.51 |
Please cite our paper in your publications if it helps your research:
@article{fang2023robust,
title={Robust grasping across diverse sensor qualities: The GraspNet-1Billion dataset},
author={Fang, Hao-Shu and Gou, Minghao and Wang, Chenxi and Lu, Cewu},
journal={The International Journal of Robotics Research},
year={2023},
publisher={SAGE Publications Sage UK: London, England}
}
@inproceedings{fang2020graspnet,
title={GraspNet-1Billion: A Large-Scale Benchmark for General Object Grasping},
author={Fang, Hao-Shu and Wang, Chenxi and Gou, Minghao and Lu, Cewu},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition(CVPR)},
pages={11444--11453},
year={2020}
}
All data, labels, code and models belong to the graspnet team, MVIG, SJTU and are freely available for free non-commercial use, and may be redistributed under these conditions. For commercial queries, please drop an email at fhaoshu at gmail_dot_com and cc lucewu at sjtu.edu.cn .
python 01_inference_from_camera.py
WARNING:root:Failed to import geometry msgs in rigid_transformations.py.
WARNING:root:Failed to import ros dependencies in rigid_transforms.py
WARNING:root:autolab_core not installed as catkin package, RigidTransform ros methods will be unavailable
Traceback (most recent call last):
File "/data/hdd1/storage/junpeng/ws_anygrasp/graspnet-baseline/01_inference_from_camera.py", line 23, in <module>
from graspnet_dataset import GraspNetDataset # 数据集
File "/data/hdd1/storage/junpeng/ws_anygrasp/graspnet-baseline/dataset/graspnet_dataset.py", line 12, in <module>
from torch._six import container_abcs
ImportError: cannot import name 'container_abcs' from 'torch._six' (/home/junpeng.hu/anaconda3/envs/anygrasp2/lib/python3.9/site-packages/torch/_six.py)
conda create -n anygrasp_base python=3.9
pip install torch==1.8.0+cu111 torchvision==0.9.0+cu111 -f https://download.pytorch.org/whl/torch_stable.html
pip install numpy Pillow scipy tqdm open3d tensorboard
cd pointnet2 python setup.py install
cd knn python setup.py install
pip install graspnetAPI
import graspnetAPI
WARNING:root:Failed to import geometry msgs in rigid_transformations.py.
WARNING:root:Failed to import ros dependencies in rigid_transforms.py
WARNING:root:autolab_core not installed as catkin package, RigidTransform ros methods will be unavailable
没问题,可以正常运行
python 01_inference_from_camera.py --checkpoint_path /data/hdd1/storage/junpeng/ws_anygrasp/graspnet-baseline/ckpt/checkpoint-rs.tar




