The MCIGP is designed to realize grasping in large-scale dense clutter scenarios. Specifically, the first part is the Monozone View Alignment (MVA), wherein we design the dynamic monozone that can align the camera view according to different objects during grasping, thereby alleviating view boundary effects and realizing grasping in large-scale dense clutter scenarios. Then, we devise the Instance-specific Grasp Detection (ISGD) to predict and optimize grasp candidates for one specific object within the monozone, ensuring an in-depth analysis of this object.
arXiv | All Experimental Videos
If you use this work, please cite (initial version):
@inproceedings{clee2025pmsgp,
title={Pyramid-Monozone Synergistic Grasping Policy in Dense Clutter},
author={Chenghao, Li and Nak Young, Chong},
booktitle={https://arxiv.org/abs/2409.06959},
year={2024}
}
Contact
Any questions or comments, contact Chenghao Li.
This code was developed with Python 3.7. Requirements can be installed by:
pip install -r requirements.txt
The code was deployed by the UFactory 850/Xarm5 Robot and Intel RealSense D435i.
- UFactory Robot API: https://github.com/xArm-Developer/xArm-Python-SDK.
- Intel Realsense API: https://github.com/IntelRealSense/librealsense.
- Download and extract the Cornell Dataset.
- Download and extract the OCID Dataset.
Has been included in this code as 'GRconvnet_RGBD_epoch_40_iou_0.52'.
Please refer to this repository https://github.com/facebookresearch/segment-anything.
Training is done by the train_network
, and predicting is done by grasp detection
.
- Real robot grasping is done by
MCIGP grasping
. - Note: Please use your own hand-eye calibration results when deploying.