Robot grasp detection based on key point estimation
Tóm tắt
Grasping is an important general ability for robots to work in aerospace and other fields. An accurate grasping detection result is the premise and key step for robots to complete grasping. To reduce the parameter quantity and computational complexity of the grasping detection model, improve the grasping detection accuracy and real-time performance, this paper proposes a grasping detection algorithm based on key point estimation. First, this model focuses on finding the center point of the grasp rectangle, and then obtains the best grasp through its position on feature heat map. Second, for RGB-D multi-modal input, an improved residual block combined with squeeze-and-excitation block is used as the feature extraction layer to explicitly learn multi-channel weight information. Different from the anchor-based detection algorithm that exhausts the possible positions of the target and needs to score the grasp candidate after classification and regression, the proposed model obtains best grasp by directly predicting the position of the center point, angle, and open width of gripper. The number of parameters in this model is only around 482 k, which is less than one-third of regular general model. The results on the Cornell Grasp Dataset show that the model achieves an accuracy of 97.75% and runs at 24.7 frames per second.
Tài liệu tham khảo
Asif U, Bennamoun M, Sohel FA (2017) RGB-D object recognition and grasp detection using hierarchical cascaded forests. IEEE Trans Rob 33(3):547–564
Bicchi A, Kumar V (2000) Robotic grasping and contact: a review. In Proceedings 2000 ICRA. Millennium conference. IEEE international conference on robotics and automation. Symposia proceedings (Cat. No. 00CH37065) (Vol 1, pp 348–353). IEEE
Bochkovskiy A, Wang CY, Liao HYM (2020) Yolov4: Optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934
Bohg J, Morales A, Asfour T, Kragic D (2013) Data-driven grasp synthesis—a survey. IEEE Trans Rob 30(2):289–309
Chu FJ, Xu R, Vela PA (2018) Real-world multiobject, multigrasp detection. IEEE Robot Autom Lett 3(4):3355–3362
Dumoulin V, Visin F (2016) A guide to convolution arithmetic for deep learning. arXiv preprint arXiv:1603.07285
Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 580–587)
Guo D, Sun F, Liu H, Kong T, Fang B, Xi N (2017) A hybrid deep architecture for robotic grasp detection. In: 2017 IEEE International Conference on Robotics and Automation (ICRA) (pp. 1609–1614). IEEE
Hara K, Vemulapalli R, Chellappa R (2017) Designing deep convolutional neural networks for continuous object orientation estimation. arXiv preprint arXiv:1702.01499
Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition (pp 7132–7141)
Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. JMLR.org.
Jiang Y, Moseson S, Saxena A (2011) Efficient grasping from rgbd images: Learning using a new rectangle representation. In: 2011 IEEE International conference on robotics and automation (pp. 3304–3311). IEEE
Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980
Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25
Kumra S, Kanan C (2017) Robotic grasp detection using deep convolutional neural networks. In: 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 769–776). IEEE
Kumra S, Joshi S, Sahin F (2020) Antipodal robotic grasping using generative residual convolutional neural network. In: 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp 9626–9633). IEEE
Lenz I, Lee H, Saxena A (2013) Deep learning for detecting robotic grasps. The International Journal of Robotics Research 34(4–5)
Miller AT, Allen PK (2004) Graspit! a versatile simulator for robotic grasping. IEEE Robot Autom Magn 11(4):110–122
Morrison D, Corke P, Leitner J (2020) Learning robust, real-time, reactive robotic grasping. Int J Robot Res 39(2–3):183–201
Pelossof, R., Miller, A., Allen, P., & Jebara, T. (2004, April). An SVM learning approach to robotic grasping. In IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA'04. 2004 (Vol. 4, pp. 3512–3518). IEEE.
Redmon, J., & Angelova, A. (2015). Real-time grasp detection using convolutional neural networks. In 2015 IEEE international conference on robotics and automation (ICRA) (pp. 1316–1322). IEEE.
Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arXiv preprint arXiv:1804.02767
Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779–788).
Ren, S., He, K., Girshick, R., & Sun, J. (2015). Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems, 28
Ruder S (2016) An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747
Wang Z, Li Z, Wang B, Liu H (2016) Robot grasp detection using multimodal deep convolutional neural networks. Adv Mech Eng 8(9):1687814016668077
Zhou X, Lan X, Zhang H, Tian Z, Zhang Y, Zheng N (2018) Fully convolutional grasp detection network with oriented anchor box. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 7223–7230). IEEE.
Zhou X, Wang D, Krähenbühl P (2019) Objects as points. arXiv preprint arXiv:1904.07850