Abstract
This paper presents an effective neural network, Residual Generative Grasping Convolutional Neural Network (ReGG-CNN), to predict reliable robotic grasps for novel objects based on an n-channel scene image. Based on Generative Grasping Convolutional Neural Network (GG-CNN) architecture, ReGG-CNN incorporates residual blocks to enhance performance while preserving the network’s lightweight and single-pass generative nature. ReGG-CNN is evaluated on two standard open-source datasets, Cornell and Jacquard grasping datasets, achieving 87% and 89% accuracy, respectively. Moreover, multiple cluttered scenes are created using novel household and adversarial objects to assess the model’s capability to generalize to most kinds of objects and make multi-grasp predictions. Even though only training on individual objects, ReGG-CNN effectively predicts grasps for diverse objects in cluttered scenes.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Redmon, J.: Real-time grasp detection using convolutional neural networks. In: 2015 IEEE International Conference on Robotics and Automation (ICRA), pp. 1316–1322. Seattle (2015)
Kumra, S.: Robotic grasp detection using deep convolutional neural networks. In: 2017 IEEE International Conference on Intelligent Robots and Systems (IROS), pp. 769–776. Vancouver (2017)
Morrison, D.: Closing the Loop for Robotic Grasping: a Real-Time, Generative Grasp Synthesis Approach (2018). arXiv preprint arXiv:1804.05172
He, K.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778. Las Vegas (2016)
Kumra, S.: Antipodal robotic grasping using generative residual convolutional neural network. In: 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 9626–9633. Las Vegas (2020)
Lenz, I.: Deep learning for detecting robotic grasps. Int. J. Robot. Res. 34(4–5), 705–724 (2015)
Jiang, Y.: Efficient grasping from RGBD images: learning using a new rectangle representation. In: IEEE International Conference on Robotics and Automation (ICRA), pp. 3304–3311. Shanghai (2011)
Depierre, A.: Jacquard: a large scale dataset for robotic grasp detection. In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 3511–3516. Madrid (2018)
Kingma, D.P.: Adam: a Method For Stochastic Optimization. arXiv preprint arXiv:1412.6980
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zhang, Z. et al. (2024). Enhancing Pixel-Wise Robotic Grasping with Residual Blocks. In: Park, J.S., Yang, L.T., Pan, Y., Park, J.J. (eds) Advances in Computer Science and Ubiquitous Computing. CUTECSA 2023. Lecture Notes in Electrical Engineering, vol 1190. Springer, Singapore. https://doi.org/10.1007/978-981-97-2447-5_35
Download citation
DOI: https://doi.org/10.1007/978-981-97-2447-5_35
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-97-2446-8
Online ISBN: 978-981-97-2447-5
eBook Packages: EngineeringEngineering (R0)