Abstract
The visual path–following technology is widely used in cutting, laser welding, painting, gluing, and other fields, which is a crucial content of robotics studies. As an important algorithm of artificial intelligence (AI), reinforcement learning provides a new insight for robots to learn path-following skills which has the ability of machine vision and decision making. In order to build a robotic agent with path-following skills, this paper proposes a visual path–following algorithm based on artificial intelligence deep reinforcement learning double deep Q-network (DDQN). The proposed approach allows the robot to learn path-following skill by itself, using a visual sensor in the Robot Operating System (ROS) simulation environment. The robot can learn paths with different textures, colors, and shapes, which enhances the flexibility for different industrial robot application scenarios. Skills acquired in simulation can be directly translated to the real world. In order to verify the performance of the path-following skill, a path randomly hand-drawn on the workpiece is tested by the six-joint robot Universal Robots 5 (UR5). The simulation and real experiment results demonstrate that robots can efficiently and accurately perform path following autonomously using visual information without the parameters of the path and without programming the path in advance.
Similar content being viewed by others
References
Ghobakhloo M (2022) Industry4.0, digitization, and opportunities for sustainability. J Clean Prod 252:119869
Duan YQ, Edwards JS, Dwivedi YK (2019) Artificial intelligence for decision making in the era of Big Data – evolution, challenges and research agenda. Int J Inform Manage 48:63–71
Pérez L, Rodríguez Í, Rodríguez N, Usamentiaga R, García DF (2016) Robot guidance using machine vision techniques in industrial environments: a comparative review. Sensors 16(335)
Wang ZG, Wang HT, She Q, Shi XS, Zhang YM (2020) Robot4.0: continual learning and spatial-temporal intelligence through edge. J Comp Res Dev 57(9)1854–1863
Mnih V, Kavukcuoglu K, Silver D, Antonoglou I, Wierstra D, Riedmiller M (2013) Playing Atari with deep reinforcement learning. Comput Sci
Neto P, Mendes N, Ara´ujo R, Pires JN, Moreira AP (2012) High-level robot programming based on CAD: dealing with unpredictable environments. Ind Robot Int J Robot Res Appl 39(3):294–303
Polden J, Pan Z, Larkin N, Van Duin S, Norrish J (2011) Offline programming for a complex welding system using DELMIA automation. In: Chen SB, Fang G (eds) Robotic welding, intelligence and automation. Springer, Berlin, pp 341–349
Mnih V, Kavukcuoglu K, Silver D (2015) Human-level control through deep reinforcement learning. Nature 518(7540):529–533
Pahič R, Lončarević Z, Gams A, Ude A (2021) Robots kill learning in latent space of a deep autoencoder neural network. Robot Auton Syst 135:103690
Bedaka AK, Vidal J, Lin CY (2019) Automatic robot path integration using three-dimensional vision and offline programming. Int J Adv Manuf Technol 102:1935–1950
Deng D, Polden JW, Dong JF, Tao PY (2018) Sensor guided robot path generation for surface repair tasks on a large-scale buoyancy module. Ieee-Asme T Mech 23:636–645
Tian YX, Liu HF, Li L, Wang WB, Feng JC, Xi FF, Yuan GJ (2020) Robust identification of weld seam based on region of interest operation. Adv Manuf 8:473–485
Shah HNM, Sulaiman M, Shukor AK, Kamis Z (2018) Butt welding joints recognition and location identification by using local thresholding. Robot Cim-Int Manuf 51:181–188
Yang L, liu YH, Peng JZ, Liang ZZ, (2020) A novel system for offline 3D seam extraction and path planning based on point cloud segmentation for arc welding robot. Robot Cim-Int Manuf 64:101929
Zou YB, Wei XZ, Chen JX (2020) Conditional generative adversarial network-based training image inpainting for laser vision seam tracking. Opt Laser Eng 134:10614
Zhou CM, Huang BD, Fränti P (2022) A review of motion planning algorithms for intelligent robots. J Intell Manuf 33:387–424
Liu NJ, Tao L, Cai YH, Wang S (2019) A review of robot manipulation skills learning methods. Acta Automatica Sinica 45(3):458–470
Moravčík M, Schmid M, Burch N, Lisý V, Morrill D, Bard N, Davis T, Waugh K, Johanson M, Bowling M (2017) DeepStack: expert-level artifcial intelligence in heads-up no-limit poker. Science 356:508–513
Silver D, Hubert T, Schrittwieser J, Antonoglou I, Lai M, Guez A, Lanctot M, Sifre L, Kumaran D, Graepel T, Lillicrap T, Simonyan K, Hassabis D (2017) Mastering chess and shogi by self-play with a general reinforcement learning algorithm
Jaderberg M, Czarnecki WM, Dunning I, Marris L, Lever G, Castañeda AG, Beattie C, Rabinowitz NC, Morcos AS, Ruderman A, Sonnerat N, Green T, Deason L, Leibo JZ, Silver D, Hassabis D, Kavukcuoglu K, Graepel T (2019) Human-level performance in 3D multiplayer games with population-based reinforcement learning. Science 364:859–865
Taitler A, Shimkin N (2017) Learning control for air hockey striking using deep reinforcement learning. Int Conf Cont Artif Intel Robot Optim IEEE
Zeng A, Song S, Welker S (2018) Learning synergies between pushing and grasping with self-supervised deep reinforcement learning. IEEE/RSJ Int Conf Intel Robot Syst (IROS) 4238–4245
Hundt A, Killeen B, Greene N, Wu HT (2020) “Good robot!”: efficient reinforcement learning for multi-step visual tasks with sim to real transfer. Ieee Robot Autom Let 5(4):6724–6731
Guo xw, Peng GZ, Meng YY (2021) A modified Q‑learning algorithm for robot path planning in a digital twin assembly system. Int J Adv Manuf Technol
Li FM, Jiang Q, Zhang SS, Wei M, Song R (2019) Robot skill acquisition in assembly process using deep reinforcement learning. Neurocomputing 345:92–102
Wen SH, Zhao YF, Yuan X, Wang ZT, Zhang D, Manfredi LG (2020) Path planning for active SLAM based on deep reinforcement learning under unknown environments. Intel Serv Robot 13:262–272
Zhang T, Xiao M, Zou YB, Xiao JD (2020) Robotic constant-force grinding control with a press-and-release model and model-based reinforcement learning. Int J Adv Manuf Technol 106:589–602
Meyes R, Tercan H, Roggendorf S, Thiele T, Büscher C, Obdenbusch M, Brecher C, Jeschke S, Meisen T (2017) Motion planning for industrial robots using reinforcement learning. The 50th CIRP Conf Manufac Syst
Hasselt HV, Guez A, Silver D (2016) Deep reinforcement learning with double Q-learning. Proceed 13th AAAI Conf Artif Intel 2094–2100
Schaul T, Quan J, Antonoglou I (2016) Prioritized experience replay. Proceed 4th Int Conf Learn Represent 322–355
Shelhamer E, Long J, Darrell T (2017) Fully convolutional networks for semantic segmentation. Ieee T Pattern Anal 39(4):640–651
Le N, Rathour VS, Yamazaki K, Luu K, Savvides M (2021) Deep reinforcement learning in computer vision: a comprehensive survey. Artif Intell Rev
Quigley M, Gerkey B, Conley K, Faust J (2009) ROS: an open-source Robot Operating System. ICRA workshop on open source software
Funding
This research is supported by the Key Laboratory Open Fund in Autonomous Region (2020520002) and the Key Research and Development Program in Autonomous Region (202003129).
Author information
Authors and Affiliations
Contributions
Guoliang Liu: investigation, conceptualization, code, software, experiment, data curation, and writing of the manuscript. Wenlei Sun: funding acquisition, conceptualization, project administration. Wenxian Xie: software, code, methodology, experiment. Yangyang Xu: ROS platform, system, code, experiment, software.
Corresponding author
Ethics declarations
Ethics approval
This manuscript has not been published elsewhere, in whole or in part. All authors have made significant contributions and agree with the content of the manuscript.
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Liu, G., Sun, W., Xie, W. et al. Learning visual path–following skills for industrial robot using deep reinforcement learning. Int J Adv Manuf Technol 122, 1099–1111 (2022). https://doi.org/10.1007/s00170-022-09800-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00170-022-09800-1