Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Optimizing robotic arm control using deep Q-learning and artificial neural networks through demonstration-based methodologies: : A case study of dynamic and static conditions

Published: 21 November 2024 Publication History

Highlights

Intelligent machines were designed to execute tasks in hazardous environments.
The robot is required to perceive and identify objects in the vicinity.
Path planning is another important aspect of mobile robot navigation.

Abstract

This paper uses robot programming techniques, such as Deep Q Network, Artificial Neural Network, and Artificial Deep Q Network, to address challenges related to controlling robotic arms through demonstration learning. Static and dynamic states of the subjects were the subjects of experiments. Each method's classification accuracy process success values and experimental condition combination were evaluated. The DQN method demonstrated favourable classification accuracy outcomes, achieving an Accuracy value of 0.64 for the fixed dice and 0.52 for the moving dice. The Response value was 0.51 for the fixed dice and 0.41 for the moving dice, indicating a moderate level. The ANN method demonstrated lower accuracy, with Accuracy values of 0.59 and 0.56 and Response values of 0.61 and 0.58, respectively. The ADQN method demonstrated superior outcomes, with Accuracy values of 0.66 and 0.59 and Response values of 0.67 and 0.61. During the initial learning iterations, ADQN demonstrated the highest success rate at 33.67 %, whereas DQN and ANN achieved 28.39 % and 20.13 % success rates, respectively. As the number of iterations increased, all methods demonstrated improvement in their results. ADQN maintained a high success rate of 97.59 %, while DQN and ANN attained 82.16 % and 88.66 %, respectively. As the number of iterations increases, the results of all methods improve, but the success rate of the Artificial Deep Q Network remains high. As the number of iterations increases, both Deep Q Network and Artificial Neural Network demonstrate the potential to achieve good results. Overall, the findings support the efficacy of robot programming techniques that incorporate demonstration learning. The Artificial Deep Q Network is the most successful and fast-converging method suitable for various robot control tasks. These findings provide a foundation for future research and large-scale, comprehensive learning applications for complex rot control.

References

[1]
S. O'Sullivan, N. Nevejans, C. Allen, A. Blyth, S. Leonard, U. Pagallo, K. Holzinger, A. Holzinger, M.I. Sajid, H. Ashrafian, Legal, regulatory, and ethical frameworks for the development of standards in artificial intelligence (AI) and autonomous robotic surgery, Int. J. Med. Robot. Comput. Assist. Surg. 15 (1) (2019),. art. no. e1968.
[2]
P.K. Donepudi, Reinforcement learning for robotic grasping and manipulation: a review, Asia Pacific J. Energy Environ. 7 (2) (2020) 69–78,.
[3]
L. Chang, L. Shan, C. Jiang, Y. Dai, Reinforcement-based mobile robot path planning with improved dynamic window approach in an unknown environment, Auton. Robot. 45 (2021) 51–76,.
[4]
Q. Wu, H. Lin, Y. Jin, Z. Chen, S. Li, D. Chen, A new fallback beetle antennae search algorithm for path planning of mobile robots with collision-free capability, Soft Comput 24 (2020) 2369–2380,.
[5]
X. Xiao, B. Liu, G. Warnell, P. Stone, Motion planning and control for mobile robot navigation using machine learning: a survey, Auton. Robot. 46 (5) (2022) 569–597,.
[6]
K. Zhang, Z. Yang, T. Başar, Multi-agent reinforcement learning: a selective overview of theories and algorithms, Handbook of Reinforcement Learning and Control, Springer, Cham, 2021, pp. 321–384,.
[7]
L.X. Shi, J.J. Lim, Y. Lee, Skill-based model-based reinforcement learning, (2022) arXiv:2207.07560.
[8]
Z. Fan, W. Zhang, W. Liu, Multi-agent deep reinforcement learning based distributed optimal generation control of DC microgrids, IEEE Trans. Smart Grid (2023),. in press.
[9]
K.H. Fu, Using Machine Learning to Learn from demonstration: Application to the AR. Drone quadrotor control, Ph.D. dissertation, Faculty of Science, School of Computer Science, University of the Witwatersrand, Witwatersrand, 2016.
[10]
L.B.D.C.S. Santos, Learning from Demonstration Using Hierarchical Inverse Reinforcement learning, Ph.D. Dissertation, Universidade do Porto, Porto, 2021.
[11]
S.M. Park, Y.G. Kim, A metaverse: taxonomy, components, applications, and open challenges, IEEE Access 10 (2022) 4209–4251,.
[12]
T. Peirelinck, H. Kazmi, B.V. Mbuwir, C. Hermans, F. Spiessens, J. Suykens, G. Deconinck, Transfer learning in demand response: a review of algorithms for data-efficient modelling and control, Energy AI 7 (2022),. art. no. 100126.
[13]
P. Budhwar, A. Malik, M.T. De Silva, P. Thevisuthan, Artificial intelligence-challenges and opportunities for international HRM: a review and research agenda, Int. J. Human Resourc. Manag. 33 (6) (2022) 1065–1097,.
[14]
F. Bounini, D. Gingras, H. Pollart, D. Gruyer, Modified artificial potential field method for online path planning applications, in: 2017 IEEE Intelligent Vehicles Symposium (IV), IEEE, Piscataway, NJ, 2017, pp. 180–185,.
[15]
M.G. Mohanan, A. Salgoankar, A survey of robotic motion planning in dynamic environments, Rob. Auton. Syst. 100 (2018) 171–185,.
[16]
A.K. Rath, D.R. Parhi, H.C. Das, P.B. Kumar, M.K. Muni, K. Salony, Path optimisation for navigation of a humanoid robot using hybridised fuzzy-genetic algorithm, Int. J. Intell. Unmanned Syst. 7 (3) (2019) 112–119,.
[17]
S.P. Subramanian, S.K. Chandrasekar, Simultaneous allocation and sequencing of orders for robotic mobile fulfillment system using reinforcement learning algorithm, Expert Syst. Appl. 239 (2024),.
[18]
L. Zhang, Z. Cai, Y. Yan, C. Yang, Y. Hu, Multi-agent policy learning-based path planning for autonomous mobile robots, Eng. Appl. Artif. Intell. 129 (2024),.
[19]
C. Chen, X.Q. Chen, F. Ma, X.J. Zeng, J. Wang, A knowledge-free path planning approach for smart ships based on reinforcement learning, Ocean Eng. 189 (2019),. art. no. 106299.
[20]
J. Jiang, J. Xin, Path planning of a mobile robot in a free-space environment using Q-learning, Prog. Artif. Intell. 8 (2019) 133–142,.
[21]
M.U. Khan, Mobile robot navigation using reinforcement learning in unknown environments, Balk. J. Electr. Comput. Eng. 7 (3) (2019) 235–244,.
[22]
F. Gul, W. Rahiman, S.S. Nazli Alhady, A comprehensive study for robot navigation techniques, Cogent Eng 6 (1) (2019),. art. no. 1632046.
[23]
E.S. Low, P. Ong, K.C. Cheah, Solving the optimal path planning of a mobile robot using improved Q-learning, Rob. Auton. Syst. 115 (2019) 143–161,.
[24]
N. Saqib, M.M. Yousuf, Design and implementation of shortest path line follower autonomous rover using decision-making algorithms, in: 2021 Asian Conference on Innovation in Technology (ASIANCON), IEEE, Piscataway, NJ, 2021, pp. 1–6,.
[25]
S.M. Khansari-Zadeh, A. Billard, Learning control Lyapunov function to ensure stability of dynamical system-based robot reaching motions, Rob. Auton. Syst. 62 (6) (2014) 752–765,.
[26]
C. Song, G. Liu, X. Zhang, X. Zang, C. Xu, J. Zhao, Robot complex motion learning based on unsupervised trajectory segmentation and movement primitives, ISA Trans. 97 (2020) 325–335,.
[27]
S. James, E. Johns, 3D simulation for robot arm control with deep Q-learning, (2016) arXiv:1609.03759.
[28]
K. Kim, J. Cho, J. Pyo, S. Kang, J. Kim, Dynamic object recognition using precise location detection and ANN for robot manipulator, in: 2017 International Conference on Control, Artificial Intelligence, Robotics & Optimization (ICCAIRO), IEEE, Piscataway, NJ, 2017, pp. 237–241,.
[29]
M. Alakuijala, G. Dulac-Arnold, J. Mairal, J. Ponce, C. Schmid, Residual reinforcement learning from demonstrations, (2021) arXiv:2106.08050.
[30]
A. Sanmorino, A study for DDOS attack classification method, J. Phys. Conf. Ser. 1175 (1) (2019),.
[31]
T. Hester, M. Vecerik, O. Pietquin, M. Lanctot, T. Schaul, B. Piot, D. Horgan, J. Quan, A. Sendonaris, I. Osband, G. Dulac-Arnold, J. Agapiou, J.Z. Leibo, A. Gruslys, Deep Q-learning from demonstrations, in: Proceedings of the AAAI Conference on Artificial Intelligence, AAAI Press, Washington, DC, 2018, pp. 3223–3230. art. no. 394.
[32]
S. Calinon, A. Billard, A probabilistic programming by demonstration framework handling constraints in joint space and task space, in: 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, IEEE, Piscataway, NJ, 2008, pp. 367–372,.
[33]
A. Vakanski, I. Mantegh, A. Irish, F. Janabi-Sharifi, Trajectory learning for robot programming by demonstration using hidden Markov model and dynamic time warping, in: IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 42, IEEE, Piscataway, NJ, 2012, pp. 1039–1052,.
[34]
A. Ly, R. Dazeley, P. Vamplew, F. Cruz, S. Aryal, Elastic step DQN: a novel multi-step algorithm to alleviate overestimation in Deep Q-Networks, Neurocomput 576 (2024),.
[35]
C. Xiangrui, K. Gang, L. Tie, P. Yi, Jie Ke versus AlphaGo: a ranking approach using decision making method for large-scale data with incomplete information, Eur. J. Oper. Res. 265 (1) (2018) 239–347,.
[36]
S. Liu, Y. Liu, L. Hu, Z. Zhou, Y. Xie, Z. Zhao, W. Li, Z. Gan, DiffSkill: improving reinforcement learning through diffusion-based skill denoiser for robotic manipulation, Knowl.-Based Syst (2024),.
[37]
Y. Yin, Z. Chen, G. Liu, J. Yin, J. Guo, Autonomous navigation of mobile robots in unknown environments using off-policy reinforcement learning with curriculum learning, Expert Syst. Appl. 247 (2024),.

Index Terms

  1. Optimizing robotic arm control using deep Q-learning and artificial neural networks through demonstration-based methodologies: A case study of dynamic and static conditions
      Index terms have been assigned to the content through auto-classification.

      Recommendations

      Comments

      Please enable JavaScript to view thecomments powered by Disqus.

      Information & Contributors

      Information

      Published In

      cover image Robotics and Autonomous Systems
      Robotics and Autonomous Systems  Volume 181, Issue C
      Nov 2024
      181 pages

      Publisher

      North-Holland Publishing Co.

      Netherlands

      Publication History

      Published: 21 November 2024

      Author Tags

      1. Robotic arm
      2. Demonstration learning
      3. Deep learning
      4. ADQN
      5. Task success

      Qualifiers

      • Research-article

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • 0
        Total Citations
      • 0
        Total Downloads
      • Downloads (Last 12 months)0
      • Downloads (Last 6 weeks)0
      Reflects downloads up to 01 Dec 2024

      Other Metrics

      Citations

      View Options

      View options

      Login options

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media