Nothing Special   »   [go: up one dir, main page]

Skip to main content

The Simulation of Adaptive Coverage Path Planning Policy for an Underwater Desilting Robot Using Deep Reinforcement Learning

  • Conference paper
  • First Online:
Robot Intelligence Technology and Applications 7 (RiTA 2022)

Part of the book series: Lecture Notes in Networks and Systems ((LNNS,volume 642))

  • 1016 Accesses

Abstract

Sewer sedimentation tanks accumulate a lot of sludge all year round and require a lot of labor and resources to clean up. Such underwater cleaning has many difficulties in cleaning works. Desilting robots are widely used in sludge cleaning to reduce labor. However, since this method is usually planned based on manual experience, the operation is inefficient, not intelligent, and requires a certain amount of labor. To enable the desilting robot to intelligently plan rational cleaning paths, it uses the complete coverage path planning (CCPP) algorithm. Existing CCPP cannot meet the requirements of various environments. Therefore, in this paper, we propose to use the deep reinforcement learning (DRL) algorithm to learn the sewer sedimentation tank environment and to find the optimal cleaning path. Experiments show that 2000 episodes are trained in a simplified simulation environment using deep Q network (DQN) and double DQN (DDQN), respectively. DQN only got a cumulative reward of 1000 in the 1228th episode, while DDQN got a cumulative reward of 8000 in the 343rd episode and completed the task. Therefore, the slag removal robot using DDQN as the control strategy can adaptively complete the CCPP problem.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Noh, D., Lee, W., Kim, H.R., Cho, I.S., Shim, I.B., Baek, S.: Adaptive coverage path planning policy for a cleaning robot with deep reinforcement learning. In: 2022 IEEE International Conference on Consumer Electronics (ICCE), pp. 1–6. IEEE (2022). https://doi.org/10.1109/ICCE53296.2022.9730307

    Chapter  Google Scholar 

  2. Mnih, V., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529–533 (2015)

    Article  Google Scholar 

  3. Galceran, E., Carreras, M.: A survey on coverage path planning for robotics. Robot. Auton. Syst. 61(12), 1258–1276 (2013)

    Article  Google Scholar 

  4. Kollar, T., Roy, N.: Trajectory optimization using reinforcement learning for map exploration. Int. J. Robot. Res. 27(2), 175–196 (2008)

    Article  Google Scholar 

  5. Lei, X., Zhang, Z., Dong, P.: Dynamic path planning of unknown environment based on deep reinforcement learning. J. Robot. 2018, 1–10 (2018)

    Article  Google Scholar 

  6. Nguyen, Q.P., Low, B.K.H., Jaillet, P.: Inverse reinforcement learning with locally consistent reward functions. Technical report (2015)

    Google Scholar 

  7. Wulfmeier, M., Rao, D., Wang, D.Z., Ondruska, P., Posner, I.: Large-scale cost function learning for path planning using deep inverse reinforcement learning. Int. J. Robot. Res. 36(10), 1073–1087 (2017)

    Article  Google Scholar 

Download references

Acknowledgment

This work was partly supported by the Technological Innovation R&D Program (S3264239) funded by the Ministry of SMEs and Startups, and the Technological Innovation R&D Program (S3154675) funded by the Ministry of SMEs and Startups (MSS, Korea).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Chang Gyoon Lim .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhao, Y., Sun, P., Lim, C. (2023). The Simulation of Adaptive Coverage Path Planning Policy for an Underwater Desilting Robot Using Deep Reinforcement Learning. In: Jo, J., et al. Robot Intelligence Technology and Applications 7. RiTA 2022. Lecture Notes in Networks and Systems, vol 642. Springer, Cham. https://doi.org/10.1007/978-3-031-26889-2_7

Download citation

Publish with us

Policies and ethics