Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3523181.3523188acmotherconferencesArticle/Chapter ViewAbstractPublication PagesasseConference Proceedingsconference-collections
research-article

A two-stage spatial pose detection method for bearing rings based on YOLOv5 and binocular vision

Published: 18 April 2022 Publication History

Abstract

The bearing ring has the characteristics of weak texture and high reflection, which easily cause errors in stereo vision matching. To address these issues, this paper proposes a two-stage spatial pose detection method for bearing rings based on YOLOv5 and binocular vision. In Stage 1 called Pre-positioning: the YOLOv5 is used to detect the unobstructed bearing ring; then binocular vision is employed to locate the unobstructed bearing ring; and a line laser projection is performed on the unobstructed bearing ring for fine positioning. In Stage 2 called Fine-positioning: the YOLOv5 is used again to locate the line laser stripes projected on the upper surface of the bearing ring, and then the detected line laser stripes are used to calculate the robot's best grasping pose. The practical application showed that the introduction of line laser solved the problem of weak texture matching, and the application of yolov5 solved the problems of high reflection on the bearing ring surfaces and strong noises after laser reflection. Experiments showed that YOLOv5 detected the unobstructed bearing rings and laser line stripes with 98.5% and 99.3% mAP, respectively, and this method had a 96.3% grasping success rate.

References

[1]
Lenz, I., Lee, H., & Saxena, A. (2015). Deep learning for detecting robotic grasps.  The International Journal of Robotics Research, 34 (4–5), 705–724. https://doi.org/10.1177/0278364914549607
[2]
Bohg, J., Morales, A., Asfour, T., & Kragic, D. (2013). Data-driven grasp synthesis—a survey.  IEEE Transactions on Robotics, 30 (2), 289-309.
[3]
Morrison, D., Corke, P., & Leitner, J. (2020). Learning robust, real-time, reactive robotic grasping.  The International journal of robotics research, 39 (2-3), 183-201.
[4]
Collet, A., Berenson, D., Srinivasa, S. S., & Ferguson, D. (2009, May). Object recognition and full pose registration from a single image for robotic manipulation. In 2009 IEEE International Conference on Robotics and Automation (pp. 48-55). IEEE.
[5]
Wang, C., Cao, Y., Yu, F., & Tang, Y. (2021, February). Dynamic Weight of Adaptive Information Density Network for Image Super-Resolution. In 2021 2nd Asia Service Sciences and Software Engineering Conference (pp. 123-129).
[6]
Xu, M., Wang, X., Zhangt, X., Bin, G., Jia, Z., & Chen, K. (2020, May). Computation-Efficient Multi-Model Deep Neural Network for Sleep Stage Classification. In  Proceedings of the 2020 Asia Service Sciences and Software Engineering Conference (pp. 1-8).
[7]
Lingyu, Z., Bin, W., & Bai, W. (2020, May). Call Attention to Stances: Detect Rumor with a Stance Attention Network. In  Proceedings of the 2020 Asia Service Sciences and Software Engineering Conference (pp. 27-34).
[8]
Chen, W., Lei, Y., Liu, M., Zhang, J., & Yao, X. (2020, October). Workpiece Detection Based on Image Processing and Convolutional Neural Network. In 2020 International Conference on Sensing, Measurement & Data Analytics in the era of Artificial Intelligence (ICSMD) (pp. 302-306). IEEE.
[9]
Lei, Y., Yao, X., Chen, W., Zhang, J., Mehnen, J., & Yang, E. (2020, July). Multiple object detection of workpieces based on fusion of deep learning and image processing. In 2020 International Joint Conference on Neural Networks (IJCNN) (pp. 1-7). IEEE.
[10]
Jiao, L., Zhang, F., Liu, F., Yang, S., Li, L., Feng, Z., & Qu, R. (2019). A survey of deep learning-based object detection.  IEEE access, 7, 128837-128868.
[11]
Zou, Z., Shi, Z., Guo, Y., & Ye, J. (2019). Object detection in 20 years: A survey.  arXiv preprint arXiv:1905.05055.
[12]
Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You only look once: Unified, real-time object detection. In  Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779-788).
[13]
Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks.  Advances in neural information processing systems, 25, 1097-1105.
[14]
Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition.  arXiv preprint arXiv:1409.1556.
[15]
Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., ... & Rabinovich, A. (2015). Going deeper with convolutions. In  Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1-9).
[16]
Ren, S., He, K., Girshick, R., & Sun, J. (2015). Faster r-cnn: Towards real-time object detection with region proposal networks.  Advances in neural information processing systems, 28, 91-99.
[17]
Redmon, J., & Farhadi, A. (2018). Yolov3: An incremental improvement.  arXiv preprint arXiv:1804.02767.

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Other conferences
ASSE' 22: 2022 3rd Asia Service Sciences and Software Engineering Conference
February 2022
202 pages
ISBN:9781450387453
DOI:10.1145/3523181
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 18 April 2022

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Funding Sources

  • Zhanjiang Science and Technology Project
  • the National Natural Science Foundation of China under grant
  • Guangdong Basic and Applied Basic Research Foundation
  • the National Natural Science Foundation of China and the Royal Society of Edinburgh

Conference

ASSE' 22

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 47
    Total Downloads
  • Downloads (Last 12 months)6
  • Downloads (Last 6 weeks)0
Reflects downloads up to 21 Nov 2024

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media