CNN-Based QR Code Reading of Package for Unmanned Aerial Vehicle
<p>The picture of the unmanned aerial vehicle.</p> "> Figure 2
<p>The roll control (<b>a</b>) and yaw control (<b>b</b>) of the UAV.</p> "> Figure 3
<p>Experimental environment and package placement angle. (<b>a</b>) UAV flight field and frame, (<b>b</b>) fill light in the cabinet, and (<b>c</b>) package type.</p> "> Figure 4
<p>UAV flight process.</p> "> Figure 5
<p>UAV roll and yaw control by PID control.</p> "> Figure 6
<p>Various rotated packages.</p> "> Figure 7
<p>Data volume and distribution. (<b>a</b>) Training and testing data volume; (<b>b</b>) data volume distribution.</p> "> Figure 8
<p>Identification process of a QR code.</p> "> Figure 9
<p>CNN architecture used in this paper.</p> "> Figure 10
<p>Training and validation of the proposed CNN, performed using SDG. (<b>a</b>) Training and validation accuracy with 92.0% accuracy, (<b>b</b>) training and validation loss, and (<b>c</b>) confusion matrix.</p> "> Figure 11
<p>Training and validation of the proposed CNN, performed using RMSprop. (<b>a</b>) Training and validation accuracy with 92.9% accuracy, (<b>b</b>) training and validation loss, and (<b>c</b>) confusion matrix.</p> "> Figure 12
<p>Training and validation of the proposed CNN, performed using Adadelta. (<b>a</b>) Training and validation accuracy with 92.9% accuracy, (<b>b</b>) training and validation loss, and (<b>c</b>) confusion matrix.</p> "> Figure 13
<p>Training and validation of the proposed CNN, performed using Adam. (<b>a</b>) Training and validation accuracy with 92.3% accuracy, (<b>b</b>) training and validation loss, and (<b>c</b>) confusion matrix.</p> "> Figure 14
<p>QR code image correction and reading. (<b>a</b>) Original image, (<b>b</b>) Sobel operator, (<b>c</b>) QR code framed by the minimum circumscribed rectangle, (<b>d</b>) perspective transformation, (<b>e</b>) mask matrix, (<b>f</b>) Gaussian blur (sharpening), (<b>g</b>) histogram equalization, (<b>h</b>) image overlay, (<b>i</b>) write the read information to CSV, (<b>j</b>) predict that the package is not 90° with a rotation and list a warning, and (<b>k</b>) success message.</p> "> Figure 14 Cont.
<p>QR code image correction and reading. (<b>a</b>) Original image, (<b>b</b>) Sobel operator, (<b>c</b>) QR code framed by the minimum circumscribed rectangle, (<b>d</b>) perspective transformation, (<b>e</b>) mask matrix, (<b>f</b>) Gaussian blur (sharpening), (<b>g</b>) histogram equalization, (<b>h</b>) image overlay, (<b>i</b>) write the read information to CSV, (<b>j</b>) predict that the package is not 90° with a rotation and list a warning, and (<b>k</b>) success message.</p> ">
Abstract
:1. Introduction
- A UAV is designed, which includes a positive cross quadcopter drone and a variety of sensors and components, such as flight controllers, single-board computers, optical flow sensors, ultrasonic sensors and cameras, etc.
- The UAV is successfully stabilized by PID control even when disturbances occurr.
- The placement angle of the package can be accurately classified by CNN. Optimization functions, such as SGD, RMSprop, Adadelta, and Adam, are applied to improve the system performance and show the recognition rates of 94%, 92%, 95%, and 93%, respectively. If the angle is not 90°, a warning will be issued to prompt the management personnel to handle it to avoid accidents or losses. In addition, image processing is required to assist in reading the QR code, including the use of Sobel edge computing, minimum circumscribed rectangle, perspective transformation, and image enhancement. Successful QR code reading is also provided.
2. Unmanned Aerial Vehicle System
3. Convolutional Neural Network
4. Experimental Results
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Lin, H.-T. Implementation of Industry 4.0 Production Line for Intelligent Manufacture-A Case Study on Production Test Line of Notebooks. Master’s Thesis, Tunghai University, Taichung City, Taiwan, 2015. [Google Scholar]
- Diversified Development of UAV Applications. Available online: https://www.moea.gov.tw/MNS/ietc/bulletin/Bulletin.aspx?kind=51&html=1&menu_id=33594&bull_id=9494 (accessed on 29 March 2023). (In Chinese)
- Tsai, Y.-S. Intelligent Warehouse with Mobile Vehicle for Maker Education. Master’s Thesis, Cheng Shiu University, Kaohsiung, Taiwan, 2021. [Google Scholar]
- Occupational Safety and Health Administration (OSHA). Warehouse Safety: A Practical Guide to Preventing Warehouse Incidents and Injuries; Occupational Safety and Health Administration (OSHA): Washington, DC, USA, 2019.
- National Institute for Occupational Safety and Health (NIOSH). Preventing Injuries and Deaths of Workers Who Operate or Work Near Forklifts; National Institute for Occupational Safety and Health (NIOSH): Washington, DC, USA, 2013.
- Fernández-Caramés, T.M. Towards an Autonomous Industry 4.0 Warehouse: A UAV and Blockchain-Based System for Inventory and Traceability Applications in Big Data-Driven Supply Chain Management. Sensors 2019, 19, 2394. [Google Scholar] [CrossRef] [PubMed]
- Ong, J.H.; Sanchez, A.; Williams, J. Multi-UAV System for Inventory Automation. In Proceedings of the 2007 1st Annual RFID Eurasia, Istanbul, Turkey, 5–6 September 2007; pp. 1–6. [Google Scholar]
- Chin, J.-Y. Research and Design of UAV Implement Automated RFID Inventory System. Master’s Thesis, National Chin-Yi University of Technology, Taichung City, Taiwan, 2017. [Google Scholar]
- Huang, C.-Y. Designing of a RFID-enabled Drone system for Autonomous Stock Counting Processes. Master’s Thesis, National Chin-Yi University of Technology, Taichung City, Taiwan, 2017. [Google Scholar]
- Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
- Hough, P.V.C. Method and Means for Recognizing Complex Patterns. U.S. Patent No 3,069,654, 18 December 1962. [Google Scholar]
- Harris, C.; Stephens, M. A combined corner and edge detector. In Proceedings of the Alvey Vision Conference, Manchester, UK, 31 August–2 September 1988; pp. 10–5244. [Google Scholar]
- Li, S. Rotation invariance neural network. arXiv 2017, arXiv:1706.05534. [Google Scholar]
- Li, J.; Yang, Z.; Liu, H.; Cai, D. Deep rotation equivariant network. Neurocomputing 2018, 290, 26–33. [Google Scholar] [CrossRef]
- Marcos, D.; Volpi, M.; Komodakis, N.; Tuia, D. Rotation equivariant vector field networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5048–5057. [Google Scholar]
- Kanezaki, A.; Matsushita, Y.; Nishida, Y. Rotationnet: Joint object categorization and pose estimation using multiviews from unsupervised viewpoints. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5010–5019. [Google Scholar]
- Shi, B.; Bai, S.; Zhou, Z.; Bai, X. Deeppano: Deep panoramic representation for 3-d shape recognition. IEEE Signal. Process. Lett. 2015, 22, 2339–2343. [Google Scholar] [CrossRef]
- Li, M.; Zhang, J.; Song, J.; Li, Z.; Lu, S. A Clinical-Oriented Non-Severe Depression Diagnosis Method Based on Cognitive Behavior of Emotional Conflict. IEEE Trans. Comput. Soc. Syst. 2023, 10, 131–141. [Google Scholar] [CrossRef]
- Zhao, H.; Zhang, P.; Zhang, R.; Yao, R.; Wu, D. A novel performance trend prediction approach using ENBLS with GWO. Meas. Sci. Technol. 2022, 34, 025018. [Google Scholar] [CrossRef]
- Zhao, H.; Yang, X.; Chen, B.; Chen, H.; Wu, D. Bearing fault diagnosis using transfer learning and optimized deep belief network. Meas. Sci. Technol. 2022, 33, 1239–1253. [Google Scholar] [CrossRef]
- Pei, Y.; Huang, Y.; Zou, Q.; Zhang, X.; Wang, S. Effects of image degradation and degradation removal to CNN-based image classification. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1239–1253. [Google Scholar] [CrossRef] [PubMed]
- Suganthi, M.; Sathiaseelan, J.G.R. An exploratory of hybrid techniques on deep learning for image classification. In Proceedings of the 2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP), Chennai, India, 28–29 September 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–4. [Google Scholar]
- Hasan, M.S. An application of pre-trained CNN for image classification. In Proceedings of the 2017 20th International Conference of Computer and Information Technology (ICCIT), Dhaka, Bangladesh, 22–24 December 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6. [Google Scholar]
- Jalal, A.; Kamal, S.; Kim, D. Shape and motion features approach for activity tracking and recognition from kinect video camera. In Proceedings of the 2015 IEEE 29th International Conference on Advanced Information Networking and Applications Workshops, Gwangju, Republic of Korea, 24–27 March 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 445–450. [Google Scholar]
- Chen, D.; Song, Z.; Guo, J.; Yang, S.; Yao, H. CNN-based image small-angle rotation angle estimation. In Proceedings of the 2022 IEEE 24th International Workshop on Multimedia Signal Processing (MMSP), Shanghai, China, 26–28 September 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–5. [Google Scholar]
- Sultana, R.; Adams, R.D.; Yan, Y.; Yanik, P.M.; Tanaka, M.L. Trash and recycled material identification using convolutional neural networks (CNN). In Proceedings of the 2020 SoutheastCon, Raleigh, NC, USA, 28–29 March 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–8. [Google Scholar]
- Lai, C.; Liu, T.; Mei, R.; Wang, H.; Hu, S. The cloud images classification based on convolutional neural network. In Proceedings of the 2019 International Conference on Meteorology Observations (ICMO), Chengdu, China, 28–31 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–4. [Google Scholar]
- Wang, D. A comparative research on clothing images classification based on neural network models. In Proceedings of the 2020 IEEE 2nd International Conference on Civil Aviation Safety and Information Technology (ICCASIT), Weihai, China, 14–16 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 495–499. [Google Scholar]
- Ibrahim, M.S.; Hamid, S.A.; Muhammad, Z.; Leh, N.A.M.; Abdullah, S.; Bakar, S.J.A.; Osman, M.K.; Fadhlullah, S.Y. CNN Comparative Study for Apple Quality Classification. In Proceedings of the 2022 IEEE 12th International Conference on Control System, Computing and Engineering (ICCSCE), Penang, Malaysia, 21–22 October 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 53–58. [Google Scholar]
- Viratkapan, V.; Mruetusatorn, S. Classification of House Categories Using Convolutional Neural Networks (CNN). In Proceedings of the 2022 7th International Conference on Business and Industrial Research (ICBIR), Bangkok, Thailand, 19–20 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 509–514. [Google Scholar]
- Li, Q.; Hu, S.; Shimasaki, K.; Ishii, I. An Active Multi-Object Ultrafast Tracking System with CNN-Based Hybrid Object Detection. Sensors 2023, 23, 4150. [Google Scholar] [CrossRef] [PubMed]
- Lin, J.-B. Design of the Fuzzy Autonomous Flight Control for Hexacopter. Master’s Thesis, Department of Electrical Engineering, National Ilan University, Yilan City, Taiwan, 2021. [Google Scholar]
- Lin, T.-L. Unmanned Aerial Vehicle Position and Sensor Fusion. Master’s Thesis, Department of Mechanical Engineering, National Yang Ming Chiao Tung University, Hsinchu City, Taiwan, 2018. [Google Scholar]
- Yue, J.; Zhang, Q. Research Progress and Key Technologies of Micro Quad-Rotor UAVs. Electron. Opt. Amp. Control. 2012, 17, 46–52. [Google Scholar]
- Wan, J.; Ai, J.-L. Design and Simulation of Fuzzy Control System of UAV Formation Flight. J. Syst. Simul. 2009, 21, 4183–4189. [Google Scholar]
- Chen, P.-J.; Yang, S.-Y.; Wang, C.-S.; Muslikhin, M.; Wang, M.-S. Development of a Chinese chess robotic system for the elderly using convolutional neural networks. Sustainability 2020, 12, 3980. [Google Scholar] [CrossRef]
- Gui, W.; Jun, T. Chinese chess recognition algorithm based on computer vision. In Proceedings of the 26th Chinese Control and Decision Conference, Changsha, China, 31 May–2 June 2014. [Google Scholar]
- Rosenblatt, F. The Perceptron: A Probabilistic Model for Information Storage and Organization in the Brain. Psychol. Rev. 1958, 65, 386–408. [Google Scholar] [CrossRef] [PubMed]
- Easy to Understand the Principle of Regularization L1 and L2. Available online: https://zhuanlan.zhihu.com/p/309373613 (accessed on 29 March 2023). (In Chinese).
10-Fold | Optimization Function | |||
---|---|---|---|---|
SGD | RMSprop | Adadelta | Adam | |
1 | 89.4% | 91.4% | 91.9% | 90.9% |
2 | 91.0% | 93.3% | 91.5% | 90.9% |
3 | 90.5% | 93.0% | 91.9% | 91.8% |
4 | 90.5% | 91.5% | 91.5% | 90.8% |
5 | 90.4% | 92.7% | 93.1% | 91.9% |
6 | 91.0% | 89.7% | 92.1% | 91.3% |
7 | 90.5% | 91.0% | 91.7% | 90.3% |
8 | 89.4% | 90.9% | 90.5% | 91.5% |
9 | 90.8% | 92.4% | 91.8% | 91.4% |
10 | 90.9% | 92.6% | 91.8% | 90.6% |
No. | Angle of Package (SGD) | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
90° | L15° | L30° | L45° | L60° | L75° | R15° | R30° | R45° | R60° | R75° | |
1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | R45° | 0 | 0 | 0 |
2 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
3 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
4 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
5 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
6 | 0 | L30° | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
7 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
8 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
9 | 0 | 0 | 0 | L60° | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
10 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
11 | 0 | 0 | 0 | L30° | 0 | 0 | R30° | 0 | 0 | 0 | 0 |
12 | L15° | 0 | 0 | 0 | L45° | 0 | 0 | 0 | 0 | 0 | 0 |
13 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | R60° | 0 | 0 |
14 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
15 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | L75° |
16 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
17 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
18 | 0 | R75° | 0 | L30° | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
19 | 0 | 0 | 0 | 0 | R60° | 0 | 0 | 0 | 0 | 0 | 0 |
20 | 0 | 0 | 0 | 0 | L45° | 0 | 0 | 0 | 0 | 0 | 0 |
Angles | Optimization Function | |||
---|---|---|---|---|
SGD | RMSprop | Adadelta | Adam | |
90° | 95% | 95% | 90% | 95% |
L15° | 90% | 85% | 95% | 90% |
L30° | 100% | 95% | 100% | 95% |
L45° | 85% | 95% | 95% | 90% |
L60° | 85% | 85% | 90% | 80% |
L75° | 100% | 95% | 100% | 100% |
R15° | 95% | 90% | 95% | 95% |
R30° | 95% | 95% | 100% | 95% |
R45° | 95% | 90% | 95% | 95% |
R60° | 100% | 100% | 95% | 95% |
R75° | 95% | 90% | 95% | 90% |
Average | 94% | 92% | 95% | 93% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yang, S.-Y.; Jan, H.-C.; Chen, C.-Y.; Wang, M.-S. CNN-Based QR Code Reading of Package for Unmanned Aerial Vehicle. Sensors 2023, 23, 4707. https://doi.org/10.3390/s23104707
Yang S-Y, Jan H-C, Chen C-Y, Wang M-S. CNN-Based QR Code Reading of Package for Unmanned Aerial Vehicle. Sensors. 2023; 23(10):4707. https://doi.org/10.3390/s23104707
Chicago/Turabian StyleYang, Szu-Yueh, Hsin-Che Jan, Chun-Yu Chen, and Ming-Shyan Wang. 2023. "CNN-Based QR Code Reading of Package for Unmanned Aerial Vehicle" Sensors 23, no. 10: 4707. https://doi.org/10.3390/s23104707
APA StyleYang, S. -Y., Jan, H. -C., Chen, C. -Y., & Wang, M. -S. (2023). CNN-Based QR Code Reading of Package for Unmanned Aerial Vehicle. Sensors, 23(10), 4707. https://doi.org/10.3390/s23104707