Abstract
Accurate detection and rejection of debris in tobacco products play an essential part in ensuring the quality of tobacco products. In recent years, the detection of detritus in the production process has been widely investigated, but there is still room for further improvement in the research of visible light-based tobacco debris detection methods. In this study, we collected visible light images from the tobacco industry production line and constructed a dataset for tobacco debris detection. In addition, a we propose a tobacco debris detection model named the FARLut. The FARLut model preprocesses images obtained from the tobacco production process based on the color and then inputs the processed images into a two-stage target detection algorithm with an attention mechanism for debris detection. The proposed model is verified experimentally. The experimental results show that the FARLut model achieves an average accuracy of 94.91% and a recall rate of 97.20% on the test dataset. Thus, the proposed detection model can effectively identify familiar clutter in tobacco production. The results of this study provide a useful reference for further research on clutter detection in the tobacco production field.
Similar content being viewed by others
Code Availability
The datasets and code generated during the current study are available from the corresponding author on reasonable request.
References
Xiao L (2013) A brief discussion on the control of trash in the filament-making line of a tobacco factory. Red Son (Mid)
Shen W (2018) Research on tobacco foreign object recognition algorithm based on the color model. Nanjing University of Aeronautics and Astronautics
Tang X, Zhang Y, Huang G, Xie W, Guang H, Li J (2004) Comparison of the tobacco sorter and tobacco scan 6000 tobacco foreign matter rejection systems. Tob Sci Technol 2:1–3
Chen W, Zhong X, Li Z, Chen J, Liu J (2003) Key technologies in online foreign matter rejection system for tobacco. Meas Control Technol 22(5):29–32
Zhang S, Dong D, Ren Y, Dai P, Kong D (2009) Typical foreign matter handling methods in tobacco foreign matter rejection systems. Tob Technol 5:22–25
Liu D (2012) Research and application of key technologies for clutter recognition based on machine vision. Master’s thesis. Nanjing University of Aeronautics and Astronautics
Tu P (2016) Research on algorithms related to tobacco foreign body rejection system. Master’s thesis, Southeast University
Xie F, Zhu D-B (2022) A review of deep learning target detection methods. Comput Syst Appl 31(2):1–12
Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779–788
Redmon J, Farhadi A (2017) Yolo9000: better, faster, stronger. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7263–7271
Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arXiv:1804.02767
Bochkovskiy A, Wang C, Liao H (2020) Yolov4: Optimal speed and accuracy of object detection. arXiv:2004.10934
Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C-Y, Berg AC (2016) Ssd: Single shot multibox detector. In: European conference on computer vision, pp 21–37. Springer
Jeong J, Park H, Kwak N (2017) Enhancement of ssd by concatenating feature maps for object detection. arXiv:1705.09587
Fu C.-Y, Liu W, Ranga A, Tyagi A, Berg AC (2017) Dssd: Deconvolutional single shot detector. arXiv:1701.06659
Li Z, Zhou F (2017) Fssd: Feature fusion single shot multibox detector. arxiv 2017. arXiv:1712.00960
Shen Z, Liu Z, Li J, Jiang Y.-G, Chen Y, Xue X (2017) Dsod: Learning deeply supervised object detectors from scratch. In: Proceedings of the IEEE international conference on computer vision, pp 1919–1927
Lin T, Goyal P, Girshick R, He K, Doll P (2017) Focal loss for dense object detection. In: Proceedings of the IEEE international conference on computer vision, pp 2980–2988
Zhao Q, Sheng T, Wang Y, Tang Z, Chen Y, Cai L, Ling H (2019) M2det: A single-shot object detector based on multi-level feature pyramid network. Proceedings of the AAAI conference on artificial intelligence 33:9259–9266
Girshick R, Donahue J, Darrell T, Malik J Rich Feature Hierarchies for accurate object detection and semantic segmentation
Li Y, Yun L, Ye Z, Wang K, Zhuo N (2021) Research on image recognition method of moldy tobacco leaves based on the convolutional neural network. Comput Eng Sci 43(3):473–479
Li J (2016) Research on automatic identification of tobacco diseases based on the convolutional neural network. PhD thesis. Shandong Agricultural University
He K, Zhang X, Ren S, Sun J (2015) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell 37(9):1904–1916
Girshick R (2015) Fast r-cnn, pp 1440–1448
Ren S, He K, Girshick R, Sun J (2015)Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in Neural Information Processing Systems 28
Lin T-Y, Dollár P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2117–2125
He K, Gkioxari G, Dollár P, Girshick R (2017) Mask r-cnn. In: Proceedings of the ieee international conference on computer vision, pp 2961–2969
Kim J-a, Sung J-Y, Park S-h (2020) Comparison of faster-rcnn, yolo, and ssd for real-time vehicle type recognition. In: 2020 IEEE international conference on consumer electronics-asia (ICCE-Asia), pp 1–4. IEEE
Yu X, Si Y, Li L (2019) Pedestrian detection based on improved faster rcnn algorithm. In: 2019 IEEE/CIC international conference on communications in china (ICCC), pp 346–351. IEEE
Wei B, Hao K, Tang X-s, Ren L (2018) Fabric defect detection based on faster rcnn. In: International conference on artificial intelligence on textile and apparel, pp 45–51. Springer
Xueshi C, Tong S, Weimin Q (2022) Improved faster rcnn-based defect detection algorithm for printed circuit boards. Journal of Jianghan University (Natural Science Edition)
Le VNT, Truong G, Alameh K (2021) Detecting weeds from crops under complex field environments based on faster rcnn. In: 2020 IEEE eighth international conference on communications and electronics (ICCE), pp 350–355.IEEE
Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the ieee conference on computer vision and pattern recognition, pp 7132–7141
Wang Q, Wu B, Zhu P, Li P, Zuo W, Hu Q (2020) Supplementary material for ‘eca-net: Efficient channel attention for deep convolutional neural networks. In: Proceedings of the 2020 IEEE/CVF conference on computer vision and pattern recognition. IEEE, Seattle, WA, USA, pp 13–19
Woo S, Park J, Lee J-Y, Kweon IS (2018) Cbam: Convolutional block attention module. In: Proceedings of the european conference on computer vision (ECCV), pp 3–19
Yosinski J, Clune J, Bengio Y, Lipson H (2014) How transferable are features in deep neural networks. Advances in Neural Information Processing Systems 27
Zhao T, Yi X, Zeng Z, Feng T (2021) Mobilenet-yolo based wildlife detection model: A case study in yunnan tongbiguan nature reserve, china. Journal of Intelligent & Fuzzy Systems (Preprint), pp 1–11
Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database.In: 2009 IEEE conference on computer vision and pattern recognition, pp 248–255. Ieee
Everingham M, Van Gool L, Williams CK, Winn J, Zisserman A (2010) The pascal visual object classes (voc) challenge. Int J Comput Vis 88(2):303–338
Lin T-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, Dollár P, Zitnick CL (2014) Microsoft coco: Common objects in context. In: European conference on computer vision, pp 740–755. Springer
Gong Y, Liu L, Yang M, Bourdev L (2014) Compressing deep convolutional networks using vector quantization. arXiv:1412.6115
Han S, Mao H, Dally WJ (2015) Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv:1510.00149
Hinton G, Vinyals O, Dean J, et al (2015) Distilling the knowledge in a neural network. 2(7). arXiv:1503.02531
Author information
Authors and Affiliations
Contributions
Jianglai Liang, Gang Huang, Tao Feng: Conception and design of the study. Gang Huang,Jianglai Liang:Acquistion of data and Labeling the training data. Jianglai Liang,Zhiyong Zeng:data curation and visualization.Jianglai Liang: Drafting the manuscript. Tao Feng ,Zhiyong Zeng:Revising the manuscript. Tao Feng, Zhiyong Zeng: Approval of the version of the manuscript to be published. All authors have read and agreed to the published version of the manuscript.
Corresponding author
Ethics declarations
Competing of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Liang, J., Zeng, Z., Huang, G. et al. FARLut: a two-stage tobacco foreign body detection model incorporating color information and attention mechanism. Multimed Tools Appl 83, 64271–64284 (2024). https://doi.org/10.1007/s11042-024-18190-3
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-024-18190-3