Nothing Special   »   [go: up one dir, main page]

Skip to main content
Log in

FARLut: a two-stage tobacco foreign body detection model incorporating color information and attention mechanism

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Accurate detection and rejection of debris in tobacco products play an essential part in ensuring the quality of tobacco products. In recent years, the detection of detritus in the production process has been widely investigated, but there is still room for further improvement in the research of visible light-based tobacco debris detection methods. In this study, we collected visible light images from the tobacco industry production line and constructed a dataset for tobacco debris detection. In addition, a we propose a tobacco debris detection model named the FARLut. The FARLut model preprocesses images obtained from the tobacco production process based on the color and then inputs the processed images into a two-stage target detection algorithm with an attention mechanism for debris detection. The proposed model is verified experimentally. The experimental results show that the FARLut model achieves an average accuracy of 94.91% and a recall rate of 97.20% on the test dataset. Thus, the proposed detection model can effectively identify familiar clutter in tobacco production. The results of this study provide a useful reference for further research on clutter detection in the tobacco production field.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Code Availability

The datasets and code generated during the current study are available from the corresponding author on reasonable request.

References

  1. Xiao L (2013) A brief discussion on the control of trash in the filament-making line of a tobacco factory. Red Son (Mid)

  2. Shen W (2018) Research on tobacco foreign object recognition algorithm based on the color model. Nanjing University of Aeronautics and Astronautics

  3. Tang X, Zhang Y, Huang G, Xie W, Guang H, Li J (2004) Comparison of the tobacco sorter and tobacco scan 6000 tobacco foreign matter rejection systems. Tob Sci Technol 2:1–3

    Google Scholar 

  4. Chen W, Zhong X, Li Z, Chen J, Liu J (2003) Key technologies in online foreign matter rejection system for tobacco. Meas Control Technol 22(5):29–32

    Google Scholar 

  5. Zhang S, Dong D, Ren Y, Dai P, Kong D (2009) Typical foreign matter handling methods in tobacco foreign matter rejection systems. Tob Technol 5:22–25

    Google Scholar 

  6. Liu D (2012) Research and application of key technologies for clutter recognition based on machine vision. Master’s thesis. Nanjing University of Aeronautics and Astronautics

  7. Tu P (2016) Research on algorithms related to tobacco foreign body rejection system. Master’s thesis, Southeast University

  8. Xie F, Zhu D-B (2022) A review of deep learning target detection methods. Comput Syst Appl 31(2):1–12

    Google Scholar 

  9. Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779–788

  10. Redmon J, Farhadi A (2017) Yolo9000: better, faster, stronger. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7263–7271

  11. Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arXiv:1804.02767

  12. Bochkovskiy A, Wang C, Liao H (2020) Yolov4: Optimal speed and accuracy of object detection. arXiv:2004.10934

  13. Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C-Y, Berg AC (2016) Ssd: Single shot multibox detector. In: European conference on computer vision, pp 21–37. Springer

  14. Jeong J, Park H, Kwak N (2017) Enhancement of ssd by concatenating feature maps for object detection. arXiv:1705.09587

  15. Fu C.-Y, Liu W, Ranga A, Tyagi A, Berg AC (2017) Dssd: Deconvolutional single shot detector. arXiv:1701.06659

  16. Li Z, Zhou F (2017) Fssd: Feature fusion single shot multibox detector. arxiv 2017. arXiv:1712.00960

  17. Shen Z, Liu Z, Li J, Jiang Y.-G, Chen Y, Xue X (2017) Dsod: Learning deeply supervised object detectors from scratch. In: Proceedings of the IEEE international conference on computer vision, pp 1919–1927

  18. Lin T, Goyal P, Girshick R, He K, Doll P (2017) Focal loss for dense object detection. In: Proceedings of the IEEE international conference on computer vision, pp 2980–2988

  19. Zhao Q, Sheng T, Wang Y, Tang Z, Chen Y, Cai L, Ling H (2019) M2det: A single-shot object detector based on multi-level feature pyramid network. Proceedings of the AAAI conference on artificial intelligence 33:9259–9266

    Article  Google Scholar 

  20. Girshick R, Donahue J, Darrell T, Malik J Rich Feature Hierarchies for accurate object detection and semantic segmentation

  21. Li Y, Yun L, Ye Z, Wang K, Zhuo N (2021) Research on image recognition method of moldy tobacco leaves based on the convolutional neural network. Comput Eng Sci 43(3):473–479

    Google Scholar 

  22. Li J (2016) Research on automatic identification of tobacco diseases based on the convolutional neural network. PhD thesis. Shandong Agricultural University

  23. He K, Zhang X, Ren S, Sun J (2015) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell 37(9):1904–1916

    Article  Google Scholar 

  24. Girshick R (2015) Fast r-cnn, pp 1440–1448

  25. Ren S, He K, Girshick R, Sun J (2015)Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in Neural Information Processing Systems 28

  26. Lin T-Y, Dollár P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2117–2125

  27. He K, Gkioxari G, Dollár P, Girshick R (2017) Mask r-cnn. In: Proceedings of the ieee international conference on computer vision, pp 2961–2969

  28. Kim J-a, Sung J-Y, Park S-h (2020) Comparison of faster-rcnn, yolo, and ssd for real-time vehicle type recognition. In: 2020 IEEE international conference on consumer electronics-asia (ICCE-Asia), pp 1–4. IEEE

  29. Yu X, Si Y, Li L (2019) Pedestrian detection based on improved faster rcnn algorithm. In: 2019 IEEE/CIC international conference on communications in china (ICCC), pp 346–351. IEEE

  30. Wei B, Hao K, Tang X-s, Ren L (2018) Fabric defect detection based on faster rcnn. In: International conference on artificial intelligence on textile and apparel, pp 45–51. Springer

  31. Xueshi C, Tong S, Weimin Q (2022) Improved faster rcnn-based defect detection algorithm for printed circuit boards. Journal of Jianghan University (Natural Science Edition)

  32. Le VNT, Truong G, Alameh K (2021) Detecting weeds from crops under complex field environments based on faster rcnn. In: 2020 IEEE eighth international conference on communications and electronics (ICCE), pp 350–355.IEEE

  33. Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the ieee conference on computer vision and pattern recognition, pp 7132–7141

  34. Wang Q, Wu B, Zhu P, Li P, Zuo W, Hu Q (2020) Supplementary material for ‘eca-net: Efficient channel attention for deep convolutional neural networks. In: Proceedings of the 2020 IEEE/CVF conference on computer vision and pattern recognition. IEEE, Seattle, WA, USA, pp 13–19

  35. Woo S, Park J, Lee J-Y, Kweon IS (2018) Cbam: Convolutional block attention module. In: Proceedings of the european conference on computer vision (ECCV), pp 3–19

  36. Yosinski J, Clune J, Bengio Y, Lipson H (2014) How transferable are features in deep neural networks. Advances in Neural Information Processing Systems 27

  37. Zhao T, Yi X, Zeng Z, Feng T (2021) Mobilenet-yolo based wildlife detection model: A case study in yunnan tongbiguan nature reserve, china. Journal of Intelligent & Fuzzy Systems (Preprint), pp 1–11

  38. Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database.In: 2009 IEEE conference on computer vision and pattern recognition, pp 248–255. Ieee

  39. Everingham M, Van Gool L, Williams CK, Winn J, Zisserman A (2010) The pascal visual object classes (voc) challenge. Int J Comput Vis 88(2):303–338

    Article  Google Scholar 

  40. Lin T-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, Dollár P, Zitnick CL (2014) Microsoft coco: Common objects in context. In: European conference on computer vision, pp 740–755. Springer

  41. Gong Y, Liu L, Yang M, Bourdev L (2014) Compressing deep convolutional networks using vector quantization. arXiv:1412.6115

  42. Han S, Mao H, Dally WJ (2015) Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv:1510.00149

  43. Hinton G, Vinyals O, Dean J, et al (2015) Distilling the knowledge in a neural network. 2(7). arXiv:1503.02531

Download references

Author information

Authors and Affiliations

Authors

Contributions

Jianglai Liang, Gang Huang, Tao Feng: Conception and design of the study. Gang Huang,Jianglai Liang:Acquistion of data and Labeling the training data. Jianglai Liang,Zhiyong Zeng:data curation and visualization.Jianglai Liang: Drafting the manuscript. Tao Feng ,Zhiyong Zeng:Revising the manuscript. Tao Feng, Zhiyong Zeng: Approval of the version of the manuscript to be published. All authors have read and agreed to the published version of the manuscript.

Corresponding author

Correspondence to Tao Feng.

Ethics declarations

Competing of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Liang, J., Zeng, Z., Huang, G. et al. FARLut: a two-stage tobacco foreign body detection model incorporating color information and attention mechanism. Multimed Tools Appl 83, 64271–64284 (2024). https://doi.org/10.1007/s11042-024-18190-3

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-024-18190-3

Keywords

Navigation