Abstract
Animal sourced protein is increasing rapidly due to the growing of population and incomes. The big data, robots and smart sensing technologies have brought the autonomous robotic system to the smart farming that enhance productivity and efficiency. Therefore, a YOLOv4-SAM was proposed to achieve high detection precision of cow body parts in long-term complex scenes. The proposed YOLOv4-SAM consists of two components: YOLOv4 is for multi-scale feature extraction, while the Spatial Attention Mechanisms (SAM) highlights the key cow biometric-related features. By doing this, visual biometric feature representation ability is enhanced for improving cow detection performance. To verify the performance of YOLOv4-SAM, a challenging dataset consisting of adult cows and calves with complex environments (e.g., day and night, occlusion, multiple target) was constructed for experimental testing. The precision, recall, mIoU and of the proposed YOLOv4-CBAM were 92.29%, 96.51%, 77.22% and 93.13%, respectively. The data shows that its overall performance was better than that of the comparison algorithm (Faster R-CNN, RetinaNet and YOLOv4). In addition, object detection based on the YOLOv4-SAM model can capture the key biometric-related features for cow visual representation and improve the performance of cow detection. In addition, the detected height difference between head and leg proved the capability in automatic identification of lame cows. The proposed deep learning-based cow detection approach provides a basis for developing an automated system for animal monitoring and management on commercial dairy farms.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Qiao, Y., et al.: Intelligent perception for cattle monitoring: a review for cattle identification, body condition score evaluation, and weight estimation. Comput. Electron. Agric. 185, 106143 (2021)
Fournel, S., Rousseau, A.N., Laberge, B.: Rethinking environment control strategy of confined animal housing systems through precision livestock farming. Biosys. Eng. 155, 96–123 (2017)
Mishra, S., Syed, D.F., Ploughe, M., Zhang, W.: Autonomous vision-guided object collection from water surfaces with a customized multirotor. IEEE/ASME Transactions on Mechatronics 26, 1914–1922 (2021)
Lim, J., Pyo, S., Kim, N., Lee, J., Lee, J.: Obstacle magnification for 2-D collision and occlusion avoidance of autonomous multirotor aerial vehicles. IEEE/ASME Trans. Mechatron. 25(5), 2428–2436 (2020)
Zhang, Z., Kayacan, E., Thompson, B., Chowdhary, G.: High precision control and deep learning-based corn stand counting algorithms for agricultural robot. Auton. Robot. 44(7), 1289–1302 (2020). https://doi.org/10.1007/s10514-020-09915-y
Ding, H., Gao, R.X., Isaksson, A.J., Landers, R.G., Parisini, T., Yuan, Y.: State of AI-based monitoring in smart manufacturing and introduction to focused section. IEEE/ASME Trans. Mechatron. 25(5), 2143–2154 (2020)
Qiao, Y., Truman, M., Sukkarieh, S.: Cattle segmentation and contour extraction based on mask R-CNN for precision livestock farming. Comput. Electron. Agric. 165, 104958 (2019)
Beggs, D., Jongman, E., Hemsworth, P., Fisher, A.: Lame cows on australian dairy farms: a comparison of farmer-identified lameness and formal lameness scoring, and the position of lame cows within the milking order. J. Dairy Sci. 102(2), 1522–1529 (2019)
Okinda, C., et al.: A review on computer vision systems in monitoring of poultry: a welfare perspective. Artif. Intell. Agric. 4, 184–208 (2020). https://doi.org/10.1016/j.aiia.2020.09.002
Li, G., et al.: Practices and applications of convolutional neural network- based computer vision systems in animal farming: a review. Sensors 21(4), 1492 (2021)
Riekert, M., Klein, A., Adrion, F., Hoffmann, C., Gallmann, E.: Automatically detecting pig position and posture by 2D camera imaging and deep learning. Comput. Electron. Agric. 174, 105391 (2020)
Shao, W., Kawakami, R., Yoshihashi, R., You, S., Kawase, H., Naemura, T.: Cattle detection and counting in UAV images based on convolutional neural networks. Int. J. Remote Sens. 41(1), 31–52 (2020)
Jiang, B., Wu, Q., Yin, X., Wu, D., Song, H., He, D.: FLYOLOv3 deep learning for key parts of dairy cow body detection. Comput. Electron. Agric. 166, 104982 (2019)
He, D., Liu, D., Zhao, K.: Review of perceiving animal information and behavior in precision livestock farming. Trans. Chin. Soc. Agric. Mach 47(5), 231–244 (2016)
Mahmud, M.S., Zahid, A., Das, A.K., Muzammil, M., Khan, M.U.: A systematic literature review on deep learning applications for precision cattle farming. Comput. Electron. Agric. 187, 106313 (2021)
Meena, S.D., Agilandeeswari, L.: Smart animal detection and counting framework for monitoring livestock in an autonomous unmanned ground vehicle using restricted supervised learning and image fusion. Neural Process. Lett. 53(2), 1253–1285 (2021)
Barbedo, J.G.A., Koenigkan, L.V., Santos, T.T., Santos, P.M.: A study on the detection of cattle in UAV images using deep learning. Sensors 19(24), 5436 (2019)
Shafiee, M.J., Chywl, B., Li, F., Wong, A.: Fast YOLO: a fast you only look once system for real-time embedded object detection in video. arXiv preprint arXiv:1709.05943 (2017)
Bochkovskiy, A., Wang, C.Y., Liao, H.Y.M: YOLOv4: optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934 (2020)
Fukui, H., Hirakawa, T., Yamashita, T., Fujiyoshi, H.: Attention branch network: learning of attention mechanism for visual explanation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10 705–10 714 (2019)
Li, X., Jia, X., Wang, Y., Yang, S., Zhao, H., Lee, J.: Industrial remaining useful life prediction by partial observation using deep learning with supervised attention. IEEE/ASME Trans. Mechatron. 25(5), 2241–2251 (2020)
Liu, S., Qi, L., Qin, H., Shi, J., Jia, J.: Path aggregation network for instance segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8759–8768 (2018)
He, K., Zhang, X., Ren, S., Sun, J.: Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 37(9), 1904–1916 (2015)
Hu, J., Zhi, X., Shi, T., Zhang, W., Cui, Y., Zhao, S.: Pag-YOLO: a portable attention-guided yolo network for small ship detection. Remote Sens. 13(16), 3059 (2021)
Redmon, J., Farhadi, A.: YOLOv3: an incremental improvement. arXiv preprint arXiv:1804.02767 (2018)
Wang, C., Luo, Q., Chen, X., Yi, B., Wang, H.: Citrus recognition based on YOLOv4 neural network. In: Journal of Physics: Conference Series, vol. 1820, no. 1, p. 012163. IOP Publishing (2021)
Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Advances in Neural Information Processing Systems, vol. 28, pp. 91–99 (2015)
Lin, T.Y., Goyal, P., Girshick, R., He, K., Dollár, P.: Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2980–2988 (2017)
Kaixuan, Z., Dongjian, H.: Target detection method for moving cows based on background subtraction. Int. J. Agric. Biol. Eng. 8(1), 42–49 (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Qiao, Y., Guo, Y., He, D. (2023). Deep Learning-Based Autonomous Cow Detection for Smart Livestock Farming. In: Yu, C., Zhou, J., Song, X., Lu, Z. (eds) Green, Pervasive, and Cloud Computing. GPC 2022. Lecture Notes in Computer Science, vol 13744. Springer, Cham. https://doi.org/10.1007/978-3-031-26118-3_19
Download citation
DOI: https://doi.org/10.1007/978-3-031-26118-3_19
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-26117-6
Online ISBN: 978-3-031-26118-3
eBook Packages: Computer ScienceComputer Science (R0)