Abstract
Deep learning models such as deep convolutional neural networks (DCNNs) image classifiers have achieved outstanding performance over the last decade. However, these models are mostly trained with high-quality images drawn from publicly available datasets such as ImageNet. Recently, many researchers have evaluated the impact of low-quality image degradations on the performance of different neural network-based image classifiers. But, most of these studies generate low-quality images by synthetic modification of the high-quality images. Besides, most of the studies employed various image processing techniques to remove the image degradations and trained the DCNNs again to achieve better performance. But it has since been discovered that such methods could not improve the classification accuracy of DCNNs. The robustness of DCNNs based image classifiers trained on low-quality images resulting from natural factors common in autonomous driving and other intelligent system settings was rarely studied over the recent years. In this paper, we proposed a canonical correlation analysis (CCA) based fusion of camera and radar features for improving the performance of DCNNs image classifiers trained on natural adverse weather data. CCA is a statistical approach that creates a highly discriminative feature vector by measuring the linear relationship between the camera and radar features. A spatial attention network was designed to re-weight the camera features before associating them with radar features in the CCA-feature fusion block. Our findings based on experimental evaluations have proven that, indeed, the performance of the DCNN models (i.e., Alex-Net and VGG-16-Net) is heavily affected by degradations arising from natural factors. Specifically, the DCNN models are more affected by the degradations arising from rainfall, foggy and nighttime conditions using Radiate and Carrada datasets. However, the proposed fusion frameworks have improved the performance of the individual sensing modalities significantly. The radar data has helped substantially in enhancing the fusion performance, mainly using rainfall data where the camera data is heavily affected.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. In: Advances of neural information processing systems, pp 1097–1105
Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: Proceedings of international conference on learning representations, pp 1–14
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of IEEE conference for computer vision and pattern recognition, pp 770–778
Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of IEEE conference for computer vision and pattern recognition, pp 580–587
Ren S, He K, Girshick R, Sun J (2017) Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell 39(6):1137–1149
Chen L-C, Papandreou G, Kokkinos I, Murphy K, Yuille AL (2018) DeepLab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans Pattern Anal Mach Intell 404:834–848
Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of IEEE conference for computer vision and pattern recognition, pp 3431–2440
Abdu FJ, Zhang Y, Fu M, Li Y, Deng Z (2021) Application of deep learning on millimeter-wave radar signals: a review. Sensors 21:1951
Li W, Wen S, Shi K, Yang Y, Huang T (2022) Neural architecture search with a lightweight transformer for text-to-image synthesis. IEEE Trans Netw Sci Eng (TNSE) 9(3):1567–1576
Li S et al (2021) Auto-FERNet: a facial expression recognition network with architecture search. IEEE Trans Netw Sci Eng 8(3):2213–2222
Cao Y, Cao Y, Wen S, Huang T, Zeng Z (2019) Passivity analysis of delayed reaction–diffusion memristor-based neural networks. Neural Netw 109:159–167
Wen S, Xiao S, Yang Y, Yan Z, Zeng Z, Huang T (2019) Adjusting learning rate of memristor-based multilayer neural networks via fuzzy method. Trans Computer Aided Design Integrated Circuit Syst 38(6):1084–1094
Cao Y, Jiang W, Wang J (2021) Anti-synchronization of delayed memristive neural networks with leakage term and reaction–diffusion terms. Knowl Based Syst 233:107539
Schmidhuber J (2014) Deep learning in neural networks: an overview. Neural Netw 61:85–117
Geiger A, Lenz P, Urtasun R (2012) Are we ready for autonomous driving? the KITTI vision benchmark suite. In: Proceedings of IEEE conference for computer vision and pattern recognition, pp 3354–3361
Cordts M, Omran M, Ramos S, Rehfeld T, Enzweiler M, Benenson R, Franke U, Roth S, Schiele B (2016) The cityscapes dataset for semantic urban scene understanding. In Proceedings of IEEE conference for computer vision and pattern recognition, pp 3213–3223
Griffin G, Holub A, Perona P ( 2007) Caltech-256 object category dataset
Everingham M, Van Gool L, Williams CK, Winn J, Zisserman A (2010) The pascal visual object classes (VOC) challenge. Int J Comput Vis 88:303–338
Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L (2009) Imagenet: a large-scale hierarchical image database. In: IEEE conference on computer vision and pattern recognition, pp 248–255
Zhang J, Cao Y, Fang S, Kang Y, Chen CW (2017) Fast haze removal for nighttime image using maximum reflectance prior. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7418–7426
Dodge S, Karam L (2017) A Study and comparison of human and deep learning recognition performance under visual distortions. In: IEEE international conference on computer communication and networks (ICCCN), pp 1–7
Azulay A, Weiss Y (2018) Why do deep convolutional networks generalize so poorly to small image transformations? arXiv preprint arXiv:1805.12177
Pei Y, Huang Y, Zou Q et al (2019) Effects of image degradation and degradation removal to CNN-based image classification. IEEE Trans Pattern Anal Mach Intell 43(4):1239–1253
Koziarski M, Cyganek B (2018) Impact of low resolution on image recognition with deep neural networks: an experimental study. Int J Appl Math Comput Sci 28(4):735–744
Pei Y, Huang Y, Zou Q et al (2018) Does haze removal help CNN-based image classification? In Proceedings of the European conference on computer vision (ECCV), pp 697–712
Dodge S, Karam L (2016) Understanding how image quality affects deep neural networks. In: International conference on quality of multimedia experience (QoMEX), pp 1–6
Roy P et al (2018) Effects of degradations on deep neural network architectures. ArXiv, pp 1807–10108
Karahan S, Kilinc Yildirum M, Kirtac K, Rende F S, Butun G, Ekenel H K (2016) How image degradations affect deep CNN-based face recognition? In: International conference of the biometrics special interest group (BIOSIG), pp 1–5
Liu D, Cheng B, Wang Z, Zhang H, Huang TS (2019) Enhance visual recognition under adverse conditions via deep networks. IEEE Trans Image Process 28(9):4401–4412
Vasiljevic I, Chakrabarti A, Shakhnarovich G (2016) Examining the impact of blur on recognition by convolutional networks. arXiv:1611.05760
Wang Y, Cao Y, Zha ZJ, Zhang J, Xiong Z (2020) Deep degradation prior for low-quality image classification. In: Proceedings of IEEE conference for computer vision pattern recognition, pp 11046–11055
Dodge S, Karam L (2019) Human and DNN classification performance on images with quality distortions: a comparative study. ACM Trans Appl Percept (TAP) 16(7):1–17
Dodge S, Karam L (2017) Quality resilient deep neural networks. arXiv preprint arXiv:1703.08119
Hotelling H (1936) Relations between two sets of variates. Bio-metrika 28:321–377
Uurtio V, Monteiro JM, Kandola J, Shawe-Taylor J, Fernandez-Reyes D, Rousu J (2018) A tutorial on canonical correlation methods. ACM Comput Surv (CSUR) 50(6):95
Jing X, Li S, Lan C, Zhang D, Yang J, Liu Q (2011) Color image canonical correlation analysis for face feature extraction and recognition. Signal Process 91(8):2132–2140
Chetana K, Shubhangi D, Suresh G, Aditya A (2019) Novel canonical correlation analysis based feature level fusion algorithm for multimodal recognition. In biometric sensor systems. Sens Lett 17:75–86
Sun QS, Zeng SG, Liu Y, Heng PA, Xia DS (2005) A new method of feature fusion and its application in image recognition. Pattern Recogn 38(12):2437–2448
Ahmad Z, Khan N (2018) Towards improved human action recognition using convolutional neural networks and multimodal fusion of depth and inertial sensor data. In: IEEE international symposium on multimedia (ISM), pp 223–230
Marcel S, Emanuele P, Saptarshi M, Alireza A, Sen W, Andrew W (2020) RADIATE: A radar dataset for automotive perception. arXiv, preprint arXiv
Ouaknine A, Newson A, Rebut J, Tupin F, Pérez P (2021) CARRADA dataset: camera and automotive radar with range-angle-doppler annotations. In: Proceedings of IEEE conference for computer vision and pattern recognition, pp 5068–5075
Navtech. Navtech radar technical specifications. https://navtechradar.com/clearway-technical-specifications/. Accessed 03 June 2021
Funding
This work was supported in part by the Science and Technology Innovation Project of Xiongan New Area (No. 2022XAGG0181), the Science and Technology Key Project of Fujian Province (No. 2020HZ020005, 2021HZ021004 and 2021H61010115), the National Natural Science Foundation of China (No.U1705263), the President’s Fund of Xiamen University for Undergraduate (No. 20720212006), and the Open Project of State Key Laboratory of Matamaterial Electromagnetic Modulation Technology (No. XM-202204-0024).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Abdu, F.J., Zhang, Y. & Deng, Z. CCA-Based Fusion of Camera and Radar Features for Target Classification Under Adverse Weather Conditions. Neural Process Lett 55, 7293–7319 (2023). https://doi.org/10.1007/s11063-023-11261-w
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11063-023-11261-w