Breast Cancer Diagnosis in Thermography Using Pre-Trained VGG16 with Deep Attention Mechanisms
<p>Model architecture.</p> "> Figure 2
<p>ROC curves of models for breast cancer. The ROC curve of the VGG16 model (<b>A</b>); the ROC curve of VGG16-SL model (<b>B</b>); the ROC curve of VGG16-HD model (<b>C</b>); the ROC curve of VGG16-SF model (<b>D</b>).</p> "> Figure 2 Cont.
<p>ROC curves of models for breast cancer. The ROC curve of the VGG16 model (<b>A</b>); the ROC curve of VGG16-SL model (<b>B</b>); the ROC curve of VGG16-HD model (<b>C</b>); the ROC curve of VGG16-SF model (<b>D</b>).</p> "> Figure 3
<p>VGG16 model with/without AMs experiment results. VGG16 model (<b>A</b>); VGG16-SL model (<b>B</b>); VGG16-HD model (<b>C</b>); VGG16-SF model (<b>D</b>).</p> "> Figure 3 Cont.
<p>VGG16 model with/without AMs experiment results. VGG16 model (<b>A</b>); VGG16-SL model (<b>B</b>); VGG16-HD model (<b>C</b>); VGG16-SF model (<b>D</b>).</p> "> Figure 4
<p>Aggregate confusion matrix of all folds for each model. VGG16 model (<b>A</b>); VGG16-SL model (<b>B</b>); VGG16-HD model (<b>C</b>); VGG16-SF model (<b>D</b>).</p> "> Figure 5
<p>Charts for comparison of CNN and VGG16 results with/without AMs. VGG16 and CNN model (<b>A</b>); VGG16-SL and CNN-SL model (<b>B</b>); VGG16-HD and CNN-HD model (<b>C</b>); VGG16-SF and CNN-SF model (<b>D</b>).</p> ">
Abstract
:1. Introduction
- This research study presents accurate models of breast cancer detection based on the pre-trained model VGG16 and AMs using thermographic images.
- This research study evaluates the performance of VGG16 with and without AMs to determine the effect of AMs on the performance of VGG16.
- This research study compares the proposed models with related studies.
2. Deep Learning and Convolutional Neural Network (CNN)
3. Deep Attention Mechanisms
- Soft Attention: The categorical distribution is computed through a set of elements, then the weights are generated by the resulting probabilities. The probabilities that result represent the importance of each element and are utilized as weights to generate a context-aware encoding that is the weighted sum of all elements. Due to the interdependence between the mechanism and the purpose of the deep neural network, it determines how much focus should be given to each input element by giving each element a weight between zero and one [8].
- Hard Attention: A subset of elements are selected from the input sequence. The weight allotted to an input element in hard AMs can be zero or one, forcing the approach to concentrate only on the critical elements and ignore the rest. Due to the input elements being either observed or not, the goal is non-differentiable [9].
- Self-Attention: The interdependence of the mechanism’s input elements is estimated because it permits the input to interact with the other “self” and identify what it should focus on more. One of the essential benefits of the self-attention layer against hard and soft mechanisms is its parallel computing capability for a lengthy input. To verify that all the same input elements are being paid attention to, this mechanism layer performs straightforward and efficiently parallelizable matrix calculations [9].
4. Related Work
5. Materials and Methods
5.1. Dataset
5.2. Data Pre-Processing
5.3. Feature Extraction
5.4. Bidirectional Long Short-Term Memory Layer
5.5. Attention Mechanisms Layer
- SL layer: Using an attention approach that considers the context of each timestamp when processing sequential data. It is implemented in this study by the package Keras-self-attention with multiplicative type, using the following formulae [26]:
- SF layer: It uses a low weight to multiply the associated feature map to discard unimportant regions. As a result, an area with high attention maintains its original value, whereas an area with low attention drifts closer to 0 (and becomes dark in the visualization). To calculate a weight for covering each sub-section of an image, we utilize the hidden state from the previous time step. To determine how much attention is being paid, we create a score using the formula [27]:
- HD layer: The weight applied to an input portion is either 0 or 1; this causes the model to concentrate only on the critical elements while disregarding others. The outcome is that the input parts are either observed or not, making the goal non-differentiable. Instead of utilizing a weighted average as in SF [27], HD is computed using as a sample rate to select one as the input to the next layer.
5.6. Fully Connected Layer
5.7. Sigmoid Layer
5.8. Classification Layer
5.9. Model Evaluation
6. Results
6.1. Experimental Settings
6.2. Classification Results
6.3. Comparison with Recent Methods
6.3.1. Comparison with Other Pre-Trained VGG16 Model
6.3.2. Comparison with Convolutional Neural Network (CNN) Models
Approaches | Accuracy | Specificity | Sensitivity (Recall) | Precision | F1-Score | AUC | Cohen’s Kappa |
---|---|---|---|---|---|---|---|
Recent study [22]: Deep multi-view VGG16 on thermal images | |||||||
VGG16- Multi-View | 99% | 100% | 98.04% | 100% | 99.01% | - | - |
VGG16- Frontal View | 94% | 92.31% | 95.83% | 92% | 93.88% | - | - |
Our previous work [24]: CNN with and without AM on thermal images | |||||||
CNN | 84.92% | 89.61% | 89.61% | 90.23% | 83.91% | 0.851 | 0.69 |
CNN-SL | 99.32% | 99.52% | 99.52% | 99.14% | 99.32% | 0.999 | 0.98 |
CNN-HD | 99.49% | 99.71% | 99.71% | 99.28% | 99.49% | 0.999 | 0.98 |
CNN-SF | 99.34% | 99.21% | 99.21% | 99.52% | 99.36% | 0.999 | 0.98 |
Proposed approach: VGG16 pre-trained DL method with and without AM on thermal images | |||||||
VGG16 | 99.18% | 98.48% | 99.85% | 98.55% | 99.20% | 0.999 | 0.983 |
VGG16-SL | 99.49% | 99.56% | 99.42% | 99.57% | 99.49% | 0.994 | 0.989 |
VGG16-HD | 99.80% | 99.75% | 99.85% | 99.76% | 99.80% | 0.999 | 0. 996 |
VGG16-SF | 99.32% | 99.46% | 99.19% | 99.48% | 99.33% | 0.999 | 0.986 |
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Rajinikanth, V.; Kadry, S.; Taniar, D.; Damaševičius, R.; Rauf, H.T. Breast-Cancer Detection Using Thermal Images with Marine-Predators-Algorithm Selected Features. In Proceedings of the 2021 Seventh International conference on Bio Signals, Images, and Instrumentation (ICBSII), Chennai, India, 25–27 March 2021; pp. 1–6. [Google Scholar]
- Singh, D.; Singh, A.K. Role of Image Thermography in Early Breast Cancer Detection-Past, Present and Future. Comput. Methods Programs Biomed. 2020, 183, 105074. [Google Scholar] [CrossRef] [PubMed]
- Parisky, Y.R.; Skinner, K.A.; Cothren, R.; DeWittey, R.L.; Birbeck, J.S.; Conti, P.S.; Rich, J.K.; Dougherty, W.R. Computerized Thermal Breast Imaging Revisited: An Adjunctive Tool to Mammography. In Proceedings of the 20th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Biomedical Engineering towards the Year 2000 and Beyond, Hong Kong, China, 1 November 1998; Volume 2, pp. 919–921. [Google Scholar]
- Wu, P.; Qu, H.; Yi, J.; Huang, Q.; Chen, C.; Metaxas, D. Deep Attentive Feature Learning for Histopathology Image Classification. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 1865–1868. [Google Scholar]
- Palminteri, S.; Pessiglione, M. Chapter Five—Reinforcement Learning and Tourette Syndrome. In International Review of Neurobiology; Martino, D., Cavanna, A.E., Eds.; Advances in the Neurochemistry and Neuropharmacology of Tourette Syndrome; Academic Press: Cambridge, MA, USA, 2013; Volume 112, pp. 131–153. [Google Scholar]
- Rezende, E.; Ruppert, G.; Carvalho, T.; Theophilo, A.; Ramos, F.; Geus, P.D. Malicious Software Classification Using VGG16 Deep Neural Network’s Bottleneck Features. In Information Technology—New Generations; Latifi, S., Ed.; Advances in Intelligent Systems and Computing; Springer International Publishing: Cham, Germany, 2018; Volume 738, pp. 51–59. ISBN 978-3-319-77027-7. [Google Scholar]
- Tian, H.; Wang, P.; Tansey, K.; Han, D.; Zhang, J.; Zhang, S.; Li, H. A Deep Learning Framework under Attention Mechanism for Wheat Yield Estimation Using Remotely Sensed Indices in the Guanzhong Plain, PR China. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102375. [Google Scholar] [CrossRef]
- Shen, T.; Zhou, T.; Long, G.; Jiang, J.; Wang, S.; Zhang, C. Reinforced Self-Attention Network: A Hybrid of Hard and Soft Attention for Sequence Modeling. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, International Joint Conferences on Artificial Intelligence Organization, Stockholm, Sweden, 5 July 2018; pp. 4345–4352. [Google Scholar]
- de Santana Correia, A.; Colombini, E.L. Attention, Please! A Survey of Neural Attention Models in Deep Learning. Artif. Intell. Rev. 2022, 55, 6037–6124. [Google Scholar] [CrossRef]
- Rashed, E.; El Seoud, M.S.A. Deep Learning Approach for Breast Cancer Diagnosis. In Proceedings of the 2019 8th International Conference on Software and Information Engineering, ACM, Cairo, Egypt, 9 April 2019; pp. 243–247. [Google Scholar]
- Platania, R.; Shams, S.; Yang, S.; Zhang, J.; Lee, K.; Park, S.-J. Automated Breast Cancer Diagnosis Using Deep Learning and Region of Interest Detection (BC-DROID). In Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics, ACM, Boston, MA, USA, 20 August 2017; pp. 536–543. [Google Scholar]
- Patil, R.S.; Biradar, N. Automated Mammogram Breast Cancer Detection Using the Optimized Combination of Convolutional and Recurrent Neural Network. Evol. Intel. 2021, 14, 1459–1474. [Google Scholar] [CrossRef]
- Santana, M.A.; Pereira, J.M.S.; Silva, F.L.; Lima, N.M.; Sousa, F.N.; Arruda, G.M.S.; de Lima, R.C.F.; Silva, W.W.A.; Santos, W.P. Breast Cancer Diagnosis Based on Mammary Thermography and Extreme Learning Machines. Res. Biomed. Eng. 2018, 34, 45–53. [Google Scholar] [CrossRef] [Green Version]
- Gogoi, U.R.; Bhowmik, M.K.; Ghosh, A.K.; Bhattacharjee, D.; Majumdar, G. Discriminative Feature Selection for Breast Abnormality Detection and Accurate Classification of Thermograms. In Proceedings of the 2017 International Conference on Innovations in Electronics, Signal Processing and Communication (IESC), Shillong, India, 6–7 April 2017; pp. 39–44. [Google Scholar]
- Ekici, S.; Jawzal, H. Breast Cancer Diagnosis Using Thermography and Convolutional Neural Networks. Med. Hypotheses 2020, 137, 109542. [Google Scholar] [CrossRef] [PubMed]
- Mishra, S.; Prakash, A.; Roy, S.K.; Sharan, P.; Mathur, N. Breast Cancer Detection Using Thermal Images and Deep Learning. In Proceedings of the 2020 7th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 12–14 March 2020; pp. 211–216. [Google Scholar]
- Mambou, S.J.; Maresova, P.; Krejcar, O.; Selamat, A.; Kuca, K. Breast Cancer Detection Using Infrared Thermal Imaging and a Deep Learning Model. Sensors 2018, 18, 2799. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Mookiah, M.R.K.; Acharya, U.R.; Ng, E.Y.K. Data Mining Technique for Breast Cancer Detection in Thermograms Using Hybrid Feature Extraction Strategy. Quant. InfraRed. Thermogr. J. 2012, 9, 151–165. [Google Scholar] [CrossRef]
- Tello-Mijares, S.; Woo, F.; Flores, F. Breast Cancer Identification via Thermography Image Segmentation with a Gradient Vector Flow and a Convolutional Neural Network. J. Healthc. Eng. 2019, 2019, e9807619. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Sánchez-Cauce, R.; Pérez-Martín, J.; Luque, M. Multi-Input Convolutional Neural Network for Breast Cancer Detection Using Thermal Images and Clinical Data. Comput. Methods Programs Biomed. 2021, 204, 106045. [Google Scholar] [CrossRef] [PubMed]
- de Freitas Oliveira Baffa, M.; Grassano Lattari, L. Convolutional Neural Networks for Static and Dynamic Breast Infrared Imaging Classification. In Proceedings of the 2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Parana, Brazil, 29 October–1 November 2018; pp. 174–181. [Google Scholar]
- Tiwari, D.; Dixit, M.; Gupta, K. Deep Multi-View Breast Cancer Detection: A Multi-View Concatenated Infrared Thermal Images Based Breast Cancer Detection System Using Deep Transfer Learning. TS 2021, 38, 1699–1711. [Google Scholar] [CrossRef]
- Deng, J.; Ma, Y.; Li, D.; Zhao, J.; Liu, Y.; Zhang, H. Classification of Breast Density Categories Based on SE-Attention Neural Networks. Comput. Methods Programs Biomed. 2020, 193, 105489. [Google Scholar] [CrossRef] [PubMed]
- Alshehri, A.; AlSaeed, D. Breast Cancer Detection in Thermography Using Convolutional Neural Networks (CNNs) with Deep Attention Mechanisms. Appl. Sci. 2022, 12, 12922. [Google Scholar] [CrossRef]
- Silva, L.F.; Saade, D.C.M.; Sequeiros, G.O.; Silva, A.C.; Paiva, A.C.; Bravo, R.S.; Conci, A. A New Database for Breast Research with Infrared Image. J. Med. Imaging Health Inform. 2014, 4, 92–100. [Google Scholar] [CrossRef]
- CyberZHG Keras-Self-Attention: Attention Mechanism for Processing Sequential Data That Considers the Context for Each Timestamp. Available online: https://pypi.org/project/keras-self-attention/ (accessed on 12 December 2022).
- Soft & Hard Attention. Available online: https://jhui.github.io/2017/03/15/Soft-and-hard-attention/ (accessed on 12 December 2022).
- Murugan, R.; Goel, T. E-DiCoNet: Extreme Learning Machine Based Classifier for Diagnosis of COVID-19 Using Deep Convolutional Network. J Ambient Intell Hum. Comput. 2021, 12, 8887–8898. [Google Scholar] [CrossRef] [PubMed]
- Luke, J.; Joseph, R.; Balaji, M. Impact of Image Size on Accuracy and Generalization of Convolutional Neural Networks. Int. J. Res. Anal. Rev. (IJRAR) 2019. Available online: https://www.researchgate.net/profile/Mahesh-Balaji/publication/332241609_IMPACT_OF_IMAGE_SIZE_ON_ACCURACY_AND_GENERALIZATION_OF_CONVOLUTIONAL_NEURAL_NETWORKS/links/5fa7a715299bf10f732fdc1c/IMPACT-OF-IMAGE-SIZE-ON-ACCURACY-AND-GENERALIZATION-OF-CONVOLUTIONAL-NEURAL-NETWORKS.pdf (accessed on 20 December 2021).
- McHugh, M.L. Interrater Reliability: The Kappa Statistic. Biochem. Med. 2012, 22, 276–282. [Google Scholar] [CrossRef]
Ref. | Approaches | Imaging Modalities | Datasets | Results |
---|---|---|---|---|
[10] | U-Net CNN | Mammography | CBIS-DDSM mass images CBIS-DDSM microcalcification images | Acc = 94.31% |
[11] | CNN | Mammography | DDSM | Acc = 93.5% AUC= 0.92315 |
[12] | CNN-RNN | Mammography | Mammogram image dataset | Acc = 90.59% Sn = 92.42% Sp = 89.88% |
[13] | Bayes Network Naïve Bayes SVM Knowledge Tree J48 MLP RF RT ELM | Thermal | University Hospital of the Federal University of Pernambuco | Acc = 76.01% |
[14] | SVM KNN DT ANN | Thermal | DBT-TU-JU DMR-IR | Acc = 84.29% Acc = 87.50% |
[15] | CNNs-Bayes optimization algorithm | Thermal | DMI | Acc: 98.95% |
[16] | DCNNs | Thermal | DMR-IR | Acc = 95.8% Sn = 99.40% Sp = 76.3% |
[17] | DNN | Thermal | DMR-IR | Conf (Sick) = 78% Conf (Healthy) = 94% |
[18] | DT Fuzzy Sugeno Naïve Bayes k-Nearest Neighbor Gaussian Mixture Model Probabilistic Neural Network | Thermal | Singapore General Hospital-NEC-Avio Thermo TVS2000 MkIIST System | Acc = 93.30% Sn = 86.70% Sp = 100% |
[19] | CNNs-GVF | Thermal | DMR-IR | Acc = 100% Sn = 100% Sp = 100% |
[20] | Multi-input CNN | Thermal | DMR-IR | Acc = 97% Sn = 83% Sp = 100% AUC = 0.99 |
[21] | CNNs | Thermal | DMR-IR | Acc(color) = 98% Acc(grayscale) = 95% Acc(color) = 95% Acc(grayscale) = 92% |
[22] | VGG16 | Thermal | DMR-IR | Acc = 99% Sn = 98.04% Sp = 100% Prec = 100% F1-score = 99.01% |
[23] | CNNs + SE-Attention | Mammography | New Benchmarking dataset | Acc = 92.17% |
[24] | CNNs + AMs | Thermal | DMR-IR | Acc = 99.49% Sn = 99.71% Sp = 99.71% Prec = 99.28% F1-score = 99.49% AUC = 0.999 |
Dataset | Size | Abnormal Breasts | Normal Breasts |
---|---|---|---|
DMR-IR-Original | 1542 | 762 | 780 |
DMR-IR-Augmented | 2604 | 1284 | 1320 |
Total | 4146 | 2046 | 2100 |
Models | Accuracy (%) | Specificity (%) | Sensitivity (Recall) (%) | Precision (%) | F1-Score (%) | AUC | Cohen’s Kappa |
---|---|---|---|---|---|---|---|
VGG16 | 99.18% | 98.48% | 99.85% | 98.55% | 99.20% | 0.999 | 0.983 |
VGG16-SL | 99.49% | 99.56% | 99.42% | 99.57% | 99.49% | 0.994 | 0.989 |
VGG16-HD | 99.80% | 99.75% | 99.85% | 99.76% | 99.80% | 0.999 | 0.996 |
VGG16-SF | 99.32% | 99.46% | 99.19% | 99.48% | 99.33% | 0.999 | 0.986 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Alshehri, A.; AlSaeed, D. Breast Cancer Diagnosis in Thermography Using Pre-Trained VGG16 with Deep Attention Mechanisms. Symmetry 2023, 15, 582. https://doi.org/10.3390/sym15030582
Alshehri A, AlSaeed D. Breast Cancer Diagnosis in Thermography Using Pre-Trained VGG16 with Deep Attention Mechanisms. Symmetry. 2023; 15(3):582. https://doi.org/10.3390/sym15030582
Chicago/Turabian StyleAlshehri, Alia, and Duaa AlSaeed. 2023. "Breast Cancer Diagnosis in Thermography Using Pre-Trained VGG16 with Deep Attention Mechanisms" Symmetry 15, no. 3: 582. https://doi.org/10.3390/sym15030582
APA StyleAlshehri, A., & AlSaeed, D. (2023). Breast Cancer Diagnosis in Thermography Using Pre-Trained VGG16 with Deep Attention Mechanisms. Symmetry, 15(3), 582. https://doi.org/10.3390/sym15030582