Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,569)

Search Parameters:
Keywords = VGG19

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 3514 KiB  
Article
Advancing Brain MRI Image Classification: Integrating VGG16 and ResNet50 with a Multi-Verse Optimization Method
by Nazanin Tataei Sarshar, Soroush Sadeghi, Mohammadreza Kamsari, Mahrokh Avazpour, Saeid Jafarzadeh Ghoushchi and Ramin Ranjbarzadeh
BioMed 2024, 4(4), 499-523; https://doi.org/10.3390/biomed4040038 - 24 Nov 2024
Viewed by 296
Abstract
Background/Objectives: The accurate categorization of brain MRI images into tumor and non-tumor categories is essential for a prompt and effective diagnosis. This paper presents a novel methodology utilizing advanced Convolutional Neural Network (CNN) designs to tackle the complexity and unpredictability present in brain [...] Read more.
Background/Objectives: The accurate categorization of brain MRI images into tumor and non-tumor categories is essential for a prompt and effective diagnosis. This paper presents a novel methodology utilizing advanced Convolutional Neural Network (CNN) designs to tackle the complexity and unpredictability present in brain MRI data. Methods: The methodology commences with an extensive preparation phase that includes image resizing, grayscale conversion, Gaussian blurring, and the delineation of the brain region for preparing the MRI images for analysis. The Multi-verse Optimizer (MVO) is utilized to optimize data augmentation parameters and refine the configuration of trainable layers in VGG16 and ResNet50. The model’s generalization capabilities are significantly improved by the MVO’s ability to effectively balance computational cost and performance. Results: The amalgamation of VGG16 and ResNet50, further refined by the MVO, exhibits substantial enhancements in classification metrics. The MVO-optimized hybrid model demonstrates enhanced performance, exhibiting a well-calibrated balance between precision and recall, rendering it exceptionally trustworthy for medical diagnostic applications. Conclusions: The results highlight the effectiveness of MVO-optimized CNN models for classifying brain tumors in MRI data. Future investigations may examine the model’s applicability to multiclass issues and its validation in practical clinical environments. Full article
29 pages, 5374 KiB  
Article
LVGG-IE: A Novel Lightweight VGG-Based Image Encryption Scheme
by Mingliang Sun, Jie Yuan, Xiaoyong Li, Dongxiao Liu and Xinghai Wei
Entropy 2024, 26(12), 1013; https://doi.org/10.3390/e26121013 - 23 Nov 2024
Viewed by 162
Abstract
Image security faces increasing challenges with the widespread application of computer science and artificial intelligence. Although chaotic systems are employed to encrypt images and prevent unauthorized access or tampering, the degradation that occurs during the binarization process in chaotic systems reduces security. The [...] Read more.
Image security faces increasing challenges with the widespread application of computer science and artificial intelligence. Although chaotic systems are employed to encrypt images and prevent unauthorized access or tampering, the degradation that occurs during the binarization process in chaotic systems reduces security. The chaos- and DNA-based image encryption schemes increases its complexity, while the integration of deep learning with image encryption is still in its infancy and has several shortcomings. An image encryption scheme with high security and efficiency is required for the protection of the image. To address these problems, we propose a novel image encryption scheme based on the lightweight VGG (LVGG), referred to as LVGG-IE. In this work, we design an LVGG network with fewer layers while maintaining a high capacity for feature capture. This network is used to generate a key seed, which is then employed to transform the plaintext image into part of the initial value of a chaotic system, ensuring that the chaos-based key generator correlates with the plaintext image. A dynamic substitution box (S-box) is also designed and used to scramble the randomly shuffled plaintext image. Additionally, a single-connected (SC) layer is combined with a convolution layer from VGG to encrypt the image, where the SC layer is dynamically constructed by the secret key and the convolution kernel is set to . The encryption efficiency is simulated, and the security is analyzed. The results show that the correlation coefficient between adjacent pixels in the proposed scheme achieves . The NPCR exceeds 0.9958, and the UACI falls within the theoretical value with a significance level of 0.05. The encryption quality, the security of the dynamic S-box and the SC layer, and the efficiency are tested. The result shows that the proposed image encryption scheme demonstrates high security, efficiency, and robustness, making it effective for image security in various applications. Full article
(This article belongs to the Section Multidisciplinary Applications)
24 pages, 21174 KiB  
Article
An Ensemble Deep CNN Approach for Power Quality Disturbance Classification: A Technological Route Towards Smart Cities Using Image-Based Transfer
by Mirza Ateeq Ahmed Baig, Naeem Iqbal Ratyal, Adil Amin, Umar Jamil, Sheroze Liaquat, Haris M. Khalid and Muhammad Fahad Zia
Future Internet 2024, 16(12), 436; https://doi.org/10.3390/fi16120436 - 22 Nov 2024
Viewed by 594
Abstract
The abundance of powered semiconductor devices has increased with the introduction of renewable energy sources into the grid, causing power quality disturbances (PQDs). This represents a huge challenge for grid reliability and smart city infrastructures. Accurate detection and classification are important for grid [...] Read more.
The abundance of powered semiconductor devices has increased with the introduction of renewable energy sources into the grid, causing power quality disturbances (PQDs). This represents a huge challenge for grid reliability and smart city infrastructures. Accurate detection and classification are important for grid reliability and consumers’ appliances in a smart city environment. Conventionally, power quality monitoring relies on trivial machine learning classifiers or signal processing methods. However, recent advancements have introduced Deep Convolution Neural Networks (DCNNs) as promising methods for the detection and classification of PQDs. These techniques have the potential to demonstrate high classification accuracy, making them a more appropriate choice for real-time operations in a smart city framework. This paper presents a voting ensemble approach to classify sixteen PQDs, using the DCNN architecture through transfer learning. In this process, continuous wavelet transform (CWT) is employed to convert one-dimensional (1-D) PQD signals into time–frequency images. Four pre-trained DCNN architectures, i.e., Residual Network-50 (ResNet-50), Visual Geometry Group-16 (VGG-16), AlexNet and SqeezeNet are trained and implemented in MATLAB, using images of four datasets, i.e., without noise, 20 dB noise, 30 dB noise and random noise. Additionally, we also tested the performance of ResNet-50 with a squeeze-and-excitation (SE) mechanism. It was observed that ResNet-50 with the SE mechanism has a better classification accuracy; however, it causes computational overheads. The classification performance is enhanced by using the voting ensemble model. The results indicate that the proposed scheme improved the accuracy (99.98%), precision (99.97%), recall (99.80%) and F1-score (99.85%). As an outcome of this work, it is demonstrated that ResNet-50 with the SE mechanism is a viable choice as a single classification model, while an ensemble approach further increases the generalized performance for PQD classification. Full article
(This article belongs to the Special Issue Artificial Intelligence and Blockchain Technology for Smart Cities)
Show Figures

Figure 1

Figure 1
<p>Power quality disturbance sources in smart grid.</p>
Full article ">Figure 2
<p>The proposed ensemble classifier based on DCNN models for PQD classification. Here, PQD and ResNet are abbreviations of power quality disturbance and residual neural network, respectively.</p>
Full article ">Figure 3
<p>An example of PQDs with 20 dB noise: (<b>a</b>) flicker; (<b>b</b>) flicker + harmonics; (<b>c</b>) flicker + sag; (<b>d</b>) flicker + swell; (<b>e</b>) harmonics; (<b>f</b>) impulsive transient; (<b>g</b>) interruption; (<b>h</b>) interruption + harmonics; (<b>i</b>) normal; (<b>j</b>) notch; (<b>k</b>) oscillatory transient; (<b>l</b>) sag; (<b>m</b>) sag + harmonics; (<b>n</b>) spike; (<b>o</b>) swell; and (<b>p</b>) swell + harmonics.</p>
Full article ">Figure 4
<p>An example of a time–frequency representation of PQDs with 20 dB noise: (<b>a</b>) flicker; (<b>b</b>) flicker + harmonics; (<b>c</b>) flicker + sag; (<b>d</b>) flicker + swell; (<b>e</b>) harmonics; (<b>f</b>) impulsive transient; (<b>g</b>) interruption; (<b>h</b>) interruption + harmonics; (<b>i</b>) normal; (<b>j</b>) notch; (<b>k</b>) oscillatory transient; (<b>l</b>) sag; (<b>m</b>) sag + harmonics; (<b>n</b>) spike; (<b>o</b>) swell; (<b>p</b>) swell + harmonics.</p>
Full article ">Figure 5
<p>ResNet-50 architecture for PQD classification.</p>
Full article ">Figure 6
<p>VGG-16 architecture for PQD classification.</p>
Full article ">Figure 7
<p>AlexNet model for PQD classification.</p>
Full article ">Figure 8
<p>SqueezeNet architecture for PQDs classification. Here, ReLU is an acronym for activation function.</p>
Full article ">Figure 9
<p>ResNet-50 training performance for noisy and noiseless datasets.</p>
Full article ">Figure 9 Cont.
<p>ResNet-50 training performance for noisy and noiseless datasets.</p>
Full article ">Figure 10
<p>ResNet-50 confusion matrices for noisy and noiseless testing datasets.</p>
Full article ">Figure 11
<p>VGG-16 training performance for noisy and noiseless datasets.</p>
Full article ">Figure 12
<p>VGG-16 confusion matrices for noisy and noiseless testing datasets.</p>
Full article ">Figure 12 Cont.
<p>VGG-16 confusion matrices for noisy and noiseless testing datasets.</p>
Full article ">Figure 13
<p>AlexNet training performance for noisy and noiseless datasets.</p>
Full article ">Figure 14
<p>AlexNet confusion matrices for noisy and noiseless testing datasets.</p>
Full article ">Figure 15
<p>ResNet-50 with SE mechanism’s training performance for noisy and noiseless datasets.</p>
Full article ">
21 pages, 6945 KiB  
Article
Automatic Segmentation and Statistical Analysis of the Foveal Avascular Zone
by Geanina Totolici, Mihaela Miron and Anisia-Luiza Culea-Florescu
Technologies 2024, 12(12), 235; https://doi.org/10.3390/technologies12120235 - 21 Nov 2024
Viewed by 475
Abstract
This study facilitates the extraction of foveal avascular zone (FAZ) metrics from optical coherence tomography angiography (OCTA) images, offering valuable clinical insights and enabling detailed statistical analysis of FAZ size and shape across three patient groups: healthy, type II diabetes mellitus and both [...] Read more.
This study facilitates the extraction of foveal avascular zone (FAZ) metrics from optical coherence tomography angiography (OCTA) images, offering valuable clinical insights and enabling detailed statistical analysis of FAZ size and shape across three patient groups: healthy, type II diabetes mellitus and both type II diabetes mellitus (DM) and high blood pressure (HBP). Additionally, it evaluates the performance of four deep learning (DL) models—U-Net, U-Net with DenseNet121, U-Net with MobileNetV2 and U-Net with VGG16—in automating the segmentation of the FAZ. Manual segmentation of the images by ophthalmological clinicians was performed initially, and data augmentation was used to enhance the dataset for robust model training and evaluation. Consequently, the original set of 103 full retina OCTA images was extended to 672 cases, including 42 images from normal patients, 357 images from DM patients, and 273 images from patients with both DM and HBP. Among the models, U-Net with DenseNet outperformed the others, achieving the highest accuracy, Intersection over Union (IoU), and Dice coefficient across all patient groups. This research is distinct in its focus on full retina OCTA images and its inclusion of patients with both hypertension and diabetes, an area that is less studied in the existing literature. Full article
(This article belongs to the Special Issue Medical Imaging & Image Processing III)
Show Figures

Figure 1

Figure 1
<p>Block diagram of the proposed framework.</p>
Full article ">Figure 2
<p>The extraction of the ground-truth mask of FAZ.</p>
Full article ">Figure 3
<p>U-Net model with encoder highlighted by dashed box.</p>
Full article ">Figure 4
<p>U-Net variant with DenseNet121 model, backbone highlighted by dashed box.</p>
Full article ">Figure 5
<p>U-Net variant with MobileNetV2, backbone highlighted by dashed box.</p>
Full article ">Figure 6
<p>U-Net variant with VGG16, backbone highlighted by dashed box.</p>
Full article ">Figure 7
<p>Boxplot analysis of FAZ metrics across patient groups with type II diabetes and type II diabetes + high blood pressure: (<b>a</b>) perimeter (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math>); (<b>b</b>) area (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>A</mi> </mrow> <mrow> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math>).</p>
Full article ">Figure 8
<p>Boxplot analysis of FAZ metrics across patient groups with type II diabetes and type II diabetes + high blood pressure: (<b>a</b>) equivalent circle perimeter (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mn>2</mn> </mrow> </msub> </mrow> </semantics></math>); (<b>b</b>) acircularity index (<span class="html-italic">A_index</span>).</p>
Full article ">Figure 9
<p>Boxplot analysis of FAZ metrics across patient groups with type II diabetes and type II diabetes + high blood pressure: (<b>a</b>) angle (θ); (<b>b</b>) axis ratio (<span class="html-italic">A_ratio</span>).</p>
Full article ">Figure 10
<p>Relationship between central vascular density (DVF C) and (<b>a</b>) perimeter (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math>); (<b>b</b>) area (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>A</mi> </mrow> <mrow> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math>).</p>
Full article ">Figure 11
<p>Relationship between central vascular density (DVF C) and (<b>a</b>) equivalent circle perimeter (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mn>2</mn> </mrow> </msub> </mrow> </semantics></math>); (<b>b</b>) acircularity index (<math display="inline"><semantics> <mrow> <mi>A</mi> <mo>_</mo> <mi>i</mi> <mi>n</mi> <mi>d</mi> <mi>e</mi> <mi>x</mi> </mrow> </semantics></math>).</p>
Full article ">Figure 12
<p>Relationship between central vascular density (DVF C) and (<b>a</b>) angle (θ); (<b>b</b>) axis ratio (A_ratio).</p>
Full article ">Figure 13
<p>Training and validation accuracy for normal group: (<b>a</b>) U-Net; (<b>b</b>) U-Net variant with DenseNet121; (<b>c</b>) U-Net variant with MobileNetV2; (<b>d</b>) U-Net variant with VGG16.</p>
Full article ">Figure 14
<p>Training and validation accuracy for type II DM group: (<b>a</b>) U-Net; (<b>b</b>) U-Net variant with DenseNet121; (<b>c</b>) U-Net variant with MobileNetV2; (<b>d</b>) U-Net variant with VGG16.</p>
Full article ">Figure 14 Cont.
<p>Training and validation accuracy for type II DM group: (<b>a</b>) U-Net; (<b>b</b>) U-Net variant with DenseNet121; (<b>c</b>) U-Net variant with MobileNetV2; (<b>d</b>) U-Net variant with VGG16.</p>
Full article ">Figure 15
<p>Training and validation accuracy for type II DM + HBP group: (<b>a</b>) U-Net; (<b>b</b>) U-Net variant with DenseNet121; (<b>c</b>) U-Net variant with MobileNetV2; (<b>d</b>) U-Net variant with VGG16.</p>
Full article ">Figure 15 Cont.
<p>Training and validation accuracy for type II DM + HBP group: (<b>a</b>) U-Net; (<b>b</b>) U-Net variant with DenseNet121; (<b>c</b>) U-Net variant with MobileNetV2; (<b>d</b>) U-Net variant with VGG16.</p>
Full article ">
19 pages, 4058 KiB  
Article
Enhanced Fault Detection in Photovoltaic Panels Using CNN-Based Classification with PyQt5 Implementation
by Younes Ledmaoui, Adila El Maghraoui, Mohamed El Aroussi and Rachid Saadane
Sensors 2024, 24(22), 7407; https://doi.org/10.3390/s24227407 - 20 Nov 2024
Viewed by 369
Abstract
Solar photovoltaic systems have increasingly become essential for harvesting renewable energy. However, as these systems grow in prevalence, the issue of the end of life of modules is also increasing. Regular maintenance and inspection are vital to extend the lifespan of these systems, [...] Read more.
Solar photovoltaic systems have increasingly become essential for harvesting renewable energy. However, as these systems grow in prevalence, the issue of the end of life of modules is also increasing. Regular maintenance and inspection are vital to extend the lifespan of these systems, minimize energy losses, and protect the environment. This paper presents an innovative explainable AI model for detecting anomalies in solar photovoltaic panels using an enhanced convolutional neural network (CNN) and the VGG16 architecture. The model effectively identifies physical and electrical changes, such as dust and bird droppings, and is implemented using the PyQt5 Python tool to create a user-friendly interface that facilitates decision-making for users. Key processes included dataset balancing through oversampling and data augmentation to expand the dataset. The model achieved impressive performance metrics: 91.46% accuracy, 98.29% specificity, and an F1 score of 91.67%. Overall, it enhances power generation efficiency and prolongs the lifespan of photovoltaic systems, while minimizing environmental risks. Full article
(This article belongs to the Special Issue Sensor Enabled Smart Energy Solutions)
Show Figures

Figure 1

Figure 1
<p>Evolution of installed solar capacity from 2004 to 2023 [<a href="#B4-sensors-24-07407" class="html-bibr">4</a>].</p>
Full article ">Figure 2
<p>PV cell (<b>a</b>), electrical schematic diagram (<b>b</b>).</p>
Full article ">Figure 3
<p>PV system fault classification.</p>
Full article ">Figure 4
<p>Classes of solar panels.</p>
Full article ">Figure 5
<p>Proposed solar panel anomaly detection and classification model.</p>
Full article ">Figure 6
<p>Methodology for the proposed architecture.</p>
Full article ">Figure 7
<p>Architecture of VGG16.</p>
Full article ">Figure 8
<p>Architecture of transfer learning.</p>
Full article ">Figure 9
<p>Correlation matrix of the solar panel dataset for the proposed model.</p>
Full article ">Figure 10
<p>Results with PyQt5 implementation.</p>
Full article ">Figure 11
<p>SPAD upload image (<b>a</b>), image preview (<b>b</b>), and prediction result implementation (<b>c</b>).</p>
Full article ">
30 pages, 2346 KiB  
Article
A Novel Method for 3D Lung Tumor Reconstruction Using Generative Models
by Hamidreza Najafi, Kimia Savoji, Marzieh Mirzaeibonehkhater, Seyed Vahid Moravvej, Roohallah Alizadehsani and Siamak Pedrammehr
Diagnostics 2024, 14(22), 2604; https://doi.org/10.3390/diagnostics14222604 - 20 Nov 2024
Viewed by 487
Abstract
Background: Lung cancer remains a significant health concern, and the effectiveness of early detection significantly enhances patient survival rates. Identifying lung tumors with high precision is a challenge due to the complex nature of tumor structures and the surrounding lung tissues. Methods: To [...] Read more.
Background: Lung cancer remains a significant health concern, and the effectiveness of early detection significantly enhances patient survival rates. Identifying lung tumors with high precision is a challenge due to the complex nature of tumor structures and the surrounding lung tissues. Methods: To address these hurdles, this paper presents an innovative three-step approach that leverages Generative Adversarial Networks (GAN), Long Short-Term Memory (LSTM), and VGG16 algorithms for the accurate reconstruction of three-dimensional (3D) lung tumor images. The first challenge we address is the accurate segmentation of lung tissues from CT images, a task complicated by the overwhelming presence of non-lung pixels, which can lead to classifier imbalance. Our solution employs a GAN model trained with a reinforcement learning (RL)-based algorithm to mitigate this imbalance and enhance segmentation accuracy. The second challenge involves precisely detecting tumors within the segmented lung regions. We introduce a second GAN model with a novel loss function that significantly improves tumor detection accuracy. Following successful segmentation and tumor detection, the VGG16 algorithm is utilized for feature extraction, preparing the data for the final 3D reconstruction. These features are then processed through an LSTM network and converted into a format suitable for the reconstructive GAN. This GAN, equipped with dilated convolution layers in its discriminator, captures extensive contextual information, enabling the accurate reconstruction of the tumor’s 3D structure. Results: The effectiveness of our method is demonstrated through rigorous evaluation against established techniques using the LIDC-IDRI dataset and standard performance metrics, showcasing its superior performance and potential for enhancing early lung cancer detection. Conclusions:This study highlights the benefits of combining GANs, LSTM, and VGG16 into a unified framework. This approach significantly improves the accuracy of detecting and reconstructing lung tumors, promising to enhance diagnostic methods and patient results in lung cancer treatment. Full article
(This article belongs to the Special Issue AI and Digital Health for Disease Diagnosis and Monitoring)
Show Figures

Figure 1

Figure 1
<p>Overview of the proposed model: In step 1, two lungs are segmented in CT images using the GAN-based model. In step 2, the tumor is detected using the second GAN-based model. After the features are extracted by VGG16, a 3D model of the tumor is reconstructed using the third GAN in step 3.</p>
Full article ">Figure 2
<p>Architecture of the U-Ne-based generator network used for lung segmentation, illustrating the flow from input CT scan through the encoder and decoder stages to the final mask output.</p>
Full article ">Figure 3
<p>Comparative visualization of original CT scans and segmented lung regions by the proposed model.</p>
Full article ">Figure 4
<p>(<b>a</b>) Optimization of the <math display="inline"><semantics> <mi>λ</mi> </semantics></math> parameter for lung segmentation model performance, (<b>b</b>) learning trajectory of the reward optimization of the agent on 100 episodes.</p>
Full article ">Figure 5
<p>Comparative analysis of tumor detection. The top row shows the ground truth, while the bottom row presents the model predictions. The black square in every sample shows the tumor extracted by the proposed model.</p>
Full article ">Figure 6
<p>Comparison of original and reconstructed 3D tumor shapes highlighting the effectiveness of the proposed reconstruction method in preserving boundary smoothness.</p>
Full article ">Figure 7
<p>Distribution of decision-making times for 3D reconstruction in RTB environments.</p>
Full article ">Figure 8
<p>Loss trends in (<b>a</b>) lung segmentation, (<b>b</b>) tumor detection, and (<b>c</b>) 3D reconstruction models over 250 epochs.</p>
Full article ">Figure 9
<p>HD metric trends in lung segmentation, tumor detection, and 3D reconstruction over 250 epochs.</p>
Full article ">
16 pages, 1799 KiB  
Article
Optimizing Fire Scene Analysis: Hybrid Convolutional Neural Network Model Leveraging Multiscale Feature and Attention Mechanisms
by Shakhnoza Muksimova, Sabina Umirzakova, Mirjamol Abdullaev and Young-Im Cho
Fire 2024, 7(11), 422; https://doi.org/10.3390/fire7110422 - 20 Nov 2024
Viewed by 398
Abstract
The rapid and accurate detection of fire scenes in various environments is crucial for effective disaster management and mitigation. Fire scene classification is a critical aspect of modern fire detection systems that directly affects public safety and property preservation. This research introduced a [...] Read more.
The rapid and accurate detection of fire scenes in various environments is crucial for effective disaster management and mitigation. Fire scene classification is a critical aspect of modern fire detection systems that directly affects public safety and property preservation. This research introduced a novel hybrid deep learning model designed to enhance the accuracy and efficiency of fire scene classification across diverse environments. The proposed model integrates advanced convolutional neural networks with multiscale feature extraction, attention mechanisms, and ensemble learning to achieve superior performance in real-time fire detection. By leveraging the strengths of pre-trained networks such as ResNet50, VGG16, and EfficientNet-B3, the model captures detailed features at multiple scales, ensuring robust detection capabilities. Including spatial and channel attention mechanisms further refines the focus on critical areas within the input images, reducing false positives and improving detection precision. Extensive experiments on a comprehensive dataset encompassing wildfires, building fires, vehicle fires, and non-fire scenes demonstrate that the proposed framework outperforms existing cutting-edge techniques. The model also exhibited reduced computational complexity and enhanced inference speed, making it suitable for deployment in real-time applications on various hardware platforms. This study sets a new benchmark for fire detection and offers a powerful tool for early warning systems and emergency response initiatives. Full article
Show Figures

Figure 1

Figure 1
<p>Hybrid convolutional neural network architecture for fire type classification.</p>
Full article ">Figure 2
<p>Examples of images from each of the four classes used in the training datasets.</p>
Full article ">Figure 2 Cont.
<p>Examples of images from each of the four classes used in the training datasets.</p>
Full article ">
27 pages, 28012 KiB  
Article
A Model Development Approach Based on Point Cloud Reconstruction and Mapping Texture Enhancement
by Boyang You and Barmak Honarvar Shakibaei Asli
Big Data Cogn. Comput. 2024, 8(11), 164; https://doi.org/10.3390/bdcc8110164 - 20 Nov 2024
Viewed by 318
Abstract
To address the challenge of rapid geometric model development in the digital twin industry, this paper presents a comprehensive pipeline for constructing 3D models from images using monocular vision imaging principles. Firstly, a structure-from-motion (SFM) algorithm generates a 3D point cloud from photographs. [...] Read more.
To address the challenge of rapid geometric model development in the digital twin industry, this paper presents a comprehensive pipeline for constructing 3D models from images using monocular vision imaging principles. Firstly, a structure-from-motion (SFM) algorithm generates a 3D point cloud from photographs. The feature detection methods scale-invariant feature transform (SIFT), speeded-up robust features (SURF), and KAZE are compared across six datasets, with SIFT proving the most effective (matching rate higher than 0.12). Using K-nearest-neighbor matching and random sample consensus (RANSAC), refined feature point matching and 3D spatial representation are achieved via antipodal geometry. Then, the Poisson surface reconstruction algorithm converts the point cloud into a mesh model. Additionally, texture images are enhanced by leveraging a visual geometry group (VGG) network-based deep learning approach. Content images from a dataset provide geometric contours via higher-level VGG layers, while textures from style images are extracted using the lower-level layers. These are fused to create texture-transferred images, where the image quality assessment (IQA) metrics SSIM and PSNR are used to evaluate texture-enhanced images. Finally, texture mapping integrates the enhanced textures with the mesh model, improving the scene representation with enhanced texture. The method presented in this paper surpassed a LiDAR-based reconstruction approach by 20% in terms of point cloud density and number of model facets, while the hardware cost was only 1% of that associated with LiDAR. Full article
Show Figures

Figure 1

Figure 1
<p>Samples from Dataset 1 (Source: <a href="https://github.com/Abhishek-Aditya-bs/MultiView-3D-Reconstruction/tree/main/Datasets" target="_blank">https://github.com/Abhishek-Aditya-bs/MultiView-3D-Reconstruction/tree/main/Datasets</a> accessed on 18 November 2024) and samples from Dataset 2.</p>
Full article ">Figure 2
<p>Demonstration of Dataset 3.</p>
Full article ">Figure 3
<p>Diagram of SFM algorithm.</p>
Full article ">Figure 4
<p>Camera imaging model.</p>
Full article ">Figure 5
<p>Coplanarity condition of photogrammetry.</p>
Full article ">Figure 6
<p>Process of surface reconstruction.</p>
Full article ">Figure 7
<p>Demonstration of isosurface.</p>
Full article ">Figure 8
<p>Demonstration of VGG network.</p>
Full article ">Figure 9
<p>Demonstration of Gram matrix.</p>
Full article ">Figure 10
<p>Style transformation architecture.</p>
Full article ">Figure 11
<p>Texture mapping process.</p>
Full article ">Figure 12
<p>Demonstration of the three kinds of feature descriptors used on Dataset 1 and Dataset 2.</p>
Full article ">Figure 13
<p>Matching rate fitting of three kinds of image descriptors.</p>
Full article ">Figure 14
<p>SIFT point matching for <span class="html-italic">CNC1</span> object under different thresholds.</p>
Full article ">Figure 15
<p>SIFT point matching for <span class="html-italic">Fountain</span> object under different thresholds.</p>
Full article ">Figure 16
<p>Matching result of Dataset 2 using RANSAC method.</p>
Full article ">Figure 17
<p>Triangulation presentation of feature points obtained from objects in Dataset 1.</p>
Full article ">Figure 18
<p>Triangulation presentation of feature points obtained from objects in Dataset 2.</p>
Full article ">Figure 19
<p>Point cloud data of objects in Dataset 1.</p>
Full article ">Figure 20
<p>Point cloud data of objects in Dataset 2.</p>
Full article ">Figure 21
<p>Normal vector presentation of the points set obtained from objects in Dataset 1.</p>
Full article ">Figure 22
<p>Normal vector of the points set obtained from objects in Dataset 2.</p>
Full article ">Figure 23
<p>Poisson surface reconstruction results of objects in Dataset 1.</p>
Full article ">Figure 24
<p>Poisson surface reconstruction results of objects in Dataset 2.</p>
Full article ">Figure 25
<p>Style transfer result of <span class="html-italic">Statue</span> object.</p>
Full article ">Figure 26
<p>Style transfer result of <span class="html-italic">Fountain</span> object.</p>
Full article ">Figure 27
<p>Style transfer result of <span class="html-italic">Castle</span> object.</p>
Full article ">Figure 28
<p>Style transfer result of <span class="html-italic">CNC1</span> object.</p>
Full article ">Figure 29
<p>Style transfer result of <span class="html-italic">CNC2</span> object.</p>
Full article ">Figure 30
<p>Style transfer result of <span class="html-italic">Robot</span> object.</p>
Full article ">Figure 31
<p>Training loss in style transfer for <b>CNC1</b> object.</p>
Full article ">Figure 32
<p>IQA assessment for <b>CNC1</b> images after style transfer.</p>
Full article ">Figure 33
<p>Results of texture mapping for Dataset 1.</p>
Full article ">Figure 34
<p>Results of texture mapping for Dataset 2.</p>
Full article ">Figure A1
<p>Results of Camera calibration.</p>
Full article ">
22 pages, 1988 KiB  
Article
Assessing the Performance of Deep Learning Predictions for Dynamic Aperture of a Hadron Circular Particle Accelerator
by Davide Di Croce, Massimo Giovannozzi, Carlo Emilio Montanari, Tatiana Pieloni, Stefano Redaelli and Frederik F. Van der Veken
Instruments 2024, 8(4), 50; https://doi.org/10.3390/instruments8040050 - 19 Nov 2024
Viewed by 392
Abstract
Understanding the concept of dynamic aperture provides essential insights into nonlinear beam dynamics, beam losses, and the beam lifetime in circular particle accelerators. This comprehension is crucial for the functioning of modern hadron synchrotrons like the CERN Large Hadron Collider and the planning [...] Read more.
Understanding the concept of dynamic aperture provides essential insights into nonlinear beam dynamics, beam losses, and the beam lifetime in circular particle accelerators. This comprehension is crucial for the functioning of modern hadron synchrotrons like the CERN Large Hadron Collider and the planning of future ones such as the Future Circular Collider. The dynamic aperture defines the extent of the region in phase space where the trajectories of charged particles are bounded over numerous revolutions, the actual number being defined by the physical application. Traditional methods for calculating the dynamic aperture depend on computationally demanding numerical simulations, which require tracking over multiple turns of numerous initial conditions appropriately distributed in phase space. Prior research has shown the efficiency of a multilayer perceptron network in forecasting the dynamic aperture of the CERN Large Hadron Collider ring, achieving a remarkable speed-up of up to 200-fold compared to standard numerical tracking tools. Building on recent advancements, we conducted a comparative study of various deep learning networks based on BERT, DenseNet, ResNet and VGG architectures. The results demonstrate substantial enhancements in the prediction of the dynamic aperture, marking a significant advancement in the development of more precise and efficient surrogate models of beam dynamics. Full article
Show Figures

Figure 1

Figure 1
<p>Stability time for a distribution of initial conditions in <math display="inline"><semantics> <mrow> <mi>x</mi> <mo>−</mo> <mi>y</mi> </mrow> </semantics></math> space for one of the LHC configurations that are part of the data set used for constructing the DL surrogate models. The reduction in the extent of the stable region for an increasing number of turns is clearly visible. This information is then used to determine the value of the angular <math display="inline"><semantics> <mi>DA</mi> </semantics></math>. As an example, the angular <math display="inline"><semantics> <mi>DA</mi> </semantics></math> for 1 × 10<sup>5</sup> turns is shown in red.</p>
Full article ">Figure 2
<p>Distribution of the angular <math display="inline"><semantics> <mi>DA</mi> </semantics></math> before (blue) and after (red) the augmentation and unbiasing pre-processing.</p>
Full article ">Figure 3
<p>Architecture of the designed MLP for the <math display="inline"><semantics> <mi>DA</mi> </semantics></math> regressor, featuring a fully connected structure.</p>
Full article ">Figure 4
<p>Architecture of the designed BERT for the <math display="inline"><semantics> <mi>DA</mi> </semantics></math> regressor.</p>
Full article ">Figure 5
<p>Architecture of the designed DenseNet-121 for the <math display="inline"><semantics> <mi>DA</mi> </semantics></math> regressor.</p>
Full article ">Figure 6
<p>Architecture of the designed ResNet-18 for the <math display="inline"><semantics> <mi>DA</mi> </semantics></math> regressor.</p>
Full article ">Figure 7
<p>Architecture of the designed VGG-16 for the <math display="inline"><semantics> <mi>DA</mi> </semantics></math> regressor.</p>
Full article ">Figure 8
<p>Architecture of the designed Hybrid network for the <math display="inline"><semantics> <mi>DA</mi> </semantics></math> regressor.</p>
Full article ">Figure 9
<p>Training and validation performance over epochs. The blue line represents the training loss (MAE), while the orange line shows the validation loss. The green line tracks the learning rate adjustments throughout the training. The vertical dashed line indicates the epoch at which the model was saved.</p>
Full article ">Figure 10
<p>Predicted angular <math display="inline"><semantics> <mi>DA</mi> </semantics></math> as a function of the computed angular <math display="inline"><semantics> <mi>DA</mi> </semantics></math> values for the test data set for: (<b>a</b>) BERT, (<b>b</b>) DenseNet-121, (<b>c</b>) ResNet-18, (<b>d</b>) MLP (baseline), (<b>e</b>) Hybrid and (<b>f</b>) VGG-16 architectures. The Pearson correlation coefficient is also shown.</p>
Full article ">Figure 10 Cont.
<p>Predicted angular <math display="inline"><semantics> <mi>DA</mi> </semantics></math> as a function of the computed angular <math display="inline"><semantics> <mi>DA</mi> </semantics></math> values for the test data set for: (<b>a</b>) BERT, (<b>b</b>) DenseNet-121, (<b>c</b>) ResNet-18, (<b>d</b>) MLP (baseline), (<b>e</b>) Hybrid and (<b>f</b>) VGG-16 architectures. The Pearson correlation coefficient is also shown.</p>
Full article ">Figure 11
<p>Computed (blue) and predicted (red) angular <math display="inline"><semantics> <mi>DA</mi> </semantics></math> distribution for the test data set for: (<b>a</b>) BERT, (<b>b</b>) DenseNet-121, (<b>c</b>) ResNet-18, (<b>d</b>) MLP (baseline), (<b>e</b>) Hybrid and (<b>f</b>) VGG-16 architectures. The outcome of the Kolmogorov–Smirnov test, used to compare the two distributions, is also reported.</p>
Full article ">Figure 11 Cont.
<p>Computed (blue) and predicted (red) angular <math display="inline"><semantics> <mi>DA</mi> </semantics></math> distribution for the test data set for: (<b>a</b>) BERT, (<b>b</b>) DenseNet-121, (<b>c</b>) ResNet-18, (<b>d</b>) MLP (baseline), (<b>e</b>) Hybrid and (<b>f</b>) VGG-16 architectures. The outcome of the Kolmogorov–Smirnov test, used to compare the two distributions, is also reported.</p>
Full article ">Figure 12
<p>The true angular <math display="inline"><semantics> <mi>DA</mi> </semantics></math> (<b>left</b>) and the BERT prediction (<b>right</b>) for four different machine configurations present on the test data set. The colours indicate the stability time in turns.</p>
Full article ">Figure 13
<p>Box plot of the MAE as a function of the number of turns for the machine configurations tracked up to 5 × 10<sup>5</sup> turns in the test data set. The box limits indicate the range of the central 50% of the data with a central line marking the mean value.</p>
Full article ">
17 pages, 1713 KiB  
Article
Simplified Knowledge Distillation for Deep Neural Networks Bridging the Performance Gap with a Novel Teacher–Student Architecture
by Sabina Umirzakova, Mirjamol Abdullaev, Sevara Mardieva, Nodira Latipova and Shakhnoza Muksimova
Electronics 2024, 13(22), 4530; https://doi.org/10.3390/electronics13224530 - 18 Nov 2024
Viewed by 359
Abstract
The rapid evolution of deep learning has led to significant achievements in computer vision, primarily driven by complex convolutional neural networks (CNNs). However, the increasing depth and parameter count of these networks often result in overfitting and elevated computational demands. Knowledge distillation (KD) [...] Read more.
The rapid evolution of deep learning has led to significant achievements in computer vision, primarily driven by complex convolutional neural networks (CNNs). However, the increasing depth and parameter count of these networks often result in overfitting and elevated computational demands. Knowledge distillation (KD) has emerged as a promising technique to address these issues by transferring knowledge from a large, well-trained teacher model to a more compact student model. This paper introduces a novel knowledge distillation method that simplifies the distillation process and narrows the performance gap between teacher and student models without relying on intricate knowledge representations. Our approach leverages a unique teacher network architecture designed to enhance the efficiency and effectiveness of knowledge transfer. Additionally, we introduce a streamlined teacher network architecture that transfers knowledge effectively through a simplified distillation process, enabling the student model to achieve high accuracy with reduced computational demands. Comprehensive experiments conducted on the CIFAR-10 dataset demonstrate that our proposed model achieves superior performance compared to traditional KD methods and established architectures such as ResNet and VGG networks. The proposed method not only maintains high accuracy but also significantly reduces training and validation losses. Key findings highlight the optimal hyperparameter settings (temperature T = 15.0 and smoothing factor α = 0.7), which yield the highest validation accuracy and lowest loss values. This research contributes to the theoretical and practical advancements in knowledge distillation, providing a robust framework for future applications and research in neural network compression and optimization. The simplicity and efficiency of our approach pave the way for more accessible and scalable solutions in deep learning model deployment. Full article
Show Figures

Figure 1

Figure 1
<p>The architecture of teacher–student models using the knowledge distillation algorithm.</p>
Full article ">Figure 2
<p>The architecture of novel teacher model.</p>
Full article ">Figure 3
<p>The result of the proposed method based on CIFAR—10 dataset, blue boxes show correct predictions while red boxes are incorrect.</p>
Full article ">
32 pages, 11565 KiB  
Article
Advanced Segmentation of Gastrointestinal (GI) Cancer Disease Using a Novel U-MaskNet Model
by Aditya Pal, Hari Mohan Rai, Mohamed Ben Haj Frej and Abdul Razaque
Life 2024, 14(11), 1488; https://doi.org/10.3390/life14111488 - 15 Nov 2024
Viewed by 478
Abstract
The purpose of this research is to contribute to the development of approaches for the classification and segmentation of various gastrointestinal (GI) cancer diseases, such as dyed lifted polyps, dyed resection margins, esophagitis, normal cecum, normal pylorus, normal Z line, polyps, and ulcerative [...] Read more.
The purpose of this research is to contribute to the development of approaches for the classification and segmentation of various gastrointestinal (GI) cancer diseases, such as dyed lifted polyps, dyed resection margins, esophagitis, normal cecum, normal pylorus, normal Z line, polyps, and ulcerative colitis. This research is relevant and essential because of the current challenges related to the absence of efficient diagnostic tools for early diagnostics of GI cancers, which are fundamental for improving the diagnosis of these common diseases. To address the above challenges, we propose a new hybrid segmentation model, U-MaskNet, which is a combination of U-Net and Mask R-CNN models. Here, U-Net is utilized for pixel-wise classification and Mask R-CNN for instance segmentation, together forming a solution for classifying and segmenting GI cancer. The Kvasir dataset, which includes 8000 endoscopic images of various GI cancers, is utilized to validate the proposed methodology. The experimental results clearly demonstrated that the novel proposed model provided superior segmentation compared to other well-known models, such as DeepLabv3+, FCN, and DeepMask, as well as improved classification performance compared to state-of-the-art (SOTA) models, including LeNet-5, AlexNet, VGG-16, ResNet-50, and the Inception Network. The quantitative analysis revealed that our proposed model outperformed the other models, achieving a precision of 98.85%, recall of 98.49%, and F1 score of 98.68%. Additionally, the novel model achieved a Dice coefficient of 94.35% and IoU of 89.31%. Consequently, the developed model increased the accuracy and reliability in detecting and segmenting GI cancer, and it was proven that the proposed model can potentially be used for improving the diagnostic process and, consequently, patient care in the clinical environment. This work highlights the benefits of integrating the U-Net and Mask R-CNN models, opening the way for further research in medical image segmentation. Full article
(This article belongs to the Section Medical Research)
Show Figures

Figure 1

Figure 1
<p>Proposed architecture of U-MaskNet used in our research for GI image segmentation.</p>
Full article ">Figure 2
<p>Dataset overview: sample GI images.</p>
Full article ">Figure 3
<p>Detailed architecture of the VGG19 convolutional neural network.</p>
Full article ">Figure 4
<p>Proposed U-MaskNet architecture used in our methodology.</p>
Full article ">Figure 5
<p>Training curves of the DeepLabv3+ model: (<b>a</b>) Dice plot, (<b>b</b>) IoU plot, (<b>c</b>) loss plot, (<b>d</b>) precision plot, and (<b>e</b>) recall plot.</p>
Full article ">Figure 6
<p>Training curves of the FCN model: (<b>a</b>) Dice plot, (<b>b</b>) IoU plot, (<b>c</b>) loss plot, (<b>d</b>) precision plot, and (<b>e</b>) recall plot.</p>
Full article ">Figure 7
<p>Training curves of the DeepMask model: (<b>a</b>) Dice plot, (<b>b</b>) IoU plot, (<b>c</b>) loss plot, (<b>d</b>) precision plot, and (<b>e</b>) recall plot.</p>
Full article ">Figure 8
<p>Training curves of the U-MaskNet model: (<b>a</b>) Dice plot, (<b>b</b>) IoU plot, (<b>c</b>) loss plot, (<b>d</b>) precision plot, and (<b>e</b>) recall plot.</p>
Full article ">Figure 9
<p>Qualitative classification results of GI diseases.</p>
Full article ">Figure 10
<p>Segmentation performance of GI cancer using U-MaskNet.</p>
Full article ">Figure 11
<p>(<b>a</b>) Training set confusion matrix, (<b>b</b>) validation set confusion matrix, and (<b>c</b>) test set confusion matrix.</p>
Full article ">Figure 11 Cont.
<p>(<b>a</b>) Training set confusion matrix, (<b>b</b>) validation set confusion matrix, and (<b>c</b>) test set confusion matrix.</p>
Full article ">Figure 12
<p>Performance comparison of segmentation models.</p>
Full article ">Figure 13
<p>Visualized performance of the proposed U-MaskNet model compared to other state-of-the-art models.</p>
Full article ">
16 pages, 4027 KiB  
Article
Detecting Botrytis Cinerea Control Efficacy via Deep Learning
by Wenlong Yi, Xunsheng Zhang, Shiming Dai, Sergey Kuzmin, Igor Gerasimov and Xiangping Cheng
Agriculture 2024, 14(11), 2054; https://doi.org/10.3390/agriculture14112054 - 14 Nov 2024
Viewed by 384
Abstract
This study proposes a deep learning-based method for monitoring the growth of Botrytis cinerea and evaluating the effectiveness of control measures. It aims to address the limitations of traditional statistical analysis methods in capturing non-linear relationships and multi-factor synergistic effects. The method integrates [...] Read more.
This study proposes a deep learning-based method for monitoring the growth of Botrytis cinerea and evaluating the effectiveness of control measures. It aims to address the limitations of traditional statistical analysis methods in capturing non-linear relationships and multi-factor synergistic effects. The method integrates colony growth environment data and images as network inputs, achieving real-time prediction of colony area through an improved RepVGG network. The innovations include (1) combining channel attention mechanism, multi-head self-attention mechanism, and multi-scale feature extractor to improve prediction accuracy and (2) introducing the Shapley value algorithm to achieve a precise quantitative analysis of environmental variables’ contribution to colony growth. Experimental results show that the validation loss of this method reaches 0.007, with a mean absolute error of 0.0148, outperforming other comparative models. This study enriches the theory of gray mold control and provides information technology for optimizing and selecting its inhibitors. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>Data collection device for the control efficacy of Botrytis cinerea.</p>
Full article ">Figure 2
<p>Semantic segmentation of Botrytis cinerea colonies.</p>
Full article ">Figure 3
<p>Network for detecting Botrytis cinerea prevention results.</p>
Full article ">Figure 4
<p>Calculation of Botrytis cinerea colony area.</p>
Full article ">Figure 5
<p>Proposed network model training and validation results. (<b>a</b>) Loss value; (<b>b</b>) MAE value.</p>
Full article ">Figure 6
<p>Results of statistical analysis.</p>
Full article ">Figure 7
<p>Analysis of Botrytis cinerea control efficacy. (<b>a</b>) Impact of varying conditions on colony growth; (<b>b</b>) Impact of single sample on the proposed model’s performance.</p>
Full article ">Figure 8
<p>Training Loss results of different improved RepVGG networks. (<b>a</b>) Loss value; (<b>b</b>) MAE value.</p>
Full article ">Figure 9
<p>Comparison of training average absolute errors for various network architecture variants.</p>
Full article ">Figure 10
<p>Comparison of different network models. (<b>a</b>) Loss value; (<b>b</b>) MAE value.</p>
Full article ">
20 pages, 578 KiB  
Review
Systematic Review of Deep Learning Techniques in Skin Cancer Detection
by Carolina Magalhaes, Joaquim Mendes and Ricardo Vardasca
BioMedInformatics 2024, 4(4), 2251-2270; https://doi.org/10.3390/biomedinformatics4040121 - 14 Nov 2024
Viewed by 462
Abstract
Skin cancer is a serious health condition, as it can locally evolve into disfiguring states or metastasize to different tissues. Early detection of this disease is critical because it increases the effectiveness of treatment, which contributes to improved patient prognosis and reduced healthcare [...] Read more.
Skin cancer is a serious health condition, as it can locally evolve into disfiguring states or metastasize to different tissues. Early detection of this disease is critical because it increases the effectiveness of treatment, which contributes to improved patient prognosis and reduced healthcare costs. Visual assessment and histopathological examination are the gold standards for diagnosing these types of lesions. Nevertheless, these processes are strongly dependent on dermatologists’ experience, with excision advised only when cancer is suspected by a physician. Multiple approaches have surfed over the last few years, particularly those based on deep learning (DL) strategies, with the goal of assisting medical professionals in the diagnosis process and ultimately diminishing diagnostic uncertainty. This systematic review focused on the analysis of relevant studies based on DL applications for skin cancer diagnosis. The qualitative assessment included 164 records relevant to the topic. The AlexNet, ResNet-50, VGG-16, and GoogLeNet architectures are considered the top choices for obtaining the best classification results, and multiclassification approaches are the current trend. Public databases are considered key elements in this area and should be maintained and improved to facilitate scientific research. Full article
Show Figures

Figure 1

Figure 1
<p>PRISMA flow diagram for the systematic review of deep learning applications for skin cancer diagnosis [<a href="#B16-biomedinformatics-04-00121" class="html-bibr">16</a>].</p>
Full article ">
27 pages, 3127 KiB  
Article
Recognition of Sheep Feeding Behavior in Sheepfolds Using Fusion Spectrogram Depth Features and Acoustic Features
by Youxin Yu, Wenbo Zhu, Xiaoli Ma, Jialei Du, Yu Liu, Linhui Gan, Xiaoping An, Honghui Li, Buyu Wang and Xueliang Fu
Animals 2024, 14(22), 3267; https://doi.org/10.3390/ani14223267 - 13 Nov 2024
Viewed by 312
Abstract
In precision feeding, non-contact and pressure-free monitoring of sheep feeding behavior is crucial for health monitoring and optimizing production management. The experimental conditions and real-world environments differ when using acoustic sensors to identify sheep feeding behaviors, leading to discrepancies and consequently posing challenges [...] Read more.
In precision feeding, non-contact and pressure-free monitoring of sheep feeding behavior is crucial for health monitoring and optimizing production management. The experimental conditions and real-world environments differ when using acoustic sensors to identify sheep feeding behaviors, leading to discrepancies and consequently posing challenges for achieving high-accuracy classification in complex production environments. This study enhances the classification performance by integrating the deep spectrogram features and acoustic characteristics associated with feeding behavior. We conducted the task of collecting sound data in actual production environments, considering noise and complex surroundings. The method included evaluating and filtering the optimal acoustic features, utilizing a customized convolutional neural network (SheepVGG-Lite) to extract Short-Time Fourier Transform (STFT) spectrograms and Constant Q Transform (CQT) spectrograms’ deep features, employing cross-spectrogram feature fusion and assessing classification performance through a support vector machine (SVM). Results indicate that the fusion of cross-spectral features significantly improved classification performance, achieving a classification accuracy of 96.47%. These findings highlight the value of integrating acoustic features with spectrogram deep features for accurately recognizing sheep feeding behavior. Full article
Show Figures

Figure 1

Figure 1
<p>Diagrammatic representation of the experimental sheep enclosure. The star indicates the center of the image acquisition area.</p>
Full article ">Figure 2
<p>Arrangement of recording apparatus on sheep. Each sheep is numbered to correspond with the recording.</p>
Full article ">Figure 3
<p>Architecture of the data acquisition and synchronization system.</p>
Full article ">Figure 4
<p>The distribution, spectral analysis, and feature visualization of sheep behavior audio data.</p>
Full article ">Figure 5
<p>The methodology we have put forth.</p>
Full article ">Figure 6
<p>SheepVGG-Lite architecture design.</p>
Full article ">Figure 7
<p>Evaluation outcomes of acoustic features using five-fold cross-validation.</p>
Full article ">Figure 8
<p>Inference time performance comparison of models.</p>
Full article ">Figure 9
<p>Model size performance comparison of models.</p>
Full article ">Figure 10
<p>Visualization of spectrogram feature activation for different sheep behaviors using EigenCAM.</p>
Full article ">Figure A1
<p>Principal component loading value analysis.</p>
Full article ">
33 pages, 13566 KiB  
Article
KOC_Net: Impact of the Synthetic Minority Over-Sampling Technique with Deep Learning Models for Classification of Knee Osteoarthritis Using Kellgren–Lawrence X-Ray Grade
by Syeda Nida Hassan, Mudassir Khalil, Humayun Salahuddin, Rizwan Ali Naqvi, Daesik Jeong and Seung-Won Lee
Mathematics 2024, 12(22), 3534; https://doi.org/10.3390/math12223534 - 12 Nov 2024
Viewed by 407
Abstract
One of the most common diseases afflicting humans is knee osteoarthritis (KOA). KOA occurs when the knee joint cartilage breaks down, and knee bones start rubbing together. The diagnosis of KOA is a lengthy process, and missed diagnosis can have serious consequences. Therefore, [...] Read more.
One of the most common diseases afflicting humans is knee osteoarthritis (KOA). KOA occurs when the knee joint cartilage breaks down, and knee bones start rubbing together. The diagnosis of KOA is a lengthy process, and missed diagnosis can have serious consequences. Therefore, the diagnosis of KOA at an initial stage is crucial which prevents the patients from Severe complications. KOA identification using deep learning (DL) algorithms has gained popularity during the past few years. By applying knee X-ray images and the Kellgren–Lawrence (KL) grading system, the objective of this study was to develop a DL model for detecting KOA. This study proposes a novel model based on CNN called knee osteoarthritis classification network (KOC_Net). The KOC_Net model contains 05 convolutional blocks, and each convolutional block has three components such as Convlotuioanl2D, ReLU, and MaxPooling 2D. The KOC_Net model is evaluated on two publicly available benchmark datasets which consist of X-ray images of KOA based on the KL grading system. Additionally, we applied contrast-limited adaptive histogram equalization (CLAHE) methods to enhance the contrast of the images and utilized SMOTE Tomek to deal with the problem of minority classes. For the diagnosis of KOA, the classification performance of the proposed KOC_Net model is compared with baseline deep networks, namely Dense Net-169, Vgg-19, Xception, and Inception-V3. The proposed KOC_Net was able to classify KOA into 5 distinct groups (including Moderate, Minimal, Severe, Doubtful, and Healthy), with an AUC of 96.71%, accuracy of 96.51%, recall of 91.95%, precision of 90.25%, and F1-Score of 96.70%. Dense Net-169, Vgg-19, Xception, and Inception-V3 have relative accuracy rates of 84.97%, 81.08%, 87.06%, and 83.62%. As demonstrated by the results, the KOC_Net model provides great assistance to orthopedics in making diagnoses of KOA. Full article
(This article belongs to the Special Issue Deep Learning Methods for Biomedical and Medical Images)
Show Figures

Figure 1

Figure 1
<p>Proposed study framework for the identification of KOA using KL grading system.</p>
Full article ">Figure 2
<p>Image Samples of KOA; (<b>a</b>) KOA images with marked abnormalities, while, (<b>b</b>) show raw samples of KOA.</p>
Full article ">Figure 3
<p>Steps of conducting pre-processing to enhance the KOA dataset.</p>
Full article ">Figure 4
<p>After applying CLAHE on KOA images.</p>
Full article ">Figure 5
<p>Synthetic image sample of KOA dataset after using SMOTE Tomek.</p>
Full article ">Figure 6
<p>Proposed KOC_Net architecture for KOA classification.</p>
Full article ">Figure 7
<p>Structure of the flatten layer for KOC_Net.</p>
Full article ">Figure 8
<p>Representation of results in terms of accuracy; (<b>a</b>) Vgg-19, (<b>b</b>) XceptionNet, (<b>c</b>) DenseNet-169, (<b>d</b>) Inception-V3, (<b>e</b>) KOC_Net model without SMOTE Tomek, and (<b>f</b>) KOC_Net model with SMOTE Tomek.</p>
Full article ">Figure 8 Cont.
<p>Representation of results in terms of accuracy; (<b>a</b>) Vgg-19, (<b>b</b>) XceptionNet, (<b>c</b>) DenseNet-169, (<b>d</b>) Inception-V3, (<b>e</b>) KOC_Net model without SMOTE Tomek, and (<b>f</b>) KOC_Net model with SMOTE Tomek.</p>
Full article ">Figure 9
<p>Representation of results in terms of AUC; (<b>a</b>) Vgg-19, (<b>b</b>) XceptionNet, (<b>c</b>) DenseNet-169, (<b>d</b>) Inception-V3, (<b>e</b>) KOC_Net model without SMOTE Tomek, and (<b>f</b>) KOC_Net model with SMOTE Tomek.</p>
Full article ">Figure 10
<p>Representation of results in terms of precision; (<b>a</b>) Vgg-19, (<b>b</b>) XceptionNet, (<b>c</b>) DenseNet-169, (<b>d</b>) Inception-V3, (<b>e</b>) Proposed KOC_Net model without SMOTE Tomek, and (<b>f</b>) Proposed KOC_Net model with SMOTE Tomek.</p>
Full article ">Figure 11
<p>Outcomes of recall; (<b>a</b>) Vgg-19, (<b>b</b>) XceptionNet, (<b>c</b>) DenseNet-169, (<b>d</b>) Inception-V3, (<b>e</b>) Proposed KOC_Net model without SMOTE Tomek, and (<b>f</b>) Proposed KOC_Net model with SMOTE Tomek.</p>
Full article ">Figure 12
<p>Comparison of the F1-Score of the proposed model with baseline models; (a) Vgg-19, (b) XceptionNet, (c) DenseNet-169, (d) Inception-V3, (e) KOC_Net model without SMOTE Tomek, and (f) KOC_Net model with SMOTE Tomek.</p>
Full article ">Figure 13
<p>Loss of proposed and baseline models; (<b>a</b>) Vgg-19, (<b>b</b>) XceptionNet, (<b>c</b>) DenseNet-169, (<b>d</b>) Inception-V3, (<b>e</b>) KOC_Net model without SMOTE Tomek, and (<b>f</b>) KOC_Net model with SMOTE Tomek.</p>
Full article ">Figure 14
<p>Representation of results in terms of ROC; (<b>a</b>) Vgg-19, (<b>b</b>) XceptionNet, (<b>c</b>) DenseNet-169, (<b>d</b>) Inception-V3, (<b>e</b>) Proposed KOC_Net model without SMOTE Tomek, and (<b>f</b>) Proposed KOC_Net model with SMOTE Tomek.</p>
Full article ">Figure 15
<p>Representation of results in terms of AU (ROC); (<b>a</b>) Vgg-19, (<b>b</b>) XceptionNet, (<b>c</b>) DenseNet-169, (<b>d</b>) Inception-V3, (<b>e</b>) Proposed KOC_Net model without SMOTE Tomek, and (<b>f</b>) Proposed KOC_Net model with SMOTE Tomek.</p>
Full article ">Figure 16
<p>Confusion Matrix; (<b>a</b>) Proposed KOC_Net model with SMOTE Tomek, (<b>b</b>) Vgg-19, (<b>c</b>) XceptionNet, (<b>d</b>) KOC_Net model without SMOTE Tomek, (<b>e</b>) Inception-V3, and (<b>f</b>) DenseNet-169.</p>
Full article ">Figure 16 Cont.
<p>Confusion Matrix; (<b>a</b>) Proposed KOC_Net model with SMOTE Tomek, (<b>b</b>) Vgg-19, (<b>c</b>) XceptionNet, (<b>d</b>) KOC_Net model without SMOTE Tomek, (<b>e</b>) Inception-V3, and (<b>f</b>) DenseNet-169.</p>
Full article ">Figure 16 Cont.
<p>Confusion Matrix; (<b>a</b>) Proposed KOC_Net model with SMOTE Tomek, (<b>b</b>) Vgg-19, (<b>c</b>) XceptionNet, (<b>d</b>) KOC_Net model without SMOTE Tomek, (<b>e</b>) Inception-V3, and (<b>f</b>) DenseNet-169.</p>
Full article ">Figure 17
<p>Visualization of the infected region of KOA using GRAD-CAM. The first rows represent the Healthy images and the remaining rows from Grade 1 to Grade 4 highlight the infected region of KOA by using the proposed KOC_net model.</p>
Full article ">
Back to TopTop