Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (12,698)

Search Parameters:
Keywords = CNN

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 1518 KiB  
Article
VAS-3D: A Visual-Based Alerting System for Detecting Drowsy Drivers in Intelligent Transportation Systems
by Hadi El Zein, Hassan Harb, François Delmotte, Oussama Zahwe and Samir Haddad
World Electr. Veh. J. 2024, 15(12), 540; https://doi.org/10.3390/wevj15120540 (registering DOI) - 21 Nov 2024
Abstract
Nowadays, driving accidents are considered one of the most crucial challenges for governments and communities that affect transportation systems and peoples lives. Unfortunately, there are many causes behind the accidents; however, drowsiness is one of the main factors that leads to a significant [...] Read more.
Nowadays, driving accidents are considered one of the most crucial challenges for governments and communities that affect transportation systems and peoples lives. Unfortunately, there are many causes behind the accidents; however, drowsiness is one of the main factors that leads to a significant number of injuries and deaths. In order to reduce its effect, researchers and communities have proposed many techniques for detecting drowsiness situations and alerting the driver before an accident occurs. Mostly, the proposed solutions are visually-based, where a camera is positioned in front of the driver to detect their facial behavior and then determine their situation, e.g., drowsy or awake. However, most of the proposed solutions make a trade-off between detection accuracy and speed. In this paper, we propose a novel Visual-based Alerting System for Detecting Drowsy Drivers (VAS-3D) that ensures an optimal trade-off between the accuracy and speed metrics. Mainly, VAS-3D consists of two stages: detection and classification. In the detection stage, we use pre-trained Haar cascade models to detect the face and eyes of the driver. Once the driver’s eyes are detected, the classification stage uses several pre-trained Convolutional Neural Network (CNN) models to classify the driver’s eyes as either open or closed, and consequently their corresponding situation, either awake or drowsy. Subsequently, we tested and compared the performance of several CNN models, such as InceptionV3, MobileNetV2, NASNetMobile, and ResNet50V2. We demonstrated the performance of VAS-3D through simulations on real drowsiness datasets and experiments on real world scenarios based on real video streaming. The obtained results show that VAS-3D can enhance the accuracy detection of drowsy drivers by at least 7.5% (the best accuracy reached was 95.5%) and the detection speed by up to 57% (average of 0.25 ms per frame) compared to other existing models. Full article
Show Figures

Figure 1

Figure 1
<p>VAS-3D architecture.</p>
Full article ">Figure 2
<p>MRL Eye Dataset screenshot.</p>
Full article ">Figure 3
<p>InceptionV3 architecture adapted in our system.</p>
Full article ">Figure 4
<p>MobileNetV2 architecture adapted in our system.</p>
Full article ">Figure 5
<p>NASNetMobile architecture adapted in our system.</p>
Full article ">Figure 6
<p>ResNet50V2 architecture adapted in VAS-3D.</p>
Full article ">Figure 7
<p>Samples of visual driver behavior detection using HCC.</p>
Full article ">Figure 8
<p>Performance evaluation regarding various scenarios: VAS-3D vs. state-of-the-art. Model1 and Model2 refer to those proposed in [<a href="#B45-wevj-15-00540" class="html-bibr">45</a>] and [<a href="#B46-wevj-15-00540" class="html-bibr">46</a>] respectively.</p>
Full article ">
24 pages, 2160 KiB  
Article
Combination of a Rabbit Optimization Algorithm and a Deep-Learning-Based Convolutional Neural Network–Long Short-Term Memory–Attention Model for Arc Sag Prediction of Transmission Lines
by Xiu Ji, Chengxiang Lu, Beimin Xie, Haiyang Guo and Boyang Zheng
Electronics 2024, 13(23), 4593; https://doi.org/10.3390/electronics13234593 (registering DOI) - 21 Nov 2024
Abstract
Arc droop presents significant challenges in power system management due to its inherent complexity and dynamic nature. To address these challenges in predicting arc sag for transmission lines, this paper proposes an innovative time–series prediction model, AROA-CNN-LSTM-Attention(AROA-CLA). The model aims to enhance arc [...] Read more.
Arc droop presents significant challenges in power system management due to its inherent complexity and dynamic nature. To address these challenges in predicting arc sag for transmission lines, this paper proposes an innovative time–series prediction model, AROA-CNN-LSTM-Attention(AROA-CLA). The model aims to enhance arc sag prediction by integrating a convolutional neural network (CNN), a long short-term memory network (LSTM), and an attention mechanism, while also utilizing, for the first time, the adaptive rabbit optimization algorithm (AROA) for CLA parameter tuning. This combination improves both the prediction performance and the generalization capability of the model. By effectively leveraging historical data and exhibiting superior time–series processing capabilities, the AROA-CLA model demonstrates excellent prediction accuracy and stability across different time scales. Experimental results show that, compared to traditional and other modern optimization models, AROA-CLA achieves significant improvements in RMSE, MAE, MedAE, and R2 metrics, particularly in reducing errors, accelerating convergence, and enhancing robustness. These findings confirm the effectiveness and applicability of the AROA-CLA model in arc droop prediction, offering novel approaches for transmission line monitoring and intelligent power system management. Full article
24 pages, 4759 KiB  
Article
Proteomic Analysis of Biomarkers Predicting Treatment Response in Patients with Head and Neck Cancers
by Emeshaw Damtew Zebene, Rita Lombardi, Biagio Pucci, Hagos Tesfay Medhin, Edom Seife, Elena Di Gennaro, Alfredo Budillon and Gurja Belay Woldemichael
Int. J. Mol. Sci. 2024, 25(23), 12513; https://doi.org/10.3390/ijms252312513 (registering DOI) - 21 Nov 2024
Abstract
Head and neck cancers (HNCs) are the sixth most commonly diagnosed cancer and the eighth leading cause of cancer-related mortality worldwide, with squamous cell carcinoma being the most prevalent type. The global incidence of HNCs is steadily increasing, projected to rise by approximately [...] Read more.
Head and neck cancers (HNCs) are the sixth most commonly diagnosed cancer and the eighth leading cause of cancer-related mortality worldwide, with squamous cell carcinoma being the most prevalent type. The global incidence of HNCs is steadily increasing, projected to rise by approximately 30% per year by 2030, a trend observed in both developed and undeveloped countries. This study involved serum proteomic profiling to identify predictive clinical biomarkers in cancer patients undergoing chemoradiotherapy (CRT). Fifteen HNC patients at Tikur Anbessa Specialized Hospital, Radiotherapy (RT) center in Addis Ababa were enrolled. Serum samples were collected before and after RT, and patients were classified as responders (R) or non-responders (NR). Protein concentrations in the serum were determined using the Bradford assay, followed by nano-HPLC–MS/MS for protein profiling. Progenesis QI for proteomics identified 55 differentially expressed proteins (DEPs) between R and NR, with a significance of p < 0.05 and a fold-change (FC) ≥ 1.5. The top five-up-regulated proteins included MAD1L1, PSMC2, TRIM29, C5, and SERPING1, while the top five-down-regulated proteins were RYR1, HEY2, HIF1A, TF, and CNN3. Notably, about 16.4% of the DEPs were involved in cellular responses to DNA damage from cancer treatments, encompassing proteins related to deoxyribonucleic acid (DNA) damage sensing, checkpoint activation, DNA repair, and apoptosis/cell cycle regulation. The analysis of the relative abundance of ten proteins with high confidence scores identified three DEPs: ADIPOQ, HEY2, and FUT10 as potential predictive biomarkers for treatment response. This study highlighted the identification of three potential predictive biomarkers—ADIPOQ, HEY2, and FUT10—through serum proteomic profiling in HNC patients undergoing RT, emphasizing their significance in predicting treatment response. Full article
(This article belongs to the Special Issue DNA Damage Response from Molecular Mechanisms to Cancer Therapy)
Show Figures

Figure 1

Figure 1
<p>Hierarchical cluster of DEPs in both Responders and Non-Responders with a fold-change ≥ 1.5 and <span class="html-italic">p</span>-value &lt;0.05. The expression level is indicated by the intensity of the color: red, high expression; blue, low expression.</p>
Full article ">Figure 2
<p>Principal component analysis highlighting the close clustering of biological replicates between Responders and Non-Responders (baseline vs end of radiotherapy). Baseline R (blue), baseline NR (purple), End RT_R (orange), End RT_NR (green).</p>
Full article ">Figure 3
<p>Gene ontology classification analysis of DEPs. The number of DEPs in the three ontology classifications: Molecular function (MF), Cellular component (CC), Biological process (BP). The horizontal axis represents the number of overlapping proteins, while the vertical axis represents the ontology classification name.</p>
Full article ">Figure 4
<p>Dot plot of the KEGG pathway enrichment analysis. The horizontal axis represents the DEP ratio, while the vertical axis represents the enrichment pathway name. The color scale indicates different thresholds of the <span class="html-italic">p</span>-value, and the size of the dot indicates the number of DEPs corresponding to each pathway.</p>
Full article ">Figure 5
<p>Human papilloma virus infection pathway with differentially expressed proteins (DEPs). The down-regulated DEPs are highlighted in a green color.</p>
Full article ">Figure 6
<p>HIF-1 signaling pathway with differentially expressed proteins (DEPs). The down-regulated DEPs are highlighted in a green color.</p>
Full article ">Figure 7
<p>Apelin signaling pathway with differentially expressed proteins (DEPs). The up-regulated DEPs are highlighted in a red color and the down-regulated DEPS are highlighted in a green color.</p>
Full article ">Figure 8
<p>WNT signaling pathway with differentially expressed proteins (DEPs). The up-regulated DEPs are highlighted in red and the down-regulated DEPs are highlighted in green.</p>
Full article ">Figure 9
<p>Protein–protein interaction (PPI) performed by Ingenuity pathway analysis (IPA). Four core hubs were included in the PPI network; the lead protein of each hub is shown by an orange color. The DEPs were highlighted as grey nodes.</p>
Full article ">Figure 10
<p>ROC curve of the DEPs. AUCs of the predictive proteins were 0.823, 0.870, and 0.823 respectively.</p>
Full article ">
19 pages, 5999 KiB  
Article
Automated Pipeline for Robust Cat Activity Detection Based on Deep Learning and Wearable Sensor Data
by Md Ariful Islam Mozumder, Tagne Poupi Theodore Armand, Rashadul Islam Sumon, Shah Muhammad Imtiyaj Uddin and Hee-Cheol Kim
Sensors 2024, 24(23), 7436; https://doi.org/10.3390/s24237436 (registering DOI) - 21 Nov 2024
Abstract
The health, safety, and well-being of household pets such as cats has become a challenging task in previous years. To estimate a cat’s behavior, objective observations of both the frequency and variability of specific behavior traits are required, which might be difficult to [...] Read more.
The health, safety, and well-being of household pets such as cats has become a challenging task in previous years. To estimate a cat’s behavior, objective observations of both the frequency and variability of specific behavior traits are required, which might be difficult to come by in a cat’s ordinary life. There is very little research on cat activity and cat disease analysis based on real-time data. Although previous studies have made progress, several key questions still need addressing: What types of data are best suited for accurately detecting activity patterns? Where should sensors be strategically placed to ensure precise data collection, and how can the system be effectively automated for seamless operation? This study addresses these questions by pointing out whether the cat should be equipped with a sensor, and how the activity detection system can be automated. Magnetic, motion, vision, audio, and location sensors are among the sensors used in the machine learning experiment. In this study, we collect data using three types of differentiable and realistic wearable sensors, namely, an accelerometer, a gyroscope, and a magnetometer. Therefore, this study aims to employ cat activity detection techniques to combine data from acceleration, motion, and magnetic sensors, such as accelerometers, gyroscopes, and magnetometers, respectively, to recognize routine cat activity. Data collecting, data processing, data fusion, and artificial intelligence approaches are all part of the system established in this study. We focus on One-Dimensional Convolutional Neural Networks (1D-CNNs) in our research, to recognize cat activity modeling for detection and classification. Such 1D-CNNs have recently emerged as a cutting-edge approach for signal processing-based systems such as sensor-based pet and human health monitoring systems, anomaly identification in manufacturing, and in other areas. Our study culminates in the development of an automated system for robust pet (cat) activity analysis using artificial intelligence techniques, featuring a 1D-CNN-based approach. In this experimental research, the 1D-CNN approach is evaluated using training and validation sets. The approach achieved a satisfactory accuracy of 98.9% while detecting the activity useful for cat well-being. Full article
(This article belongs to the Special Issue Advances in Sensing-Based Animal Biomechanics)
Show Figures

Figure 1

Figure 1
<p>Housing, monitoring, and husbandry environment of the cats.</p>
Full article ">Figure 2
<p>Wearable sensors with internal features.</p>
Full article ">Figure 3
<p>Data collection procedure. (<b>A</b>) Server room for real-time monitoring and storing data, (<b>B</b>) sensor device, (<b>C</b>) sensor device on the cat’s neck, (<b>D</b>) cat living space, including surveillance cameras, (<b>E</b>) transferring sensor data to the server.</p>
Full article ">Figure 4
<p>Data distribution of activity detection.</p>
Full article ">Figure 5
<p>Samples of bio-signals from the wearable devices on the cats.</p>
Full article ">Figure 6
<p>The deep learning model architecture of our experimental research work.</p>
Full article ">Figure 7
<p>Classification of the five activities.</p>
Full article ">Figure 8
<p>The complete process of the automated pipeline.</p>
Full article ">Figure 9
<p>Confusion matrix without normalization using the test dataset.</p>
Full article ">Figure 10
<p>Confusion matrix with normalization using the test dataset.</p>
Full article ">Figure 11
<p>Accuracy graph for the validation and training.</p>
Full article ">Figure 12
<p>Loss graph for the validation and training.</p>
Full article ">Figure 13
<p>Receiver operating characteristic (ROC) curves and AUCs for each class.</p>
Full article ">
20 pages, 13179 KiB  
Article
A Study on the Monitoring of Floating Marine Macro-Litter Using a Multi-Spectral Sensor and Classification Based on Deep Learning
by Youchul Jeong, Jisun Shin, Jong-Seok Lee, Ji-Yeon Baek, Daniel Schläpfer, Sin-Young Kim, Jin-Yong Jeong and Young-Heon Jo
Remote Sens. 2024, 16(23), 4347; https://doi.org/10.3390/rs16234347 (registering DOI) - 21 Nov 2024
Abstract
Increasing global plastic usage has raised critical concerns regarding marine pollution. This study addresses the pressing issue of floating marine macro-litter (FMML) by developing a novel monitoring system using a multi-spectral sensor and drones along the southern coast of South Korea. Subsequently, a [...] Read more.
Increasing global plastic usage has raised critical concerns regarding marine pollution. This study addresses the pressing issue of floating marine macro-litter (FMML) by developing a novel monitoring system using a multi-spectral sensor and drones along the southern coast of South Korea. Subsequently, a convolutional neural network (CNN) model was utilized to classify four distinct marine litter materials: film, fiber, fragment, and foam. Automatic atmospheric correction with the drone data atmospheric correction (DROACOR) method, which is specifically designed for currently available drone-based sensors, ensured consistent reflectance across altitudes in the FMML dataset. The CNN models exhibited promising performance, with precision, recall, and F1 score values of 0.9, 0.88, and 0.89, respectively. Furthermore, gradient-weighted class activation mapping (Grad-CAM), an object recognition technique, allowed us to interpret the classification performance. Overall, this study will shed light on successful FMML identification using multi-spectral observations for broader applications in diverse marine environments. Full article
(This article belongs to the Special Issue Recent Progress in UAV-AI Remote Sensing II)
Show Figures

Figure 1

Figure 1
<p>The overall workflow shows the processes that led to the classification of FMML using drone-acquired data and deep learning models. We performed three steps: (1) FMML exploration; (2) data processing for the deep learning models; and (3) deep learning to process FMML classification and visualization.</p>
Full article ">Figure 2
<p>The study location on Gadeok Island in South Korea and the data acquisition location of the drone surveys in the study area in drone-based imagery (red rectangle). Maps of the study area and a Pix4Dmapper image were used to illustrate the data acquisition.</p>
Full article ">Figure 3
<p>FMML dataset of images captured by the drone in the study area.</p>
Full article ">Figure 4
<p>CNN architecture for the classification of FMML. The training, validation, and test sets comprised FMML datasets as input. The input image size was 128 × 128 × 5. The output was labeled as film, fiber, fragment, and foam for the FMML. This network consisted of input, feature learning, classification, and output.</p>
Full article ">Figure 5
<p>Reflectance analysis of flight altitude through atmospheric correction. (<b>a</b>) A multi-spectral image was obtained on 29 March 2023 (true color RGB; R: 668 nm; G: 560 nm; B: 475 nm; a 51 m flight altitude). Images for atmospheric correction were acquired at altitudes of 23, 51, 70, 101, 127, 146, and 170 m. (<b>b</b>) The image values for each altitude of the orange film buoy image before atmospheric correction were compared. (<b>c</b>) The reflectance for each altitude of the orange film buoy image using a DROACOR atmospheric correction processor were compared.</p>
Full article ">Figure 6
<p>Spectra of all FMML lists in the dataset from the DROACOR-calculated reflectance.</p>
Full article ">Figure 7
<p>A confusion matrix of the CNN-3 model (<span class="html-italic">x</span>-axis: recall; <span class="html-italic">y</span>-axis: precision). The green box indicates correct classification by the model, and the red box indicates incorrect classification.</p>
Full article ">Figure 8
<p>Visualization of FMML using Grad-CAM on CNN-3 model. (<b>a</b>–<b>d</b>) Confident detections of FMML dataset labels. (<b>e</b>–<b>h</b>) Unconfident detections of FMML dataset labels.</p>
Full article ">Figure 9
<p>The well-classified and misclassified results of each category in the CNN-3 Model. All the images are Micasense multi-spectral images of band five. (<b>a</b>–<b>d</b>) Classified as fiber. (<b>e</b>–<b>h</b>) Classified as film. (<b>i</b>–<b>l</b>) Classified as foam. (<b>m</b>–<b>p</b>) Classified as fragment. Green and red circles indicate well-classified and misclassified results, respectively.</p>
Full article ">
21 pages, 10351 KiB  
Article
TSViT: A Time Series Vision Transformer for Fault Diagnosis of Rotating Machinery
by Shouhua Zhang, Jiehan Zhou, Xue Ma, Susanna Pirttikangas and Chunsheng Yang
Appl. Sci. 2024, 14(23), 10781; https://doi.org/10.3390/app142310781 - 21 Nov 2024
Viewed by 65
Abstract
Efficient and accurate fault diagnosis of rotating machinery is extremely important. Fault diagnosis methods using vibration signals based on convolutional neural networks (CNNs) have become increasingly mature. They often struggle with capturing the temporal dynamics of vibration signals. To overcome this, the application [...] Read more.
Efficient and accurate fault diagnosis of rotating machinery is extremely important. Fault diagnosis methods using vibration signals based on convolutional neural networks (CNNs) have become increasingly mature. They often struggle with capturing the temporal dynamics of vibration signals. To overcome this, the application of Transformer-based Vision Transformer (ViT) methods to fault diagnosis is gaining attraction. Nonetheless, these methods typically require extensive preprocessing, which increases computational complexity, potentially reducing the efficiency of the diagnosis process. Addressing this gap, this paper presents the Time Series Vision Transformer (TSViT), tailored for effective fault diagnosis. The TSViT incorporates a convolutional layer to extract local features from vibration signals alongside a transformer encoder to discern long-term temporal patterns. A thorough experimental comparison of three diverse datasets demonstrates the TSViT’s effectiveness and adaptability. Moreover, the paper delves into the influence of hyperparameter tuning on the model’s performance, computational demand, and parameter count. Remarkably, the TSViT achieves an unprecedented 100% average accuracy on two of the test sets and 99.99% on the other, showcasing its exceptional fault diagnosis capabilities for rotating machinery. The implementation of this model will bring significant economic benefits. Full article
(This article belongs to the Special Issue Signal Acquisition and Processing for Measurement and Testing)
Show Figures

Figure 1

Figure 1
<p>The architecture of the TSViT.</p>
Full article ">Figure 2
<p>The experimental bench of each dataset (<b>a</b>) PBR dataset (<b>b</b>) CWRU dataset (<b>c</b>) XJTU dataset.</p>
Full article ">Figure 3
<p>Time domain waveforms of vibration signal samples in the PBR dataset (<b>a</b>) NC (<b>b</b>) PL (<b>c</b>) BBF (<b>d</b>) RU.</p>
Full article ">Figure 4
<p>Time domain waveforms of vibration signal samples in the CWRU dataset (<b>a</b>) NC (<b>b</b>) F1 (<b>c</b>) F2 (<b>d</b>) F3 (<b>e</b>) F4 (<b>f</b>) F5 (<b>g</b>) F6 (<b>h</b>) F7 (<b>i</b>) F8 (<b>j</b>) F9.</p>
Full article ">Figure 4 Cont.
<p>Time domain waveforms of vibration signal samples in the CWRU dataset (<b>a</b>) NC (<b>b</b>) F1 (<b>c</b>) F2 (<b>d</b>) F3 (<b>e</b>) F4 (<b>f</b>) F5 (<b>g</b>) F6 (<b>h</b>) F7 (<b>i</b>) F8 (<b>j</b>) F9.</p>
Full article ">Figure 5
<p>Time domain waveforms of vibration signal samples in the XJTU dataset (<b>a</b>) Cage (<b>b</b>) Inner race (<b>c</b>) Outer race (<b>d</b>) Composite.</p>
Full article ">Figure 6
<p>The training flow of the TSViT.</p>
Full article ">Figure 7
<p>The trend of loss changes on the PBR dataset (<b>a</b>) downward trend on the training dataset (<b>b</b>) downward trend on the test dataset (<b>c</b>) downward trend in average loss over 10 trials.</p>
Full article ">Figure 8
<p>The trend of accuracy changes in the PBR dataset (<b>a</b>) upward trend on the training dataset (<b>b</b>) upward trend on the test dataset (<b>c</b>) upward trend in average accuracy over 10 trials.</p>
Full article ">Figure 9
<p>The trend of loss changes in the CWRU dataset (<b>a</b>) downward trend on the training dataset (<b>b</b>) downward trend on the test dataset (<b>c</b>) downward trend in average loss over 10 trials.</p>
Full article ">Figure 10
<p>The trend of accuracy changes in the CWRU dataset (<b>a</b>) upward trend on the training dataset (<b>b</b>) upward trend on the test dataset (<b>c</b>) upward trend in average accuracy over 10 trials.</p>
Full article ">Figure 11
<p>The trend of loss changes on XJTU dataset (<b>a</b>) downward trend on the training dataset (<b>b</b>) downward trend on the test dataset (<b>c</b>) downward trend in average loss over 10 trials.</p>
Full article ">Figure 12
<p>The trend of accuracy changes on XJTU dataset (<b>a</b>) upward trend on the training dataset (<b>b</b>) upward trend on the test dataset (<b>c</b>) upward trend in average accuracy over 10 trials.</p>
Full article ">Figure 13
<p>Confusion matrix of the minimum optimal model of the CWRU test set.</p>
Full article ">Figure 14
<p>Accuracies on three noise-added test sets.</p>
Full article ">Figure 15
<p>Comparison of the influence of different hyperparameters on accuracy based on three datasets (<b>a</b>) patch size (<b>b</b>) dimension of time series patch embeddings (<b>c</b>) number of heads in MSA (<b>d</b>) number of blocks (<b>e</b>) dimension of linear transformation in MLP.</p>
Full article ">Figure 15 Cont.
<p>Comparison of the influence of different hyperparameters on accuracy based on three datasets (<b>a</b>) patch size (<b>b</b>) dimension of time series patch embeddings (<b>c</b>) number of heads in MSA (<b>d</b>) number of blocks (<b>e</b>) dimension of linear transformation in MLP.</p>
Full article ">Figure 16
<p>Feature visualization in different TSViT layers through t-SNE on the PBR test set (<b>a</b>) raw signals (<b>b</b>) cnn (<b>c</b>) the first block (<b>d</b>) the second block (<b>e</b>) the third block (<b>f</b>) the fourth block (<b>g</b>) the fifth block (<b>h</b>) the sixth block (<b>i</b>) the seventh block (<b>j</b>) the eighth block (<b>k</b>) classification layer.</p>
Full article ">Figure 17
<p>Feature visualization in different TSViT layers through t-SNE on the CWRU test set (<b>a</b>) raw signals (<b>b</b>) cnn (<b>c</b>) the first block (<b>d</b>) the second block (<b>e</b>) the third block (<b>f</b>) the fourth block (<b>g</b>) the fifth block (<b>h</b>) the sixth block (<b>i</b>) the seventh block (<b>j</b>) the eighth block (<b>k</b>) classification layer.</p>
Full article ">Figure 18
<p>Feature visualization in different TSViT layers through t-SNE on the XJTU test set (<b>a</b>) raw signals (<b>b</b>) cnn (<b>c</b>) the first block (<b>d</b>) the second block (<b>e</b>) the third block (<b>f</b>) the fourth block (<b>g</b>) the fifth block (<b>h</b>) the sixth block (<b>i</b>) the seventh block (<b>j</b>) the eighth block (<b>k</b>) classification layer.</p>
Full article ">
16 pages, 5582 KiB  
Article
Evaluating Brain Tumor Detection with Deep Learning Convolutional Neural Networks Across Multiple MRI Modalities
by Ioannis Stathopoulos, Luigi Serio, Efstratios Karavasilis, Maria Anthi Kouri, Georgios Velonakis, Nikolaos Kelekis and Efstathios Efstathopoulos
J. Imaging 2024, 10(12), 296; https://doi.org/10.3390/jimaging10120296 - 21 Nov 2024
Viewed by 94
Abstract
Central Nervous System (CNS) tumors represent a significant public health concern due to their high morbidity and mortality rates. Magnetic Resonance Imaging (MRI) has emerged as a critical non-invasive modality for the detection, diagnosis, and management of brain tumors, offering high-resolution visualization of [...] Read more.
Central Nervous System (CNS) tumors represent a significant public health concern due to their high morbidity and mortality rates. Magnetic Resonance Imaging (MRI) has emerged as a critical non-invasive modality for the detection, diagnosis, and management of brain tumors, offering high-resolution visualization of anatomical structures. Recent advancements in deep learning, particularly convolutional neural networks (CNNs), have shown potential in augmenting MRI-based diagnostic accuracy for brain tumor detection. In this study, we evaluate the diagnostic performance of six fundamental MRI sequences in detecting tumor-involved brain slices using four distinct CNN architectures enhanced with transfer learning techniques. Our dataset comprises 1646 MRI slices from the examinations of 62 patients, encompassing both tumor-bearing and normal findings. With our approach, we achieved a classification accuracy of 98.6%, underscoring the high potential of CNN-based models in this context. Additionally, we assessed the performance of each MRI sequence across the different CNN models, identifying optimal combinations of MRI modalities and neural networks to meet radiologists’ screening requirements effectively. This study offers critical insights into the integration of deep learning with MRI for brain tumor detection, with implications for improving diagnostic workflows in clinical settings. Full article
Show Figures

Figure 1

Figure 1
<p>Six different MRI sequences of a normal brain examination. From left to right and top to bottom: T1, T2, FLAIR, T1+C, Diffusion, apparent diffusion coefficient (ADC) map.</p>
Full article ">Figure 2
<p>Six different MRI sequences of a verified Brain Tumor examination. From left to right and top to bottom: T1, T2, FLAIR, T1+C, Diffusion, and ADC.</p>
Full article ">Figure 3
<p>Image representation of the preprocessing steps.</p>
Full article ">Figure 4
<p>One normal and two tumor examinations are shown for all six MRI sequences. In all images, the original image is displayed on the left, and the overlap with the heatmap produced from the last convolutional layer of the VGG16 model is displayed on the right. In the titles, N represents the Normal class, and T represents the Tumor class, both followed by the prediction probability for the respective class. Misclassified cases are highlighted in red.</p>
Full article ">Scheme 1
<p>ROCs for FLAIR sequence.</p>
Full article ">Scheme 2
<p>ROCs for T1+C sequence.</p>
Full article ">Scheme 3
<p>ROCs for ADC sequence.</p>
Full article ">Scheme 4
<p>ROCs for T1 sequence.</p>
Full article ">Scheme 5
<p>ROCs for Diffusion sequence.</p>
Full article ">Scheme 6
<p>ROCs for T2 sequence.</p>
Full article ">Scheme 7
<p>(<b>Left</b>): The evaluation metrics results of the experiment are in the whole dataset. (<b>Right</b>): the corresponding ROC curve.</p>
Full article ">
19 pages, 5046 KiB  
Article
A Convolutional Neural Network for the Removal of Simultaneous Ocular and Myogenic Artifacts from EEG Signals
by Maryam Azhar, Tamoor Shafique and Anas Amjad
Electronics 2024, 13(22), 4576; https://doi.org/10.3390/electronics13224576 - 20 Nov 2024
Viewed by 338
Abstract
Electroencephalography (EEG) is a non-invasive technique widely used in neuroscience to diagnose neural disorders and analyse brain activity. However, ocular and myogenic artifacts from eye movements and facial muscle activity often contaminate EEG signals, compromising signal analysis accuracy. While deep learning models are [...] Read more.
Electroencephalography (EEG) is a non-invasive technique widely used in neuroscience to diagnose neural disorders and analyse brain activity. However, ocular and myogenic artifacts from eye movements and facial muscle activity often contaminate EEG signals, compromising signal analysis accuracy. While deep learning models are a popular choice for denoising EEG signals, most focus on removing either ocular or myogenic artifacts independently. This paper introduces a novel EEG denoising model capable of handling the simultaneous occurrence of both artifacts. The model uses convolutional layers to extract spatial features and a fully connected layer to reconstruct clean signals from learned features. The model integrates the Adam optimiser, average pooling, and ReLU activation to effectively capture and restore clean EEG signals. It demonstrates superior performance, achieving low training and validation losses with a significantly reduced RRMSE value of 0.35 in both the temporal and spectral domains. A high cross-correlation coefficient of 0.94 with ground-truth EEG signals confirms the model’s fidelity. Compared to the existing architectures and models (FPN, UNet, MCGUNet, LinkNet, MultiResUNet3+, Simple CNN, Complex CNN) across a range of signal-to-noise ratio values, the model shows superior performance for artifact removal. It also mitigates overfitting, underscoring its robustness in artifact suppression. Full article
Show Figures

Figure 1

Figure 1
<p>Artifacts in EEG: (<b>a</b>) eye movement, (<b>b</b>) eye blinks, and (<b>c</b>) muscle tension [<a href="#B18-electronics-13-04576" class="html-bibr">18</a>].</p>
Full article ">Figure 2
<p>Framework for simultaneous EOG-EMG artifact removal.</p>
Full article ">Figure 3
<p>Noisy EEG signal synthesis.</p>
Full article ">Figure 4
<p>Example segment of simultaneous EOG- and EMG-corrupted EEG signal and ground-truth EEG signal.</p>
Full article ">Figure 5
<p>Network structure for the denoising model.</p>
Full article ">Figure 6
<p>EEG signal dimensions in each layer.</p>
Full article ">Figure 7
<p>Training and validation loss curves for the proposed model.</p>
Full article ">Figure 8
<p>Training and validation loss curves for Complex CNN and Simple CNN.</p>
Full article ">Figure 9
<p>Power ratios for various frequency bands for denoised, EOG-EMG-contaminated, and clean EEG signals.</p>
Full article ">Figure 10
<p>Temporal representation of denoised, EOG-EMG-contaminated, and clean EEG signals.</p>
Full article ">Figure 11
<p>Spectral representation of denoised, EOG-EMG-contaminated, and clean EEG signals.</p>
Full article ">Figure 12
<p>A comparison of estimated performance metrics (<math display="inline"><semantics> <mrow> <mi>C</mi> <mi>C</mi> <mo>,</mo> <mo> </mo> <mi>R</mi> <mi>R</mi> <mi>M</mi> <mi>S</mi> <mi>E</mi> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>R</mi> <mi>R</mi> <mi>M</mi> <mi>S</mi> <mi>E</mi> </mrow> </semantics></math> in time and frequency domains) across different <math display="inline"><semantics> <mrow> <mi>S</mi> <mi>N</mi> <mi>R</mi> </mrow> </semantics></math> values.</p>
Full article ">Figure 13
<p>Comparison of performance between the proposed model and the existing models.</p>
Full article ">
19 pages, 4058 KiB  
Article
Enhanced Fault Detection in Photovoltaic Panels Using CNN-Based Classification with PyQt5 Implementation
by Younes Ledmaoui, Adila El Maghraoui, Mohamed El Aroussi and Rachid Saadane
Sensors 2024, 24(22), 7407; https://doi.org/10.3390/s24227407 (registering DOI) - 20 Nov 2024
Viewed by 233
Abstract
Solar photovoltaic systems have increasingly become essential for harvesting renewable energy. However, as these systems grow in prevalence, the issue of the end of life of modules is also increasing. Regular maintenance and inspection are vital to extend the lifespan of these systems, [...] Read more.
Solar photovoltaic systems have increasingly become essential for harvesting renewable energy. However, as these systems grow in prevalence, the issue of the end of life of modules is also increasing. Regular maintenance and inspection are vital to extend the lifespan of these systems, minimize energy losses, and protect the environment. This paper presents an innovative explainable AI model for detecting anomalies in solar photovoltaic panels using an enhanced convolutional neural network (CNN) and the VGG16 architecture. The model effectively identifies physical and electrical changes, such as dust and bird droppings, and is implemented using the PyQt5 Python tool to create a user-friendly interface that facilitates decision-making for users. Key processes included dataset balancing through oversampling and data augmentation to expand the dataset. The model achieved impressive performance metrics: 91.46% accuracy, 98.29% specificity, and an F1 score of 91.67%. Overall, it enhances power generation efficiency and prolongs the lifespan of photovoltaic systems, while minimizing environmental risks. Full article
(This article belongs to the Special Issue Sensor Enabled Smart Energy Solutions)
Show Figures

Figure 1

Figure 1
<p>Evolution of installed solar capacity from 2004 to 2023 [<a href="#B4-sensors-24-07407" class="html-bibr">4</a>].</p>
Full article ">Figure 2
<p>PV cell (<b>a</b>), electrical schematic diagram (<b>b</b>).</p>
Full article ">Figure 3
<p>PV system fault classification.</p>
Full article ">Figure 4
<p>Classes of solar panels.</p>
Full article ">Figure 5
<p>Proposed solar panel anomaly detection and classification model.</p>
Full article ">Figure 6
<p>Methodology for the proposed architecture.</p>
Full article ">Figure 7
<p>Architecture of VGG16.</p>
Full article ">Figure 8
<p>Architecture of transfer learning.</p>
Full article ">Figure 9
<p>Correlation matrix of the solar panel dataset for the proposed model.</p>
Full article ">Figure 10
<p>Results with PyQt5 implementation.</p>
Full article ">Figure 11
<p>SPAD upload image (<b>a</b>), image preview (<b>b</b>), and prediction result implementation (<b>c</b>).</p>
Full article ">
13 pages, 860 KiB  
Article
Multi-Scale 3D Cephalometric Landmark Detection Based on Direct Regression with 3D CNN Architectures
by Chanho Song, Yoosoo Jeong, Hyungkyu Huh, Jee-Woong Park, Jun-Young Paeng, Jaemyung Ahn, Jaebum Son and Euisung Jung
Diagnostics 2024, 14(22), 2605; https://doi.org/10.3390/diagnostics14222605 - 20 Nov 2024
Viewed by 274
Abstract
Background: Cephalometric analysis is important in diagnosing and planning treatments for patients, traditionally relying on 2D cephalometric radiographs. With advancements in 3D imaging, automated landmark detection using deep learning has gained prominence. However, 3D imaging introduces challenges due to increased network complexity and [...] Read more.
Background: Cephalometric analysis is important in diagnosing and planning treatments for patients, traditionally relying on 2D cephalometric radiographs. With advancements in 3D imaging, automated landmark detection using deep learning has gained prominence. However, 3D imaging introduces challenges due to increased network complexity and computational demands. This study proposes a multi-scale 3D CNN-based approach utilizing direct regression to improve the accuracy of maxillofacial landmark detection. Methods: The method employs a coarse-to-fine framework, first identifying landmarks in a global context and then refining their positions using localized 3D patches. A clinical dataset of 150 CT scans from maxillofacial surgery patients, annotated with 30 anatomical landmarks, was used for training and evaluation. Results: The proposed method achieved an average RMSE of 2.238 mm, outperforming conventional 3D CNN architectures. The approach demonstrated consistent detection without failure cases. Conclusions: Our multi-scale-based 3D CNN framework provides a reliable method for automated landmark detection in maxillofacial CT images, showing potential for other clinical applications. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

Figure 1
<p>Diagram of multi-scale cephalometric landmark detection architecture. Stage 1 involves coarse detection to identify coordinates and classes from the entire input volume, followed by generating local volumes through 3D region of interest (ROI) processing. Stage 2 focuses on fine localization using these local volumes and classes.</p>
Full article ">Figure 2
<p>Steps for data preprocessing to convert volumetric CT data into normalized voxel data.</p>
Full article ">Figure 3
<p>Convolutional Neural Network (CNN) architectures including ResNet, DenseNet, Inception, and InceptionResNet.</p>
Full article ">Figure 4
<p>DenseNet169-based multi-output learning model for coarse detection.</p>
Full article ">Figure 5
<p>DenseNet169-based multi-input learning model for fine localization.</p>
Full article ">
29 pages, 8399 KiB  
Article
Automatic Modulation Recognition Based on Multimodal Information Processing: A New Approach and Application
by Wenna Zhang, Kailiang Xue, Aiqin Yao and Yunqiang Sun
Electronics 2024, 13(22), 4568; https://doi.org/10.3390/electronics13224568 - 20 Nov 2024
Viewed by 274
Abstract
Automatic modulation recognition (AMR) has wide applications in the fields of wireless communications, radar systems, and intelligent sensor networks. The existing deep learning-based modulation recognition models often focus on temporal features while overlooking the interrelations and spatio-temporal relationships among different types of signals. [...] Read more.
Automatic modulation recognition (AMR) has wide applications in the fields of wireless communications, radar systems, and intelligent sensor networks. The existing deep learning-based modulation recognition models often focus on temporal features while overlooking the interrelations and spatio-temporal relationships among different types of signals. To overcome these limitations, a hybrid neural network based on a multimodal parallel structure, called the multimodal parallel hybrid neural network (MPHNN), is proposed to improve the recognition accuracy. The algorithm first preprocesses the data by parallelly processing the multimodal forms of the modulated signals before inputting them into the network. Subsequently, by combining Convolutional Neural Networks (CNN) and Bidirectional Gated Recurrent Unit (Bi-GRU) models, the CNN is used to extract spatial features of the received signals, while the Bi-GRU transmits previous state information of the time series to the current state to capture temporal features. Finally, the Convolutional Block Attention Module (CBAM) and Multi-Head Self-Attention (MHSA) are introduced as two attention mechanisms to handle the temporal and spatial correlations of the signals through an attention fusion mechanism, achieving the calibration of the signal feature maps. The effectiveness of this method is validated using various datasets, with the experimental results demonstrating that the proposed approach can fully utilize the information of multimodal signals. The experimental results show that the recognition accuracy of MPHNN on multiple datasets reaches 93.1%, and it has lower computational complexity and fewer parameters than other models. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Visualization of instantaneous amplitude, instantaneous phase, instantaneous frequency, and IQ time-domain plots for 11 modulation modes.</p>
Full article ">Figure 2
<p>Overall architecture of the MPHNN.</p>
Full article ">Figure 3
<p>Structure of CBAM.</p>
Full article ">Figure 4
<p>Working mechanism of the CBAM.</p>
Full article ">Figure 5
<p>Structure of the Multi-Head Self-Attention (MHSA) module.</p>
Full article ">Figure 6
<p>Scaled dot-product attention.</p>
Full article ">Figure 7
<p>Structure of attention fusion mechanism.</p>
Full article ">Figure 8
<p>Bi-GRU information flow transfer diagram.</p>
Full article ">Figure 9
<p>Changes during training: (<b>a</b>) accuracy and (<b>b</b>) loss values.</p>
Full article ">Figure 10
<p>Recognition accuracy of the dataset RadioML2016.10A on several models.</p>
Full article ">Figure 11
<p>Confusion matrix at an SNR of 18 dB for (<b>a</b>) 1D-CNN, (<b>b</b>) 2D-CNN, (<b>c</b>) CLDNN, (<b>d</b>) DenseNet, (<b>e</b>) LSTM, (<b>f</b>) ResNet, and (<b>g</b>) proposed model.</p>
Full article ">Figure 12
<p>Confusion matrix at full SNR: (<b>a</b>) 1D-CNN, (<b>b</b>) 2D-CNN, (<b>c</b>) CLDNN, (<b>d</b>) DenseNet, (<b>e</b>) LSTM, (<b>f</b>) ResNet, and (<b>g</b>) proposed model.</p>
Full article ">Figure 13
<p>Recognition accuracy for each modulated signal in the range of −20~18 db for all seven methods. (<b>a</b>) The recognition accuracy of each modulated signal using 1D-CNN. (<b>b</b>) The recognition accuracy of each modulated signal using 2D-CNN. (<b>c</b>) The recognition accuracy of CLDNN for each modulated signal. (<b>d</b>) The recognition accuracy of each modulated signal using DenseNet. (<b>e</b>) The recognition accuracy of each modulated signal using LSTM. (<b>f</b>) The recognition accuracy of each modulated signal using ResNET. (<b>g</b>) The recognition accuracy of the proposed model in this paper for each modulated signal.</p>
Full article ">Figure 13 Cont.
<p>Recognition accuracy for each modulated signal in the range of −20~18 db for all seven methods. (<b>a</b>) The recognition accuracy of each modulated signal using 1D-CNN. (<b>b</b>) The recognition accuracy of each modulated signal using 2D-CNN. (<b>c</b>) The recognition accuracy of CLDNN for each modulated signal. (<b>d</b>) The recognition accuracy of each modulated signal using DenseNet. (<b>e</b>) The recognition accuracy of each modulated signal using LSTM. (<b>f</b>) The recognition accuracy of each modulated signal using ResNET. (<b>g</b>) The recognition accuracy of the proposed model in this paper for each modulated signal.</p>
Full article ">Figure 13 Cont.
<p>Recognition accuracy for each modulated signal in the range of −20~18 db for all seven methods. (<b>a</b>) The recognition accuracy of each modulated signal using 1D-CNN. (<b>b</b>) The recognition accuracy of each modulated signal using 2D-CNN. (<b>c</b>) The recognition accuracy of CLDNN for each modulated signal. (<b>d</b>) The recognition accuracy of each modulated signal using DenseNet. (<b>e</b>) The recognition accuracy of each modulated signal using LSTM. (<b>f</b>) The recognition accuracy of each modulated signal using ResNET. (<b>g</b>) The recognition accuracy of the proposed model in this paper for each modulated signal.</p>
Full article ">Figure 13 Cont.
<p>Recognition accuracy for each modulated signal in the range of −20~18 db for all seven methods. (<b>a</b>) The recognition accuracy of each modulated signal using 1D-CNN. (<b>b</b>) The recognition accuracy of each modulated signal using 2D-CNN. (<b>c</b>) The recognition accuracy of CLDNN for each modulated signal. (<b>d</b>) The recognition accuracy of each modulated signal using DenseNet. (<b>e</b>) The recognition accuracy of each modulated signal using LSTM. (<b>f</b>) The recognition accuracy of each modulated signal using ResNET. (<b>g</b>) The recognition accuracy of the proposed model in this paper for each modulated signal.</p>
Full article ">Figure 14
<p>Validation on other datasets: (<b>a</b>) RadioML2016.10B and (<b>b</b>) RadioML2018.01A-sample.</p>
Full article ">
18 pages, 7440 KiB  
Article
Energy Consumption Prediction for Drilling Pumps Based on a Long Short-Term Memory Attention Method
by Chengcheng Wang, Zhi Yan, Qifeng Li, Zhaopeng Zhu and Chengkai Zhang
Appl. Sci. 2024, 14(22), 10750; https://doi.org/10.3390/app142210750 - 20 Nov 2024
Viewed by 245
Abstract
In the context of carbon neutrality and emission reduction goals, energy consumption optimization in the oil and gas industry is crucial for reducing carbon emissions and improving energy efficiency. As a key component in drilling operations, optimizing the energy consumption of drilling pumps [...] Read more.
In the context of carbon neutrality and emission reduction goals, energy consumption optimization in the oil and gas industry is crucial for reducing carbon emissions and improving energy efficiency. As a key component in drilling operations, optimizing the energy consumption of drilling pumps has significant potential for energy savings. However, due to the complex and variable geological conditions, diverse operational parameters, and inherent nonlinear relationships in the drilling process, accurately predicting energy consumption presents considerable challenges. This study proposes a novel Long Short-Term Memory Attention model for precise prediction of drilling pump energy consumption. By integrating Long Short-Term Memory (LSTM) networks with the Attention mechanism, the model effectively captures complex nonlinear relationships and long-term dependencies in energy consumption data. Comparative experiments with traditional LSTM and Convolutional Neural Network (CNN) models demonstrate that the LSTM-Attention model outperforms these models across multiple evaluation metrics, significantly reducing prediction errors and enhancing robustness and adaptability. The proposed model achieved Mean Absolute Error (MAE) values ranging from 5.19 to 10.20 and R2 values close to one (0.95 to 0.98) in four test scenarios, demonstrating excellent predictive performance under complex conditions. The high-precision prediction of drilling pump energy consumption based on this method can support energy optimization and provide guidance for field operations. Full article
(This article belongs to the Special Issue Development and Application of Intelligent Drilling Technology)
Show Figures

Figure 1

Figure 1
<p>The structure of the LSTM.</p>
Full article ">Figure 2
<p>Attention mechanism.</p>
Full article ">Figure 3
<p>Application of the RANSAC mechanism.</p>
Full article ">Figure 4
<p>Independent sliding window approach for real-time pump power data.</p>
Full article ">Figure 5
<p>LSTM-Attention model.</p>
Full article ">Figure 6
<p>Pump power prediction experiments: (<b>a</b>) Test 1; (<b>b</b>) Test 2; (<b>c</b>) Test 3; (<b>d</b>) Test 4.</p>
Full article ">Figure 7
<p>Trends and fluctuations in Test 1.</p>
Full article ">Figure 8
<p>Trends and fluctuations in Test 2.</p>
Full article ">Figure 9
<p>Trends and fluctuations in Test 3.</p>
Full article ">Figure 10
<p>Trends and fluctuations in Test 4.</p>
Full article ">Figure 11
<p>Performance comparison with different indices: (<b>a</b>) MAE; (<b>b</b>) RMSE; (<b>c</b>) MSE; (<b>d</b>) R<sup>2</sup>.</p>
Full article ">
11 pages, 1225 KiB  
Article
AI-Driven Prediction of Symptom Trajectories in Cancer Care: A Deep Learning Approach for Chemotherapy Management
by Joseph Finkelstein, Aref Smiley, Christina Echeverria and Kathi Mooney
Bioengineering 2024, 11(11), 1172; https://doi.org/10.3390/bioengineering11111172 - 20 Nov 2024
Viewed by 229
Abstract
This study presents an advanced method for predicting symptom escalation in chemotherapy patients using Long Short-Term Memory (LSTM) networks and Convolutional Neural Networks (CNNs). The accurate prediction of symptom escalation is critical in cancer care to enable timely interventions and improve symptom management [...] Read more.
This study presents an advanced method for predicting symptom escalation in chemotherapy patients using Long Short-Term Memory (LSTM) networks and Convolutional Neural Networks (CNNs). The accurate prediction of symptom escalation is critical in cancer care to enable timely interventions and improve symptom management to enhance patients’ quality of life during treatment. The analytical dataset consists of daily self-reported symptom logs from chemotherapy patients, including a wide range of symptoms, such as nausea, fatigue, and pain. The original dataset was highly imbalanced, with approximately 84% of the data containing no symptom escalation. The data were resampled into varying interval lengths to address this imbalance and improve the model’s ability to detect symptom escalation (n = 3 to n = 7 days). This allowed the model to predict significant changes in symptom severity across these intervals. The results indicate that shorter intervals (n = 3 days) yielded the highest overall performance, with the CNN model achieving an accuracy of 81%, precision of 87%, recall of 80%, and an F1 score of 83%. This was an improvement over the LSTM model, which had an accuracy of 79%, precision of 85%, recall of 79%, and an F1 score of 82%. The model’s accuracy and recall declined as the interval length increased, though precision remained relatively stable. The findings demonstrate that both CNN’s temporospatial feature extraction and LSTM’s temporal modeling effectively capture escalation patterns in symptom progression. By integrating these predictive models into digital health systems, healthcare providers can offer more personalized and proactive care, enabling earlier interventions that may reduce symptom burden and improve treatment adherence. Ultimately, this approach has the potential to significantly enhance the overall quality of life for chemotherapy patients by providing real-time insights into symptom trajectories and guiding clinical decision making. Full article
Show Figures

Figure 1

Figure 1
<p>Flowchart of the developed algorithm to evaluate data. Final evaluation of compiled data.</p>
Full article ">Figure 2
<p>Comparison of the ROC curves for different n (3–7).</p>
Full article ">
16 pages, 5881 KiB  
Article
Projection of Changes in Stream Water Use Due to Climate Change
by Young-Ho Seo, Junehyeong Park, Byung-Sik Kim and Jang Hyun Sung
Sustainability 2024, 16(22), 10120; https://doi.org/10.3390/su162210120 - 20 Nov 2024
Viewed by 258
Abstract
This study investigates the impact of rising temperatures on residential water use (RWU) in Seoul from 2015 to 2021, addressing the challenges of urban water sustainability under climate change. Using advanced models—convolutional neural networks (CNNs), long short-term memory (LSTM) Networks, eXtreme Gradient Boosting [...] Read more.
This study investigates the impact of rising temperatures on residential water use (RWU) in Seoul from 2015 to 2021, addressing the challenges of urban water sustainability under climate change. Using advanced models—convolutional neural networks (CNNs), long short-term memory (LSTM) Networks, eXtreme Gradient Boosting (XGBoost), and Bayesian Neural Networks (BNNs)—we examined RWU prediction accuracy and incorporated a method to quantify prediction uncertainties. As a result, the BNN model emerged as a robust alternative, demonstrating competitive predictive accuracy and the capability to account for uncertainties in predictions. Recent studies highlight a strong correlation between rising temperatures and increased RWU, especially during summer, with tropical nights (with temperatures above 25 °C) becoming more common; Seoul experienced a record 26 consecutive tropical nights in July 2024, underscoring a trend toward higher RWU. To capture these dynamics, we employed Shared Socioeconomic Pathway (SSP) scenarios and downscaled the KACE-1-0-G Global Climate Model (GCM) for Seoul, projecting a progressive increase in RWU: 0.49% in the F1 period (2011–2040), 1.53% in F2 (2041–2070), and 2.95% in F3 (2071–2100), with significant rises in maximum RWU across these intervals. Our findings highlight an urgent need for proactive measures to secure water resources in response to the anticipated increase in urban water demand driven by climate change. Full article
(This article belongs to the Section Sustainable Water Management)
Show Figures

Figure 1

Figure 1
<p>Pungnap water intake Facility and Paldang Dam.</p>
Full article ">Figure 2
<p>RWU and temperatures in Seoul.</p>
Full article ">Figure 3
<p>The structures of the deep learning models.</p>
Full article ">Figure 4
<p>Training and validation histories of the artificial neural networks: (<b>a</b>) CNN, (<b>b</b>) LSTM, and (<b>c</b>) BNN.</p>
Full article ">Figure 5
<p>RWU Predictions with the regression models: (<b>a</b>) CNN, (<b>b</b>) LSTM, (<b>c</b>) XGBoost, and (<b>d</b>) BNN.</p>
Full article ">Figure 6
<p>The regression models’ predictions with observations across temperature Ranges.</p>
Full article ">Figure 7
<p>BNN ensemble of RWU: (<b>a</b>) monthly and (<b>b</b>) seasonal predictions.</p>
Full article ">Figure 8
<p>Projection of RWU changes during the future period.</p>
Full article ">
17 pages, 812 KiB  
Article
Enhancing Direction-of-Arrival Estimation with Multi-Task Learning
by Simone Bianco, Luigi Celona, Paolo Crotti, Paolo Napoletano, Giovanni Petraglia and Pietro Vinetti
Sensors 2024, 24(22), 7390; https://doi.org/10.3390/s24227390 - 20 Nov 2024
Viewed by 206
Abstract
There are numerous methods in the literature for Direction-of-Arrival (DOA) estimation, including both classical and machine learning-based approaches that jointly estimate the Number of Sources (NOS) and DOA. However, most of these methods do not fully leverage the potential synergies between these two [...] Read more.
There are numerous methods in the literature for Direction-of-Arrival (DOA) estimation, including both classical and machine learning-based approaches that jointly estimate the Number of Sources (NOS) and DOA. However, most of these methods do not fully leverage the potential synergies between these two tasks, which could yield valuable shared information. To address this limitation, in this article, we present a multi-task Convolutional Neural Network (CNN) capable of simultaneously estimating both the NOS and the DOA of the signal. Through experiments on simulated data, we demonstrate that our proposed model surpasses the performance of state-of-the-art methods, especially in challenging environments characterized by high noise levels and dynamic conditions. Full article
Show Figures

Figure 1

Figure 1
<p>Uniform Linear Array (ULA); <span class="html-italic">d</span> is the distance between the sensors; <math display="inline"><semantics> <msub> <mi>θ</mi> <mi>i</mi> </msub> </semantics></math> is the angle of arrival of the impinging signal, and <span class="html-italic">M</span> is the number of sensor array antennas.</p>
Full article ">Figure 2
<p>Architecture of the proposed multi-task CNN for DOA estimation. The network processes the signal covariance matrix through the backbone. The resulting feature vector is passed to two branches: the Number-of-Source estimator predicts the Number of Sources <math display="inline"><semantics> <mi mathvariant="bold-italic">b</mi> </semantics></math> (i.e., a binarized version of the logits <math display="inline"><semantics> <mi mathvariant="bold-italic">s</mi> </semantics></math>); the Direction-of-Arrival estimator provides multiple angles of arrival, denoted as <math display="inline"><semantics> <mi mathvariant="bold-italic">d</mi> </semantics></math>, corresponding to the number of angles <math display="inline"><semantics> <mi mathvariant="bold-italic">b</mi> </semantics></math> predicted by the other branch. A compound loss <math display="inline"><semantics> <mi mathvariant="script">L</mi> </semantics></math> is used to optimize the model based on the two task-specific losses.</p>
Full article ">Figure 3
<p>Ball chart reporting the RMSE versus accuracy. The size of each ball corresponds to the number of model parameters.</p>
Full article ">Figure 4
<p>The DOA estimation performance for the T1 test set at varying SNRs divided by (<b>a</b>) one signal only, (<b>b</b>) two signals, and (<b>c</b>) three signals.</p>
Full article ">Figure 5
<p>Boxplots showing CRLB index distributions for (<b>a</b>) test sets T1, T3, and T4, (<b>b</b>) various snapshots within T2, and (<b>c</b>) various snapshots within T5.</p>
Full article ">Figure 6
<p>Scenario-independent total performance comparison.</p>
Full article ">
Back to TopTop