Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (8,751)

Search Parameters:
Keywords = recognition performance

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
28 pages, 6900 KiB  
Article
A New Approach to Recognize Faces Amidst Challenges: Fusion Between the Opposite Frequencies of the Multi-Resolution Features
by Regina Lionnie, Julpri Andika and Mudrik Alaydrus
Algorithms 2024, 17(11), 529; https://doi.org/10.3390/a17110529 (registering DOI) - 17 Nov 2024
Abstract
This paper proposes a new approach to pixel-level fusion using the opposite frequency from the discrete wavelet transform with Gaussian or Difference of Gaussian. The low-frequency from discrete wavelet transform sub-band was fused with the Difference of Gaussian, while the high-frequency sub-bands were [...] Read more.
This paper proposes a new approach to pixel-level fusion using the opposite frequency from the discrete wavelet transform with Gaussian or Difference of Gaussian. The low-frequency from discrete wavelet transform sub-band was fused with the Difference of Gaussian, while the high-frequency sub-bands were fused with Gaussian. The final fusion was reconstructed using an inverse discrete wavelet transform into one enhanced reconstructed image. These enhanced images were utilized to improve recognition performance in the face recognition system. The proposed method was tested against benchmark face datasets such as The Database of Faces (AT&T), the Extended Yale B Face Dataset, the BeautyREC Face Dataset, and the FEI Face Dataset. The results showed that our proposed method was robust and accurate against challenges such as lighting conditions, facial expressions, head pose, 180-degree rotation of the face profile, dark images, acquisition with time gap, and conditions where the person uses attributes such as glasses. The proposed method is comparable to state-of-the-art methods and generates high recognition performance (more than 99% accuracy). Full article
(This article belongs to the Section Algorithms for Multidisciplinary Applications)
Show Figures

Figure 1

Figure 1
<p>Examples of images inside each dataset: (<b>a</b>) AT&amp;T [<a href="#B40-algorithms-17-00529" class="html-bibr">40</a>], (<b>b</b>) BeautyREC [<a href="#B41-algorithms-17-00529" class="html-bibr">41</a>], (<b>c</b>) EYB [<a href="#B42-algorithms-17-00529" class="html-bibr">42</a>,<a href="#B43-algorithms-17-00529" class="html-bibr">43</a>], (<b>d</b>) EYB-Dark [<a href="#B42-algorithms-17-00529" class="html-bibr">42</a>,<a href="#B43-algorithms-17-00529" class="html-bibr">43</a>], (<b>e</b>) FEI [<a href="#B44-algorithms-17-00529" class="html-bibr">44</a>], (<b>f</b>) FEI-FE [<a href="#B44-algorithms-17-00529" class="html-bibr">44</a>].</p>
Full article ">Figure 2
<p>The flowchart of our proposed method.</p>
Full article ">Figure 3
<p>The MRA-DWT sub-bands (from <b>left</b> to <b>right</b>): approximation, horizontal, vertical, diagonal sub-bands with Haar and one level of decomposition.</p>
Full article ">Figure 4
<p>The illustration of the scaling function (<b>left</b>) and wavelet function (<b>right</b>) from the Haar wavelet.</p>
Full article ">Figure 5
<p>Results from Gaussian filtering and the Difference of Gaussian (from <b>left</b> to <b>right</b>): original image, Gaussian filtered image with <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>σ</mi> </mrow> <mrow> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math>, Gaussian filtered image with <span class="html-italic">σ</span><sub>2</sub>, Difference of Gaussian.</p>
Full article ">Figure 6
<p>Example of results from proposed fusion (from <b>top</b> to <b>bottom</b>): <span class="html-italic">AL</span>, <span class="html-italic">HG</span>, <span class="html-italic">VG</span>, <span class="html-italic">DG</span> with image fusion DWT/IDWT-IF using the mean-mean rule.</p>
Full article ">Figure 7
<p>The comparison of processing times for the AT&amp;T Face Dataset; Exp. 5; Exp. 6 using <span class="html-italic">db2</span> in DWT/IDWT-IF with levels of decomposition: one (Exp. 6a); three (Exp. 6b); five (Exp. 6c); and seven (Exp. 6d).</p>
Full article ">Figure 8
<p>Accuracy results (%) for the AT&amp;T Face Dataset (proposed method) using different wavelet families in MRA-DWT/IDWT with one level of decomposition: (<b>a</b>) Experiment 5; (<b>b</b>) Experiment 6.</p>
Full article ">Figure 9
<p>Accuracy results (%) for AT&amp;T Face Dataset from Experiment 6 (proposed method) using <span class="html-italic">db2</span> wavelet in DWT/IDWT-IF and <span class="html-italic">bior3.3</span> in MRA-DWT/IDWT with variations in the level of decomposition.</p>
Full article ">Figure 10
<p>Accuracy results (%) for AT&amp;T Face Dataset from Experiment 6 (proposed method) using various wavelet families in DWT/IDWT-IF with five levels of decomposition and <span class="html-italic">bior3.3</span> in MRA-DWT/IDWT.</p>
Full article ">Figure 11
<p>Accuracy results (%) for the EYB Face Dataset for Experiments 2, 4, 5, and 6.</p>
Full article ">Figure 12
<p>Accuracy results (%) for the EYB-Dark Face Dataset for Experiments 2, 4, 5, and 6.</p>
Full article ">Figure 13
<p>Accuracy results (%) for the EYB-Dark Face Dataset for Experiment 6 using fusion rules: mean-mean, min-max, and max-min.</p>
Full article ">Figure 14
<p>Fusion results of DWT/IDWT-IF with d2 and five levels of decomposition (from left to right) top: original image, using min-max rule, max-min rule, and mean-mean rule; bottom: fusion results but scaled based on the pixel value range.</p>
Full article ">Figure 15
<p>Accuracy results (%) for the EYB-Dark Face Dataset for Experiment 6 with the mean-mean fusion rule using different wavelet families for MRA-DWT/IDWT.</p>
Full article ">Figure 16
<p>Accuracy results (%) for the BeautyREC Dataset from Exp. 5 and 6 with variations of employing 1820 images and all (3000) images.</p>
Full article ">Figure 17
<p>Accuracy results (%) for the BeautyREC Dataset: Exp. 5, LP-IF with MRA-DWT/IDWT (a) <span class="html-italic">haar</span>, (b) <span class="html-italic">db2</span>, (c) <span class="html-italic">sym2</span>, (d) <span class="html-italic">bior2.6</span>, (e) <span class="html-italic">bior3.3</span>; Exp. 6, DWT/IDWT-IF with MRA-DWT/IDWT (a) <span class="html-italic">haar</span>, (b) <span class="html-italic">db2</span>, (c) <span class="html-italic">sym2</span>, (d) <span class="html-italic">bior2.6</span>, (e) <span class="html-italic">bior3.3</span>; Exp. 6, DWT/IDWT-IF with <span class="html-italic">haar</span> for MRA-DWT/IDWT and <span class="html-italic">db2</span> wavelet with total level of decomposition (f) one, (g) three, (h) seven; Exp. 6, DWT/IDWT-IF with <span class="html-italic">haar</span> for MRA-DWT/IDWT and five levels of decomposition using wavelets (i) <span class="html-italic">haar</span>, (j) <span class="html-italic">sym2</span>, (k) <span class="html-italic">bior 2.6</span>; Exp. 6, DWT/IDWT-IF using fusion rule (l) min-max, (m) max-min. All results came from SVM with the cubic kernel.</p>
Full article ">Figure 18
<p>Example of high variations for one person inside the BeautyREC Face Dataset.</p>
Full article ">Figure 19
<p>Accuracy results (%) for the FEI Face Database from Exp. 5 and 6.</p>
Full article ">Figure 20
<p>Accuracy results (%) for the FEI-FE Face Database from Exp. 5 and 6.</p>
Full article ">
26 pages, 1128 KiB  
Review
Food Waste as Feedstock for Anaerobic Mono-Digestion Process
by Wirginia Tomczak, Monika Daniluk and Anna Kujawska
Appl. Sci. 2024, 14(22), 10593; https://doi.org/10.3390/app142210593 (registering DOI) - 17 Nov 2024
Abstract
There is a growing recognition that food waste (FW) comprises a significant amount of unused energy. Indeed, FW shows great potential to produce methane (CH4)-rich biogas via an anaerobic digestion (AD) process. Nevertheless, to ensure high AD process performance, deepening the [...] Read more.
There is a growing recognition that food waste (FW) comprises a significant amount of unused energy. Indeed, FW shows great potential to produce methane (CH4)-rich biogas via an anaerobic digestion (AD) process. Nevertheless, to ensure high AD process performance, deepening the knowledge of FW characteristics is required. Furthermore, the biogas yield is strongly influenced by several operational parameters. Taking into account the above, in the current study, based on the data in the literature, the physicochemical parameters of FW generated throughout the world are presented and discussed. In addition, the performance profile of the single-stage anaerobic mono-digestion process with the use of FW as a feedstock was investigated. The performed analysis clearly demonstrated that FW is characterized by significant variations in several parameters, such as pH, the total solid (TS) and volatile solid (VS) contents, the volatile solids to total solids ratio (VS/TS), soluble chemical oxygen demand (sCOD), the concentrations of VFAs and ammonium nitrogen (NH4+-N), and the carbon-to-nitrogen ratio (C/N). Moreover, it was shown that the selected operational parameters, such as temperature, pH, the ratio of food waste to inoculum (I) (FW/I), and the organic loading rate (OLR), may have the most significant impact on the performance of the single-stage anaerobic mono-digestion process. In addition, it was found that most of the experimental investigations presented in the literature were conducted on a laboratory scale. Hence, in future research, more effort should be made to determine the biogas yield with the use of full-scale systems. To summarize, it should be clearly highlighted that the analysis presented in this study may have important implications for the management and application of FW as feedstock for an anaerobic mono-digestion process on an industrial scale. Full article
(This article belongs to the Special Issue Advances in Bioprocess Monitoring and Control)
16 pages, 558 KiB  
Article
The Usefulness of Carotid Artery Doppler Measurement as a Predictor of Early Death in Sepsis Patients Admitted to the Emergency Department
by Su-Il Kim, Yun-Deok Jang, Jae-Gu Ji, Yong-Seok Kim, In-Hye Kang, Seong-Ju Kim, Seong-Min Han and Min-Seok Choi
J. Clin. Med. 2024, 13(22), 6912; https://doi.org/10.3390/jcm13226912 (registering DOI) - 16 Nov 2024
Viewed by 313
Abstract
Background: This study aims to verify whether the blood flow velocity and the diameter size, measured through intra-carotid artery Doppler measurements performed on sepsis patients visiting the emergency department, are useful as tools for predicting the risk of early death. Methods: As [...] Read more.
Background: This study aims to verify whether the blood flow velocity and the diameter size, measured through intra-carotid artery Doppler measurements performed on sepsis patients visiting the emergency department, are useful as tools for predicting the risk of early death. Methods: As a prospective study, this research was performed on sepsis patients who visited a local emergency medical center from August 2021 to February 2023. The sepsis patients’ carotid artery was measured using Doppler imaging, and they were divided into patients measured for the size of systolic and diastolic mean blood flow velocity and diameter size: those measured for their qSOFA (quick sequential organ failure assessment) score and those measured using the SIRS (systemic inflammatory response syndrome) criteria. By measuring and comparing their mortality prediction accuracies, this study sought to verify the usefulness of blood flow velocity and the diameter size of the intra-carotid artery as tools to predict early death. Results: This study was conducted on 1026 patients, excluding 45 patients out of the total of 1071 patients. All sepsis patients were measured using systolic and diastolic blood flow velocity and diameter by Doppler imaging of the intra-carotid artery, assessed using qSOFA and evaluated using SIRS criteria. The results of the analysis performed to compare the mortality prediction accuracy were as follows. First, the hazard ratio (95% CI) of the intra-carotid artery was significant (p < 0.05), at 1.020 (1.004–1.036); the hazard ratio (95% CI) of qSOFA was significant (p < 0.05), at 3.871 (2.526–5.931); and the hazard ratio (95% CI) of SIRS showed no significant difference, at 1.002 (0.995–1.009). After 2 h of infusion treatment, the diameter size was 4.72 ± 1.23, showing a significant difference (p < 0.05). After 2 h of fluid treatment, the blood flow velocity was 101 m/s ± 21.12, which showed a significant difference (p < 0.05). Conclusions: Measuring the mean blood flow velocity in the intra-carotid arteries of sepsis patients who visit the emergency department is useful for predicting the risk of death at an early stage. And this study showed that Doppler measurement of the diameter size of the carotid artery significantly increased after performing fluid treatment after early recognition. Full article
(This article belongs to the Special Issue Emergency Ultrasound: State of the Art and Perspectives)
18 pages, 1642 KiB  
Article
Crouch Gait Recognition in the Anatomical Space Using Synthetic Gait Data
by Juan-Carlos Gonzalez-Islas, Omar Arturo Dominguez-Ramirez, Omar Lopez-Ortega and Jonatan Pena Ramirez
Appl. Sci. 2024, 14(22), 10574; https://doi.org/10.3390/app142210574 (registering DOI) - 16 Nov 2024
Viewed by 264
Abstract
Crouch gait, also referred to as flexed knee gait, is an abnormal walking pattern, characterized by an excessive flexion of the knee, and sometimes also with anomalous flexion in the hip and/or the ankle, during the stance phase of gait. Due to the [...] Read more.
Crouch gait, also referred to as flexed knee gait, is an abnormal walking pattern, characterized by an excessive flexion of the knee, and sometimes also with anomalous flexion in the hip and/or the ankle, during the stance phase of gait. Due to the fact that the amount of clinical data related to crouch gait are scarce, it is difficult to find studies addressing this problem from a data-based perspective. Consequently, in this paper we propose a gait recognition strategy using synthetic data that have been obtained using a polynomial based-generator. Furthermore, though this study, we consider datasets that correspond to different levels of crouch gait severity. The classification of the elements of the datasets into the different levels of abnormality is achieved by using different algorithms like k-nearest neighbors (KNN) and Naive Bayes (NB), among others. On the other hand, to evaluate the classification performance we consider different metrics, including accuracy (Acc) and F measure (FM). The obtained results show that the proposed strategy is able to recognize crouch gait with an accuracy of more than 92%. Thus, it is our belief that this recognition strategy may be useful during the diagnosis phase of crouch gait disease. Finally, the crouch gait recognition approach introduced here may be extended to identify other gait abnormalities. Full article
(This article belongs to the Section Biomedical Engineering)
Show Figures

Figure 1

Figure 1
<p>Framework for crouch gait recognition in anatomical space using synthetic gait data.</p>
Full article ">Figure 2
<p>8 DoF gait kinematics model. (<b>Right</b>) (skeletal model), (<b>left</b>) (open kinematic chain).</p>
Full article ">Figure 3
<p>Workflow stages of the crouch gait recognition in the anatomical space.</p>
Full article ">Figure 4
<p>Gait joint angles for the 8 movements for the 5 gaits. Normal (green), crouch-1 (blue), crouch-2 (red), crouch-3 (black), and crouch-4 (magenta). Limit between the stance phase and swing phase of the gait cycle (cyan).</p>
Full article ">Figure 5
<p>Flowchart of the synthetic gait joint angles generator algorithm.</p>
Full article ">Figure 6
<p>Statistical distribution of the dataset of 5 gait classes for each joint angle.</p>
Full article ">Figure 7
<p>Joint angles of hip flexo-extension (<math display="inline"><semantics> <msub> <mi>q</mi> <mrow> <mn>3</mn> <mi>R</mi> </mrow> </msub> </semantics></math>), knee flexo-extension (<math display="inline"><semantics> <msub> <mi>q</mi> <mrow> <mn>4</mn> <mi>R</mi> </mrow> </msub> </semantics></math>), and ankle dorsi-plantar flexion (<math display="inline"><semantics> <msub> <mi>q</mi> <mrow> <mn>5</mn> <mi>R</mi> </mrow> </msub> </semantics></math>). (<b>First column</b>): joint reference angle (black) and real joint angle (blue); (<b>second column</b>): joint reference angle (black) and synthetic joint angle (blue) and (<b>third column</b>); joint reference angle (black) and average of dataset of this joint angle (blue). Limit between the stance phase and swing phase of the gait cycle (cyan).</p>
Full article ">
14 pages, 325 KiB  
Article
Multimodal Framework for Long-Tailed Recognition
by Jian Chen, Jianyin Zhao, Jiaojiao Gu, Yufeng Qin and Hong Ji
Appl. Sci. 2024, 14(22), 10572; https://doi.org/10.3390/app142210572 (registering DOI) - 16 Nov 2024
Viewed by 192
Abstract
Long-tailed data distribution (i.e., minority classes occupy most of the data, while most classes have very few samples) is a common problem in image classification. In this paper, we propose a novel multimodal framework for long-tailed data recognition. In the first stage, long-tailed [...] Read more.
Long-tailed data distribution (i.e., minority classes occupy most of the data, while most classes have very few samples) is a common problem in image classification. In this paper, we propose a novel multimodal framework for long-tailed data recognition. In the first stage, long-tailed data are used for visual-semantic contrastive learning to obtain good features, while in the second stage, class-balanced data are used for classifier training. The proposed framework leverages the advantages of multimodal models and mitigates the problem of class imbalance in long-tailed data recognition. Experimental results demonstrate that the proposed framework achieves competitive performance on the CIFAR-10-LT, CIFAR-100-LT, ImageNet-LT, and iNaturalist2018 datasets for image classification. Full article
Show Figures

Figure 1

Figure 1
<p>An overview of the proposed framework. The dashed line above in the figure represents the image–text branch, while the dashed line below represents the image-classifier branch. The term cls-text denotes the textual database describing each category. The nums-cls graph illustrates several classes in the training set along with their corresponding sample counts, displaying a long-tailed distribution curve. The image encoder and text encoder are utilized separately to extract features from their respective modalities. Due to potential noise and interference in the textual descriptions for each category, such as ambiguous or entirely incorrect descriptions, in the second phase of training, we employ filters to retain the most representative textual features for each class. Each small square in the figure represents a feature extracted by the feature extractor for each sample, forming a vector. Different colors indicate different categories. For example, the three yellow squares represent three distinct samples, but they share the same class label.</p>
Full article ">Figure 2
<p>A description of the feat-filter module. The function of this module is to select the features most relevant to the category from the textual features while filtering out irrelevant or noisy ones. We named it “feat-filter” because its primary purpose is to filter out features that are unrelated to the category. Through this module, we ensure that only the most discriminative textual features are retained for the subsequent multimodal classification tasks.</p>
Full article ">
31 pages, 64773 KiB  
Article
Versatility Evaluation of Landslide Risk with Window Sizes and Sampling Techniques Based on Deep Learning
by Fudong Ren and Koichi Isobe
Appl. Sci. 2024, 14(22), 10571; https://doi.org/10.3390/app142210571 (registering DOI) - 16 Nov 2024
Viewed by 251
Abstract
Across the globe, landslides cause significant loss of life, injuries, and widespread damage to homes and infrastructure. Therefore, assessing and analyzing landslide hazards is crucial to human, environmental, cultural, economic, and social sustainability. This study utilizes ArcGIS 10.8 and Python 3.9 to create [...] Read more.
Across the globe, landslides cause significant loss of life, injuries, and widespread damage to homes and infrastructure. Therefore, assessing and analyzing landslide hazards is crucial to human, environmental, cultural, economic, and social sustainability. This study utilizes ArcGIS 10.8 and Python 3.9 to create landslide databases for Niigata Prefecture (NIG), Iwate and Miyagi Prefectures (IWT-MYG), and Hokkaido (HKD), drawing on data obtained from the National Research Institute for Earth Science and Disaster Resilience, Japan. A distinguishing feature of this study is the application of a Convolutional Neural Network (CNN), which significantly outperforms traditional machine learning models in image-based pattern recognition by extracting contextual information from surrounding areas, a distinct advantage in image and pattern recognition tasks. Unlike conventional methods that often require manual feature selection and engineering, CNNs automate feature extraction, enabling a more nuanced understanding of complex patterns. By experimenting with CNN input window sizes ranging from 3 × 3 to 27 × 27 pixels and employing diverse sampling techniques, we demonstrate that larger windows enhance the model’s predictive accuracy by capturing a wider range of environmental interactions critical for effective landslide modeling. CNN models with 19 × 19 pixel windows typically yield the best overall performance, with CNN-19 achieving an AUC of 0.950, 0.982 and 0.969 for NIG, HKD, and IWT-MYG, respectively. Furthermore, we improve prediction reliability using oversampling and a random window-moving method. For instance, in the NIG region, the AUC of the oversampling CNN-19 is 0.983, while the downsampling AUC is 0.950). These techniques, less commonly applied in traditional machine learning approaches to landslide detection, help address the issue of data imbalance often seen in landslide datasets, where instances of landslides are far outnumbered by non-landslide occurrences. While challenges remain in enhancing the model’s generalization, this research makes significant progress in developing more robust and adaptable tools for landslide prediction, which are vital for ensuring environmental and societal resilience. Full article
(This article belongs to the Special Issue Novel Technology in Landslide Monitoring and Risk Assessment)
Show Figures

Figure 1

Figure 1
<p>Study areas and distributions of historical landslides: (<b>A</b>) NIG study area, (<b>B</b>) IWT-MYG study area, and (<b>C</b>) HKD study area .</p>
Full article ">Figure 2
<p>Distribution of landslide and non-landslide data.</p>
Full article ">Figure 3
<p>Maps of landslide features.</p>
Full article ">Figure 3 Cont.
<p>Maps of landslide features.</p>
Full article ">Figure 3 Cont.
<p>Maps of landslide features.</p>
Full article ">Figure 4
<p>Flowchart of this study.</p>
Full article ">Figure 5
<p>Schematic diagram of shifting method.</p>
Full article ">Figure 6
<p>Information gain ratio of the landslide feature.</p>
Full article ">Figure 7
<p>Distribution frequency of landslides with different features: (<b>a</b>) Elevation, (<b>b</b>) Rainfall, (<b>c</b>) Distance to river, (<b>d</b>) Land use.</p>
Full article ">Figure 8
<p>Variation of landslide areas across landslide scale samples and AUC of CNN-i models across landslide scale.</p>
Full article ">Figure 9
<p>Average AUC and Pearson correlation coefficients of CNN models for landslide scale.</p>
Full article ">Figure 10
<p>ROC curve in downsampling and oversampling in HKD based on CNN-i.</p>
Full article ">Figure 11
<p>ROC curve in downsampling and oversampling in NIG based on CNN-i.</p>
Full article ">Figure 12
<p>ROC curve in downsampling and oversampling in IWT based on CNN-i.</p>
Full article ">Figure 13
<p>LSM, CM, and LIM in CNN-15 (upper: downsampling, lower: oversampling).</p>
Full article ">Figure 14
<p>LSM, CM, and LIM of HKD in CNN-19 (upper: downsampling, lower: oversampling).</p>
Full article ">Figure 15
<p>LSM, CM, and LIM of HKD in CNN-23 (upper: downsampling, lower: oversampling).</p>
Full article ">Figure 16
<p>LSM, CM, and LIM of NIG in CNN-15 (upper: downsampling, lower: oversampling).</p>
Full article ">Figure 17
<p>LSM, CM, and LIM of NIG in CNN-19 (upper: downsampling, lower: oversampling).</p>
Full article ">Figure 18
<p>LSM, CM, and LIM of NIG in CNN-23 (upper: downsampling, lower: oversampling).</p>
Full article ">Figure 19
<p>LSM, CM, and LIM of IWT in CNN-15 (upper: downsampling, lower: oversampling).</p>
Full article ">Figure 20
<p>LSM, CM, and LIM of IWT in CNN-19 (upper: downsampling, lower: oversampling).</p>
Full article ">Figure 21
<p>LSM, CM, and LIM of IWT in CNN-23 (upper: downsampling, lower: oversampling).</p>
Full article ">Figure 22
<p>CM based on the CNN-19 with the oversampling method in HKD. Left to right: IWT, HKD, and NIG.</p>
Full article ">Figure 23
<p>CM based on the CNN-19 with the oversampling method in NIG. Left to right: IWT, HKD, and NIG.</p>
Full article ">Figure 24
<p>CM based on the CNN-19 with the oversampling method in IWT. Left to right: IWT, HKD, and NIG.</p>
Full article ">Figure A1
<p>Maps showing parts of landslide features in NIG area: (<b>a</b>) Land use, (<b>b</b>) Soil, (<b>c</b>) Lithology, and (<b>d</b>) Vegetation.</p>
Full article ">Figure A2
<p>Maps showing parts of landslide features in HKD area: (<b>a</b>) Land use, (<b>b</b>) Soil, (<b>c</b>) Lithology, and (<b>d</b>) Vegetation.</p>
Full article ">Figure A3
<p>Maps showing parts of landslide features in IWT-MYG area: (<b>a</b>) Land use, (<b>b</b>) Soil, (<b>c</b>) Lithology, and (<b>d</b>) Vegetation.</p>
Full article ">Figure A3 Cont.
<p>Maps showing parts of landslide features in IWT-MYG area: (<b>a</b>) Land use, (<b>b</b>) Soil, (<b>c</b>) Lithology, and (<b>d</b>) Vegetation.</p>
Full article ">
17 pages, 2888 KiB  
Article
Research on Fault Diagnosis of Agricultural IoT Sensors Based on Improved Dung Beetle Optimization–Support Vector Machine
by Sicheng Liang, Pingzeng Liu, Ziwen Zhang and Yong Wu
Sustainability 2024, 16(22), 10001; https://doi.org/10.3390/su162210001 (registering DOI) - 16 Nov 2024
Viewed by 224
Abstract
The accuracy of data perception in Internet of Things (IoT) systems is fundamental to achieving scientific decision-making and intelligent control. Given the frequent occurrence of sensor failures in complex environments, a rapid and accurate fault diagnosis and handling mechanism is crucial for ensuring [...] Read more.
The accuracy of data perception in Internet of Things (IoT) systems is fundamental to achieving scientific decision-making and intelligent control. Given the frequent occurrence of sensor failures in complex environments, a rapid and accurate fault diagnosis and handling mechanism is crucial for ensuring the stable operation of the system. Addressing the challenges of insufficient feature extraction and sparse sample data that lead to low fault diagnosis accuracy, this study explores the construction of a fault diagnosis model tailored for agricultural sensors, with the aim of accurately identifying and analyzing various sensor fault modes, including but not limited to bias, drift, accuracy degradation, and complete failure. This study proposes an improved dung beetle optimization–support vector machine (IDBO-SVM) diagnostic model, leveraging the optimization capabilities of the former to finely tune the parameters of the Support Vector Machine (SVM) to enhance fault recognition under conditions of limited sample data. Case analyses were conducted using temperature and humidity sensors in air and soil, with comprehensive performance comparisons made against mainstream algorithms such as the Backpropagation (BP) neural network, Sparrow Search Algorithm–Support Vector Machine (SSA-SVM), and Elman neural network. The results demonstrate that the proposed model achieved an average diagnostic accuracy of 94.91%, significantly outperforming other comparative models. This finding fully validates the model’s potential in enhancing the stability and reliability of control systems. The research results not only provide new ideas and methods for fault diagnosis in IoT systems but also lay a foundation for achieving more precise, efficient intelligent control and scientific decision-making. Full article
Show Figures

Figure 1

Figure 1
<p>IoT sensing device.</p>
Full article ">Figure 2
<p>Sensor fault waveform characteristics diagram.</p>
Full article ">Figure 3
<p>Performance comparison chart of optimization algorithms.</p>
Full article ">Figure 4
<p>IDBO-SVM troubleshooting flow.</p>
Full article ">Figure 5
<p>(<b>a</b>) Confusion matrix for classification of temperature sensor fault prediction. (<b>b</b>) Confusion matrix for classification of humidity sensor fault prediction. (<b>c</b>) Confusion matrix for classification of soil temperature sensor fault prediction. (<b>d</b>) Confusion matrix for classification of soil humidity sensor fault prediction.</p>
Full article ">Figure 6
<p>Fault diagnosis model accuracy comparison.</p>
Full article ">
22 pages, 20035 KiB  
Article
Methodology for Object-Level Change Detection in Post-Earthquake Building Damage Assessment Based on Remote Sensing Images: OCD-BDA
by Zhengtao Xie, Zifan Zhou, Xinhao He, Yuguang Fu, Jiancheng Gu and Jiandong Zhang
Remote Sens. 2024, 16(22), 4263; https://doi.org/10.3390/rs16224263 (registering DOI) - 15 Nov 2024
Viewed by 267
Abstract
Remote sensing and computer vision technologies are increasingly leveraged for rapid post-disaster building damage assessment, becoming a crucial and practical approach. In this context, the accuracy of employing various AI models in pixel-level change detection methods is significantly dependent on the consistency between [...] Read more.
Remote sensing and computer vision technologies are increasingly leveraged for rapid post-disaster building damage assessment, becoming a crucial and practical approach. In this context, the accuracy of employing various AI models in pixel-level change detection methods is significantly dependent on the consistency between pre- and post-disaster building images, particularly regarding variations in resolution, viewing angle, and lighting conditions; in object-level feature recognition methods, the low richness of semantic details of damaged buildings in images leads to a poor detection accuracy. This paper proposes a novel method, OCD-BDA (Object-Level Change Detection for Post-Disaster Building Damage Assessment), as an alternative to pixel-level change detection and object-level feature recognition methods. Inspired by human cognitive processes, this method incorporates three key steps: an efficient sample acquisition for object localization, labeling via HGC (Hierarchical and Gaussian Clustering), and model training and prediction for classification. Furthermore, this study establishes a change detection dataset based on Google Earth imagery of regions in Hatay Province before and after the Turkish earthquake. This dataset is characterized by pixel inconsistency and significant differences in photographic angles and lighting conditions between pre- and post-disaster images, making it a valuable test dataset for other studies. As a result, in the experiments of comparative generalization capabilities, OCD-BDA demonstrated a significant improvement, achieving an accuracy of 71%, which is twice that of the second-ranking method. Moreover, OCD-BDA exhibits superior performance in tasks with small sample amounts and rapid training time. With only 1% of the training samples, it achieves a prediction accuracy exceeding that of traditional transfer learning methods with 60% of samples. Additionally, it completes assessments across a large disaster area (450 km²) with 93% accuracy in under 23 min. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>The design concept of the object-level post-disaster building assessment methodology.</p>
Full article ">Figure 2
<p>Building damage assessment criteria.</p>
Full article ">Figure 3
<p>The overall architecture of OCD-BDA: YOLOv8 Lite for localization; HGC for labeling; and R50M2Net for classification.</p>
Full article ">Figure 4
<p>The network structure of YOLOv8 Lite.</p>
Full article ">Figure 5
<p>Flowchart of HGC.</p>
Full article ">Figure 6
<p>Comparison between HGC and traditional labeling methods.</p>
Full article ">Figure 7
<p>The overall architecture of R50M2Net.</p>
Full article ">Figure 8
<p>The regions contained in the Turkish earthquake dataset.</p>
Full article ">Figure 9
<p>Detection results of generalization capability experiments: (<b>a</b>) Samandag; (<b>b</b>) Antakia; (<b>c</b>) Klrlkhan; and (<b>d</b>) Iskenderun. Green represents Grade A, yellow represents Grade B, and red represents Grade C.</p>
Full article ">Figure 10
<p>Details of the comparative study.</p>
Full article ">Figure 11
<p>Accuracy comparison results under different training sample ratios: Ex1 represents traditional transfer learning; Ex2 refers to OCD-BDA (same below).</p>
Full article ">Figure 12
<p>The test results of two experiments: green represents Grade A, yellow represents Grade B, and red represents Grade C.</p>
Full article ">Figure 12 Cont.
<p>The test results of two experiments: green represents Grade A, yellow represents Grade B, and red represents Grade C.</p>
Full article ">Figure 12 Cont.
<p>The test results of two experiments: green represents Grade A, yellow represents Grade B, and red represents Grade C.</p>
Full article ">
15 pages, 3034 KiB  
Article
Polycrystalline Diamond Film Growth on Gallium Nitride with Low Boundary Thermal Resistance
by Ying Wang, Jiahao Yao, Yong Yang, Qian Fan, Xianfeng Ni and Xing Gu
Coatings 2024, 14(11), 1457; https://doi.org/10.3390/coatings14111457 - 15 Nov 2024
Viewed by 260
Abstract
As the demand for high-frequency and high-power electronic devices has increased, gallium nitride (GaN), particularly in the context of high-electron mobility transistors (HEMTs), has attracted considerable attention. However, the ‘self-heating effect’ of GaN HEMTs represents a significant limitation regarding both performance and reliability. [...] Read more.
As the demand for high-frequency and high-power electronic devices has increased, gallium nitride (GaN), particularly in the context of high-electron mobility transistors (HEMTs), has attracted considerable attention. However, the ‘self-heating effect’ of GaN HEMTs represents a significant limitation regarding both performance and reliability. Diamond, renowned for its exceptional thermal conductivity, represents an optimal material for thermal management in HEMTs. This paper proposes a novel method for directly depositing diamond films onto N-polar GaN (NP-GaN) epitaxial layers. This eliminates the complexities of the traditional diamond growth process and the need for temporary substrate steps. Given the relative lag in the development of N-polar material growth technologies, which are marked by surface roughness issues, and the recognition that the thermal boundary resistance (TBRGaN/diamond) represents a critical factor constraining efficient heat transfer, our study has introduced a series of optimizations to enhance the quality of the diamond nucleation layer while ensuring that the integrity of the GaN buffer layer remains intact. Moreover, chemical mechanical polishing (CMP) technology was employed to effectively reduce the surface roughness of the NP-GaN base, thereby providing a more favorable foundation for diamond growth. The optimization trends observed in the thermal performance test results are encouraging. Integrating diamond films onto highly smooth NP-GaN epitaxial layers demonstrates a reduction in TBRGaN/diamond compared to that of diamond layers deposited onto NP-GaN with higher surface roughness that had undergone no prior process treatment. Full article
Show Figures

Figure 1

Figure 1
<p>Integration of GaN and diamond based on NP-GaN epitaxial wafers.</p>
Full article ">Figure 2
<p>(<b>a</b>) OM images of seed crystal distributions after spin-coating for seed crystals with a diameter of 5 nm (too small) and SEM cross-sectional images of seed crystals with a diameter of 2 μm (too large) after growth; (<b>b</b>) OM images of seed crystal distributions after spin-coating for seed crystals with a diameter of 100 nm and suspension mass fractions of 0.1%, 0.3%, and 0.5%, along with corresponding SEM cross-sectional images of the diamond layers after growth.</p>
Full article ">Figure 3
<p>A cross-sectional SEM analysis of the PCD under the combined effects of the CH<sub>4</sub> concentration and plasma power density.</p>
Full article ">Figure 4
<p>(<b>a</b>) OM images of the NP-GaN surface morphology without MOCVD process optimization, after MOCVD process optimization, and after additional CMP; (<b>b</b>) AFM comparison images of the NP-GaN surface after MOCVD process optimization and subsequent CMP. Insets: Surface morphology height profiles along the direction indicated by the blue line.</p>
Full article ">Figure 5
<p>(<b>a</b>) OM images of diamond seed crystal distribution after spin-coating; (<b>b</b>) SEM images of cross-sectional and surface morphology of the grown diamond on NP-GaN bases without MOCVD process optimization, after MOCVD process optimization, and after additional CMP.</p>
Full article ">Figure 6
<p>(<b>a</b>) XRD; (<b>b</b>) Raman (normalized based on the characteristic peak of diamond) spectrum of grown diamond on NP-GaN bases without MOCVD process optimization, after MOCVD process optimization, and after additional CMP.</p>
Full article ">Figure 7
<p>TBR<sub>GaN/diamond</sub> and its variation across NP-GaN bases with different surface morphologies. The maximum, minimum, median, and mean values are depicted in the box plots. Errors were extracted using a statistic method with the lower/upper limit determined by the 25th/75th percentiles of the corresponding average values of each sample. Inset: A plot of TBR<sub>GaN/diamond</sub> versus κ<sub>Diamond</sub> after diamond growth on the NP-GaN base with a relatively smooth surface morphology achieved through optimized MOCVD process parameters.</p>
Full article ">
18 pages, 5616 KiB  
Article
Hyperspectral Imaging Combined with Deep Learning for the Early Detection of Strawberry Leaf Gray Mold Disease
by Yunmeng Ou, Jingyi Yan, Zhiyan Liang and Baohua Zhang
Agronomy 2024, 14(11), 2694; https://doi.org/10.3390/agronomy14112694 - 15 Nov 2024
Viewed by 206
Abstract
The presence of gray mold can seriously affect the yield and quality of strawberries. Due to their susceptibility and the rapid spread of this disease, it is important to develop early, accurate, rapid, and non-destructive disease identification strategies. In this study, the early [...] Read more.
The presence of gray mold can seriously affect the yield and quality of strawberries. Due to their susceptibility and the rapid spread of this disease, it is important to develop early, accurate, rapid, and non-destructive disease identification strategies. In this study, the early detection of strawberry leaf diseases was performed using hyperspectral imaging combining multi-dimensional features like spectral fingerprints and vegetation indices. Firstly, hyperspectral images of healthy and early affected leaves (24 h) were acquired using a hyperspectral imaging system. Then, spectral reflectance (616) and vegetation index (40) were extracted. Next, the CARS algorithm was used to extract spectral fingerprint features (17). Pearson correlation analysis combined with the SPA method was used to select five significant vegetation indices. Finally, we used five deep learning methods (LSTMs, CNNs, BPFs, and KNNs) to build disease detection models for strawberries based on individual and fusion characteristics. The results showed that the accuracy of the recognition model based on fused features ranged from 88.9% to 96.6%. The CNN recognition model based on fused features performed best, with a recognition accuracy of 96.6%. Overall, the fused feature-based model can reduce the dimensionality of the classification data and effectively improve the predicting accuracy and precision of the classification algorithm. Full article
(This article belongs to the Special Issue AI, Sensors and Robotics for Smart Agriculture—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>The schematic diagram of the hyperspectral imaging system.</p>
Full article ">Figure 2
<p>Healthy and gray leaf mold.</p>
Full article ">Figure 3
<p>Flowchart of the work.</p>
Full article ">Figure 4
<p>Spectral behaviors of different types of strawberry leaves: (<b>a</b>) the hyperspectral cube of the gray mold-infected strawberry leaf; (<b>b</b>) spectra of gray mold-infected strawberry leaves samples; (<b>c</b>) spectra of healthy strawberry leaves samples; and (<b>d</b>) the comparison of original spectra of healthy and disease leaves.</p>
Full article ">Figure 5
<p>(<b>a</b>) Regression coefficients of each variable; (<b>b</b>) spectral fingerprint feature distribution.</p>
Full article ">Figure 6
<p>(<b>a</b>) The correlation coefficients diagram of 40 vegetation indices; (<b>b</b>) the detail of the correlation coefficients diagram.</p>
Full article ">Figure 7
<p>The COSS of 21 VIs obtained by SPA.</p>
Full article ">Figure 8
<p>Classification accuracy comparison of various machine learning models based on different input features. (<b>a</b>) Full wavelength and fingerprint features; (<b>b</b>) full wavelength and significant vegetation index; (<b>c</b>) full wavelength and full vegetation index; and (<b>d</b>) fingerprint feature, significance, and fusion feature.</p>
Full article ">Figure 9
<p>The five models are based on the confusion matrix of mixed features.</p>
Full article ">
23 pages, 5517 KiB  
Article
Research on an Eye Control Method Based on the Fusion of Facial Expression and Gaze Intention Recognition
by Xiangyang Sun and Zihan Cai
Appl. Sci. 2024, 14(22), 10520; https://doi.org/10.3390/app142210520 - 15 Nov 2024
Viewed by 259
Abstract
With the deep integration of psychology and artificial intelligence technology and other related technologies, eye control technology has achieved certain results at the practical application level. However, it is found that the accuracy of the current single-modal eye control technology is still not [...] Read more.
With the deep integration of psychology and artificial intelligence technology and other related technologies, eye control technology has achieved certain results at the practical application level. However, it is found that the accuracy of the current single-modal eye control technology is still not high, which is mainly caused by the inaccurate eye movement detection caused by the high randomness of eye movements in the process of human–computer interaction. Therefore, this study will propose an intent recognition method that fuses facial expressions and eye movement information and expects to complete an eye control method based on the fusion of facial expression and eye movement information based on the multimodal intent recognition dataset, including facial expressions and eye movement information constructed in this study. Based on the self-attention fusion strategy, the fused features are calculated, and the multi-layer perceptron is used to classify the fused features, so as to realize the mutual attention between different features, and improve the accuracy of intention recognition by enhancing the weight of effective features in a targeted manner. In order to solve the problem of inaccurate eye movement detection, an improved YOLOv5 model was proposed, and the accuracy of the model detection was improved by adding two strategies: a small target layer and a CA attention mechanism. At the same time, the corresponding eye movement behavior discrimination algorithm was combined for each eye movement action to realize the output of eye behavior instructions. Finally, the experimental verification of the eye–computer interaction scheme combining the intention recognition model and the eye movement detection model showed that the accuracy of the eye-controlled manipulator to perform various tasks could reach more than 95 percent based on this scheme. Full article
Show Figures

Figure 1

Figure 1
<p>The technical route of this paper’s research.</p>
Full article ">Figure 2
<p>Face image dataset example.</p>
Full article ">Figure 3
<p>This eye movement intent detection flow chart describes the conversion of eye movement data to intent classification.</p>
Full article ">Figure 4
<p>Integration framework based on attention mechanism.</p>
Full article ">Figure 5
<p>Comparison of performance in single-mode and multimodal prediction.</p>
Full article ">Figure 6
<p>Line charts of five indicators of different models.</p>
Full article ">Figure 7
<p>Loss function curve of Anchor method before and after improvement.</p>
Full article ">Figure 8
<p>Structure diagram of the CA attention mechanism [<a href="#B9-applsci-14-10520" class="html-bibr">9</a>].</p>
Full article ">Figure 9
<p>Improved YOLOv5 model structure.</p>
Full article ">Figure 10
<p>Improved loss variation diagram for the YOLOv5 model.</p>
Full article ">Figure 10 Cont.
<p>Improved loss variation diagram for the YOLOv5 model.</p>
Full article ">Figure 11
<p>The average accuracy (AP) curve of the improved model.</p>
Full article ">Figure 12
<p>The F1 score curve of the improved model.</p>
Full article ">Figure 13
<p>Test results before and after improvement.</p>
Full article ">Figure 14
<p>Human–computer interaction experiment platform.</p>
Full article ">Figure 15
<p>The overall flow chart of the experiment.</p>
Full article ">Figure 16
<p>Comparison of calculation efficiency indicators.</p>
Full article ">Figure 17
<p>Complete human–computer interaction process.</p>
Full article ">Figure 18
<p>Test results.</p>
Full article ">Figure 19
<p>Test results for different tasks.</p>
Full article ">
23 pages, 4323 KiB  
Article
LIMUNet: A Lightweight Neural Network for Human Activity Recognition Using Smartwatches
by Liangliang Lin, Junjie Wu, Ran An, Song Ma, Kun Zhao and Han Ding
Appl. Sci. 2024, 14(22), 10515; https://doi.org/10.3390/app142210515 - 15 Nov 2024
Viewed by 386
Abstract
The rise of mobile communication, low-power chips, and the Internet of Things has made smartwatches increasingly popular. Equipped with inertial measurement units (IMUs), these devices can recognize user activities through artificial intelligence (AI) analysis of sensor data. However, most existing AI-based activity recognition [...] Read more.
The rise of mobile communication, low-power chips, and the Internet of Things has made smartwatches increasingly popular. Equipped with inertial measurement units (IMUs), these devices can recognize user activities through artificial intelligence (AI) analysis of sensor data. However, most existing AI-based activity recognition algorithms require significant computational power and storage, making them unsuitable for low-power devices like smartwatches. Additionally, discrepancies between training data and real-world data often hinder model generalization and performance. To address these challenges, we propose LIMUNet and its smaller variant LIMUNet-Tiny—lightweight neural networks designed for human activity recognition on smartwatches. LIMUNet utilizes depthwise separable convolutions and residual blocks to reduce computational complexity and parameter count. It also incorporates a dual attention mechanism specifically tailored to smartwatch sensor data, improving feature extraction without sacrificing efficiency. Experiments on the PAMAP2 and LIMU datasets show that the LIMUNet improves recognition accuracy by 2.9% over leading lightweight models while reducing parameters by 88.3% and computational load by 58.4%. Compared to other state-of-the-art models, LIMUNet achieves a 9.6% increase in accuracy, with a 60% reduction in parameters and a 57.8% reduction in computational cost. LIMUNet-Tiny further reduces parameters by 75% and computational load by 80%, making it even more suitable for resource-constrained devices. Full article
(This article belongs to the Special Issue Mobile Computing and Intelligent Sensing)
Show Figures

Figure 1

Figure 1
<p>Random signal frames before and after filtering.</p>
Full article ">Figure 2
<p>LIMUNet network architecture.</p>
Full article ">Figure 3
<p>Residual bottleneck layer.</p>
Full article ">Figure 4
<p>Channel attention mechanism.</p>
Full article ">Figure 5
<p>Dual attention mechanism.</p>
Full article ">Figure 6
<p>Correspondence between activities and waveforms in the LIMU dataset.</p>
Full article ">Figure 7
<p>Distribution of data in LIMU across different types of behaviors and users.</p>
Full article ">Figure 8
<p>Training curves for different datasets.</p>
Full article ">Figure 9
<p>Confusion matrices for different datasets.</p>
Full article ">Figure 10
<p>The impact of window size. (<b>a</b>) The impact of window size on accuracy in the PAMAPL2 and LIMU datasets. (<b>b</b>) The impact of window size <span class="html-italic">L</span> on FLOPS for the LIMU datasets.</p>
Full article ">Figure 11
<p>LIMUNet (N = 2), LIMUNet-Tiny (N = 1), and LIMUNet-More (N = 3) the accuracy and degree of lightweight design.</p>
Full article ">
11 pages, 7620 KiB  
Article
Ultrathin, Stretchable, and Twistable Ferroelectret Nanogenerator for Facial Muscle Detection
by Ziling Song, Xianfa Cai, Zhi Chen, Ziying Zhu, Yunqi Cao and Wei Li
Nanoenergy Adv. 2024, 4(4), 344-354; https://doi.org/10.3390/nanoenergyadv4040021 - 15 Nov 2024
Viewed by 398
Abstract
Ferroelectret nanogenerators (FENGs) have garnered attention due to their unique porous structure and excellent piezoelectric performance. However, most existing FENGs lack sufficient stretchability and flexibility, limiting their application in the field of wearable electronics. In this regard, we have focused on the development [...] Read more.
Ferroelectret nanogenerators (FENGs) have garnered attention due to their unique porous structure and excellent piezoelectric performance. However, most existing FENGs lack sufficient stretchability and flexibility, limiting their application in the field of wearable electronics. In this regard, we have focused on the development of an ultrathin, stretchable, and twistable ferroelectret nanogenerator (UST-FENG) based on Ecoflex, which is made up of graphene, Ecoflex, and anhydrous ethanol, with controllable pore shape and density. The UST-FENG has a thickness of only 860 µm, a fracture elongation rate of up to 574%, and a Young’s modulus of only 0.2 MPa, exhibiting outstanding thinness and excellent stretchability. Its quasi-static piezoelectric coefficient is approximately 38 pC/N. Utilizing this UST-FENG device can enable the recognition of facial muscle movements such as blinking and speaking, thereby helping to monitor people’s facial conditions and improve their quality of life. The successful application of the UST-FENG in facial muscle recognition represents an important step forward in the field of wearable systems for the human face. Full article
Show Figures

Figure 1

Figure 1
<p>Preparation process of the UST-FENG.</p>
Full article ">Figure 2
<p>Structure diagram of the UST-FENG. (<b>a</b>) Structure of the UST-FENG. (<b>b</b>) SEM image of the graphene flexible electrode. (<b>c</b>) SEM image of the cross-sectional view of the UST-FENG.</p>
Full article ">Figure 3
<p>Simulation of the UST-FENG performance. (<b>a</b>) Potential distribution diagram of the UST-FENG. (<b>b</b>) Displacement change diagram of the UST-FENG after stress application. (<b>c</b>) Stress distribution diagram of the UST-FENG. (<b>d</b>) Cross-section diagram of the electric potential and stress distribution of the UST-FENG.</p>
Full article ">Figure 4
<p>Piezoelectric signals generated by the UST-FENG. (<b>a</b>) Test photograph of the UST-FENG under positive electrode in the initial state. (<b>b</b>) Test photograph of the UST-FENG under reverse electrode in the initial state. (<b>c</b>) Open-circuit voltage diagram produced by the UST-FENG under stress. (<b>d</b>) After polarity switching, open-circuit voltage diagram produced by the UST-FENG under stress.</p>
Full article ">Figure 5
<p>Characteristics of the UST-FENG. (<b>a</b>) Stress–strain curve of the UST-FENG. (<b>b</b>) Relationship diagram of the surface potential and charging voltage. (<b>c</b>) Relationship diagram of quasi-static <span class="html-italic">d</span><sub>33</sub> and charging voltage. (<b>d</b>) Relationship diagram of applied pressure and generated charge. (<b>e</b>) Relationship diagram of quasi-static <span class="html-italic">d</span><sub>33</sub> and applied pressure; the inset is the amplified diagram of the signal measured from 0 to 4.5 kPa. (<b>f</b>) Relationship diagram of dynamic <span class="html-italic">d</span><sub>33</sub> with frequency.</p>
Full article ">Figure 6
<p>Application of the UST-FENG in the detection of facial muscle movement. (<b>a</b>) Transmission diagram of the UST-FENG-generated signal. (<b>b</b>) Photograph of the UST-FENG sticking on an eye. (<b>c</b>) Detection of blink signals by the UST-FENG. (<b>d</b>) Photograph of the UST-FENG sticking on a mouth. (<b>e</b>) Detection of speech signals by the UST-FENG.</p>
Full article ">
27 pages, 3743 KiB  
Article
Performance Analysis and Improvement of Machine Learning with Various Feature Selection Methods for EEG-Based Emotion Classification
by Sherzod Abdumalikov, Jingeun Kim and Yourim Yoon
Appl. Sci. 2024, 14(22), 10511; https://doi.org/10.3390/app142210511 - 14 Nov 2024
Viewed by 568
Abstract
Emotion classification is a challenge in affective computing, with applications ranging from human–computer interaction to mental health monitoring. In this study, the classification of emotional states using electroencephalography (EEG) data were investigated. Specifically, the efficacy of the combination of various feature selection methods [...] Read more.
Emotion classification is a challenge in affective computing, with applications ranging from human–computer interaction to mental health monitoring. In this study, the classification of emotional states using electroencephalography (EEG) data were investigated. Specifically, the efficacy of the combination of various feature selection methods and hyperparameter tuning of machine learning algorithms for accurate and robust emotion recognition was studied. The following feature selection methods were explored: filter (SelectKBest with analysis of variance (ANOVA) F-test), embedded (least absolute shrinkage and selection operator (LASSO) tuned using Bayesian optimization (BO)), and wrapper (genetic algorithm (GA)) methods. We also executed hyperparameter tuning of machine learning algorithms using BO. The performance of each method was assessed. Two different EEG datasets, EEG Emotion and DEAP Dataset, containing 2548 and 160 features, respectively, were evaluated using random forest (RF), logistic regression, XGBoost, and support vector machine (SVM). For both datasets, the experimented three feature selection methods consistently improved the accuracy of the models. For EEG Emotion dataset, RF with LASSO achieved the best result among all the experimented methods increasing the accuracy from 98.78% to 99.39%. In the DEAP dataset experiment, XGBoost with GA showed the best result, increasing the accuracy by 1.59% and 2.84% for valence and arousal. We also show that these results are superior to those by the previous other methods in the literature. Full article
(This article belongs to the Special Issue Advances in Biosignal Processing)
Show Figures

Figure 1

Figure 1
<p>EEG brainwave dataset training.</p>
Full article ">Figure 2
<p>Flowchart of GA.</p>
Full article ">Figure 3
<p>Violin plots of statistical features in the EEG Emotion dataset: (<b>a</b>) mean, (<b>b</b>) mean difference (computed between windows), (<b>c</b>) min, (<b>d</b>) min difference (computed between windows), (<b>e</b>) min difference (computed for each quarter window), (<b>f</b>) max, (<b>g</b>) max difference (computed between windows), (<b>h</b>) max difference (computed for each quarter window), (<b>i</b>) standard deviation, (<b>j</b>) standard deviation difference (computed between windows), (<b>k</b>) log, (<b>l</b>) correlation, (<b>m</b>) entropy, (<b>n</b>) FFT.</p>
Full article ">Figure 4
<p>Violin plot of ten randomly selected features included in the DEAP dataset.</p>
Full article ">Figure 5
<p>FFT-based frequency analysis of the EEG dataset: randomly selected FFT of a sample with (<b>a</b>) positive and (<b>b</b>) negative emotion levels; emotion level analysis of the DEAP dataset: (<b>c</b>) neutral labels from the EEG Emotion dataset, (<b>d</b>) valence level, and (<b>e</b>) arousal level from the DEAP dataset.</p>
Full article ">Figure 6
<p>Graph comparing the four performance indicators of feature selection methods on the EEG Emotion dataset: (<b>a</b>) filter-based feature selection method; (<b>b</b>) embedded-based feature selection method; (<b>c</b>) wrapper-based feature selection method.</p>
Full article ">Figure 6 Cont.
<p>Graph comparing the four performance indicators of feature selection methods on the EEG Emotion dataset: (<b>a</b>) filter-based feature selection method; (<b>b</b>) embedded-based feature selection method; (<b>c</b>) wrapper-based feature selection method.</p>
Full article ">Figure 7
<p>Graph comparing the four performance indicators of feature selection methods on the DEAP dataset: (<b>a</b>) filter-based feature selection method; (<b>b</b>) embedded-based feature selection method; (<b>c</b>) wrapper-based feature selection method.</p>
Full article ">Figure 8
<p>Correlation heatmaps: (<b>a</b>) before feature selection, (<b>b</b>) after feature selection for the EEG Emotion dataset, (<b>c</b>) before feature selection for the DEAP dataset, (<b>d</b>) after feature selection for the valence label in the DEAP dataset, and (<b>e</b>) after feature selection for the arousal label in the DEAP dataset.</p>
Full article ">
18 pages, 12032 KiB  
Article
Advanced Modulation Formats for 400 Gbps Optical Networks and AI-Based Format Recognition
by Zhou He, Hao Huang, Fanjian Hu, Jiawei Gong, Binghua Shi, Jia Guo and Xiaoran Peng
Sensors 2024, 24(22), 7291; https://doi.org/10.3390/s24227291 - 14 Nov 2024
Viewed by 418
Abstract
The integration of communication and sensing (ICAS) in optical networks is an inevitable trend in building intelligent, multi-scenario, application-converged communication systems. However, due to the impact of nonlinear effects, co-fiber transmission of sensing signals and communication signals can cause interference to the communication [...] Read more.
The integration of communication and sensing (ICAS) in optical networks is an inevitable trend in building intelligent, multi-scenario, application-converged communication systems. However, due to the impact of nonlinear effects, co-fiber transmission of sensing signals and communication signals can cause interference to the communication signals, leading to an increased bit error rate (BER). This paper proposes a noncoherent solution based on the alternate polarization chirped return-to-zero frequency shift keying (Apol-CRZ-FSK) modulation format to realize a 4 × 100 Gbps dense wavelength division multiplexing (DWDM) optical network. Simulation results show that compared to traditional modulation formats, such as chirped return-to-zero frequency shift keying (CRZ-FSK) and differential quadrature phase shift keying (DQPSK), this solution demonstrates superior resistance to nonlinear effects, enabling longer transmission distances and better transmission performance. Moreover, to meet the transmission requirements and signal sensing and recognition needs in future optical networks, this study employs the Inception-ResNet-v2 convolutional neural network model to identify three modulation formats. Compared with six deep learning methods including AlexNet, ResNet50, GoogleNet, SqueezeNet, Inception-v4, and Xception, it achieves the highest performance. This research provides a low-cost, low-complexity, and high-performance solution for signal transmission and signal recognition in high-speed optical networks designed for integrated communication and sensing. Full article
(This article belongs to the Section Optical Sensors)
Show Figures

Figure 1

Figure 1
<p>Architecture of a 4 × 100 Gbps Apol-CRZ-FSK signal transmission system for optical networks.</p>
Full article ">Figure 2
<p>Spectral diagram of a 4 × 100 Gbps signals: (<b>a</b>) Apol-CRZ-FSK; (<b>b</b>) CRZ-FSK; (<b>c</b>) DQPSK.</p>
Full article ">Figure 3
<p>The relation among SMF length, Q-factor, and launch power for the four wavelength channels of 4 × 100 Gbps Apol-CRZ-FSK signal transmission: (<b>a</b>) first channel; (<b>b</b>) second channel; (<b>c</b>) third channel; (<b>d</b>) last channel.</p>
Full article ">Figure 4
<p>The relation among SMF length, Q-factor and launch power for the four wavelength channels of 4 × 100 Gbps CRZ-FSK signal transmission: (<b>a</b>) first channel; (<b>b</b>) second channel; (<b>c</b>) third channel; (<b>d</b>) last channel.</p>
Full article ">Figure 5
<p>The relation among SMF length, Q-factor and launch power for the four wavelength channels of 4 × 100 Gbps DQPSK signal transmission: (<b>a</b>) first channel; (<b>b</b>) second channel; (<b>c</b>) third channel; (<b>d</b>) last channel.</p>
Full article ">Figure 6
<p>Performance analysis and comparison of three signals in different distances.</p>
Full article ">Figure 7
<p>Eye diagrams of the four-channel signals for the three types of signals at the launch power of 6 dBm and transmission distance of 1500 km: (<b>a</b>) first channel; (<b>b</b>) second channel; (<b>c</b>) third channel; (<b>d</b>) last channel.</p>
Full article ">Figure 8
<p>Model of the MFI method based on the Inception-ResNet-v2.</p>
Full article ">Figure 9
<p>Loss values for training and test sets.</p>
Full article ">Figure 10
<p>MFI confusion matrix for training and testing sets: (<b>a</b>) training set output confusion matrix; (<b>b</b>) testing set output confusion matrix.</p>
Full article ">Figure 11
<p>Effect of different factors on model MFI: (<b>a</b>) accuracy of the model at different number of rounds; (<b>b</b>) effect of different transmission distances on MFI; (<b>c</b>) effect of different signal-to-noise ratios on MFI.</p>
Full article ">Figure 12
<p>Comparative analysis of different modulation format recognition methods: (<b>a</b>) accuracy; (<b>b</b>) precision; (<b>c</b>) recall; (<b>d</b>) F1 score.</p>
Full article ">
Back to TopTop