Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (194)

Search Parameters:
Keywords = removal of background noise

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 4873 KiB  
Article
An Ensemble Approach for Speaker Identification from Audio Files in Noisy Environments
by Syed Shahab Zarin, Ehzaz Mustafa, Sardar Khaliq uz Zaman, Abdallah Namoun and Meshari Huwaytim Alanazi
Appl. Sci. 2024, 14(22), 10426; https://doi.org/10.3390/app142210426 - 13 Nov 2024
Viewed by 304
Abstract
Automatic noise-robust speaker identification is essential in various applications, including forensic analysis, e-commerce, smartphones, and security systems. Audio files containing suspect speech often include background noise, as they are typically not recorded in soundproof environments. To this end, we address the challenges of [...] Read more.
Automatic noise-robust speaker identification is essential in various applications, including forensic analysis, e-commerce, smartphones, and security systems. Audio files containing suspect speech often include background noise, as they are typically not recorded in soundproof environments. To this end, we address the challenges of noise robustness and accuracy in speaker identification systems. An ensemble approach is proposed combining two different neural network architectures including an RNN and DNN using softmax. This approach enhances the system’s ability to identify speakers even in noisy environments accurately. Using softmax, we combine voice activity detection (VAD) with a multilayer perceptron (MLP). The VAD component aims to remove noisy frames from the recording. The softmax function addresses these residual traces by assigning a higher probability to the speaker’s voice compared to the noise. We tested our proposed solution on the Kaggle speaker recognition dataset and compared it to two baseline systems. Experimental results show that our approach outperforms the baseline systems, achieving a 3.6% and 5.8% increase in test accuracy. Additionally, we compared the proposed MLP system with Long Short-Term Memory (LSTM) and Bidirectional LSTM (BiLSTM) classifiers. The results demonstrate that the MLP with VAD and softmax outperforms the LSTM by 23.2% and the BiLSTM by 6.6% in test accuracy. Full article
(This article belongs to the Special Issue Advances in Intelligent Information Systems and AI Applications)
Show Figures

Figure 1

Figure 1
<p>The proposed framework.</p>
Full article ">Figure 2
<p>Illustration of recurrent neural network.</p>
Full article ">Figure 3
<p>Illustration of deep neural network.</p>
Full article ">Figure 4
<p>The proposed MLP classifier.</p>
Full article ">Figure 5
<p>The LSTM network used.</p>
Full article ">Figure 6
<p>The BiLSTM model.</p>
Full article ">Figure 7
<p>The proposed framework compared with baselines in terms of spectrogram features.</p>
Full article ">Figure 8
<p>The proposed framework compared with baselines in terms of MFCC features.</p>
Full article ">Figure 9
<p>MLP model loss with different features.</p>
Full article ">Figure 10
<p>MLP model validation loss.</p>
Full article ">Figure 11
<p>MLP model accuracy.</p>
Full article ">Figure 12
<p>MLP model validation accuracy.</p>
Full article ">Figure 13
<p>Model accuracy of the three models.</p>
Full article ">Figure 14
<p>Validation accuracy comparison.</p>
Full article ">Figure 15
<p>Model loss of the three models.</p>
Full article ">Figure 16
<p>Validation loss of the three models.</p>
Full article ">Figure 17
<p>Model MSE of the three models.</p>
Full article ">Figure 18
<p>Validation MSE of the three models.</p>
Full article ">
13 pages, 3614 KiB  
Article
Automatic Defects Recognition of Lap Joint of Unequal Thickness Based on X-Ray Image Processing
by Dazhao Chi, Ziming Wang and Haichun Liu
Materials 2024, 17(22), 5463; https://doi.org/10.3390/ma17225463 - 8 Nov 2024
Viewed by 308
Abstract
It is difficult to automatically recognize defects using digital image processing methods in X-ray radiographs of lap joints made from plates of unequal thickness. The continuous change in the wall thickness of the lap joint workpiece causes very different gray levels in an [...] Read more.
It is difficult to automatically recognize defects using digital image processing methods in X-ray radiographs of lap joints made from plates of unequal thickness. The continuous change in the wall thickness of the lap joint workpiece causes very different gray levels in an X-ray background image. Furthermore, due to the shape and fixturing of the workpiece, the distribution of the weld seam in the radiograph is not vertical which results in an angle between the weld seam and the vertical direction. This makes automatic defect detection and localization difficult. In this paper, a method of X-ray image correction based on invariant moments is presented to solve the problem. In addition, a novel background removal method based on image processing is introduced to reduce the difficulty of defect recognition caused by variations in grayscale. At the same time, an automatic defect detection method combining image noise suppression, image segmentation, and mathematical morphology is adopted. The results show that the proposed method can effectively recognize the gas pores in an automatic welded lap joint of unequal thickness, making it suitable for automatic detection. Full article
Show Figures

Figure 1

Figure 1
<p>Preparation for weld specimen: (<b>a</b>) Geometric form of the joint, (<b>b</b>) Appearance of the weld specimen.</p>
Full article ">Figure 1 Cont.
<p>Preparation for weld specimen: (<b>a</b>) Geometric form of the joint, (<b>b</b>) Appearance of the weld specimen.</p>
Full article ">Figure 2
<p>Overall testing system and defect testing methods.</p>
Full article ">Figure 3
<p>Positioning of the weld under testing.</p>
Full article ">Figure 4
<p>Image correction steps.</p>
Full article ">Figure 5
<p>Digital image processing for defect detection.</p>
Full article ">Figure 6
<p>Background removal. (<b>a</b>) Cross-section of the lap joint. (<b>b</b>) Grayscale distribution of the radiograph. (<b>c</b>) Linear grayscale distribution without defect. (<b>d</b>) Linear grayscale distribution with defect. (<b>e</b>) Linear grayscale distribution of background. (<b>f</b>) Linear grayscale distribution of foreground.</p>
Full article ">Figure 7
<p>Image corrections: (<b>a</b>) Original radiograph, (<b>b</b>) Contour extraction, (<b>c</b>) Image correction, (<b>d</b>) Image corrected.</p>
Full article ">Figure 8
<p>Defect detection images: (<b>a</b>) Noise suppression, (<b>b</b>) Background image, (<b>c</b>) Foreground image, (<b>d</b>) Image segmentation, (<b>e</b>) Mathematical morphology.</p>
Full article ">
26 pages, 284813 KiB  
Article
Automatic Method for Detecting Deformation Cracks in Landslides Based on Multidimensional Information Fusion
by Bo Deng, Qiang Xu, Xiujun Dong, Weile Li, Mingtang Wu, Yuanzhen Ju and Qiulin He
Remote Sens. 2024, 16(21), 4075; https://doi.org/10.3390/rs16214075 - 31 Oct 2024
Viewed by 622
Abstract
As cracks are a precursor landslide deformation feature, they can provide forecasting information that is useful for the early identification of landslides and determining motion instability characteristics. However, it is difficult to solve the size effect and noise-filtering problems associated with the currently [...] Read more.
As cracks are a precursor landslide deformation feature, they can provide forecasting information that is useful for the early identification of landslides and determining motion instability characteristics. However, it is difficult to solve the size effect and noise-filtering problems associated with the currently available automatic crack detection methods under complex conditions using single remote sensing data sources. This article uses multidimensional target scene images obtained by UAV photogrammetry as the data source. Firstly, under the premise of fully considering the multidimensional image characteristics of different crack types, this article accomplishes the initial identification of landslide cracks by using six algorithm models with indicators including the roughness, slope, eigenvalue rate of the point cloud and pixel gradient, gray value, and RGB value of the images. Secondly, the initial extraction results are processed through a morphological repair task using three filtering algorithms (calculating the crack orientation, length, and frequency) to address background noise. Finally, this article proposes a multi-dimensional information fusion method, the Bayesian probability of minimum risk methods, to fuse the identification results derived from different models at the decision level. The results show that the six tested algorithm models can be used to effectively extract landslide cracks, providing Area Under the Curve (AUC) values between 0.6 and 0.85. After the repairing and filtering steps, the proposed method removes complex noise and minimizes the loss of real cracks, thus increasing the accuracy of each model by 7.5–55.3%. Multidimensional data fusion methods solve issues associated with the spatial scale effect during crack identification, and the F-score of the fusion model is 0.901. Full article
(This article belongs to the Topic Landslides and Natural Resources)
Show Figures

Figure 1

Figure 1
<p>Location of study area: (<b>a</b>) the location and traffic conditions of the study area on satellite images; (<b>b</b>) optical image of the landslide (photographed by UAV in May 2021); (<b>c</b>) main deformation area and DSM at the landslide site.</p>
Full article ">Figure 2
<p>Flight route and terrain products of the UAV operation in WuLiPo: (<b>a</b>) planned flight plane route and checkpoint positions; (<b>b</b>) FeiMa D200 drone; (<b>c</b>) terrain-following flight route; (<b>d</b>) DOM; (<b>e</b>) 3D point cloud.</p>
Full article ">Figure 3
<p>Results of the field investigation at Wulipo: (<b>a</b>) section A–A’ and material composition characteristics; (<b>b</b>) manual survey results of cracks; (<b>c</b>) on-site photos of the main cracks (numbers correspond to the shooting range of the black rectangular frame in (<b>b</b>)).</p>
Full article ">Figure 4
<p>Flow chart showing the automatic landslide crack detection process utilizing multidimensional data fusion.</p>
Full article ">Figure 5
<p>Schematic diagram representing the image pre-processing method.</p>
Full article ">Figure 6
<p>2D and 3D characteristics of landslide cracks with different scales: (<b>a</b>,<b>b</b>) the texture and morphology of the same landslide crack in the image and point cloud (the same numbered frames represent crack comparisons at the same location); (<b>c</b>) schematic diagram of the tensile crack formation; (<b>d</b>) schematic diagram of the shear crack formation. The base map of c and d is digitized from [<a href="#B3-remotesensing-16-04075" class="html-bibr">3</a>].</p>
Full article ">Figure 7
<p>Schematic diagram of the principle by which a K-D tree is used to search the local neighborhood in the point cloud and generate various crack-extraction indicators.</p>
Full article ">Figure 8
<p>Schematic diagram showing the crack edge threshold segmentation principle in which grayscale images and the Sobel gradient map are used: (<b>a</b>) grayscale image of a crack; (<b>b</b>) local grayscale feature of the crack and the background surface; (<b>c</b>) gradient feature of the crack image processed by the Sobel operator; and (<b>d</b>) edge binarization effect of the crack image.</p>
Full article ">Figure 9
<p>Object classification results derived based on maximum likelihood supervision: (<b>a</b>) image map and sampling points; (<b>b</b>) distribution of objects after the classification process; (<b>c</b>,<b>d</b>) spatial distribution and categories of sample pixels before and after the classification process, respectively.</p>
Full article ">Figure 10
<p>Process by which the crack binary image is repaired using the morphological closure operation: (<b>a</b>,<b>b</b>) cracks and background surfaces after binarization, respectively; (<b>c</b>,<b>d</b>) expanding and corroding effects of local crack pixels, respectively; and (<b>e</b>) repaired crack.</p>
Full article ">Figure 11
<p>Schematic diagram showing the principles of the crack-filtering method: (<b>a</b>) local image of crack binary image orientational filtering convolution (from the rectangular box in (<b>d</b>)); (<b>b</b>) eigenvalues after orientational filter convolution; (<b>c</b>) original image with cracks; (<b>d</b>) preliminary automatically extracted crack image; (<b>e</b>) crack identification image after orientation, frequency, and length filtering; (<b>f</b>) principle by which a single crack is clustered using DBSCAN; (<b>g</b>) local characteristics of cracks after clustering (from the rectangular box in (<b>d</b>)); and (<b>h</b>) local characteristics of cracks after orientation and frequency filtering (from the rectangular box in (<b>e</b>)).</p>
Full article ">Figure 12
<p>WuLiPo orthophoto image and 3D point cloud pre-processing results.</p>
Full article ">Figure 13
<p>Recognition results of the WuLiPo cracks obtained by each model: (<b>a</b>–<b>c</b>) calculation results of the point cloud roughness, eigenvalue ratios, and slope, respectively; (<b>d</b>–<b>f</b>) grid conversion results corresponding to panels (<b>a</b>–<b>c</b>); (<b>g</b>) binary image transformed by Sobel; (<b>h</b>) preprocessed grayscale image.</p>
Full article ">Figure 14
<p>WuLiPo image classification results derived based on maximum likelihood supervised learning: (<b>a</b>) Orthophoto and manually selected sampling locations; (<b>b</b>) distribution of sample categories after prediction.</p>
Full article ">Figure 15
<p>Crack pixel binary classification confusion matrix.</p>
Full article ">Figure 16
<p>ROC curve test results of each crack identification and classification model.</p>
Full article ">Figure 17
<p>Semantic segmentation results of cracks in WuLiPo derived using various models.</p>
Full article ">Figure 18
<p>Effects of the repairing and filtering processes on the initial extraction crack results of each model (The red areas in the image are the identified crack pixels).</p>
Full article ">Figure 19
<p>Statistical chart of TPR, FPR, and precision metrics of the crack extraction models before and after repair filtering.</p>
Full article ">Figure 20
<p>Crack identification results of Wulipo: (<b>a</b>) automatic detection results of gradient value segmentation and slope segmentation model; (<b>b</b>) manual investigation results.</p>
Full article ">Figure 21
<p>Distribution of the image fusion features and the posterior probability comparison results derived for WuLiPo cracks: (<b>a</b>) distribution of 64 fusion feature samples; (<b>b</b>) posterior probability values of 64 fusion feature samples, and the reference line with red font indicates that the fusion result has equal probability of cracks and non-cracks.</p>
Full article ">Figure 22
<p>Bayesian probability fusion results derived under different risk factors: (<b>a</b>–<b>f</b>) are the results of crack fusion recognition when <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>λ</mi> </mrow> <mrow> <mn>1</mn> <mo>,</mo> <mn>1</mn> </mrow> </msub> <mo>:</mo> <msub> <mrow> <mi>λ</mi> </mrow> <mrow> <mn>1,2</mn> </mrow> </msub> <mo>:</mo> <msub> <mrow> <mi>λ</mi> </mrow> <mrow> <mn>2</mn> <mo>,</mo> <mn>1</mn> </mrow> </msub> <mo>:</mo> <msub> <mrow> <mi>λ</mi> </mrow> <mrow> <mn>2</mn> <mo>,</mo> <mn>2</mn> </mrow> </msub> </mrow> </semantics></math>= (0:1.5:1:0), (0:1.2:1:0), (0:1:1:0), (0:1:2:0), (0:1:4:0), and (0:1:7:0), respectively. The red areas in the image are the identified crack pixels.</p>
Full article ">Figure 23
<p>Changes in model evaluation indicators under different risk ratios based on Bayesian probability fusion: (<b>a</b>–<b>d</b>) represent the changes of TPR, FPR, Precision, and F-score under different <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>λ</mi> </mrow> <mrow> <mn>12</mn> </mrow> </msub> <mo>:</mo> <msub> <mrow> <mi>λ</mi> </mrow> <mrow> <mn>21</mn> </mrow> </msub> </mrow> </semantics></math>, respectively.</p>
Full article ">
20 pages, 10803 KiB  
Article
Improved Early-Stage Maize Row Detection Using Unmanned Aerial Vehicle Imagery
by Lulu Xue, Minfeng Xing and Haitao Lyu
ISPRS Int. J. Geo-Inf. 2024, 13(11), 376; https://doi.org/10.3390/ijgi13110376 - 29 Oct 2024
Viewed by 424
Abstract
Monitoring row centerlines during early growth stages is essential for effective production management. However, detection becomes more challenging due to weed interference and crop row intersection in images. This study proposed an enhanced Region of Interest (ROI)-based approach for detecting early-stage maize rows. [...] Read more.
Monitoring row centerlines during early growth stages is essential for effective production management. However, detection becomes more challenging due to weed interference and crop row intersection in images. This study proposed an enhanced Region of Interest (ROI)-based approach for detecting early-stage maize rows. It integrated a modified green vegetation index with a dual-threshold algorithm for background segmentation. The median filtering algorithm was also selected to effectively remove most noise points. Next, an improved ROI-based feature point extraction method was used to eliminate residual noises and extract feature points. Finally, the least square method was employed to fit the row centerlines. The detection accuracy of the proposed method was evaluated using the unmanned aerial vehicle (UAV) image data set containing both regular and intersecting crop rows. The average detection accuracy of the proposed approach was between 0.456° and 0.789° (the angle between the fitted centerline and the expert line), depending on whether crop rows were regular/intersecting. Compared to the Hough Transform (HT) algorithm, the results demonstrated that the proposed method achieved higher accuracy and robustness in detecting regular and intersecting crop rows. The proposed method in this study is helpful for refined agricultural management such as fertilization and irrigation. Additionally, it can detect the missing-seedling regions and replenish seedings in time to increase crop yields. Full article
Show Figures

Figure 1

Figure 1
<p>The location of the study area.</p>
Full article ">Figure 2
<p>Crop row images under different intersecting situations: (<b>a</b>) regular crop row; (<b>b</b>–<b>d</b>) intersecting crop rows of different types.</p>
Full article ">Figure 3
<p>The determination of micro-ROIs. The green box is a micro-ROI. The green cross point, the red cross point at the bottom, and the red cross point at the top denote the centroid, the starting point of this micro-ROI, and the starting point of the next micro-ROI.</p>
Full article ">Figure 4
<p>The flowchart of the improved ROI-based feature point extraction method.</p>
Full article ">Figure 5
<p>The counting process of intersecting crop rows: (<b>a</b>) grayscale images obtained with <span class="html-italic">ExGG</span>; (<b>b</b>) binary images; (<b>c</b>) the horizontal projection graph, where the red line represents the line with a threshold of T/2; (<b>d</b>) the threshold curve graph, where the red curve represents the threshold curve.</p>
Full article ">Figure 6
<p>The process for extracting intersecting points and detecting intersecting crop rows: (<b>a</b>) the division of horizontal strips, where the red box represents the horizontal strip; (<b>b</b>) the determination of intersecting points for every intersecting crop row; (<b>c</b>) the location of all intersecting points in the image, where pentagram points represent intersecting points; (<b>d</b>) the elimination of interference pixels between adjacent crop rows; (<b>e</b>) the area division and crop row detection, where green lines represent crop row centerlines fitted.</p>
Full article ">Figure 7
<p>The flowchart of the missing-seedling application: (<b>a</b>) the missing-seedling region detection method; (<b>b</b>) the seeding replenishment method.</p>
Full article ">Figure 8
<p>Image preprocessing: (<b>a</b>) grayscale image obtained with <span class="html-italic">ExGG</span>; (<b>b</b>) binary image obtained by OTSU algorithm; (<b>c</b>) binary image obtained by PSO-OTSU algorithm; (<b>d</b>) image obtained by median filtering algorithm.</p>
Full article ">Figure 9
<p>Crop row centerlines fitted by ROI-based feature point extraction method. Red lines are expert lines. Green cross points and yellow lines were obtained using ROI-based feature point extraction method. Points in the red boxes represent more significant noise points.</p>
Full article ">Figure 10
<p>The comparison of crop row centerlines before and after the improvement method. Green cross points were obtained by the proposed method. Red, green, and yellow lines denote expert lines, ROI-based feature point extraction method outcomes, and the proposed method results, respectively.</p>
Full article ">Figure 11
<p>The evaluation index increase chart. The <math display="inline"><semantics> <mrow> <mi>θ</mi> </mrow> </semantics></math>, <span class="html-italic">X<sub>mean</sub></span>, and <span class="html-italic">R</span><sup>2</sup> denote the detection accuracy, the average lateral position deviation, and the determination coefficient of crop row centerlines fitted, respectively. <span class="html-italic">A</span>–<span class="html-italic">G</span> denote row centerlines fitted. The <span class="html-italic">Average</span> and <span class="html-italic">SD</span> denote evaluation index parameters.</p>
Full article ">Figure 12
<p>Intersecting crop row centerlines: (<b>a</b>–<b>c</b>) different types of intersecting crop row centerlines. Red, green, and black lines are expert lines, the proposed method results, and the ROI-based feature point extraction method outcomes, respectively.</p>
Full article ">Figure 13
<p>The detection result of different types of crop rows: (<b>a</b>) regular crop row centerlines; (<b>b</b>–<b>d</b>) intersecting crop row centerlines of different types. Red, green, and yellow lines are expert lines, the proposed method results, and the HT algorithm outcomes, respectively.</p>
Full article ">Figure 14
<p>Accuracy line charts of crop rows in <a href="#ijgi-13-00376-f013" class="html-fig">Figure 13</a>. (<b>a</b>–<b>d</b>) represent accuracy line diagram of the fitted crop row centerlines in <a href="#ijgi-13-00376-f013" class="html-fig">Figure 13</a>a–d, respectively. The <span class="html-italic">Average</span>, <span class="html-italic">SD</span>, and <span class="html-italic">Max</span> denote evaluation indicators.</p>
Full article ">Figure 15
<p>The missing-seedling regions and the seeding replenishment locations. Green lines, red lines, and red cross points indicate row centerlines obtained by the proposed method, the missing-seedling regions, and the replanting positions of the seedlings, respectively.</p>
Full article ">
27 pages, 6293 KiB  
Article
Lightweight Advanced Deep Neural Network (DNN) Model for Early-Stage Lung Cancer Detection
by Isha Bhatia, Aarti, Syed Immamul Ansarullah, Farhan Amin and Amerah Alabrah
Diagnostics 2024, 14(21), 2356; https://doi.org/10.3390/diagnostics14212356 - 22 Oct 2024
Viewed by 597
Abstract
Background: Lung cancer, also known as lung carcinoma, has a high mortality rate; however, an early prediction helps to reduce the risk. In the current literature, various approaches have been developed for the prediction of lung carcinoma (at an early stage), but these [...] Read more.
Background: Lung cancer, also known as lung carcinoma, has a high mortality rate; however, an early prediction helps to reduce the risk. In the current literature, various approaches have been developed for the prediction of lung carcinoma (at an early stage), but these still have various issues, such as low accuracy, high noise, low contrast, poor recognition rates, and a high false-positive rate, etc. Thus, in this research effort, we have proposed an advanced algorithm and combined two different types of deep neural networks to make it easier to spot lung melanoma in the early phases. Methods: We have used WDSI (weakly supervised dense instance-level lung segmentation) for laborious pixel-level annotations. In addition, we suggested an SS-CL (deep continuous learning-based deep neural network) that can be applied to the labeled and unlabeled data to improve efficiency. This work intends to evaluate potential lightweight, low-memory deep neural net (DNN) designs for image processing. Results: Our experimental results show that, by combining WDSI and LSO segmentation, we can achieve super-sensitive, specific, and accurate early detection of lung cancer. For experiments, we used the lung nodule (LUNA16) dataset, which consists of the patients’ 3D CT scan images. We confirmed that our proposed model is lightweight because it uses less memory. We have compared them with state-of-the-art models named PSNR and SSIM. The efficiency is 32.8% and 0.97, respectively. The proposed lightweight deep neural network (DNN) model archives a high accuracy of 98.2% and also removes noise more effectively. Conclusions: Our proposed approach has a lot of potential to help medical image analysis to help improve the accuracy of test results, and it may also prove helpful in saving patients’ lives. Full article
(This article belongs to the Special Issue Artificial Intelligence in Cancers—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Block schematic of the proposed model.</p>
Full article ">Figure 2
<p>Dataset utilization for cancer detection.</p>
Full article ">Figure 3
<p>Dataset Utilization for Cancerous Image. (<b>a</b>) Large-Cell Carcinoma; (<b>b</b>) Squamous Cell Carcinoma.</p>
Full article ">Figure 4
<p>Sample images in the proposed model. (<b>A</b>,<b>B</b>) The top and bottom images show the original CT sample images and the contrast-stretched images. (<b>C</b>,<b>D</b>) The top and bottom images show the edge enhancement image and segmentation. (<b>E</b>) The images show the classified output (large-cell cancer, squamous cell cancer, and normal).</p>
Full article ">Figure 4 Cont.
<p>Sample images in the proposed model. (<b>A</b>,<b>B</b>) The top and bottom images show the original CT sample images and the contrast-stretched images. (<b>C</b>,<b>D</b>) The top and bottom images show the edge enhancement image and segmentation. (<b>E</b>) The images show the classified output (large-cell cancer, squamous cell cancer, and normal).</p>
Full article ">Figure 5
<p>Left and right with cancer and highlighted cancerous image.</p>
Full article ">Figure 6
<p>Nodule Detection Confusion Matrix.</p>
Full article ">Figure 7
<p>Risk Assessment Confusion Matrix.</p>
Full article ">Figure 8
<p>Performance Measures for Risk Assessment and Nodule Detection.</p>
Full article ">Figure 9
<p>Comparative analysis of the nodule detection accuracy of the proposed model with different techniques.</p>
Full article ">Figure 10
<p>Comparative analysis of the risk assessment accuracy in the proposed model with different techniques.</p>
Full article ">Figure 11
<p>The assessment of the proposed and new models with SSIM performance measures.</p>
Full article ">Figure 12
<p>Evaluation of the suggested models and recent ones using the performance metrics ET.</p>
Full article ">Figure 13
<p>Evaluation of the suggested models and recent ones using the performance metrics (PSNR).</p>
Full article ">
22 pages, 2339 KiB  
Article
Signal Acquisition and Algorithm Design for Bioimpedance-Based Heart Rate Estimation from the Wrist
by Didzis Lapsa, Margus Metshein, Andrei Krivošei, Rims Janeliukstis, Olev Märtens and Atis Elsts
Appl. Sci. 2024, 14(21), 9632; https://doi.org/10.3390/app14219632 - 22 Oct 2024
Viewed by 697
Abstract
Background: Heart rate (HR) is a critical biomarker that provides insights into overall health, stress levels, and the autonomic nervous system. Pulse wave signals contain valuable information about the cardiovascular system and heart status. However, signal acquisition in wearables poses challenges, particularly when [...] Read more.
Background: Heart rate (HR) is a critical biomarker that provides insights into overall health, stress levels, and the autonomic nervous system. Pulse wave signals contain valuable information about the cardiovascular system and heart status. However, signal acquisition in wearables poses challenges, particularly when using electrical sensors, due to factors like the distance from the heart, body movement, and suboptimal electrode placement. Methods: Electrical bioimpedance (EBI) measurements using bipolar and tetrapolar electrode systems were employed for pulse wave signal acquisition from the wrist in both perpendicular and distal configurations. Signal preprocessing techniques, including baseline removal via Hankel matrix methods, normalization, cross-correlation, and peak detection, were applied to improve signal quality. This study describes the combination of sensor-level signal acquisition and processing for accurate wearable HR estimation. Results: The bipolar system was shown to produce larger ΔZ(t), while the tetrapolar system demonstrated higher sensitivity. Distal placement of the electrodes yielded greater ΔZ(t) (up to 0.231 Ω) when targeting both wrist arteries. Bandpass filtering resulted in a better signal-to-noise ratio (SNR), achieving 3.6 dB for the best bipolar setup and 4.8 dB for the tetrapolar setup, compared to 2.6 and 3.3 dB SNR, respectively, with the Savitzky–Golay filter. The custom HR estimation algorithm presented in this paper demonstrated improved accuracy over a reference method, achieving an average error of 1.8 beats per minute for the best bipolar setup, with a mean absolute percentage error (MAPE) of 8%. Conclusions: The analysis supports the feasibility of using bipolar electrode setups on the wrist and highlights the importance of electrode positioning relative to the arteries. The proposed signal processing method, featuring a preprocessing pipeline and HR estimation algorithm, provides a proof-of-concept demonstration for HR estimation from EBI signals acquired at the wrist. Full article
(This article belongs to the Special Issue Robotics, IoT and AI Technologies in Bioengineering)
Show Figures

Figure 1

Figure 1
<p>Origin of the pulse wave-related change in the case of measuring EBI of a single artery, illustrating the pulse wave as a bubble of blood.</p>
Full article ">Figure 2
<p>Electrode contact area sizes (10 mm width on the left and 17 mm width on the right).</p>
Full article ">Figure 3
<p>Chosen electrode placements on the forearm: perpendicular placement of electrodes with 10 mm width (<b>a</b>) and with 17 mm width (<b>b</b>); distal placement of electrodes with 10 mm width (<b>c</b>) and with 17 mm width (<b>d</b>).</p>
Full article ">Figure 4
<p>EBI-based pulse waveform with illustrated definition of the modulation depth.</p>
Full article ">Figure 5
<p>Flow diagram describing the data processing steps for accurate heart rate estimation.</p>
Full article ">Figure 6
<p>The raw EBI signal (<b>top panel</b>) and the results after applying a band-pass and a Savitzky–Golay filter (<b>the bottom panel</b>).</p>
Full article ">Figure 7
<p>Example of EBI signal partition for the window-based peak detection (<b>upper plot</b>) and the peaks of the detected pulse waves (<b>bottom plot</b>).</p>
Full article ">Figure 8
<p>Examples of calculated BPM variations from good quality (high SNR) electrical bioimpedance (EBI) signals (<b>left plot</b>) and calculated BPM values from noisy (low SNR) EBI signals (<b>right plot</b>) while using the PPG signals for reference information.</p>
Full article ">Figure 9
<p>Calculated mean values of S in the cases of the chosen electrode configurations.</p>
Full article ">Figure 10
<p>Signal quality estimation results shown as a scatter plot in BPM difference—SNR plane for both signal filtering methods—band-pass signal and Savitzky–Golay. (<b>a</b>) The ‘findpeaks’ BPM estimation method; (<b>b</b>) the ‘BPM DL’ BPM estimation method. Diamond symbols show centroids of point clouds for both filtering methods and straight lines show the fitted linear regression model according to the color of each dataset.</p>
Full article ">Figure 11
<p>SNR [dB] colormap across all electrode configurations for three test subjects. The table visualizes SNR where higher values are in green and lower values in red (with a gradient transition for intermediate values between these extremes).</p>
Full article ">Figure 12
<p>Absolute BPM estimation errors, as quantified by the <math display="inline"><semantics> <msub> <mi>DIFF</mi> <mi>BPM</mi> </msub> </semantics></math> metric. Shows results across all electrode configurations for three test subjects. Higher errors are shown in red and lower differences in green, using a color gradient for values in-between.</p>
Full article ">Figure 13
<p>Relative BPM estimation errors, as quantified by the MAPE metric. Shows results across all electrode configurations for three test subjects. Higher errors are shown in red and lower differences in green, using a color gradient for values in-between.</p>
Full article ">
21 pages, 12855 KiB  
Article
Noise Study Auralization of an Open-Rotor Engine
by Qing Zhang, Siyi Jiang, Xiaojun Yang, Yongjia Xu and Maosheng Zhu
Aerospace 2024, 11(10), 857; https://doi.org/10.3390/aerospace11100857 - 17 Oct 2024
Viewed by 620
Abstract
Based on the performance and acoustic data files of reduced-size open-rotor engines in low-speed wind tunnels, the static sound pressure level was derived by converting the 1-foot lossless spectral density into sound-pressure-level data, the background noise was removed, and the results were corrected [...] Read more.
Based on the performance and acoustic data files of reduced-size open-rotor engines in low-speed wind tunnels, the static sound pressure level was derived by converting the 1-foot lossless spectral density into sound-pressure-level data, the background noise was removed, and the results were corrected according to the environmental parameters of the low-speed wind tunnels. In accordance with the requirements of Annex 16 of the Convention on International Civil Aviation Organization and Part 36 of the Civil Aviation Regulations of China on noise measurement procedures, the takeoff trajectory was physically modeled; the static noise source was mapped onto the takeoff trajectory to simulate the propagation process of the noise during takeoff; and the 24 one-third-octave center frequencies that corresponded to the SPL data were corrected for geometrical dispersion, atmospheric absorption, and Doppler effects, so that the takeoff noise could be corrected to represent a real environment. In addition, the audible processing of noise data with a 110° source pointing angle was achieved, which can be useful for enabling practical observers to analyze the noise characteristics. Full article
(This article belongs to the Section Aeronautics)
Show Figures

Figure 1

Figure 1
<p>The new-generation open-rotor engine configuration.</p>
Full article ">Figure 2
<p>Aircraft-noise-monitoring points for noise airworthiness requirements.</p>
Full article ">Figure 3
<p>Full- and reduced-thrust takeoff trajectories.</p>
Full article ">Figure 4
<p>Example of reduced-thrust takeoff trajectory noise source localization.</p>
Full article ">Figure 5
<p>Attenuation curve of the geometric dispersion effect.</p>
Full article ">Figure 6
<p>Atmospheric absorption sound attenuation curve.</p>
Full article ">Figure 7
<p>Doppler effect curve (The intersection of the blue dotted line and the Doppler effect curve is the angle at which the sound pressure level attenuation is 0).</p>
Full article ">Figure 8
<p>Source synthesis, propagation path and receiver setting.</p>
Full article ">Figure 9
<p>Broadband and monophonic filtering results.</p>
Full article ">Figure 10
<p>Broadband synthesis.</p>
Full article ">Figure 11
<p>Flight path simulated in a 3D virtual environment.</p>
Full article ">Figure 12
<p>The 110° angle noise-data audio realization point (at the red dot).</p>
Full article ">Figure 13
<p>Wind tunnel environment simulation.</p>
Full article ">Figure 14
<p>Acoustic measurement position.</p>
Full article ">Figure 15
<p>Mixed-reality environment visualization.</p>
Full article ">Figure 16
<p>Labeling of the test.</p>
Full article ">Figure 17
<p>Dynamic bar graph of the mixed-reality environment.</p>
Full article ">
22 pages, 749 KiB  
Article
Improving Distantly Supervised Relation Extraction with Multi-Level Noise Reduction
by Wei Song and Zijiang Yang
AI 2024, 5(3), 1709-1730; https://doi.org/10.3390/ai5030084 - 23 Sep 2024
Viewed by 723
Abstract
Background: Distantly supervised relation extraction (DSRE) aims to identify semantic relations in large-scale texts automatically labeled via knowledge base alignment. It has garnered significant attention due to its high efficiency, but existing methods are plagued by noise at both the word and [...] Read more.
Background: Distantly supervised relation extraction (DSRE) aims to identify semantic relations in large-scale texts automatically labeled via knowledge base alignment. It has garnered significant attention due to its high efficiency, but existing methods are plagued by noise at both the word and sentence level and fail to address these issues adequately. The former level of noise arises from the large proportion of irrelevant words within sentences, while noise at the latter level is caused by inaccurate relation labels for various sentences. Method: We propose a novel multi-level noise reduction neural network (MLNRNN) to tackle both issues by mitigating the impact of multi-level noise. We first build an iterative keyword semantic aggregator (IKSA) to remove noisy words, and capture distinctive features of sentences by aggregating the information of keywords. Next, we implement multi-objective multi-instance learning (MOMIL) to reduce the impact of incorrect labels in sentences by identifying the cluster of correctly labeled instances. Meanwhile, we leverage mislabeled sentences with cross-level contrastive learning (CCL) to further enhance the classification capability of the extractor. Results: Comprehensive experimental results on two DSRE benchmark datasets demonstrated that the MLNRNN outperformed state-of-the-art methods for distantly supervised relation extraction in almost all cases. Conclusions: The proposed MLNRNN effectively addresses both word- and sentence-level noise, providing a significant improvement in relation extraction performance under distant supervision. Full article
(This article belongs to the Section AI Systems: Theory and Applications)
Show Figures

Figure 1

Figure 1
<p>Overall framework of the proposed multi-level noise reduction model.The dashed arrows indicate the utilization of representations from other bags in CCL. Different colors are used to distinguish between the variations in the input sentence embeddings.</p>
Full article ">Figure 2
<p>The detailed structure of the IKSA, illustrating the procedure for handling a sentence, and the dashed part represents the iterative step of this module.</p>
Full article ">Figure 3
<p>Overview of multi-objective multi-instance learning. The sentence representations highlighted in blue represent the selected true instances, while the others are false instances.</p>
Full article ">Figure 4
<p>PR curves of the MLNRNN and the competitors for the dataset NYT-10.</p>
Full article ">Figure 5
<p>PR curves of the MLNRNN and the competitors for the dataset NYT-16.</p>
Full article ">Figure 6
<p>PR curves of the MLNRNN and the competitors for the dataset Wiki-20m.</p>
Full article ">Figure 7
<p>PR curves of the MLNRNN and three ablation methods on the NYT-10 dataset.</p>
Full article ">Figure 8
<p>The performance with different thresholds.</p>
Full article ">
18 pages, 5556 KiB  
Article
Paper-Recorded ECG Digitization Method with Automatic Reference Voltage Selection for Telemonitoring and Diagnosis
by Liang-Hung Wang, Chao-Xin Xie, Tao Yang, Hong-Xin Tan, Ming-Hui Fan, I-Chun Kuo, Zne-Jung Lee, Tsung-Yi Chen, Pao-Cheng Huang, Shih-Lun Chen and Patricia Angela R. Abu
Diagnostics 2024, 14(17), 1910; https://doi.org/10.3390/diagnostics14171910 - 29 Aug 2024
Viewed by 721
Abstract
In electrocardiograms (ECGs), multiple forms of encryption and preservation formats create difficulties for data sharing and retrospective disease analysis. Additionally, photography and storage using mobile devices are convenient, but the images acquired contain different noise interferences. To address this problem, a suite of [...] Read more.
In electrocardiograms (ECGs), multiple forms of encryption and preservation formats create difficulties for data sharing and retrospective disease analysis. Additionally, photography and storage using mobile devices are convenient, but the images acquired contain different noise interferences. To address this problem, a suite of novel methodologies was proposed for converting paper-recorded ECGs into digital data. Firstly, this study ingeniously removed gridlines by utilizing the Hue Saturation Value (HSV) spatial properties of ECGs. Moreover, this study introduced an innovative adaptive local thresholding method with high robustness for foreground–background separation. Subsequently, an algorithm for the automatic recognition of calibration square waves was proposed to ensure consistency in amplitude, rather than solely in shape, for digital signals. The original signal reconstruction algorithm was validated with the MIT–BIH and PTB databases by comparing the difference between the reconstructed and the original signals. Moreover, the mean of the Pearson correlation coefficient was 0.97 and 0.98, respectively, while the mean absolute errors were 0.324 and 0.241, respectively. The method proposed in this study converts paper-recorded ECGs into a digital format, enabling direct analysis using software. Automated techniques for acquiring and restoring ECG reference voltages enhance the reconstruction accuracy. This innovative approach facilitates data storage, medical communication, and remote ECG analysis, and minimizes errors in remote diagnosis. Full article
(This article belongs to the Special Issue Recent Advances in Cardiac Imaging: 2024)
Show Figures

Figure 1

Figure 1
<p>Scheme of paper-recorded ECG digitization system for telemonitoring and diagnosis.</p>
Full article ">Figure 2
<p>Targeted color adjustment algorithm: the original ECG image is converted and split into hue (H), saturate (S), and value (V) channels. The difference between the grid and the ECG curves is expanded at the most moderate level.</p>
Full article ">Figure 3
<p>The image was divided into 4 × 4 equal parts and each part has a local threshold value.</p>
Full article ">Figure 4
<p>The recognition of the reference square wave based on the Butterworth filter. (<b>a</b>) The original signal; (<b>b</b>) The signal after second-order Butterworth high-pass filter with a cutoff frequency of 100 Hz; (<b>c</b>) The signal after threshold processing and peak detection.</p>
Full article ">Figure 5
<p>Schematic of benchmark square pulse. The states 1, 2, 3, and 4 refer to the first low-level voltage segment, the high-level voltage segment, the second low-level voltage segment of the square wave, and the ECG signal segment, respectively.</p>
Full article ">Figure 6
<p>Results for geometric distortion correction. (<b>a</b>) ECG photo taken at random; (<b>b</b>) ECG photo taken with white background; (<b>c</b>) ECG photo taken in a complex background.</p>
Full article ">Figure 7
<p>Results for ECG image binarization. The image is taken in different environments. (<b>a</b>) photo of common ECG paper saved in patient; (<b>b</b>) photo taken from a cardiology book.</p>
Full article ">Figure 8
<p>Results for ECG image binarization. The image is taken in different environments. (<b>a</b>) photo with strong uneven illumination interference; (<b>b</b>) photo of clinical 12-lead ECG thermal paper.</p>
Full article ">Figure 9
<p>Validation scheme for 1D ECG signal reconstruction algorithm.</p>
Full article ">Figure 10
<p>Result of comparison between reconstruction signal and original signal.</p>
Full article ">
21 pages, 13401 KiB  
Article
Virtual Restoration of Ancient Mold-Damaged Painting Based on 3D Convolutional Neural Network for Hyperspectral Image
by Sa Wang, Yi Cen, Liang Qu, Guanghua Li, Yao Chen and Lifu Zhang
Remote Sens. 2024, 16(16), 2882; https://doi.org/10.3390/rs16162882 - 7 Aug 2024
Cited by 1 | Viewed by 1253
Abstract
Painted cultural relics hold significant historical value and are crucial in transmitting human culture. However, mold is a common issue for paper or silk-based relics, which not only affects their preservation and longevity but also conceals the texture, patterns, and color information, hindering [...] Read more.
Painted cultural relics hold significant historical value and are crucial in transmitting human culture. However, mold is a common issue for paper or silk-based relics, which not only affects their preservation and longevity but also conceals the texture, patterns, and color information, hindering cultural value and heritage. Currently, the virtual restoration of painting relics primarily involves filling in the RGB based on neighborhood information, which might cause color distortion and other problems. Another approach considers mold as noise and employs maximum noise separation for its removal; however, eliminating the mold components and implementing the inverse transformation often leads to more loss of information. To effectively acquire virtual restoration for mold removal from ancient paintings, the spectral characteristics of mold were analyzed. Based on the spectral features of mold and the cultural relic restoration philosophy of maintaining originality, a 3D CNN artifact restoration network was proposed. This network is capable of learning features in the near-infrared spectrum (NIR) and spatial dimensions to reconstruct the reflectance of visible spectrum, achieving the virtual restoration for mold removal of calligraphic and art relics. Using an ancient painting from the Qing Dynasty as a test subject, the proposed method was compared with the Inpainting, Criminisi, and inverse MNF transformation methods across three regions. Visual analysis, quantitative evaluation (the root mean squared error (RMSE), mean absolute percentage error (MAPE), mean absolute error (MEA), and a classification application were used to assess the restoration accuracy. The visual results and quantitative analyses demonstrated that the proposed 3D CNN method effectively removes or mitigates mold while restoring the artwork to its authentic color in various backgrounds. Furthermore, the color classification results indicated that the images restored with 3D CNN had the highest classification accuracy, with overall accuracies of 89.51%, 92.24%, and 93.63%, and Kappa coefficients of 0.88, 0.91, and 0.93, respectively. This research provides technological support for the digitalization and restoration of cultural artifacts, thereby contributing to the preservation and transmission of cultural heritage. Full article
Show Figures

Figure 1

Figure 1
<p>The original true-color visualizations of hyperspectral data of (<b>a</b>) Lady’s picture, (<b>b</b>) Clothes’ picture, and (<b>c</b>) Branch’s picture (R: 650.02 nm, G: 539.59 nm, B: 477.88 nm, 400 × 400 pixels, the region inside the yellow box is where mold was mainly concentrated, and in figure (<b>d</b>–<b>f</b>), the red dots indicate selected mold and the black dots indicate selected non-mold in <a href="#sec2dot2-remotesensing-16-02882" class="html-sec">Section 2.2</a>).</p>
Full article ">Figure 2
<p>HS-VN/SW2500CR Heritage Spectral Imaging System.</p>
Full article ">Figure 3
<p>The reflectance figures of (<b>a</b>–<b>d</b>) Lady’s picture, (<b>e</b>–<b>h</b>) Clothes’ picture, (<b>i</b>–<b>l</b>) Branch’s picture at 534 nm, 720 nm, 830 nm and 930 nm (the region inside the yellow box is where mold was mainly concentrated).</p>
Full article ">Figure 3 Cont.
<p>The reflectance figures of (<b>a</b>–<b>d</b>) Lady’s picture, (<b>e</b>–<b>h</b>) Clothes’ picture, (<b>i</b>–<b>l</b>) Branch’s picture at 534 nm, 720 nm, 830 nm and 930 nm (the region inside the yellow box is where mold was mainly concentrated).</p>
Full article ">Figure 4
<p>Spectral characteristics (<b>a</b>,<b>c</b>) and removed enveloped reflectance (<b>b</b>,<b>d</b>) of mold and background in Region 1 Lady’s picture, Region 2 Clothes’ picture, and Region 3 Branch’s picture.</p>
Full article ">Figure 5
<p>Flowchart of the 3D CNN network for restoration of the mold regions.</p>
Full article ">Figure 6
<p>The results of Region 1 Lady’s picture (<b>a</b>) original, (<b>b</b>) mold region, (<b>c</b>) Inpainting (<b>d</b>) Criminisi (<b>e</b>) Inverse MNF, and (<b>f</b>) the proposed 3D CNN.</p>
Full article ">Figure 7
<p>The results of Region 2 Clothes’ picture (<b>a</b>) original, (<b>b</b>) mold region, (<b>c</b>) Inpainting (<b>d</b>) Criminisi (<b>e</b>) Inverse MNF, and (<b>f</b>) the proposed 3D CNN.</p>
Full article ">Figure 8
<p>The results of Region 3 Branch’s picture (<b>a</b>) original, (<b>b</b>) mold region, (<b>c</b>) Inpainting (<b>d</b>) Criminisi (<b>e</b>) Inverse MNF, and (<b>f</b>) the proposed 3D CNN.</p>
Full article ">Figure 9
<p>The classification results of Region 1 Lady’s picture are (<b>a</b>) original, (<b>b</b>) Inverse MNF, (<b>c</b>) Inpainting (<b>d</b>) Criminisi, and (<b>e</b>) the proposed 3D CNN methods.</p>
Full article ">Figure 10
<p>The classification results of region 2 Clothes’ picture (<b>a</b>) original, (<b>b</b>) Inverse MNF, (<b>c</b>) Inpainting (<b>d</b>) Criminisi, and (<b>e</b>) the proposed 3D CNN methods.</p>
Full article ">Figure 11
<p>The classification results of Region 3 Branch’s picture are (<b>a</b>) original, (<b>b</b>) Inverse MNF, (<b>c</b>) Inpainting (<b>d</b>) Criminisi, and (<b>e</b>) the proposed 3D CNN methods.</p>
Full article ">
27 pages, 10826 KiB  
Article
CRLNet: A Multimodal Peach Detection Network Based on Cooperative Asymptotic Enhancement and the Fusion of Granularity Refinement
by Jiahao Liu, Chaoying He, Mingfang Wang, Yichu Jiang, Manman Sun, Miying Yan and Mingfang He
Plants 2024, 13(14), 1980; https://doi.org/10.3390/plants13141980 - 19 Jul 2024
Viewed by 779
Abstract
Accurate peach detection is essential for automated agronomic management, such as mechanical peach harvesting. However, ubiquitous occlusion makes identifying peaches from complex backgrounds extremely challenging. In addition, it is difficult to capture fine-grained peach features from a single RGB image, which can suffer [...] Read more.
Accurate peach detection is essential for automated agronomic management, such as mechanical peach harvesting. However, ubiquitous occlusion makes identifying peaches from complex backgrounds extremely challenging. In addition, it is difficult to capture fine-grained peach features from a single RGB image, which can suffer from light and noise in scenarios with dense small target clusters and extreme light. To solve these problems, this study proposes a multimodal detector, called CRLNet, based on RGB and depth images. First, YOLOv9 was extended to design a backbone network that can extract RGB and depth features in parallel from an image. Second, to address the problem of information fusion bias, the Rough–Fine Hybrid Attention Fusion Module (RFAM) was designed to combine the advantageous information of different modes while suppressing the hollow noise at the edge of the peach. Finally, a Transformer-based Local–Global Joint Enhancement Module (LGEM) was developed to jointly enhance the local and global features of peaches using information from different modalities in order to enhance the percentage of information about the target peaches and remove the interference of redundant background information. CRLNet was trained on the Peach dataset and evaluated against other state-of-the-art methods; the model achieved an mAP50 of 97.1%. In addition, CRLNet also achieved an mAP50 of 92.4% in generalized experiments, validating its strong generalization capability. These results provide valuable insights for peach and other outdoor fruit multimodal detection. Full article
Show Figures

Figure 1

Figure 1
<p>Difficult scenes for peach detection: (<b>a</b>) with normal light, (<b>b</b>) with dark light at night, and (<b>c</b>) with the presence of glare. (<b>d</b>–<b>f</b>) are depth images of the same scene in (<b>a</b>–<b>c</b>).</p>
Full article ">Figure 2
<p>(<b>a</b>) The overall structure of CRLNet. The red line indicates the RGB information flow, the blue line indicates the depth information flow, the black line indicates the fusion information flow, the solid line indicates the main branch information flow, and the dashed line indicates the auxiliary branch information flow. (<b>b</b>) Overall structure of the LGEM. (<b>c</b>) Overall structure of the RFAM. (<b>d</b>) Workflow for using CRLNet in orchards.</p>
Full article ">Figure 3
<p>The distribution of data in the training dataset. (<b>a</b>) The number and size of different categories of peaches in the dataset. (<b>b</b>) Modeling of the correlation between tags using the target detection algorithm during training, where the darker the color, the higher the correlation.</p>
Full article ">Figure 4
<p>Comparison of the detection results for different combinations of modes. (<b>a</b>) normal brightness; (<b>b</b>) dense peach; (<b>c</b>) glare interference; (<b>d</b>) dark light.</p>
Full article ">Figure 5
<p>Extracted feature maps with different input information, where warmer colored regions indicate that the network is focusing more attention on that region and the opposite is true for cooler colored regions.</p>
Full article ">Figure 6
<p>P and R for different classes of CRLNet at different resolutions.</p>
Full article ">Figure 7
<p>Comparison of P and R metrics for different sizes of detectors.</p>
Full article ">Figure 8
<p><span class="html-italic">P</span>, <span class="html-italic">R</span>, and <span class="html-italic">F1</span> transformation curves for different methods at different confidence levels. where the horizontal coordinate of each subplot indicates the specific value of the confidence level and the vertical coordinate indicates the specific value of the indicator. Different colors indicate different categories, and the data in the legend indicate the best indicator at the best confidence level.</p>
Full article ">Figure 9
<p>Plot of the training loss variation for different data combinations and different module combinations.</p>
Full article ">Figure 10
<p>Comparison of the accuracy (P) and recall (R) of each detection model for different occlusion categories of peaches. The four color regions from left to right in each sub-figure correspond to depth-only, IR-only, RGB-only, and mixed.</p>
Full article ">Figure 11
<p>Visualization of the state-of-the-art single- and dual-stream detection algorithms on the peach dataset: (<b>a</b>) normal brightness; (<b>b</b>) dense peach; (<b>c</b>) dark light; (<b>d</b>) glare interference.</p>
Full article ">
17 pages, 10905 KiB  
Article
Complementary-View SAR Target Recognition Based on One-Shot Learning
by Benteng Chen, Zhengkang Zhou, Chunyu Liu and Jia Zheng
Remote Sens. 2024, 16(14), 2610; https://doi.org/10.3390/rs16142610 - 17 Jul 2024
Cited by 2 | Viewed by 740
Abstract
The consistent speckle noise in SAR images easily interferes with the semantic information of the target. Additionally, the limited quantity of supervisory information available in one-shot learning leads to poor performance. To address the aforementioned issues, we creatively propose an SAR target recognition [...] Read more.
The consistent speckle noise in SAR images easily interferes with the semantic information of the target. Additionally, the limited quantity of supervisory information available in one-shot learning leads to poor performance. To address the aforementioned issues, we creatively propose an SAR target recognition model based on one-shot learning. This model incorporates a background noise removal technique to eliminate the interference caused by consistent speckle noise in the image. Then, a global and local complementary strategy is employed to utilize the data’s inherent a priori information as a supplement to the supervisory information. The experimental results show that our approach achieves a recognition performance of 70.867% under the three-way one-shot condition, which attains a minimum improvement of 7.467% compared to five state-of-the-art one-shot learning methods. The ablation studies demonstrate the efficacy of each design introduced in our model. Full article
Show Figures

Figure 1

Figure 1
<p>Schematic of the background noise cancellation phase in the proposed model.</p>
Full article ">Figure 2
<p>Schematic representation of the meta-training phase in the proposed model.</p>
Full article ">Figure 3
<p>Gradcam diagrams for comparison methods.</p>
Full article ">Figure 4
<p>Confusion matrix for different methods.</p>
Full article ">Figure 5
<p>Gradcam diagrams for ablation experiments.</p>
Full article ">Figure 6
<p>Classification accuracy results of varying parameter <math display="inline"><semantics> <mi>α</mi> </semantics></math>.</p>
Full article ">Figure 7
<p>Gradcam diagram for different augmentation.</p>
Full article ">
27 pages, 4322 KiB  
Article
Adaptive Filtering with Fitted Noise Estimate (AFFiNE): Blink Artifact Correction in Simulated and Real P300 Data
by Kevin E. Alexander, Justin R. Estepp and Sherif M. Elbasiouny
Bioengineering 2024, 11(7), 707; https://doi.org/10.3390/bioengineering11070707 - 12 Jul 2024
Viewed by 777
Abstract
(1) Background: The electroencephalogram (EEG) is frequently corrupted by ocular artifacts such as saccades and blinks. Methods for correcting these artifacts include independent component analysis (ICA) and recursive-least-squares (RLS) adaptive filtering (-AF). Here, we introduce a new method, AFFiNE, that applies Bayesian adaptive [...] Read more.
(1) Background: The electroencephalogram (EEG) is frequently corrupted by ocular artifacts such as saccades and blinks. Methods for correcting these artifacts include independent component analysis (ICA) and recursive-least-squares (RLS) adaptive filtering (-AF). Here, we introduce a new method, AFFiNE, that applies Bayesian adaptive regression spline (BARS) fitting to the adaptive filter’s reference noise input to address the known limitations of both ICA and RLS-AF, and then compare the performance of all three methods. (2) Methods: Artifact-corrected P300 morphologies, topographies, and measurements were compared between the three methods, and to known truth conditions, where possible, using real and simulated blink-corrupted event-related potential (ERP) datasets. (3) Results: In both simulated and real datasets, AFFiNE was successful at removing the blink artifact while preserving the underlying P300 signal in all situations where RLS-AF failed. Compared to ICA, AFFiNE resulted in either a practically or an observably comparable error. (4) Conclusions: AFFiNE is an ocular artifact correction technique that is implementable in online analyses; it can adapt to being non-stationarity and is independent of channel density and recording duration. AFFiNE can be utilized for the removal of blink artifacts in situations where ICA may not be practically or theoretically useful. Full article
(This article belongs to the Section Biosignal Processing)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Block diagram of the AFFiNE method. The reference noise input, r<sub>v</sub>[<span class="html-italic">n</span>], is conditioned by BARS to be free of bidirectional contamination from EEG sources close to the recording VEOG electrodes, thus resulting in a more ideal reference noise input signal for RLS-AF. The resulting adaptive filter is a FIR of length M, which operates on the conditioned reference noise input signal and is then linearly subtracted from the recorded EEG signal, s[<span class="html-italic">n</span>], to produce an EEG signal that is free of blink artifacts, e[<span class="html-italic">n</span>].</p>
Full article ">Figure 2
<p>This flowchart graphically outlines the main components of the current study with the relevant section numbers provided in parenthesis. Those sections primarily pertaining to the simulated data are shown in grey, and those pertaining primarily to the real data are in black; sections common to both datasets are shown in white.</p>
Full article ">Figure 3
<p>Stimulus timing sequence for the visual oddball paradigm. Each stimulus appeared on the screen for 100 ms, with inter-stimulus intervals varying on a uniform distribution between 2.0 and 2.5 s. During the inter-stimulus interval, a fixation cross appeared on the screen.</p>
Full article ">Figure 4
<p>Example signals used in the creation of the simulated dataset. (<b>A</b>) Example standard and oddball ERPs (grey and black, respectively) at electrode Pz, and (<b>B</b>) the spatial weights used to project the signal to all simulated electrodes. (<b>C</b>) An example simulated blink signal at FPz, and (<b>D</b>) its spatial weights used to project the signal across the scalp.</p>
Full article ">Figure 5
<p>Example time segment of the simulated signals and the summed (‘Total’) simulated signal at electrode Pz. The total simulated signal consisted of summed ‘Noise’, ‘ERP’, and ‘Blink’ signals. The blink samples in the blink signal were placed such that they coincided with half of the ERP samples.</p>
Full article ">Figure 6
<p>Histogram showing trial counts for each participant contained in the sample of real data used to compare the performance of the correction methods. A total of 17 participants were included with trial counts ranging from 18 (2 participants) to 33 (1 participant).</p>
Full article ">Figure 7
<p>Standard, oddball, and difference wave ERPs calculated from the simulated dataset for all Artifact by Correction Method combinations at FPz and Pz. Because these are simulated data, ‘Truth’ is the actual ERP calculated in each condition using only the simulated cortical source (i.e., task-relevant and -irrelevant) data. Data are shown with y-axis breaks in FPz/Blink Present to visualize the Uncorrected blink artifact maximum amplitude without detracting from the lower-amplitude features of the Truth and Corrected (ICA, RLS-AF, and AFFiNE) waveforms.</p>
Full article ">Figure 8
<p>Topographic MAE from each of the standard, oddball, and difference wave ERPs, both Blink Present and Blink Absent Artifact conditions, and the ICA, RLS, and AFFiNE levels of the Correction Method. The ‘Uncorrected’ level of the Correction Method is omitted here because no correction is reasonably attempted in the Unfiltered condition.</p>
Full article ">Figure 9
<p>Simulated data MAE values at FPz and Pz electrodes over 30 oddball and 30 standard trials in the Blink Present (<b>A</b>) and Blink Absent (<b>B</b>) Artifact conditions. Individual MAE values are plotted as a ‘swarmplot’ to aid in the visualization of the discrete data by offsetting data points in the x-dimension; this offset in the x-dimension is for visualization only and does not encode any information about the distribution of the presented data. Grey circles indicate distribution means.</p>
Full article ">Figure 10
<p>Real data, original and corrected, grand-averaged ERP waveforms at electrodes FPz and Pz. Pz Oddball and difference wave ERP measurement windows are indicated as shaded portions of the x-axis. Data are shown with y-axis breaks in FPz/Blink Present to visualize the Uncorrected blink artifact maximum amplitude without detracting from the lower-amplitude features of the Truth and Corrected (ICA, RLS-AF, and AFFiNE) waveforms.</p>
Full article ">Figure 11
<p>Real data mean participant MAEs. Errors were calculated using the Uncorrected data in the Blink Absent Artifact condition as the “truth data”.</p>
Full article ">Figure 12
<p>Oddball and difference wave ERP amplitude (<b>A</b>) and latency (<b>B</b>) measure absolute errors after performing each correction method on the Blink Absent trials. Individual absolute errors are plotted as a ‘swarmplot’ to aid in the visualization of the discrete data by offsetting data points in the x-dimension; this offset in the x-dimension is for visualization only and does not encode any information about the distribution of the presented data. Grey circles indicate distribution means. Practically speaking, latency error was either zero or one sample point (i.e., 3.9 ms), with rare exceptions, across all Correction Methods.</p>
Full article ">Figure 13
<p>Oddball and Difference wave ERP amplitude (<b>A</b>) and latency (<b>B</b>) measure absolute errors after performing each correction method on the Blink Present trials. Individual absolute errors are plotted as a ‘swarmplot’ to aid in the visualization of the discrete data by offsetting data points in the x-dimension; this offset in the x-dimension is for visualization only and does not encode any information about the distribution of the presented data. Grey circles indicate distribution means.</p>
Full article ">Figure 14
<p>Real data from a single participant; uncorrected and corrected ERP waveforms at electrodes FPz and Pz. Pz Oddball and difference wave ERP measurement windows are indicated as shaded portions of the x-axis. Data are shown with y-axis breaks in FPz/Blink Present to visualize Uncorrected blink artifact maximum amplitudes without detracting from the lower-amplitude features of the Truth and Corrected (ICA, RLS-AF, and AFFiNE) waveforms.</p>
Full article ">Figure 15
<p>Real data from an additional, single participant; uncorrected and corrected ERP waveforms at electrodes FPz and Pz. Pz Oddball and difference wave ERP measurement windows are indicated as shaded portions of the x-axis. Data are shown with y-axis breaks in FPz/Blink Present to visualize Uncorrected blink artifact maximum amplitudes without detracting from the lower-amplitude features of the Truth and Corrected (ICA, RLS-AF, and AFFiNE) waveforms.</p>
Full article ">
15 pages, 5081 KiB  
Article
A Novel Noise Reduction Approach of Acoustic Emission (AE) Signals in the SiC Lapping Process on Fixed Abrasive Pads
by Jie Lin, Jiapeng Chen, Wenkun Lin, Anjie He, Xiaodong Hao, Zhenlin Jiang, Wenjun Wang, Baoxiu Wang, Kerong Wang, Ying Wei and Tao Sun
Micromachines 2024, 15(7), 900; https://doi.org/10.3390/mi15070900 - 10 Jul 2024
Viewed by 734
Abstract
Acoustic emission (AE) technology has been widely utilized to monitor the SiC wafer lapping process. The root-mean-square (RMS) of the time–domain eigenvalues of the AE signal has a linear relationship with the material removal rate (MRR). However, the existence of background noise severely [...] Read more.
Acoustic emission (AE) technology has been widely utilized to monitor the SiC wafer lapping process. The root-mean-square (RMS) of the time–domain eigenvalues of the AE signal has a linear relationship with the material removal rate (MRR). However, the existence of background noise severely reduces signal monitoring accuracy. Noise interference often leads to increased RMS deviation and signal distortion. In the study presented in this manuscript, a frequency threshold noise reduction approach was developed by combining and improving wavelet packet noise reduction and spectral subtraction noise reduction techniques. Three groups of SiC lapping experiments were conducted on a fixed abrasive pad, and the lapping acoustic signals were processed using three different noise reduction approaches: frequency threshold, wavelet packet, and spectral subtraction. The results show that the noise reduction method using the frequency threshold is the most effective, with the best coefficient of determination (R2) for the linear fit of the RMS to the MRR. Full article
(This article belongs to the Section D:Materials and Processing)
Show Figures

Figure 1

Figure 1
<p>Schematics of the device for AE monitoring of SiC wafer lapping on fixed abrasive pads.</p>
Full article ">Figure 2
<p>Initial surface morphology of the Si-face of 4H-SiC.</p>
Full article ">Figure 3
<p>Wavelet packet 3-layer decomposition (<span class="html-italic">f</span> is the sampling frequency).</p>
Full article ">Figure 4
<p>Flowchart of frequency threshold noise reduction.</p>
Full article ">Figure 5
<p>MRR for each segment in the three groups.</p>
Full article ">Figure 6
<p>Surface morphology of SiC wafers after the Group One experiment.</p>
Full article ">Figure 7
<p>Surface micromorphology of fixed abrasive pads: (<b>a</b>) before the Group One experiment; (<b>b</b>) after the Group One experiment; (<b>c</b>) after the Group Two experiment; (<b>d</b>) after the Group Three experiment; macro-morphology of Group One (<b>e</b>) and Group Three (<b>f</b>).</p>
Full article ">Figure 8
<p>RMS of samples per noise segment; (<b>a</b>) average of RMS and (<b>b</b>) polarization of RMS.</p>
Full article ">Figure 9
<p>Original AE RMS and MRR.</p>
Full article ">Figure 10
<p>Group One AE RMS after noise reduction and MRR.</p>
Full article ">Figure 11
<p>Group Two AE RMS after noise reduction and MRR.</p>
Full article ">Figure 12
<p>Group Three AE RMS after noise reduction and MRR.</p>
Full article ">Figure 13
<p>Groups One and Three of frequency domain features.</p>
Full article ">
18 pages, 8143 KiB  
Article
Fuzzy Classification of the Maturity of the Orange (Citrus × sinensis) Using the Citrus Color Index (CCI)
by Marcos J. Villaseñor-Aguilar, Miroslava Cano-Lara, Adolfo R. Lopez, Horacio Rostro-Gonzalez, José Alfredo Padilla-Medina and Alejandro Israel Barranco-Gutiérrez
Appl. Sci. 2024, 14(13), 5953; https://doi.org/10.3390/app14135953 - 8 Jul 2024
Cited by 1 | Viewed by 967
Abstract
The orange (Citrus sinensis) is a fruit of the Citrus genus, which is part of the Rutaceae family. The orange has gained considerable importance due to its extensive range of applications, including the production of juices, jams, sweets, and extracts. The [...] Read more.
The orange (Citrus sinensis) is a fruit of the Citrus genus, which is part of the Rutaceae family. The orange has gained considerable importance due to its extensive range of applications, including the production of juices, jams, sweets, and extracts. The consumption of oranges confers several nutritional benefits, including flavonoids, vitamin C, potassium, beta-carotene, and dietary fiber. It is crucial to acknowledge that the primary quality criterion employed by consumers and producers is maturity, which is correlated with the visual quality associated with the color of the epicarp. This study proposes the implementation of a computer vision system that estimates the degree of ripeness of oranges Valencia using fuzzy logic (FL); the soluble solids content was determined by refractometry, while the firmness of the fruit was evaluated through the fruit firmness test. The proposed method was divided into five distinct steps. The initial stage involved the acquisition of RGB images. The second stage presents the segmentation of the fruit, which entails the removal of extraneous noise and backgrounds. The third and fourth steps involve determining the centroid of the fruit, and five regions of interest were obtained in the centroid of the fruit of the Citrus Color Index (CII), ranging from 3 × 3 to 11 × 11 pixels. Finally, in the fifth step, a model was created to estimate maturity, °Brix, and firmness using Matlab 2024 and the Fuzzy Logic Designer and Neuro-Fuzzy Designer applications. Consequently, a statistically significant correlation was established between maturity, degree Brix, and firmness, with a value greater than 0.9, using the Citrus Color Index (CII), which reflects the physical–chemical changes that occur in the orange. Full article
(This article belongs to the Special Issue Advances in Machine Vision for Industry and Agriculture)
Show Figures

Figure 1

Figure 1
<p>Samples of oranges in different degrees of °Brix.</p>
Full article ">Figure 2
<p>Samples of oranges in different degrees of firmness.</p>
Full article ">Figure 3
<p>Samples of oranges Valencia in different degrees of maturity.</p>
Full article ">Figure 4
<p>Samples mapped in RGB color space.</p>
Full article ">Figure 5
<p>Samples mapped in CIE-Lab color space.</p>
Full article ">Figure 6
<p>Samples mapped in HSV space.</p>
Full article ">Figure 7
<p>The proposed method for determining the Brix degree and firmness of orange.</p>
Full article ">Figure 8
<p>The proposed method in orange fruit for the removal of small blobs from the image of the captured sample: (<b>i</b>) real image; (<b>ii</b>) binarization of the image; and (<b>iii</b>) segmentation of the sample and discrimination of areas.</p>
Full article ">Figure 9
<p>Representation of the sub-regions of the Citrus Color Index (CCI) of Valencia orange fruit.</p>
Full article ">Figure 10
<p>Proposed fuzzy inference system.</p>
Full article ">Figure 11
<p>Membership functions used in the proposed model maturity of oranges Valencia. Low Citrus Color Index: (<b>a</b>) 3 × 3, (<b>b</b>) 5 × 5, (<b>c</b>) 11 × 11, (<b>d</b>) 21 × 21, and (<b>e</b>) 31 × 31 data.</p>
Full article ">Figure 12
<p>Membership functions used in the proposed model degree Brix of oranges Valencia: (<b>a</b>) membership functions for Low Citrus Color Index 3 × 3 data; (<b>b</b>) membership functions for Low Citrus Color Index 5 × 5 data; (<b>c</b>) membership functions for Low Citrus Color Index 11 × 11 data; (<b>d</b>) membership functions for Low Citrus Color Index 21 × 21 data; and (<b>e</b>) membership functions for Low Citrus Color Index 31 × 31 data.</p>
Full article ">Figure 13
<p>Membership functions used in the proposed model firmness or oranges Valencia: (<b>a</b>) membership functions for Low Citrus Color Index 3 × 3 data; (<b>b</b>) membership functions for Low Citrus Color Index 5 × 5 data; (<b>c</b>) membership functions for Low Citrus Color Index 11 × 11 data; (<b>d</b>) membership functions for Low Citrus Color Index 21 × 21 data; and (<b>e</b>) membership functions for Low Citrus Color Index 31 × 31 data.</p>
Full article ">Figure 14
<p>Takagi–Sugeno for defuzzification.</p>
Full article ">Figure 15
<p>Predictions of the fuzzy inference systems of maturity.</p>
Full article ">Figure 16
<p>Mean square error of the predictions of the fuzzy inference systems of maturity.</p>
Full article ">Figure 17
<p>Predictions of the fuzzy inference systems of degree Brix.</p>
Full article ">Figure 18
<p>Mean square error of the predictions of the fuzzy inference systems of degree Brix.</p>
Full article ">Figure 19
<p>Predictions of the fuzzy inference systems of firmness.</p>
Full article ">Figure 20
<p>Mean square error of the predictions of the fuzzy inference systems of firmness.</p>
Full article ">
Back to TopTop