Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (207)

Search Parameters:
Keywords = removal of background noise

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
26 pages, 13026 KiB  
Article
Unified Spatial-Frequency Modeling and Alignment for Multi-Scale Small Object Detection
by Jing Liu, Ying Wang, Yanyan Cao, Chaoping Guo, Peijun Shi and Pan Li
Symmetry 2025, 17(2), 242; https://doi.org/10.3390/sym17020242 - 6 Feb 2025
Viewed by 412
Abstract
Small object detection in aerial imagery remains challenging due to sparse feature representation, limited spatial resolution, and complex background interference. Current deep learning approaches enhance detection performance through multi-scale feature fusion, leveraging convolutional operations to expand the receptive field or self-attention mechanisms for [...] Read more.
Small object detection in aerial imagery remains challenging due to sparse feature representation, limited spatial resolution, and complex background interference. Current deep learning approaches enhance detection performance through multi-scale feature fusion, leveraging convolutional operations to expand the receptive field or self-attention mechanisms for global context modeling. However, these methods primarily rely on spatial-domain features, while self-attention introduces high computational costs, and conventional fusion strategies (e.g., concatenation or addition) often result in weak feature correlation or boundary misalignment. To address these challenges, we propose a unified spatial-frequency modeling and multi-scale alignment fusion framework, termed USF-DETR, for small object detection. The framework comprises three key modules: the Spatial-Frequency Interaction Backbone (SFIB), the Dual Alignment and Balance Fusion FPN (DABF-FPN), and the Efficient Attention-AIFI (EA-AIFI). The SFIB integrates the Scharr operator for spatial edge and detail extraction and FFT/IFFT for capturing frequency-domain patterns, achieving a balanced fusion of global semantics and local details. The DABF-FPN employs bidirectional geometric alignment and adaptive attention to enhance the significance expression of the target area, suppress background noise, and improve feature asymmetry across scales. The EA-AIFI streamlines the Transformer attention mechanism by removing key-value interactions and encoding query relationships via linear projections, significantly boosting inference speed and contextual modeling. Experiments on the VisDrone and TinyPerson datasets demonstrate the effectiveness of USF-DETR, achieving improvements of 2.3% and 1.4% mAP over baselines, respectively, while balancing accuracy and computational efficiency. The framework outperforms state-of-the-art methods in small object detection. Full article
(This article belongs to the Special Issue Symmetry and Asymmetry Study in Object Detection)
Show Figures

Figure 1

Figure 1
<p>Comparison between RT-DETR and the proposed USF-DETR method. The feature maps generated by USF-DETR (bottom row) exhibit sharper edges and richer details due to the SFIB and EA-AIFI modules. After multi-scale alignment fusion through the DABF-FPN Encoder, USF-DETR produces more accurate heatmaps, effectively highlighting small objects and improving detection results with fewer missed detections and false positives, as demonstrated by the red bounding boxes.</p>
Full article ">Figure 2
<p>Architecture of the proposed USF-DETR, which includes three modules: SFIB, EA-AIFI, and DABF-FPN. The top part illustrates the pipeline of USF-DETR, while the bottom part presents the module flowchart.</p>
Full article ">Figure 3
<p>The pipeline of SFIB consists of four stages, with each stage including a Conv layer and a SFI block. The SFI block, shown in the lower left figure, is connected across layers using the CSP concept; As depicted in the lower right image, the SFI extracts spatial and frequency domain features of the image and then fuses them.</p>
Full article ">Figure 4
<p>The overall structure of the DABF-FPN integrates bidirectional feature fusion to enhance small object detection and outputs multi-scale features (P2, N3, N4, and N5) for further processing.</p>
Full article ">Figure 5
<p>The structure of the DABF module involves high-level semantic features and low-level detailed features being adaptively processed to extract mutual representations. Two DABF blocks facilitate comprehensive information exchange and enhance feature fusion quality.</p>
Full article ">Figure 6
<p>EA-AIFI module. (<b>a</b>) Through input embedding and positional encoding, combined with enhanced representation of contextual information, further internal feature interaction and optimization are carried out through a FFN. (<b>b</b>) Efficient Additive Attention eliminates key value interactions and relies solely on linear projections.</p>
Full article ">Figure 7
<p>Bounding box distribution. (<b>a</b>) VisDrone2019-DET Dataset. (<b>b</b>) TinyPerson Dataset. The vertical axis represents the categories of annotated bounding boxes, while the horizontal axis depicts the square root of the bounding box area, measured in pixels.</p>
Full article ">Figure 8
<p>Visualization of feature maps. (<b>a</b>) Input image. (<b>b</b>) Feature map generated without using the SFI module in the baseline model. (<b>c</b>) Feature map generated with the SFI module in USF-DETR.</p>
Full article ">Figure 9
<p>Visualizing the detection results and heatmap on TinyPerson. The highlighted area represents the region of network attention, demonstrating the outstanding performance of USF-DETR in detecting small objects.</p>
Full article ">Figure 10
<p>Detection results of the USF-DETR on the VisDrone dataset. Boxes of different colors represent different target categories.</p>
Full article ">Figure 11
<p>A comparison of detection results between USF-DETR and the baseline model is presented. Green boxes indicate correct detections, blue boxes represent false positives, and red boxes denote missed detections.</p>
Full article ">Figure 12
<p>A comparison of detection performance between the two methods. The first row represents USF-DETR, while the second row shows the baseline method. USF-DETR significantly reduces false positives (blue) and false negatives (red).</p>
Full article ">Figure 13
<p>Comparison of detection performance between USF-DETR and popular methods. The yellow circle shows the outstanding detection effect of USF-DETR.</p>
Full article ">
18 pages, 3690 KiB  
Article
Text Removal for Trademark Images Based on Self-Prompting Mechanisms and Multi-Scale Texture Aggregation
by Wenchao Zhou, Xiuhui Wang, Boxiu Zhou and Longwen Li
Appl. Sci. 2025, 15(3), 1553; https://doi.org/10.3390/app15031553 - 4 Feb 2025
Viewed by 489
Abstract
With the rapid development of electronic business, there has been a surge in incidents of trademark infringement, making it imperative to improve the accuracy of trademark retrieval systems as a key measure to combat such illegal behaviors. Evidently, the textual information encompassed within [...] Read more.
With the rapid development of electronic business, there has been a surge in incidents of trademark infringement, making it imperative to improve the accuracy of trademark retrieval systems as a key measure to combat such illegal behaviors. Evidently, the textual information encompassed within trademarks substantially influences the precision of search results. Considering the diversity of trademark text and the complexity of its design elements, accurately locating and analyzing this text poses a considerable challenge. Against this background, this research has developed an original self-prompting text removal model, denoted as “Self-prompting Trademark Text Removal Based on Multi-scale Texture Aggregation” (abbreviated as MTF-STTR). This model astutely applies a text detection network to automatically generate the required input cues for the Segment Anything Model (SAM) while incorporating the technological benefits of diffusion models to attain a finer level of trademark text removal. To further elevate the performance of the model, we introduce two innovative architectures to the text detection network: the Integrated Differentiating Feature Pyramid (IDFP) and the Texture Fusion Module (TFM). These mechanisms are capable of efficiently extracting multilevel features and multiscale textual information, which enhances the model’s stability and adaptability in complex scenarios. The experimental validation has demonstrated that the trademark text erasure model designed in this paper achieves a peak signal-to-noise ratio as high as 40.1 dB on the SCUT-Syn dataset, which is an average improvement of 11.3 dB compared with other text erasure models. Furthermore, the text detection network component of the designed model attains an accuracy of up to 89.9% on the CTW1500 dataset, representing an average enhancement of 10 percentage points over other text detection networks. Full article
Show Figures

Figure 1

Figure 1
<p>Comparison of scene text and trademark text.</p>
Full article ">Figure 2
<p>Architecture of the proposed MTF-STTR network.</p>
Full article ">Figure 3
<p>Architecture of the IDFP module.</p>
Full article ">Figure 4
<p>Architecture of the TFM module.</p>
Full article ">Figure 5
<p>Architecture of the LDM module.</p>
Full article ">Figure 6
<p>Examples from the CTW1500 dataset.</p>
Full article ">Figure 7
<p>Examples from the SCUT-Syn dataset.</p>
Full article ">Figure 8
<p>Actual-scene application effect.</p>
Full article ">
14 pages, 5487 KiB  
Article
Automated Quantification of Rebar Mesh Inspection in Hidden Engineering Structures via Deep Learning
by Yalong Xie, Xianhui Nie, Hongliang Liu, Yifan Shen and Yuming Liu
Appl. Sci. 2025, 15(3), 1063; https://doi.org/10.3390/app15031063 - 22 Jan 2025
Viewed by 679
Abstract
This paper presents an in-depth study of the automated recognition and geometric information quantification of rebar meshes, proposing a deep learning-based method for rebar mesh detection and segmentation. By constructing a diverse rebar mesh image dataset, an improved Unet-based model was developed, incorporating [...] Read more.
This paper presents an in-depth study of the automated recognition and geometric information quantification of rebar meshes, proposing a deep learning-based method for rebar mesh detection and segmentation. By constructing a diverse rebar mesh image dataset, an improved Unet-based model was developed, incorporating residual modules to enhance the network’s feature extraction capabilities and training efficiency. The study found that the improved model maintains high segmentation accuracy and robustness even in the presence of complex backgrounds and noise. To achieve the precise measurement of rebar spacing, a rebar intersection detection algorithm based on convolution operations was designed, and the IQR (Interquartile Range) algorithm was applied to remove outliers, ensuring the accuracy and reliability of spacing calculations. The experimental results demonstrate that the proposed model and methods effectively and efficiently accomplish the automated recognition and geometric information extraction of rebar meshes, providing reliable technical support for the automated detection and geometric data analysis of rebar meshes in practical engineering applications. Full article
Show Figures

Figure 1

Figure 1
<p>Rebar mesh images with complex backgrounds: (<b>a</b>,<b>b</b>) overlapping rebars; (<b>c</b>,<b>d</b>) color differences among different types of Rebars; (<b>e</b>,<b>f</b>) obvious shadows behind the Rebars.</p>
Full article ">Figure 2
<p>Data augmentation modes: (<b>1</b>)~(<b>5</b>) data augmentation results of the original image under different methods.</p>
Full article ">Figure 3
<p>Schematic diagram of the Unet network framework.</p>
Full article ">Figure 4
<p>Comparison of model performance during training, validation, and testing: (<b>a</b>) comparison of loss values of different models during training; (<b>b</b>) comparison of loss values of different models during validation; (<b>c</b>) comparison of MIoU of different models during testing.</p>
Full article ">Figure 5
<p>Image thinning process.</p>
Full article ">Figure 6
<p>Feature detection performance of different filters under regular rebar mesh distribution: (<b>a</b>) Filter Traversal Process Diagram; (<b>b</b>) Square Filter; (<b>c</b>) Circular Filter; (<b>d</b>) Image Features After Filtering with a Square Filter; (<b>e</b>) Image Features After Filtering with a Circular Filter.</p>
Full article ">Figure 7
<p>Feature detection performance of different filters under irregular rebar mesh distribution: (<b>a</b>) Filter Traversal Process Diagram; (<b>b</b>) Square Filter; (<b>c</b>) Circular Filter; (<b>d</b>) Image Features After Filtering with a Square Filter; (<b>e</b>) Image Features After Filtering with a Circular Filter.</p>
Full article ">Figure 8
<p>Schematic diagram of duplicate suppression mask.</p>
Full article ">Figure 9
<p>Schematic diagram of intersection detection: (<b>a</b>) Process of Detecting Neighboring Intersections (Nearest Intersection); (<b>b</b>) Extracted Intersection Image (Red Box Indicates Identified Intersections).</p>
Full article ">Figure 10
<p>Actual detection results of rebar mesh spacing: (<b>a</b>–<b>f</b>) Comparison of Detection Results and True Values for Different Rebar Specimen Spacings.</p>
Full article ">
22 pages, 6259 KiB  
Article
3D Seismic Attribute Conditioning Using Multiscale Sheet-Enhancing Filtering
by Taiyin Zhao, Yuehua Yue, Tian Chen and Feng Qian
Remote Sens. 2025, 17(2), 278; https://doi.org/10.3390/rs17020278 - 14 Jan 2025
Viewed by 511
Abstract
Seismic coherence attributes are valuable for identifying structural features, but they often face challenges due to significant background noise and non-feature-related stratigraphic discontinuities. To address this, it is necessary to apply attribute conditioning to the coherence to enhance the visibility of these structures. [...] Read more.
Seismic coherence attributes are valuable for identifying structural features, but they often face challenges due to significant background noise and non-feature-related stratigraphic discontinuities. To address this, it is necessary to apply attribute conditioning to the coherence to enhance the visibility of these structures. The primary challenge of attribute conditioning lies in finding a concise structural representation that isolates only the true interpretive features while effectively removing noise and stratigraphic interference. In this study, we choose sheet-like structures as this concise structural representation, as faults are typically characterized by their thin and narrow profiles. Inspired by multiscale Hessian-based filtering (MHF) and its application on vascular structure detection, we propose a method called anisotropic multiscale Hessian-based sheet-enhancing filtering (AMHSF). This method is specifically designed to extract and magnify sheet-like structures from noisy coherence images, with a novel enhancement function distinct from those traditionally used in vascular enhancement. The effectiveness of our AMHSF is demonstrated through experiments on both synthetic and real datasets, showcasing its potential to improve the identification of structural features in coherence images. Full article
Show Figures

Figure 1

Figure 1
<p>The relation between the eigenvalues of the Hessian matrix and image structure orientations, including a sheet, a tube, a blob, and noise. Note that our interpretation features have sheet-like structures, while the vascular features fuse tubular structures.</p>
Full article ">Figure 2
<p>The 3D cubes of attribute conditioning results on the synthetic 3D fault dataset. (<b>a</b>–<b>d</b>) Seismic amplitudes, (<b>e</b>–<b>h</b>) variance coherence, (<b>i</b>–<b>l</b>) attribute condition (results) obtained by AMHSF, (<b>m</b>–<b>p</b>) attribute condition (results) obtained by ant tracking, and (<b>q</b>,<b>r</b>) the ground truth faults.</p>
Full article ">Figure 3
<p>The effect of the 20th inline slices of attribute conditioning results on the synthetic 3D fault dataset. (<b>a</b>–<b>d</b>) Seismic amplitudes, (<b>e</b>–<b>h</b>) variance coherence, (<b>i</b>–<b>l</b>) attribute condition (results) obtained by AMHSF, (<b>m</b>–<b>p</b>) attribute condition (results) obtained by ant tracking, and (<b>q</b>,<b>r</b>) the ground truth faults.</p>
Full article ">Figure 4
<p>The quantitative comparison of the results between our AMHSF and ant tracking.</p>
Full article ">Figure 5
<p>The intermediate attribute conditioning results on the synthetic 3D fault dataset. (<b>a</b>) Anisotropic Hessian, (<b>b</b>) eigenvalue, (<b>c</b>–<b>g</b>) intermediate results using Equation (<a href="#FD14-remotesensing-17-00278" class="html-disp-formula">14</a>), and (<b>h</b>–<b>k</b>) intermediate results using Equation (<a href="#FD15-remotesensing-17-00278" class="html-disp-formula">15</a>).</p>
Full article ">Figure 6
<p>The impact of parameter selection. Corresponding parameters are given in <a href="#remotesensing-17-00278-t003" class="html-table">Table 3</a>.</p>
Full article ">Figure 7
<p>The 3D cubes of attribute conditioning results using Opunake-3D.</p>
Full article ">Figure 8
<p>The 50th crossline slices of attribute conditioning results using Opunake-3D.</p>
Full article ">Figure 9
<p>The 3D cubes of attribute conditioning results using Parihaka-3D.</p>
Full article ">Figure 10
<p>The 50th timeline slices of attribute conditioning results using Parihaka-3D.</p>
Full article ">
17 pages, 2421 KiB  
Article
Determining Water Pipe Leakage Using an RP-CNN Model to Identify the Causes and Improve Poor-Accuracy Cases
by Muhammad Anshari Caronge, Taichi Shibuya, Yasuhiro Arai, Xinyi Dong, Takaharu Kunizane and Akira Koizumi
Acoustics 2025, 7(1), 2; https://doi.org/10.3390/acoustics7010002 - 3 Jan 2025
Viewed by 640
Abstract
This study aimed to assess and improve the accuracy of a water leakage detection model proposed in preliminary research. The poor results for water leakage sound (recall) and background noise (specificity) were clarified using countermeasures in accordance with each condition. Additionally, frequency amplification [...] Read more.
This study aimed to assess and improve the accuracy of a water leakage detection model proposed in preliminary research. The poor results for water leakage sound (recall) and background noise (specificity) were clarified using countermeasures in accordance with each condition. Additionally, frequency amplification in the range of 500–600 Hz, the attenuation of weak components, and a band-stop filter were used to remove the 50 Hz component and harmonics. Pre-processing was carried out in the form of amplification, with weak noise removed using a band-stop filter. The results showed that the application of the proposed model improved the detection accuracy by 80% at the observation points that initially had poor accuracy. Thus, the proposed method was effective at improving the performance of the Recurrence Plot-Convolutional Neural Network (RP-CNN) model for detecting water leakages. Full article
(This article belongs to the Special Issue Duct Acoustics)
Show Figures

Figure 1

Figure 1
<p>Flowchart of water leakage determination using the RP-CNN model.</p>
Full article ">Figure 2
<p>Learning and assessment data for 9- and 10-point models.</p>
Full article ">Figure 3
<p>Box plot of acoustic data converted into 15 dimensions.</p>
Full article ">Figure 4
<p>Positioning results based on principal component scores.</p>
Full article ">Figure 5
<p>FFT spectrum of point 1-B.</p>
Full article ">Figure 6
<p>Changes in RP due to amplification process of point 4-B (RP number 2000 to 2002).</p>
Full article ">Figure 7
<p>Results of applying the proposed method (15-dimensional data).</p>
Full article ">Figure 8
<p>Changes in RP before and after pre-processing of points 3-B and 4-B.</p>
Full article ">Figure 9
<p>Comparison of confusion matrix for points 3-B and 4-B.</p>
Full article ">
17 pages, 8228 KiB  
Article
Application of Enhanced Weighted Least Squares with Dark Background Image Fusion for Inhomogeneity Noise Removal in Brain Tumor Hyperspectral Images
by Jiayue Yan, Chenglong Tao, Yuan Wang, Jian Du, Meijie Qi, Zhoufeng Zhang and Bingliang Hu
Appl. Sci. 2025, 15(1), 321; https://doi.org/10.3390/app15010321 - 31 Dec 2024
Viewed by 626
Abstract
The inhomogeneity of spectral pixel response is an unavoidable phenomenon in hyperspectral imaging, which is mainly manifested by the existence of inhomogeneity banding noise in the acquired hyperspectral data. It must be carried out to get rid of this type of striped noise [...] Read more.
The inhomogeneity of spectral pixel response is an unavoidable phenomenon in hyperspectral imaging, which is mainly manifested by the existence of inhomogeneity banding noise in the acquired hyperspectral data. It must be carried out to get rid of this type of striped noise since it is frequently uneven and densely distributed, which negatively impacts data processing and application. By analyzing the source of the instrument noise, this work first created a novel non-uniform noise removal method for a spatial dimensional push sweep hyperspectral imaging system. Clean and clear medical hyperspectral brain tumor tissue images were generated by combining scene-based and reference-based non-uniformity correction denoising algorithms, providing a strong basis for further diagnosis and classification. The precise procedure entails gathering the reference dark background image for rectification and the actual medical hyperspectral brain tumor image. The original hyperspectral brain tumor image is then smoothed using a weighted least squares algorithm model embedded with bilateral filtering (BLF-WLS), followed by a calculation and separation of the instrument fixed-mode fringe noise component from the acquired reference dark background image. The purpose of eliminating non-uniform fringe noise is achieved. In comparison to other common image denoising methods, the evaluation is based on the subjective effect and unreferenced image denoising evaluation indices. The approach discussed in this paper, according to the experiments, produces the best results in terms of the subjective effect and unreferenced image denoising evaluation indices (MICV and MNR). The image processed by this method has almost no residual non-uniform noise, the image is clear, and the best visual effect is achieved. It can be concluded that different denoising methods designed for different noises have better denoising effects on hyperspectral images. The non-uniformity denoising method designed in this paper based on a spatial dimension push-sweep hyperspectral imaging system can be widely used. Full article
Show Figures

Figure 1

Figure 1
<p>Spatial dimensions push sweep hyperspectral imaging system noise removal method.</p>
Full article ">Figure 2
<p>Spatial dimensional push sweep hyperspectral brain tumor image acquisition system.</p>
Full article ">Figure 3
<p>Hyperspectral image of the original brain tumor.</p>
Full article ">Figure 4
<p>Denoised hyperspectral image of brain tumor.</p>
Full article ">Figure 5
<p>Gray distribution space.</p>
Full article ">Figure 6
<p>Comparison of single-band images.</p>
Full article ">Figure 7
<p>Comparison of the denoising results and details in Data1 and Data2.</p>
Full article ">Figure 8
<p>Comparison of the denoising results and details in Data3 and Data4.</p>
Full article ">Figure 9
<p>Column mean curve.</p>
Full article ">
20 pages, 5692 KiB  
Article
Combining UAV Remote Sensing with Ensemble Learning to Monitor Leaf Nitrogen Content in Custard Apple (Annona squamosa L.)
by Xiangtai Jiang, Lutao Gao, Xingang Xu, Wenbiao Wu, Guijun Yang, Yang Meng, Haikuan Feng, Yafeng Li, Hanyu Xue and Tianen Chen
Agronomy 2025, 15(1), 38; https://doi.org/10.3390/agronomy15010038 - 27 Dec 2024
Viewed by 456
Abstract
One of the most important nutrients needed for fruit tree growth is nitrogen. For orchards to get targeted, well-informed nitrogen fertilizer, accurate, large-scale, real-time monitoring, and assessment of nitrogen nutrition is essential. This study examines the Leaf Nitrogen Content (LNC) of the custard [...] Read more.
One of the most important nutrients needed for fruit tree growth is nitrogen. For orchards to get targeted, well-informed nitrogen fertilizer, accurate, large-scale, real-time monitoring, and assessment of nitrogen nutrition is essential. This study examines the Leaf Nitrogen Content (LNC) of the custard apple tree, a noteworthy fruit tree that is extensively grown in China’s Yunnan Province. This study uses an ensemble learning technique based on multiple machine learning algorithms to effectively and precisely monitor the leaf nitrogen content in the tree canopy using multispectral canopy footage of custard apple trees taken via Unmanned Aerial Vehicle (UAV) across different growth phases. First, canopy shadows and background noise from the soil are removed from the UAV imagery by using spectral shadow indices across growth phases. The noise-filtered imagery is then used to extract a number of vegetation indices (VIs) and textural features (TFs). Correlation analysis is then used to determine which features are most pertinent for LNC estimation. A two-layer ensemble model is built to quantitatively estimate leaf nitrogen using the stacking ensemble learning (Stacking) principles. Random Forest (RF), Adaptive Boosting (ADA), Gradient Boosting Decision Trees (GBDT), Linear Regression (LR), and Extremely Randomized Trees (ERT) are among the basis estimators that are integrated in the first layer. By detecting and eliminating redundancy among base estimators, the Least Absolute Shrinkage and Selection Operator regression (Lasso)model used in the second layer improves nitrogen estimation. According to the analysis results, Lasso successfully finds redundant base estimators in the suggested ensemble learning approach, which yields the maximum estimation accuracy for the nitrogen content of custard apple trees’ leaves. With a root mean square error (RMSE) of 0.059 and a mean absolute error (MAE) of 0.193, the coefficient of determination (R2) came to 0. 661. The significant potential of UAV-based ensemble learning techniques for tracking nitrogen nutrition in custard apple leaves is highlighted by this work. Additionally, the approaches investigated might offer insightful information and a point of reference for UAV remote sensing applications in nitrogen nutrition monitoring for other crops. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>Geographical location of the study area.</p>
Full article ">Figure 2
<p>DJI Mavic 3M.</p>
Full article ">Figure 3
<p>(<b>a</b>–<b>f</b>) show a comparison between the original images and the images with shadows and soil removed across three growth stages.</p>
Full article ">Figure 4
<p>Design of an Ensemble Learning Workflow for Estimating LNC in Custard Apple.</p>
Full article ">Figure 5
<p>Presents the heatmap of spectral features with moderate or stronger correlations with the LNC of custard apple.</p>
Full article ">Figure 6
<p>Heatmap of the correlation between custard apple LNC and the optimal input variables.</p>
Full article ">Figure 7
<p>This illustrates the remote sensing estimation of custard apple leaf nitrogen content using different learning methods. (<b>a</b>–<b>e</b>) represent the fitting curves of the base models RF, GBDT, ADA, ERT, and LR, respectively. (<b>f</b>) shows the fitting curve of the meta-model, and (<b>g</b>) presents the Lasso weight graph.</p>
Full article ">Figure 8
<p>Remote Sensing Monitoring of Custard Apple Leaf Nitrogen Content Based on UAV Multispectral Imagery. (<b>a</b>–<b>c</b>) represent the remote sensing monitoring images of leaf nitrogen content in May, August, and November, respectively.</p>
Full article ">
20 pages, 11848 KiB  
Article
A Lightweight Small Target Detection Algorithm for UAV Platforms
by Yanhui Lv, Bo Tian, Qichao Guo and Deyu Zhang
Appl. Sci. 2025, 15(1), 12; https://doi.org/10.3390/app15010012 - 24 Dec 2024
Viewed by 654
Abstract
The targets in the aerial view of UAVs are small, scenes are complex, and background noise is strong. Additionally, the low computational capability of UAVs is challenged when trying to meet the requirements of large neural networks. Therefore, a lightweight object detection algorithm [...] Read more.
The targets in the aerial view of UAVs are small, scenes are complex, and background noise is strong. Additionally, the low computational capability of UAVs is challenged when trying to meet the requirements of large neural networks. Therefore, a lightweight object detection algorithm tailored for UAV platforms, called RSG-YOLO, is proposed. The algorithm introduces an attention module constructed with receptive field attention and coordinate attention, which helps reduce background noise interference while improving long-range information dependency. It also introduces and refines a fine-grained downsampling structure to minimize the loss of target information during the downsampling process. A general efficient layer aggregation network enhances the base feature extraction module, improving gradient flow information. Additionally, a detection layer rich in small target information is added, while redundant large object detection layers are removed, achieving a lightweight design while enhancing detection accuracy. Experimental results show that, compared to the baseline algorithm, the improved algorithm increases the P, R, [email protected], and [email protected]:0.95 by 6.9%, 7.2%, 8.4%, 5.8%, respectively, on the VisDrone 2019 dataset, and by 5.7%, 9%, 9.3%, 3.6%, respectively, on the TinyPerson dataset, while reducing the number of parameters by 23.3%. This significantly enhances the model’s detection performance and robustness, making it highly suitable for object detection tasks on low-computing-power UAV platforms. Full article
Show Figures

Figure 1

Figure 1
<p>The network structure of RSG-YOLO.</p>
Full article ">Figure 2
<p>Comparison between CA and RFCAConv. (<b>a</b>) The CA calculation flowchart; (<b>b</b>) the RFCAConv calculation flowchart.</p>
Full article ">Figure 3
<p>The SPD operation details.</p>
Full article ">Figure 4
<p>The CAMSPD operation details.</p>
Full article ">Figure 5
<p>The C2GELAN module structure.</p>
Full article ">Figure 6
<p>Comparison chart of indicator changes in the training process of the improved model versus the basic model.</p>
Full article ">Figure 7
<p>Comparison of heatmap visualizations between basic model and improved model.</p>
Full article ">Figure 7 Cont.
<p>Comparison of heatmap visualizations between basic model and improved model.</p>
Full article ">Figure 8
<p>Comparison of actual detection performance in different scenes of VisDrone2019.</p>
Full article ">Figure 9
<p>Comparison of actual detection performance in different scenes of TinyPerson.</p>
Full article ">Figure 10
<p>These are the detection effects in the actual shooting scenes of UAVs. These two pictures do not belong to the VisDrone2019 dataset or the TinyPerson dataset.</p>
Full article ">Figure 11
<p>Changes in training indicators of each comparison model.</p>
Full article ">Figure 12
<p>The actual scenario detection situations for each comparison model.</p>
Full article ">
19 pages, 4272 KiB  
Article
Two-Level Supervised Network for Small Ship Target Detection in Shallow Thin Cloud-Covered Optical Satellite Images
by Fangjian Liu, Fengyi Zhang, Mi Wang and Qizhi Xu
Appl. Sci. 2024, 14(24), 11558; https://doi.org/10.3390/app142411558 - 11 Dec 2024
Viewed by 547
Abstract
Ship detection under cloudy and foggy conditions is a significant challenge in remote sensing satellite applications, as cloud cover often reduces contrast between targets and backgrounds. Additionally, ships are small and affected by noise, making them difficult to detect. This paper proposes a [...] Read more.
Ship detection under cloudy and foggy conditions is a significant challenge in remote sensing satellite applications, as cloud cover often reduces contrast between targets and backgrounds. Additionally, ships are small and affected by noise, making them difficult to detect. This paper proposes a Cloud Removal and Target Detection (CRTD) network to detect small ships in images with thin cloud cover. The process begins with a Thin Cloud Removal (TCR) module for image preprocessing. The preprocessed data are then fed into a Small Target Detection (STD) module. To improve target–background contrast, we introduce a Target Enhancement module. The TCR and STD modules are integrated through a dual-stage supervision network, which hierarchically processes the detection task to enhance data quality, minimizing the impact of thin clouds. Experiments on the GaoFen-4 satellite dataset show that the proposed method outperforms existing detectors, achieving an average precision (AP) of 88.9%. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Challenges in ship detection under thin cloud cover. (<b>a</b>) Low target–background contrast: minimal visual difference between ship and background weakens target discernibility. (<b>b</b>) Cloud–ship similarity: similar shapes, textures, and brightness cause false detections. (<b>c</b>) Small target size: ships, occupying 2–6 pixels, limit detail for accurate detection.</p>
Full article ">Figure 2
<p>Illustration of module connections in detection frameworks. (<b>a</b>) One-stage framework: cloud removal and detection modules fuse features directly. (<b>b</b>) Multi-stage framework: each module has separate supervision, lacking inter-module information exchange. (<b>c</b>) Hierarchical supervised framework: TCR and STD modules are integrated with hierarchical supervision for enhanced ship detection in thin clouds.</p>
Full article ">Figure 3
<p>Overview of the proposed CRTD ship detection method.</p>
Full article ">Figure 4
<p>Details regarding DehazeFormer (DF) block: RescaleNorm, SoftReLU, and W-MHSA for enhanced thin cloud removal.</p>
Full article ">Figure 5
<p>Details regarding Group Attention Refine Module (GARM): Depth-wise Conv, Batch Norm, ReLU, and grouped attention calculation for feature enhancement.</p>
Full article ">Figure 6
<p>Details regarding Compact Inverted Block (CCIB): low-cost depth-wise and point-wise convolutions for efficient small target detection.</p>
Full article ">Figure 7
<p>Schematic diagram of training samples for ship detection. The dataset includes a wide range of ship targets, varying in type (cargo ships, warships, and passenger ships) and size (from small vessels occupying 2 to 6 pixels to larger ones). It covers diverse scenarios, such as busy ports with complex backgrounds and open seas with varied lighting and sea conditions. The dataset features both static ships (anchored or docked) and dynamic ships in motion, providing a comprehensive resource for training robust ship detection models.</p>
Full article ">Figure 8
<p>Visualization of ship detection results: (<b>a</b>) Cloud-Connected Target, where the ship is partially obscured by clouds, demonstrating the algorithm’s ability to detect targets in cloud-interference scenarios; (<b>b</b>) Densely Distributed Ships in Port, showing detection results in a port with high ship density, highlighting the algorithm’s performance in complex, crowded environments; (<b>c</b>) Detection of Ships at Different Scales, showcasing the detection of ships of various sizes, from small vessels to larger ones, evaluating the algorithm’s scalability and accuracy across different target sizes.</p>
Full article ">Figure 9
<p>Precision–recall curve for ship detection: comparison of different methods. This curve plots precision (%) against recall (%), where higher precision indicates fewer false positives, and higher recall reflects better identification of true ship targets. The area under the curve (AUC) serves as a key metric for model performance, with a larger AUC indicating better detection accuracy across thresholds.</p>
Full article ">Figure 10
<p>Comparison of ship detection results using different methods. This figure shows the detection results for various methods: Faster R-CNN (<b>a</b>), SSD (<b>b</b>), YOLO (<b>c</b>), RT-DETR (<b>d</b>), and our proposed method (<b>e</b>). Green boxes indicate true ship targets, red boxes show detected regions, and blue boxes highlight incorrect detections (false positives or false negatives). Subfigure (<b>f</b>) displays the ground truth for comparison, providing a visual analysis of each method’s accuracy and performance in detecting ship targets across different scenarios.</p>
Full article ">
10 pages, 2102 KiB  
Article
Research on an Echo-Signal-Detection Algorithm for Weak and Small Targets Based on GM-APD Remote Active Single-Photon Technology
by Shengwen Yin, Sining Li, Xin Zhou, Jianfeng Sun, Dongfang Guo, Jie Lu and Hong Zhao
Photonics 2024, 11(12), 1158; https://doi.org/10.3390/photonics11121158 - 9 Dec 2024
Viewed by 777
Abstract
Geiger-mode avalanche photodiode (GM-APD) is a single-photon-detection device characterized by high sensitivity and fast response, which enables it to detect echo signals of distant targets effectively. Given that weak and small targets possess relatively small volumes and occupy only a small number of [...] Read more.
Geiger-mode avalanche photodiode (GM-APD) is a single-photon-detection device characterized by high sensitivity and fast response, which enables it to detect echo signals of distant targets effectively. Given that weak and small targets possess relatively small volumes and occupy only a small number of pixels, relying solely on neighborhood information for target reconstruction proves to be difficult. Furthermore, during long-distance detection, the optical reflection cross-section is small, making signal photons highly susceptible to being submerged by noise. In this paper, a noise fitting and removal algorithm (NFRA) is proposed. This algorithm can detect the position of the echo signal from the photon statistical histogram submerged by noise and facilitate the reconstruction of weak and small targets. To evaluate the NFRA method, this paper establishes an optical detection system for remotely detecting active single-photon weak and small targets based on GM-APD. Taking unmanned aerial vehicles (UAVs) as weak and small targets for detection, this paper compares the target reconstruction effects of the peak-value method and the neighborhood method. It is thereby verified that under the conditions of a 7 km distance and a signal-to-background ratio (SBR) of 0.0044, the NFRA method can effectively detect the weak echo signal of the UAV. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Fine histogram of raw data; (<b>b</b>) fitting curve based on fine histogram; (<b>c</b>) down-sampling the fine histogram; (<b>d</b>) down-sampling the fitting curve; (<b>e</b>) residuals histogram; (<b>f</b>) target signal position interval.</p>
Full article ">Figure 2
<p>Schematic diagram of long-range active single-photon UAV detection.</p>
Full article ">Figure 3
<p>(<b>a</b>) Experimental scene, (<b>b</b>) the DJI Air 3 UAV with a length of 258.8 mm, a width of 326 mm, and a height of 105.8 mm; (<b>c</b>) the DJI Phantom 4 UAV with a length of 430 mm, a width of 430 mm, and a height of 370 mm.</p>
Full article ">Figure 4
<p>(<b>a</b>) Array statistical histogram; (<b>b</b>) single-pixel statistical histogram; (<b>c</b>–<b>j</b>) reconstruction results of the DJI Air 3 UAV at a distance of 1 km using different methods.</p>
Full article ">Figure 5
<p>(<b>a</b>) Array statistical histogram; (<b>b</b>) single-pixel statistical histogram; (<b>c</b>–<b>j</b>) reconstruction results of the DJI Phantom 4 UAV at a distance of 7 km using different methods.</p>
Full article ">
36 pages, 8015 KiB  
Article
A Robust Tuberculosis Diagnosis Using Chest X-Rays Based on a Hybrid Vision Transformer and Principal Component Analysis
by Sameh Abd El-Ghany, Mohammed Elmogy, Mahmood A. Mahmood and A. A. Abd El-Aziz
Diagnostics 2024, 14(23), 2736; https://doi.org/10.3390/diagnostics14232736 - 5 Dec 2024
Viewed by 1039
Abstract
Background: Tuberculosis (TB) is a bacterial disease that mainly affects the lungs, but it can also impact other parts of the body, such as the brain, bones, and kidneys. The disease is caused by a bacterium called Mycobacterium tuberculosis and spreads through [...] Read more.
Background: Tuberculosis (TB) is a bacterial disease that mainly affects the lungs, but it can also impact other parts of the body, such as the brain, bones, and kidneys. The disease is caused by a bacterium called Mycobacterium tuberculosis and spreads through the air when an infected person coughs or sneezes. TB can be inactive or active; in its active state, noticeable symptoms appear, and it can be transmitted to others. There are ongoing challenges in fighting TB, including resistance to medications, co-infections, and limited resources in areas heavily affected by the disease. These issues make it challenging to eradicate TB. Objective: Timely and precise diagnosis is essential for effective control, especially since TB often goes undetected and untreated, particularly in remote and under-resourced locations. Chest X-ray (CXR) images are commonly used to diagnose TB. However, difficulties can arise due to unusual findings on X-rays and a shortage of radiologists in high-infection areas. Method: To address these challenges, a computer-aided diagnosis (CAD) system that uses the vision transformer (ViT) technique has been developed to accurately identify TB in CXR images. This innovative hybrid CAD approach combines ViT with Principal Component Analysis (PCA) and machine learning (ML) techniques for TB classification, introducing a new method in this field. In the hybrid CAD system, ViT is used for deep feature extraction as a base model, PCA is used to reduce feature dimensions, and various ML methods are used to classify TB. This system allows for quickly identifying TB, enabling timely medical action and improving patient outcomes. Additionally, it streamlines the diagnostic process, reducing time and costs for patients and lessening the workload on healthcare professionals. The TB chest X-ray dataset was utilized to train and evaluate the proposed CAD system, which underwent pre-processing techniques like resizing, scaling, and noise removal to improve diagnostic accuracy. Results: The performance of our CAD model was assessed against existing models, yielding excellent results. The model achieved remarkable metrics: an average precision of 99.90%, recall of 99.52%, F1-score of 99.71%, accuracy of 99.84%, false negative rate (FNR) of 0.48%, specificity of 99.52%, and negative predictive value (NPV) of 99.90%. Conclusions: This evaluation highlights the superior performance of our model compared to the latest available classifiers. Full article
Show Figures

Figure 1

Figure 1
<p>Some examples of the TB and normal CXR images: (<b>a</b>–<b>d</b>) normal cases and (<b>e</b>–<b>h</b>) TB cases.</p>
Full article ">Figure 2
<p>The proposed model architecture for TB diagnosis based on a hybrid vision transformer and principal component analysis with machine learning.</p>
Full article ">Figure 3
<p>The multi-head self-attention process.</p>
Full article ">Figure 4
<p>The ViT architecture.</p>
Full article ">Figure 5
<p>The proposed ViT model’s architecture.</p>
Full article ">Figure 6
<p>The training and validation loss for the Six CNNs.</p>
Full article ">Figure 7
<p>The accuracy of the six CNNs for the binary classification.</p>
Full article ">Figure 8
<p>The precision–recall curves for the six CNN models for the binary classification.</p>
Full article ">Figure 9
<p>The ROC curves for the six CNN models for the binary classification.</p>
Full article ">Figure 10
<p>The accuracies of the five ML models.</p>
Full article ">Figure 11
<p>The specificity of the five ML models.</p>
Full article ">Figure 12
<p>The FNRs of the five ML models.</p>
Full article ">Figure 13
<p>The NPVs of the five ML models.</p>
Full article ">Figure 14
<p>The precisions of the five ML models.</p>
Full article ">Figure 15
<p>The recalls of the five ML models.</p>
Full article ">Figure 16
<p>The F1-score of the five ML models.4.4. Discussion of the Proposed CAD Model.</p>
Full article ">
17 pages, 2430 KiB  
Article
PyAMARES, an Open-Source Python Library for Fitting Magnetic Resonance Spectroscopy Data
by Jia Xu, Michael Vaeggemose, Rolf F. Schulte, Baolian Yang, Chu-Yu Lee, Christoffer Laustsen and Vincent A. Magnotta
Diagnostics 2024, 14(23), 2668; https://doi.org/10.3390/diagnostics14232668 - 27 Nov 2024
Viewed by 948
Abstract
Background/Objectives: Magnetic resonance spectroscopy (MRS) is a valuable tool for studying metabolic processes in vivo. While numerous quantification methods exist, the advanced method for accurate, robust, and efficient spectral fitting (AMARES) is among the most used. This study introduces pyAMARES, an open-source [...] Read more.
Background/Objectives: Magnetic resonance spectroscopy (MRS) is a valuable tool for studying metabolic processes in vivo. While numerous quantification methods exist, the advanced method for accurate, robust, and efficient spectral fitting (AMARES) is among the most used. This study introduces pyAMARES, an open-source Python implementation of AMARES, addressing the need for a flexible, user-friendly, and versatile MRS quantification tool within the Python ecosystem. Methods: PyAMARES was developed as a Python library, implementing the AMARES algorithm with additional features such as multiprocessing capabilities and customizable objective functions. The software was validated against established AMARES implementations (OXSA and jMRUI) using both simulated and in vivo MRS data. Monte Carlo simulations were conducted to assess robustness and accuracy across various signal-to-noise ratios and parameter perturbations. Results: PyAMARES utilizes spreadsheet-based prior knowledge and fitting parameter settings, enhancing flexibility and ease of use. It demonstrated comparable performance to existing software in terms of accuracy, precision, and computational efficiency. In addition to conventional AMARES fitting, pyAMARES supports fitting without prior knowledge, frequency-selective AMARES, and metabolite residual removal from mobile macromolecule (MM) spectra. Utilizing multiple CPU cores significantly enhances the performance of pyAMARES. Conclusions: PyAMARES offers a robust, flexible, and user-friendly solution for MRS quantification within the Python ecosystem. Its open-source nature, comprehensive documentation, and integration with popular data science tools enhance reproducibility and collaboration in MRS research. PyAMARES bridges the gap between traditional MRS fitting methods and modern machine learning frameworks, potentially accelerating advancements in metabolic studies and clinical applications. Full article
Show Figures

Figure 1

Figure 1
<p>Flowchart of pyAMARES. The workflow starts with importing prior knowledge from spreadsheets (1a) and loading the FID signal (1b) to establish initial values and constraints for fitting (3). If the initial parameters are far from the actual values, users can optionally employ Hankel singular value decomposition (HSVD) or Levenberg–Marquardt (LM) initializers to optimize these starting values (2a). The FID signal can be processed directly or optionally filtered using MPFIR to focus on specific spectral regions (2b). The non-linear least-squares minimization (4) using either trust region reflective (TRR) or LM, with either default or user-defined objective functions (1c). The fitting process can be iterative—the output can be fine-tuned and used as initial parameters for subsequent iterations (7). The Cramér–Rao lower bound (CRLB) estimation (5) integrates information from both the fitting results and the linear relationships between parameters (2b). These relationships include constraints like fixed amplitude ratios or chemical shift differences between multiplet peaks. The final output (6) includes fitted parameters, their uncertainties (CRLB), and signal-to-noise ratios. Solid arrows indicate the main workflow, while dashed arrows and boxes represent optional processing steps.</p>
Full article ">Figure 2
<p>PyAMARES plotting outputs. The default output figure from the <span class="html-italic">plotAMARES</span> function shows the fit of (<b>A</b>) an in vivo brain <sup>31</sup>P MRS spectrum acquired at 7T [<a href="#B34-diagnostics-14-02668" class="html-bibr">34</a>], (<b>B</b>) a voxel of hyperpolarized <sup>129</sup>Xe MRSI acquired from healthy porcine lungs at 3T, and (<b>D</b>) a voxel of in vivo brain <sup>2</sup>H 3D MRSI spectra acquired at 3T. In (<b>A</b>,<b>B</b>,<b>D</b>), the top panels display the original spectrum (gray), the fitted spectrum (red), and the residual (green dash), with individual fitted components shown in the bottom panels. Panel (<b>A</b>) is shown with phase correction applied (<span class="html-italic">ifphase = True</span> for the <span class="html-italic">plotAMARES</span> function) for display purposes, while (<b>B</b>,<b>D</b>) are not phased. The prior knowledge for the fitting (<b>A</b>) is in <a href="#diagnostics-14-02668-t001" class="html-table">Table 1</a>. The fitting results for <sup>31</sup>P MRS (<b>A</b>), including metabolite concentrations and their respective Cramér–Rao lower bounds (CRLBs), are presented in (<b>C</b>), where green grows indicate reliable fits with CRLB &lt; 20% and red rows indicate less reliable fits. The fitting results of (<b>B</b>,<b>D</b>) are shown in <a href="#app1-diagnostics-14-02668" class="html-app">Figure S2</a>. Abbreviations: RBC, red blood cells; DHO, deuterated water; Glx, combined signals of glutamate and glutamine; PCr: phosphocreatine; PE: phosphoethenolamine; GPE: glycerophosphoethanolamine; GPC: glycerophosphocholine; Pi: inorganic phosphate; NAD, nicotinamide adenine dinucleotide; UDPG, uridine diphosphoglucose.</p>
Full article ">Figure 3
<p>Comparison of Monte Carlo simulated single-peak spectra fitting using OXSA and pyAMARES. (<b>A</b>) Ground truth for spectra simulation with fixed (red) and 3000 perturbed (various colors) parameters. Gaussian noise is omitted for clarity. (<b>B</b>) Relative bias of fitted amplitude compared to ground truth at different SNR levels. (<b>C</b>) Bias of fitted chemical shift compared to ground truth at different SNRs. (<b>D</b>) CRLB of fitted amplitude at each SNR, with the 20% threshold indicated by a green dashed line. In (<b>B</b>–<b>D</b>), blue and red represent pyAMARES and OXSA fitted results, respectively; solid patterns indicate results from spectra simulated with perturbed parameters, while hatched patterns show results from spectra simulated with fixed parameters.</p>
Full article ">Figure 4
<p>Comparison of Monte Carlo simulated in vivo human brain <sup>31</sup>P MRS spectra fitting at 7T using OXSA and different algorithms implemented in pyAMARES. (<b>A</b>) Ground truth for spectra simulation with slightly perturbed parameters. Gaussian noise is omitted for clarity. (<b>B</b>) Relative bias of peak amplitude quantification compared to ground truth. (<b>C</b>) CRLB of fitted amplitude for each peak, with the 20% threshold indicated by a green dashed line. (<b>D</b>) Pearson’s correlation coefficient (R) between OXSA and pyAMARES quantified amplitudes. Abbreviations: LM: Levenberg–Marquardt algorithm; TRR: trust region reflective algorithm; Init: Initializer using LM; PCr: phosphocreatine; PE: phosphoethenolamine; GPE: glycerophosphoethanolamine; GPC: glycerophosphocholine; Pi: inorganic phosphate; NAD, nicotinamide adenine dinucleotide; UDPG, uridine diphosphoglucose.</p>
Full article ">Figure 5
<p>Multiprocessing fitting of dynamic unlocalized <sup>31</sup>P MRS spectra of the tibialis anterior muscle at 3T using pyAMARES and comparison to OXSA. (<b>A</b>). Representative fitting results from pyAMARES (blue solid line) and OXSA (red dash line), with the differences between them shown as green dashed line. The metabolites of interest (PCr and Pi) are labeled. (<b>B</b>). Linear correlations between fitted amplitudes (a.u.), linewidths (Hz), and CRLBs obtained by pyAMARES and OXSA. Pearson’s R and the <span class="html-italic">p</span>-value for each dataset are shown in the plots. (<b>C</b>,<b>D</b>) Time courses of PCr (blue) and Pi (orange) amplitudes fitted by pyAMARES (<b>C</b>) and OXSA (<b>D</b>). The time points at which exercise and recovery start are indicated by dotted and dashed vertical lines, respectively. (<b>E</b>,<b>F</b>) Mono-exponential fitting of the PCr recovery kinetics using pyAMARES (<b>E</b>) and OXSA (<b>F</b>). The fitted equations are PC<sub>recover</sub> = 0.435 − 0.173 × e<sup>−time/44.171</sup>, R<sup>2</sup> = 0.914 for pyAMARES, and PC<sub>recover</sub> = 0.435 − 0.165 × e<sup>−time/42.523</sup>, R<sup>2</sup> = 0.928 for OXSA.</p>
Full article ">Figure 6
<p>Using AMARES for post-processing: Removal of metabolite residuals from a short echo time (TE) <sup>1</sup>H MR spectrum at 9.4T. (<b>A</b>) Upper panel: Fitting of residual metabolites (red) and the resulting macromolecule (MM) spectrum after subtraction of residual metabolite signals from the original spectrum (green). Lower panel: AMARES modeling of residual metabolite signals. (<b>B</b>) Comparison of metabolite-free MM spectra obtained by jMRUI (red) and pyAMARES (blue), showing identical results as confirmed by the flat difference spectrum (black).</p>
Full article ">
20 pages, 21356 KiB  
Article
Utilizing Dual Polarized Array GPR System for Shallow Urban Road Pavement Foundation in Environmental Studies: A Case Study
by Lilong Zou, Ying Li and Amir M. Alani
Remote Sens. 2024, 16(23), 4396; https://doi.org/10.3390/rs16234396 - 24 Nov 2024
Viewed by 1070
Abstract
Maintaining the integrity of urban road pavements is vital for public safety, transportation efficiency, and economic stability. However, aging infrastructure and limited budgets make it challenging to detect subsurface defects that can lead to pavement collapses. Traditional inspection methods are often inadequate for [...] Read more.
Maintaining the integrity of urban road pavements is vital for public safety, transportation efficiency, and economic stability. However, aging infrastructure and limited budgets make it challenging to detect subsurface defects that can lead to pavement collapses. Traditional inspection methods are often inadequate for identifying such underground anomalies. Ground Penetrating Radar (GPR), especially dual-polarized array systems, offers a non-destructive, high-resolution solution for subsurface inspection. Despite its potential, effectively detecting and analyzing areas at risk of collapse in urban pavements remains a challenge. This study employed a dual-polarized array GPR system to inspect road pavements in London. The research involved comprehensive field testing, including data acquisition, signal processing, calibration, background noise removal, and 3D migration for enhanced imaging. Additionally, Short-Fourier Transform Spectrum (SFTS) analysis was applied to detect moisture-related anomalies. The results show that dual-polarized GPR systems effectively detect subsurface issues like voids, cracks, and moisture-induced weaknesses. The ability to capture data in multiple polarizations improves resolution and depth, enabling the identification of collapse-prone areas, particularly in regions with moisture infiltration. This study demonstrates the practical value of dual-polarized GPR technology in urban pavement inspection, offering a reliable tool for early detection of subsurface defects and contributing to the longevity and safety of road infrastructure. Full article
Show Figures

Figure 1

Figure 1
<p>Investigated potential collapse of city road pavement situated in Ealing, London, UK: (<b>a</b>) Google Map; (<b>b</b>) on-site photograph.</p>
Full article ">Figure 2
<p>Dual-polarized array GPR system for investigation of potential collapse areas: (<b>a</b>) RIS Hi-BrigHT GPR system; (<b>b</b>) antenna configuration.</p>
Full article ">Figure 3
<p>Flowchart of signal processing with dual-polarized array GPR data.</p>
Full article ">Figure 4
<p>Dual-polarized array GPR system calibration: (<b>a</b>) antenna direct coupling measurement; (<b>b</b>) phase delay measurement of different channels.</p>
Full article ">Figure 5
<p>Metal plate reflections of HH and VV channels.</p>
Full article ">Figure 6
<p>B-scan reflection profiles acquired by the dual-polarized Array GPR system (HH, VV, and PCF filter).</p>
Full article ">Figure 7
<p>Migration profiles acquired by the dual-polarized Array GPR system (HH, VV, and PCF filter).</p>
Full article ">Figure 8
<p>Migrated profile at 0.1 m; cross-survey direction.</p>
Full article ">Figure 9
<p>GPR peak frequency division profile at 0.1 m; cross-survey direction.</p>
Full article ">Figure 10
<p>Migrated profile at 1 m; cross-survey direction.</p>
Full article ">Figure 11
<p>Migrated profile at 2 m; cross-survey direction.</p>
Full article ">Figure 12
<p>Migrated profile at 2.9 m; cross-survey direction.</p>
Full article ">Figure 13
<p>GPR peak frequency division profile at 1 m; cross-survey direction.</p>
Full article ">Figure 14
<p>GPR peak frequency division profile at 2 m; cross-survey direction.</p>
Full article ">Figure 15
<p>GPR peak frequency division profile at 2.9 m; cross-survey direction.</p>
Full article ">Figure 16
<p>Migrated horizontal slices at 0.21 m depth.</p>
Full article ">Figure 17
<p>Migrated horizontal slices at 0.36 m depth.</p>
Full article ">
17 pages, 4873 KiB  
Article
An Ensemble Approach for Speaker Identification from Audio Files in Noisy Environments
by Syed Shahab Zarin, Ehzaz Mustafa, Sardar Khaliq uz Zaman, Abdallah Namoun and Meshari Huwaytim Alanazi
Appl. Sci. 2024, 14(22), 10426; https://doi.org/10.3390/app142210426 - 13 Nov 2024
Viewed by 708
Abstract
Automatic noise-robust speaker identification is essential in various applications, including forensic analysis, e-commerce, smartphones, and security systems. Audio files containing suspect speech often include background noise, as they are typically not recorded in soundproof environments. To this end, we address the challenges of [...] Read more.
Automatic noise-robust speaker identification is essential in various applications, including forensic analysis, e-commerce, smartphones, and security systems. Audio files containing suspect speech often include background noise, as they are typically not recorded in soundproof environments. To this end, we address the challenges of noise robustness and accuracy in speaker identification systems. An ensemble approach is proposed combining two different neural network architectures including an RNN and DNN using softmax. This approach enhances the system’s ability to identify speakers even in noisy environments accurately. Using softmax, we combine voice activity detection (VAD) with a multilayer perceptron (MLP). The VAD component aims to remove noisy frames from the recording. The softmax function addresses these residual traces by assigning a higher probability to the speaker’s voice compared to the noise. We tested our proposed solution on the Kaggle speaker recognition dataset and compared it to two baseline systems. Experimental results show that our approach outperforms the baseline systems, achieving a 3.6% and 5.8% increase in test accuracy. Additionally, we compared the proposed MLP system with Long Short-Term Memory (LSTM) and Bidirectional LSTM (BiLSTM) classifiers. The results demonstrate that the MLP with VAD and softmax outperforms the LSTM by 23.2% and the BiLSTM by 6.6% in test accuracy. Full article
(This article belongs to the Special Issue Advances in Intelligent Information Systems and AI Applications)
Show Figures

Figure 1

Figure 1
<p>The proposed framework.</p>
Full article ">Figure 2
<p>Illustration of recurrent neural network.</p>
Full article ">Figure 3
<p>Illustration of deep neural network.</p>
Full article ">Figure 4
<p>The proposed MLP classifier.</p>
Full article ">Figure 5
<p>The LSTM network used.</p>
Full article ">Figure 6
<p>The BiLSTM model.</p>
Full article ">Figure 7
<p>The proposed framework compared with baselines in terms of spectrogram features.</p>
Full article ">Figure 8
<p>The proposed framework compared with baselines in terms of MFCC features.</p>
Full article ">Figure 9
<p>MLP model loss with different features.</p>
Full article ">Figure 10
<p>MLP model validation loss.</p>
Full article ">Figure 11
<p>MLP model accuracy.</p>
Full article ">Figure 12
<p>MLP model validation accuracy.</p>
Full article ">Figure 13
<p>Model accuracy of the three models.</p>
Full article ">Figure 14
<p>Validation accuracy comparison.</p>
Full article ">Figure 15
<p>Model loss of the three models.</p>
Full article ">Figure 16
<p>Validation loss of the three models.</p>
Full article ">Figure 17
<p>Model MSE of the three models.</p>
Full article ">Figure 18
<p>Validation MSE of the three models.</p>
Full article ">
13 pages, 3614 KiB  
Article
Automatic Defects Recognition of Lap Joint of Unequal Thickness Based on X-Ray Image Processing
by Dazhao Chi, Ziming Wang and Haichun Liu
Materials 2024, 17(22), 5463; https://doi.org/10.3390/ma17225463 - 8 Nov 2024
Viewed by 668
Abstract
It is difficult to automatically recognize defects using digital image processing methods in X-ray radiographs of lap joints made from plates of unequal thickness. The continuous change in the wall thickness of the lap joint workpiece causes very different gray levels in an [...] Read more.
It is difficult to automatically recognize defects using digital image processing methods in X-ray radiographs of lap joints made from plates of unequal thickness. The continuous change in the wall thickness of the lap joint workpiece causes very different gray levels in an X-ray background image. Furthermore, due to the shape and fixturing of the workpiece, the distribution of the weld seam in the radiograph is not vertical which results in an angle between the weld seam and the vertical direction. This makes automatic defect detection and localization difficult. In this paper, a method of X-ray image correction based on invariant moments is presented to solve the problem. In addition, a novel background removal method based on image processing is introduced to reduce the difficulty of defect recognition caused by variations in grayscale. At the same time, an automatic defect detection method combining image noise suppression, image segmentation, and mathematical morphology is adopted. The results show that the proposed method can effectively recognize the gas pores in an automatic welded lap joint of unequal thickness, making it suitable for automatic detection. Full article
Show Figures

Figure 1

Figure 1
<p>Preparation for weld specimen: (<b>a</b>) Geometric form of the joint, (<b>b</b>) Appearance of the weld specimen.</p>
Full article ">Figure 1 Cont.
<p>Preparation for weld specimen: (<b>a</b>) Geometric form of the joint, (<b>b</b>) Appearance of the weld specimen.</p>
Full article ">Figure 2
<p>Overall testing system and defect testing methods.</p>
Full article ">Figure 3
<p>Positioning of the weld under testing.</p>
Full article ">Figure 4
<p>Image correction steps.</p>
Full article ">Figure 5
<p>Digital image processing for defect detection.</p>
Full article ">Figure 6
<p>Background removal. (<b>a</b>) Cross-section of the lap joint. (<b>b</b>) Grayscale distribution of the radiograph. (<b>c</b>) Linear grayscale distribution without defect. (<b>d</b>) Linear grayscale distribution with defect. (<b>e</b>) Linear grayscale distribution of background. (<b>f</b>) Linear grayscale distribution of foreground.</p>
Full article ">Figure 7
<p>Image corrections: (<b>a</b>) Original radiograph, (<b>b</b>) Contour extraction, (<b>c</b>) Image correction, (<b>d</b>) Image corrected.</p>
Full article ">Figure 8
<p>Defect detection images: (<b>a</b>) Noise suppression, (<b>b</b>) Background image, (<b>c</b>) Foreground image, (<b>d</b>) Image segmentation, (<b>e</b>) Mathematical morphology.</p>
Full article ">
Back to TopTop