Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,527)

Search Parameters:
Keywords = denoising

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 5880 KiB  
Article
Nonlinear Seismic Signal Denoising Using Template Matching with Time Difference Detection Method
by Rongwei Xu, Bo Feng, Huazhong Wang, Chengliang Wu and Zhenbo Nie
Remote Sens. 2025, 17(4), 674; https://doi.org/10.3390/rs17040674 (registering DOI) - 16 Feb 2025
Abstract
As seismic exploration shifts towards areas with more complex surface and subsurface structures, the complexity of the geological conditions often results in seismic data with low signal-to-noise ratio. It is therefore essential to implement denoising in order to enhance the signal-to-noise ratio of [...] Read more.
As seismic exploration shifts towards areas with more complex surface and subsurface structures, the complexity of the geological conditions often results in seismic data with low signal-to-noise ratio. It is therefore essential to implement denoising in order to enhance the signal-to-noise ratio of the seismic data. At present, the prevailing denoising techniques are based on the assumption that the signal adheres to linear model. However, this assumption is frequently invalid in complex geological conditions. The main challenge lies in the fact that linear models, which are foundational to traditional signal processing, fail to capture the nonlinear components of seismic signals. The objective of this paper is to present a methodology for the detection of nonlinear signal structures, with a particular focus on nonlinear time differences. We propose a method for detecting nonlinear time differences based on template matching, wherein the seismic wavelet is treated as the template. Template matching, a fundamental pattern recognition technique, plays a key role in identifying nonlinear structures within signals. By employing a local signal as a template, the template matching technique can identify all the structure of the signal, thereby enabling the detection of nonlinear features. By employing template matching, the nonlinear time differences in the signal are identified and corrected, thus enabling the signal to align with the assumption of linearity. Subsequently, linear denoising methods are employed to effectively remove noise and enhance the signal-to-noise ratio. The results of numerical experiments demonstrate that the proposed template matching method is highly accurate in detecting nonlinear time differences. Furthermore, the method’s efficacy in removing random noise from real seismic data is evident, underscoring its superiority. Full article
25 pages, 6991 KiB  
Article
A Comprehensive AI Framework for Superior Diagnosis, Cranial Reconstruction, and Implant Generation for Diverse Cranial Defects
by Mamta Juneja, Ishaan Singla, Aditya Poddar, Nitin Pandey, Aparna Goel, Agrima Sudhir, Pankhuri Bhatia, Gurzafar Singh, Maanya Kharbanda, Amanpreet Kaur, Ira Bhatia, Vipin Gupta, Sukhdeep Singh Dhami, Yvonne Reinwald, Prashant Jindal and Philip Breedon
Bioengineering 2025, 12(2), 188; https://doi.org/10.3390/bioengineering12020188 (registering DOI) - 16 Feb 2025
Abstract
Cranioplasty enables the restoration of cranial defects caused by traumatic injuries, brain tumour excisions, or decompressive craniectomies. Conventional methods rely on Computer-Aided Design (CAD) for implant design, which requires significant resources and expertise. Recent advancements in Artificial Intelligence (AI) have improved Computer-Aided Diagnostic [...] Read more.
Cranioplasty enables the restoration of cranial defects caused by traumatic injuries, brain tumour excisions, or decompressive craniectomies. Conventional methods rely on Computer-Aided Design (CAD) for implant design, which requires significant resources and expertise. Recent advancements in Artificial Intelligence (AI) have improved Computer-Aided Diagnostic systems for accurate and faster cranial reconstruction and implant generation procedures. However, these face inherent limitations, including the limited availability of diverse datasets covering different defect shapes spanning various locations, absence of a comprehensive pipeline integrating the preprocessing of medical images, cranial reconstruction, and implant generation, along with mechanical testing and validation. The proposed framework incorporates a robust preprocessing pipeline for easier processing of Computed Tomography (CT) images through data conversion, denoising, Connected Component Analysis (CCA), and image alignment. At its core is CRIGNet (Cranial Reconstruction and Implant Generation Network), a novel deep learning model rigorously trained on a diverse dataset of 2160 images, which was prepared by simulating cylindrical, cubical, spherical, and triangular prism-shaped defects across five skull regions, ensuring robustness in diagnosing a wide variety of defect patterns. CRIGNet achieved an exceptional reconstruction accuracy with a Dice Similarity Coefficient (DSC) of 0.99, Jaccard Similarity Coefficient (JSC) of 0.98, and Hausdorff distance (HD) of 4.63 mm. The generated implants showed superior geometric accuracy, load-bearing capacity, and gap-free fitment in the defected skull compared to CAD-generated implants. Also, this framework reduced the implant generation processing time from 40–45 min (CAD) to 25–30 s, suggesting its application for a faster turnaround time, enabling decisive clinical support systems. Full article
(This article belongs to the Section Biosignal Processing)
Show Figures

Figure 1

Figure 1
<p>Proposed end-to-end framework for cranial reconstruction and implant generation and its assessment: (<b>a</b>) data preprocessing, (<b>b</b>) defect generation, (<b>c</b>) AI modelling, and (<b>d</b>) assessment of CRIGNet generated implants.</p>
Full article ">Figure 2
<p>Regions for defect generation.</p>
Full article ">Figure 3
<p>Architecture of the proposed AI model CRIGNet.</p>
Full article ">Figure 4
<p>Results of various preprocessing steps: (<b>a</b>) RAW DICOM data, (<b>b</b>) noisy NRRD data found in MUG500+ dataset, (<b>c</b>) denoised NRRD file containing extraneous artefacts, (<b>d</b>) clean NRRD file with artefacts removed through CCA, and (<b>e</b>) aligned NRRD file.</p>
Full article ">Figure 5
<p>Visualisations of the various defect types created using four defect mask shapes in the five identified regions on a representative case from the MUG500+ dataset. (<b>a</b>) Masks, (<b>b</b>) Regions.</p>
Full article ">Figure 6
<p>Visual results of the CRIGNet-based reconstruction and implant generation across different regions on five arbitrarily chosen test cases: (<b>a</b>) defected skull (<b>b</b>), reconstructed skull, (<b>c</b>) ground truth implant, and (<b>d</b>) predicted implant.</p>
Full article ">Figure 7
<p>Box plots for quantitative comparison of CRIGNet with existing AI models for cranial reconstruction based on (<b>a</b>) DSC, (<b>b</b>) JSC, (<b>c</b>) HD, (<b>d</b>) precision, (<b>e</b>) recall, and (<b>f</b>) specificity. * Represents the modified architecture as described in <a href="#sec3dot5-bioengineering-12-00188" class="html-sec">Section 3.5</a>.</p>
Full article ">Figure 7 Cont.
<p>Box plots for quantitative comparison of CRIGNet with existing AI models for cranial reconstruction based on (<b>a</b>) DSC, (<b>b</b>) JSC, (<b>c</b>) HD, (<b>d</b>) precision, (<b>e</b>) recall, and (<b>f</b>) specificity. * Represents the modified architecture as described in <a href="#sec3dot5-bioengineering-12-00188" class="html-sec">Section 3.5</a>.</p>
Full article ">Figure 8
<p>CAD-generated implant (Cubical Left).</p>
Full article ">Figure 9
<p>Different views of the skull (grey) with CAD-generated implant (white) and four linear fixture plates.</p>
Full article ">Figure 10
<p>Different views of the skull (grey) with CRIGNet-generated implant (green) and four linear fixture plates.</p>
Full article ">Figure 11
<p>Cavity (red bound) of the defected skull.</p>
Full article ">Figure 12
<p>Edge gap between the boundaries of cavity of defected skull for (<b>a</b>) CAD-generated implant and (<b>b</b>) CRIGNet-generated implant.</p>
Full article ">Figure 13
<p>Simulation of CRIGNet-generated implant in ANSYS: (<b>a</b>) equivalent von Mises stress distribution and (<b>b</b>) total deformation.</p>
Full article ">
29 pages, 3663 KiB  
Article
Advanced Tax Fraud Detection: A Soft-Voting Ensemble Based on GAN and Encoder Architecture
by Masad A. Alrasheedi, Samia Ijaz, Ayed M. Alrashdi and Seung-Won Lee
Mathematics 2025, 13(4), 642; https://doi.org/10.3390/math13040642 (registering DOI) - 16 Feb 2025
Abstract
The world prevalence of the two types of authorized and fraudulent transactions makes it difficult to distinguish between the two operations. The small percentage of fraudulent transactions, in turn, gives rise to the class imbalance problem. Hence, an adequately robust fraud detection mechanism [...] Read more.
The world prevalence of the two types of authorized and fraudulent transactions makes it difficult to distinguish between the two operations. The small percentage of fraudulent transactions, in turn, gives rise to the class imbalance problem. Hence, an adequately robust fraud detection mechanism must exist for tax systems to avoid their collapse. It has become significantly difficult to obtain any dataset, specifically a tax return dataset, because of the rising importance of privacy in a society where people generally feel squeamish about sharing personal information. Because of this, we arrive at the decision to synthesize our dataset by employing publicly available data, as well as enhance them through Correlational Generative Adversarial Networks (CGANs) and the Synthetic Minority Oversampling Technique (SMOTE). The proposed method includes a preprocessing stage to denoise the data and identify anomalies, outliers, and dimensionality reduction. Then the data have undergone enhancement using the SMOTE and the proposed CGAN techniques. A unique encoder design has been proposed, which serves the purpose of exposing the hidden patterns among legitimate and fraudulent records. This research found anomalous deductions, income inconsistencies, recurrent transaction manipulations, and irregular filing practices that distinguish fraudulent from valid tax records. These patterns are identified by encoder-based feature extraction and synthetic data augmentation. Several machine learning classifiers, along with a voting ensemble technique, have been used both with and without data augmentation. Experimental results have shown that the proposed Soft-Voting technique outperformed the original without an ensemble method. Full article
(This article belongs to the Special Issue Application of Artificial Intelligence in Decision Making)
20 pages, 8383 KiB  
Article
Self-Supervised Time-Series Preprocessing Framework for Maritime Applications
by Shengli Dong, Jilong Liu, Bing Han, Shengzheng Wang, Hong Zeng and Meng Zhang
Electronics 2025, 14(4), 765; https://doi.org/10.3390/electronics14040765 (registering DOI) - 16 Feb 2025
Abstract
This study proposes a novel self-supervised data-preprocessing framework for time-series forecasting in complex ship systems. The framework integrates an improved Learnable Wavelet Packet Transform (L-WPT) for adaptive denoising and a correlation-based Uniform Manifold Approximation and Projection (UMAP) approach for dimensionality reduction. The enhanced [...] Read more.
This study proposes a novel self-supervised data-preprocessing framework for time-series forecasting in complex ship systems. The framework integrates an improved Learnable Wavelet Packet Transform (L-WPT) for adaptive denoising and a correlation-based Uniform Manifold Approximation and Projection (UMAP) approach for dimensionality reduction. The enhanced L-WPT incorporates Reversible Instance Normalization to improve training efficiency while preserving denoising performance, especially for low-frequency sporadic noise. The UMAP dimensionality reduction, combined with a modified K-means clustering using correlation coefficients, enhances the computational efficiency and interpretability of the reduced data. Experimental results validate that state-of-the-art time-series models can effectively forecast the data processed by this framework, achieving promising MSE and MAE metrics. Full article
Show Figures

Figure 1

Figure 1
<p>L-WPT networks Structure.</p>
Full article ">Figure 2
<p>Noise Features in Ship Lubricating Oil Temperature Data.</p>
Full article ">Figure 3
<p>Spectral Analysis of Ship Lubricating Oil Temperature Data: (<b>a</b>) Power Spectral Density Analysis and (<b>b</b>) Frequency Spectrum Analysis.</p>
Full article ">Figure 4
<p>Detailed View of High Signal-to-Noise Ratio Region in Lubricating Oil Temperature Measurements.</p>
Full article ">Figure 5
<p>Comparison of Different Denoising Methods.</p>
Full article ">Figure 6
<p>SNR Before and After Denoising for the Entire Dataset.</p>
Full article ">Figure 7
<p>Loss trend of ship lubricating oil temperature data processed by L-WPT.</p>
Full article ">Figure 8
<p>Loss trend of ship lubricating oil temperature data processed by L-WPT-RevIn.</p>
Full article ">Figure 9
<p>Cluster Analysis Results: (<b>a</b>) Silhouette Score Analysis and (<b>b</b>) Elbow Method for Optimal Cluster Determination.</p>
Full article ">Figure 10
<p>UMAP Principal Components of Raw Time-Series Data Before Denoising: (<b>a</b>) First Principal Component Trajectory and (<b>b</b>) Second Principal Component Trajectory.</p>
Full article ">Figure 11
<p>UMAP Principal Components of Denoised Time-Series Data: (<b>a</b>) First Principal Component Trajectory and (<b>b</b>) Second Principal Component Trajectory After L-WPT Processing.</p>
Full article ">Figure 12
<p>Minimum principal components for 99.5% variance per cluster.</p>
Full article ">Figure 13
<p>LCR results for each cluster after dimensionality reduction.</p>
Full article ">Figure 14
<p>MSE and MAE Error Distribution Across Features in TimeMixer Model.</p>
Full article ">Figure 15
<p>Error Curves of TimeMixer Model’s Principal Component(PC) Features After UMAP Reduction.</p>
Full article ">
15 pages, 1300 KiB  
Article
PyMAP: Python-Based Data Analysis Package with a New Image Cleaning Method to Enhance the Sensitivity of MACE Telescope
by Mani Khurana, Kuldeep Kumar Yadav, Pradeep Chandra, Krishna Kumar Singh, Atul Pathania and Chinmay Borwankar
Galaxies 2025, 13(1), 14; https://doi.org/10.3390/galaxies13010014 (registering DOI) - 15 Feb 2025
Abstract
Observations of Very High Energy (VHE) gamma ray sources using the ground-based Imaging Atmospheric Cherenkov Telescopes (IACTs) play a pivotal role in understanding the non-thermal energetic phenomena and acceleration processes under extreme astrophysical conditions. However, detection of the VHE gamma ray signal from [...] Read more.
Observations of Very High Energy (VHE) gamma ray sources using the ground-based Imaging Atmospheric Cherenkov Telescopes (IACTs) play a pivotal role in understanding the non-thermal energetic phenomena and acceleration processes under extreme astrophysical conditions. However, detection of the VHE gamma ray signal from the astrophysical sources is very challenging, as these telescopes detect the photons indirectly by measuring the flash of Cherenkov light from the Extensive Air Showers (EAS) initiated by the cosmic gamma rays in the Earth’s atmosphere. This requires fast detection systems, along with advanced data acquisition and analysis techniques to measure the development of extensive air showers and the subsequent segregation of gamma ray events from the huge cosmic ray background, followed by the physics analysis of the signal. Here, we report the development of a python-based package for analyzing the data from the Major Atmospheric Cherenkov Experiment (MACE), which is operational at Hanle in India. The Python-based MACE data Analysis Package (PyMAP) analyzes data by using advanced methods and machine learning algorithms. Data recorded by the MACE telescope are passed through different utilities developed in the PyMAP to extract the gamma ray signal from a given source direction. We also propose a new image cleaning method called DIOS (Denoising Image of Shower) and compare its performance with the standard image cleaning method. The working performance of DIOS indicates an advantage over the standard method with an improvement of ≈25% in the sensitivity of MACE. Full article
Show Figures

Figure 1

Figure 1
<p>Representation of the Hillas Parameters on the camera plane.</p>
Full article ">Figure 2
<p>(<b>a</b>) Raw simulated EAS image on the camera plane. (<b>b</b>) Cleaned image obtained after the implementation of image cleaning tools. The image extracted is parameterized using the Hillas parameterization technique.</p>
Full article ">Figure 3
<p>The hscore distribution for <math display="inline"><semantics> <mi>γ</mi> </semantics></math> and hadron events, green-shaded region represents hscore value for gamma like events and red for the cosmic ray events. The gamma rays were simulated with a differential energy spectrum given by <math display="inline"><semantics> <mrow> <mi>d</mi> <mi>N</mi> <mo>/</mo> <mi>d</mi> <mi>E</mi> <mo>∝</mo> <msup> <mi>E</mi> <mrow> <mo>−</mo> <mn>2.59</mn> </mrow> </msup> </mrow> </semantics></math> in the energy range of 10 GeV to 20 TeV, whereas the cosmic ray protons were simulated in the energy range of 20 GeV to 20 TeV with a spectral index of 2.7.</p>
Full article ">Figure 4
<p>Pie chart showing feature importance of various Hillas parameters for gamma/hadron classification.</p>
Full article ">Figure 5
<p>Working schematic for a new image cleaning method DIOS. Different scenarios and cases are illustrated to explain the procedure behind selecting pixels that are part of an image. The blue color pixel (not yet checked), orange colour pixel (under consideration), green is “ok pixels” (will be a part of the cleaned image), red pixels are “rejected pixels” (will not be a part of a cleaned image).</p>
Full article ">Figure 6
<p>The given figure illustrates an example of a simulated gamma ray image in the camera plane where (<b>a</b>) represents an image after sky background removal, (<b>b</b>) is a cleaned image using standard cleaning, and (<b>c</b>) is a cleaned image using DIOS cleaning method.</p>
Full article ">Figure 7
<p>Alpha distribution of all simulated gamma rays events (cuts applied) represented by solid line and cosmic ray background events (cuts applied) represented by dotted line, for two different image cleaning techniques. Different numbers of data samples are used for background and gamma rays. The green is for DIOS and red is for standard method. A total of 30% of total simulated events are used to generate the alpha distribution.</p>
Full article ">Figure 8
<p>The normalized plots for the Hillas parameter distributions—log10 (size), distance, length, and width are shown. The blue distributions represent the observed off data, while the red distributions correspond to the simulated off data.</p>
Full article ">Figure 9
<p>DIOS Alpha parameter distribution, the green-shaded region is from the Crab source direction and the blue-shaded region represents the background obtained from the Off region. Signal estimation is done using ON-OFF analysis. The signal region considered here is up to 25 degrees.</p>
Full article ">Figure 10
<p>(<b>a</b>) Plot is the effective collection area of the MACE after applying two image cleaning methods. The red line is for standard cleaning and the blue is for the DIOS method. (<b>b</b>) Plot is the differential rate curves for gamma rays using DIOS and standard cleaning method. We have used a power law spectrum given in Equation (<a href="#FD7-galaxies-13-00014" class="html-disp-formula">7</a>). The two differential rates corresponding to power law spectrum of CRAB nebula peak at ∼80 GeV with the standard method and ∼60 GeV with the DIOS method.</p>
Full article ">Figure 11
<p>The differential sensitivity plot for both image cleaning methods is expressed in Crab Units. The red solid line represents the standard method, while the blue line corresponds to the DIOS method.</p>
Full article ">Figure A1
<p>Dataflow of various utilities in PyMAP. Light green italic represents the concerned utility name under PyMAP.</p>
Full article ">
22 pages, 4878 KiB  
Article
Application of Unmanned Aerial Vehicle 3D Reconstruction for Quantitative Measurements of the Tailing Pond Dry Beach
by Yani Li, Guangyin Lu, Bei Cao, Xudong Zhu and Zicheng Yang
Appl. Sci. 2025, 15(4), 2014; https://doi.org/10.3390/app15042014 - 14 Feb 2025
Abstract
Dry beach parameters are important indicators for measuring the safe operation of tailings ponds. The Unmanned Aerial Vehicle (UAV) is widely used in construction, disaster monitoring, and agriculture. In this paper, a new semi-automatic method is proposed that measures the length and slope [...] Read more.
Dry beach parameters are important indicators for measuring the safe operation of tailings ponds. The Unmanned Aerial Vehicle (UAV) is widely used in construction, disaster monitoring, and agriculture. In this paper, a new semi-automatic method is proposed that measures the length and slope of the dry beach in a tailings pond using the point cloud obtained through the 3D reconstruction of UAVs. The programs of dry beach boundary identification, point cloud profile denoising, dry beach length, and slope measurement were established. The proposed method was applied to a case study to monitor the length and slope of a tailings pond dry beach. The results show that the maximum error of the dry beach length measurement is 2.25%, and the error of the slope measurement is 2.67%. The error is within a reasonable range, which can be accepted in practical applications, demonstrating the reliability and accuracy of the proposed method. Compared with traditional methods and image-based dry-beach-monitoring methods, the point cloud model provides comprehensive coverage of the entire dry beach. It enables the measurement of the length and slope of the dry beach at any point while offering an intuitive representation of its surface morphology and topographic features. Full article
18 pages, 1041 KiB  
Article
SCConv-Denoising Diffusion Probabilistic Model Anomaly Detection Based on TimesNet
by Jingquan Zhou, Xinhe Yang and Zhu Ren
Electronics 2025, 14(4), 746; https://doi.org/10.3390/electronics14040746 - 14 Feb 2025
Abstract
Time series anomaly detection is a significant challenge due to the inherent complexity and diversity of time series data. Traditional methods for time series anomaly detection (TAD) often struggle to effectively address the intricate nature of a complex time series and the composite [...] Read more.
Time series anomaly detection is a significant challenge due to the inherent complexity and diversity of time series data. Traditional methods for time series anomaly detection (TAD) often struggle to effectively address the intricate nature of a complex time series and the composite characteristics of diverse anomalies. In this paper, we propose SCConv-Denoising Diffusion Probabilistic Model Anomaly Detection Based on TimesNet (SDADT), a novel framework that integrates the Spatial and Channel Reconstruction Convolution (SCConv) module and Denoising Diffusion Probabilistic Models (DDPMs) to address these challenges. By transforming 1D time series into 2D tensors via TimesNet, our method captures intra- and inter-period variations, achieving state-of-the-art performance across three real-world datasets: 85.39% F1-score on SMD, 92.76% on SWaT, and 97.36% on PSM, outperforming nine baseline models including Transformers and LSTM. Ablation studies confirm the necessity of both modules, with performance dropping significantly when either SCConv or DDPMs are removed. In conclusion, this paper proposes a novel alternative solution for anomaly detection in the Cyber Physical Systems (CPSs) domain. Full article
Show Figures

Figure 1

Figure 1
<p>Framework of SDADT.</p>
Full article ">Figure 2
<p>The multi-periodicity and temporal two-dimensional variation of a time series.</p>
Full article ">Figure 3
<p>A univariate example for illustrating the 2D structure in a time series.</p>
Full article ">Figure 4
<p>Framework of TimesNet.</p>
Full article ">Figure 5
<p>Framework of SCConv.</p>
Full article ">Figure 6
<p>Comparison of F1-scores in ablation experiments.</p>
Full article ">Figure 7
<p>Performance of F1-score with varying k values.</p>
Full article ">Figure 8
<p>Performance of F1-score with varying image_size values.</p>
Full article ">Figure 9
<p>Performance of F1-score with varying seq_len values.</p>
Full article ">
22 pages, 11164 KiB  
Article
Acoustic Emission-Based Pipeline Leak Detection and Size Identification Using a Customized One-Dimensional DenseNet
by Faisal Saleem, Zahoor Ahmad, Muhammad Farooq Siddique, Muhammad Umar and Jong-Myon Kim
Sensors 2025, 25(4), 1112; https://doi.org/10.3390/s25041112 - 12 Feb 2025
Abstract
Effective leak detection and leak size identification are essential for maintaining the operational safety, integrity, and longevity of industrial pipelines. Traditional methods often suffer from high noise sensitivity, limited adaptability to non-stationary signals, and excessive computational costs, which limits their feasibility for real-time [...] Read more.
Effective leak detection and leak size identification are essential for maintaining the operational safety, integrity, and longevity of industrial pipelines. Traditional methods often suffer from high noise sensitivity, limited adaptability to non-stationary signals, and excessive computational costs, which limits their feasibility for real-time monitoring applications. This study presents a novel acoustic emission (AE)-based pipeline monitoring approach, integrating Empirical Wavelet Transform (EWT) for adaptive frequency decomposition with customized one-dimensional DenseNet architecture to achieve precise leak detection and size classification. The methodology begins with EWT-based signal segmentation, which isolates meaningful frequency bands to enhance leak-related feature extraction. To further improve signal quality, adaptive thresholding and denoising techniques are applied, filtering out low-amplitude noise while preserving critical diagnostic information. The denoised signals are processed using a DenseNet-based deep learning model, which combines convolutional layers and densely connected feature propagation to extract fine-grained temporal dependencies, ensuring the accurate classification of leak presence and severity. Experimental validation was conducted on real-world AE data collected under controlled leak and non-leak conditions at varying pressure levels. The proposed model achieved an exceptional leak detection accuracy of 99.76%, demonstrating its ability to reliably differentiate between normal operation and multiple leak severities. This method effectively reduces computational costs while maintaining robust performance across diverse operating environments. Full article
(This article belongs to the Special Issue Feature Papers in Fault Diagnosis & Sensors 2025)
Show Figures

Figure 1

Figure 1
<p>Graphical workflow of the proposed methodology.</p>
Full article ">Figure 2
<p>Flowchart of the signal preprocessing steps.</p>
Full article ">Figure 3
<p>Intrinsic mode functions for (<b>a</b>) non-leak signal and (<b>b</b>) leak signal.</p>
Full article ">Figure 4
<p>One-dimensional CNN architecture.</p>
Full article ">Figure 5
<p>DenseNet architecture.</p>
Full article ">Figure 6
<p>Experimental setup for pipeline leak detection.</p>
Full article ">Figure 7
<p>Pipeline architecture for the experiment.</p>
Full article ">Figure 8
<p>AE signals at 13-bar pressure: (<b>a</b>) normal; (<b>b</b>) leak.</p>
Full article ">Figure 9
<p>AE signals at 18-bar pressure: (<b>a</b>) normal; (<b>b</b>) leak.</p>
Full article ">Figure 10
<p>Confusion matrices for leak detection of (<b>a</b>) proposed method; (<b>b</b>) 1D CNN; (<b>c</b>) LSTM; and (<b>d</b>) XGBoost.</p>
Full article ">Figure 10 Cont.
<p>Confusion matrices for leak detection of (<b>a</b>) proposed method; (<b>b</b>) 1D CNN; (<b>c</b>) LSTM; and (<b>d</b>) XGBoost.</p>
Full article ">Figure 11
<p>Confusion matrices for leak size identification of (<b>a</b>) proposed method; (<b>b</b>) 1D CNN; (<b>c</b>) LSTM; and (<b>d</b>) XGBoost.</p>
Full article ">Figure 12
<p>t-SNE plots for leak detection of (<b>a</b>) proposed method; (<b>b</b>) 1D CNN; (<b>c</b>) LSTM; and (<b>d</b>) XGBoost.</p>
Full article ">Figure 13
<p>t-SNE plots for leak size identification of (<b>a</b>) proposed method; (<b>b</b>) 1D CNN; (<b>c</b>) LSTM; and (<b>d</b>) XGBoost.</p>
Full article ">Figure 13 Cont.
<p>t-SNE plots for leak size identification of (<b>a</b>) proposed method; (<b>b</b>) 1D CNN; (<b>c</b>) LSTM; and (<b>d</b>) XGBoost.</p>
Full article ">
21 pages, 5199 KiB  
Article
Enhanced U-Net with Multi-Module Integration for High-Exposure-Difference Image Restoration
by Bo-Lin Jian, Hong-Li Chang and Chieh-Li Chen
Sensors 2025, 25(4), 1105; https://doi.org/10.3390/s25041105 - 12 Feb 2025
Abstract
Machine vision systems have become key unmanned vehicle (UAV) sensing systems. However, under different weather conditions, the lighting direction and the selection of exposure parameters often lead to insufficient or missing object features in images, which could fail to perform various tasks. As [...] Read more.
Machine vision systems have become key unmanned vehicle (UAV) sensing systems. However, under different weather conditions, the lighting direction and the selection of exposure parameters often lead to insufficient or missing object features in images, which could fail to perform various tasks. As a result, images need to be restored to secure information that is accessible when facing a light exposure difference environment. Many applications require real-time and high-quality images; therefore, efficiently restoring images is also important for subsequent tasks. This study adopts supervised learning to solve the problem of images under lighting discrepancies using a U-Net as our main architecture of the network and adding suitable modules to its encoder and decoder, such as inception-like blocks, dual attention units, selective kernel feature fusion, and denoising blocks. In addition to the ablation study, we also compared the quality of image light restoration with other network models using BAID and considered the overall trainable parameters of the model to construct a lightweight, high-exposure-difference image restoration model. The performance of the proposed network was demonstrated by enhancing image detection and recognition. Full article
(This article belongs to the Special Issue AI-Based Computer Vision Sensors & Systems)
Show Figures

Figure 1

Figure 1
<p>The proposed CNN architecture for image restoration with high exposure difference.</p>
Full article ">Figure 2
<p>Inception-like block [<a href="#B14-sensors-25-01105" class="html-bibr">14</a>].</p>
Full article ">Figure 3
<p>Dual attention unit [<a href="#B6-sensors-25-01105" class="html-bibr">6</a>]. C denotes the concatenation operation. d is the feature descriptor. <math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mi mathvariant="normal">d</mi> </mrow> <mo stretchy="false">^</mo> </mover> </mrow> </semantics></math> generates activations. M, and f is the feature map. <math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mi mathvariant="normal">f</mi> </mrow> <mo stretchy="false">^</mo> </mover> </mrow> </semantics></math> is the spatial attention map.</p>
Full article ">Figure 4
<p>Selective kernel feature fusion [<a href="#B6-sensors-25-01105" class="html-bibr">6</a>].</p>
Full article ">Figure 5
<p>Denoising block.</p>
Full article ">Figure 6
<p>Randomly selected images of HED images and corresponding clear images from BAID. (<b>a</b>). Clear images; (<b>b</b>) HED images.</p>
Full article ">Figure 7
<p>Before-and-after indoor image restoration in the BAID test dataset.</p>
Full article ">Figure 8
<p>Before-and-after outdoor image restoration in the BAID test dataset.</p>
Full article ">Figure 9
<p>Comparison of restoration results [<a href="#B25-sensors-25-01105" class="html-bibr">25</a>].</p>
Full article ">Figure 10
<p>Before-and-after image restoration.</p>
Full article ">Figure 11
<p>Object detection before and after restoration.</p>
Full article ">Figure 12
<p>Additional object detection before and after restoration.</p>
Full article ">
15 pages, 3221 KiB  
Article
Domain Adaptation Based on Human Feedback for Enhancing Image Denoising in Generative Models
by Hyun-Cheol Park, Dat Ngo and Sung Ho Kang
Mathematics 2025, 13(4), 598; https://doi.org/10.3390/math13040598 - 12 Feb 2025
Abstract
How can human feedback be effectively integrated into generative models? This study addresses this question by proposing a method to enhance image denoising and achieve domain adaptation using human feedback. Deep generative models, while achieving remarkable performance in image denoising within training domains, [...] Read more.
How can human feedback be effectively integrated into generative models? This study addresses this question by proposing a method to enhance image denoising and achieve domain adaptation using human feedback. Deep generative models, while achieving remarkable performance in image denoising within training domains, often fail to generalize to unseen domains. To overcome this limitation, we introduce a novel approach that fine-tunes a denoising model using human feedback without requiring labeled target data. Our experiments demonstrate a significant improvement in denoising performance. For example, on the Fashion-MNIST test set, the peak signal-to-noise ratio (PSNR) increased by 94%, with an average improvement of 1.61 ± 2.78 dB and a maximum increase of 18.21 dB. Additionally, the proposed method effectively prevents catastrophic forgetting, as evidenced by the consistent performance on the original MNIST domain. By leveraging a reward model trained on human preferences, we show that the quality of denoised images can be significantly improved, even when applied to unseen target data. This work highlights the potential of human feedback for efficient domain adaptation in generative models, presenting a scalable and data-efficient solution for enhancing performance in diverse domains. Full article
(This article belongs to the Special Issue The Application of Deep Neural Networks in Image Processing)
Show Figures

Figure 1

Figure 1
<p>Overview of adaptation training: Step 1 involves pre-training the generator, Step 2 uses a reward model trained with human feedback, and Step 3 fine-tunes the generator for domain adaptation.</p>
Full article ">Figure 2
<p>The training data set ‘0’ is sourced from the MNIST data set, while the new domain data set is the Fashion-MNIST data set. The <math display="inline"><semantics> <msub> <mi>G</mi> <mi>s</mi> </msub> </semantics></math> model in Step 2 is trained on the MNIST data set during Step 1. Subsequently, the <math display="inline"><semantics> <msub> <mi>G</mi> <mi>s</mi> </msub> </semantics></math> model in Step 3 is fine-tuned using the reward model based on human feedback.</p>
Full article ">Figure 3
<p>Flow diagram of the final objective loss functions: <math display="inline"><semantics> <msub> <mi>L</mi> <mi>r</mi> </msub> </semantics></math> drives domain adaptation, while <math display="inline"><semantics> <msub> <mi>L</mi> <mi>p</mi> </msub> </semantics></math> prevents catastrophic forgetting, and <math display="inline"><semantics> <msub> <mi>L</mi> <mi>n</mi> </msub> </semantics></math> ensures regularization.</p>
Full article ">Figure 4
<p>Visual results for adaptation. The PSNR and SSIM values for each image are calculated with respect to the ground truth. <math display="inline"><semantics> <msub> <mi>G</mi> <mi>s</mi> </msub> </semantics></math> represents the model pre-trained on MNIST, while <math display="inline"><semantics> <msub> <mi>G</mi> <mi>t</mi> </msub> </semantics></math> represents the model fine-tuned from <math display="inline"><semantics> <msub> <mi>G</mi> <mi>s</mi> </msub> </semantics></math> using human feedback. (<b>a</b>) Sample images with the most significant increase in PSNR from <math display="inline"><semantics> <msub> <mi>G</mi> <mi>s</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>G</mi> <mi>t</mi> </msub> </semantics></math> output. (<b>b</b>) Most decreased PSNR images.</p>
Full article ">Figure 5
<p>Boxplot of PSNR for each generator <math display="inline"><semantics> <msub> <mi>G</mi> <mi>s</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>G</mi> <mi>t</mi> </msub> </semantics></math> on the experimental data set. Even after fine-tuning <math display="inline"><semantics> <msub> <mi>G</mi> <mi>t</mi> </msub> </semantics></math> on unseen data, we observe that <math display="inline"><semantics> <msub> <mi>G</mi> <mi>t</mi> </msub> </semantics></math> produces results without PSNR degradation in the pre-training domain. This finding demonstrates the effectiveness of our proposed method, which utilizes human feedback to mitigate catastrophic forgetting. (<b>a</b>) MNIST test set (10k). (<b>b</b>) Fashion-MNIST test set (10k). (<b>c</b>) Fashion-MNIST train set (60k).</p>
Full article ">Figure 6
<p>Comparison of image quality with and without the auxiliary loss. (<b>d</b>–<b>g</b>) Results with different auxiliary loss conditions. Each condition improves the image quality compared to (<b>c</b>), but there are noticeable differences in details such as texture and artifacts. (<b>a</b>) Input image with noise. (<b>b</b>) Ground truth. (<b>c</b>) Denoised images by <math display="inline"><semantics> <msub> <mi>G</mi> <mi>s</mi> </msub> </semantics></math>. (<b>d</b>) Denoised images by <math display="inline"><semantics> <msub> <mi>G</mi> <mi>t</mi> </msub> </semantics></math>. (<b>e</b>) Denoised images by <math display="inline"><semantics> <msub> <mi>G</mi> <mi>t</mi> </msub> </semantics></math> without the <math display="inline"><semantics> <msub> <mi>L</mi> <mi>p</mi> </msub> </semantics></math> term. (<b>f</b>) Denoised images by <math display="inline"><semantics> <msub> <mi>G</mi> <mi>t</mi> </msub> </semantics></math> without the <math display="inline"><semantics> <msub> <mi>L</mi> <mi>n</mi> </msub> </semantics></math> term. (<b>g</b>) Denoised images by <math display="inline"><semantics> <msub> <mi>G</mi> <mi>t</mi> </msub> </semantics></math> using only the <math display="inline"><semantics> <msub> <mi>L</mi> <mi>r</mi> </msub> </semantics></math> term.</p>
Full article ">
13 pages, 1142 KiB  
Technical Note
Clustering and Vectorizing Acoustic Emission Events of Large Infrastructures’ Normal Operation
by Theocharis Tsenis and Vassilios Kappatos
Infrastructures 2025, 10(2), 38; https://doi.org/10.3390/infrastructures10020038 - 11 Feb 2025
Abstract
The detection of acoustic emission events from various failing mechanisms, such as plastic deformations, is a critical element in the monitoring and timely detection of structural failures in infrastructures. This study focuses on the detection of such failures in metal gates at rivers’ [...] Read more.
The detection of acoustic emission events from various failing mechanisms, such as plastic deformations, is a critical element in the monitoring and timely detection of structural failures in infrastructures. This study focuses on the detection of such failures in metal gates at rivers’ lifting dams aiming to increase the reliability of river transport compared to the current situation, thereby, increasing the resilience of transport corridors. During our study, we used lifting dams in both France and Italy where river transport is thriving. A methodology was developed, processing corresponding acoustic emission recordings originating from lifting dams’ metal gates, using advanced denoising—preprocessing, various decompositions, and spectral embeddings associated with various latest nonlinear processing clustering techniques—thus providing a detailed cluster label morphology and profile of water gates’ normal operating area. Latest machine learning outlier detection algorithms, like One-Class Support Vector Machine, Variational Auto-Encoder, and others, were incorporated, producing a vector of confidence on upcoming out-of-the-normal gate operation and failure prediction, achieving detection contrast enhancement on out-of-the-normal operation points up to 400%. Full article
Show Figures

Figure 1

Figure 1
<p>Our methodology layout from accepting the AE signals up to feature extraction and representation in a multidimensional space.</p>
Full article ">Figure 2
<p>Starting from the referenced and inspected multidimensional space producing the OPV.</p>
Full article ">Figure 3
<p>PZT sensors used by the Mistras micro-SHM system, from top to bottom with 60 kHz (R6a), 150 kHz (R15a), and 400 kHz (wideband) useful bandwidth.</p>
Full article ">Figure 4
<p>Placement of sensors (the used AE logger Micro-SHM in the top left corner) at a sample waterlock gate at Conda di Baricetta waterlock.</p>
Full article ">Figure 5
<p>LLE dimension reduction: (<b>left</b>) silhouette values and (<b>right</b>) clustering (800 clusters) 2D projection, with x-axis the 1st feature (dimension) of the new reduced space due to LLE dimensionality reduction and y-axis the 2nd feature dimension.</p>
Full article ">Figure 6
<p>Deviation in the outlier profiles in comparison to the original gate’s normal operation, with the horizontal line displaying the 27 indices as described in <a href="#sec2dot2-infrastructures-10-00038" class="html-sec">Section 2.2</a>.</p>
Full article ">
25 pages, 12377 KiB  
Article
Exploiting Weighted Multidirectional Sparsity for Prior Enhanced Anomaly Detection in Hyperspectral Images
by Jingjing Liu, Jiashun Jin, Xianchao Xiu, Wanquan Liu and Jianhua Zhang
Remote Sens. 2025, 17(4), 602; https://doi.org/10.3390/rs17040602 - 10 Feb 2025
Abstract
Anomaly detection (AD) is an important topic in remote sensing, aiming to identify unusual or abnormal features within the data. However, most existing low-rank representation methods usually use the nuclear norm for background estimation, and do not consider the different contributions of different [...] Read more.
Anomaly detection (AD) is an important topic in remote sensing, aiming to identify unusual or abnormal features within the data. However, most existing low-rank representation methods usually use the nuclear norm for background estimation, and do not consider the different contributions of different singular values. Besides, they overlook the spatial relationships of abnormal regions, particularly failing to fully leverage the 3D structured information of the data. Moreover, noise in practical scenarios can disrupt the low-rank structure of the background, making it challenging to separate anomaly from the background and ultimately reducing detection accuracy. To address these challenges, this paper proposes a weighted multidirectional sparsity regularized low-rank tensor representation method (WMS-LRTR) for AD. WMS-LRTR uses the weighted tensor nuclear norm for background estimation to characterize the low-rank property of the background. Considering the correlation between abnormal pixels across different dimensions, the proposed method introduces a novel weighted multidirectional sparsity (WMS) by unfolding anomaly into multimodal to better exploit the sparsity of the anomaly. In order to improve the robustness of AD, we further embed a user-friendly plug-and-play (PnP) denoising prior to optimize the background modeling under low-rank structure and facilitate the separation of sparse anomalous regions. Furthermore, an effective iterative algorithm using alternate direction method of multipliers (ADMM) is introduced, whose subproblems can be solved quickly by fast solvers or have closed-form solutions. Numerical experiments on various datasets show that WMS-LRTR outperforms state-of-the-art AD methods, demonstrating its better detection ability. Full article
Show Figures

Figure 1

Figure 1
<p>The illustration of the proposed method.</p>
Full article ">Figure 2
<p>The schematic diagram of the proposed weighted multidirectional sparsity.</p>
Full article ">Figure 3
<p>The detection maps on (<b>a</b>) Airport-1, (<b>b</b>) Airport-2, (<b>c</b>) Airport-3, (<b>d</b>) Airport-4, (<b>e</b>) Beach-1, (<b>f</b>) Beach-2, (<b>g</b>) Beach-3, (<b>h</b>) Beach-4, (<b>i</b>) Urban-1, (<b>j</b>) Urban-2, (<b>k</b>) Urban-3, (<b>l</b>) Urban-4, and (<b>m</b>) Urban-5.</p>
Full article ">Figure 4
<p>The <math display="inline"><semantics> <msub> <mi>AUC</mi> <mrow> <mo>(</mo> <mrow> <mi mathvariant="normal">D</mi> <mo>,</mo> <mi mathvariant="normal">F</mi> </mrow> <mo>)</mo> </mrow> </msub> </semantics></math> values of (<b>a</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>/</mo> <mi>μ</mi> </mrow> </semantics></math> on Urban-3, (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>/</mo> <mi>μ</mi> </mrow> </semantics></math> on Urban-4, (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>/</mo> <mi>μ</mi> </mrow> </semantics></math> on Urban-5, (<b>d</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>/</mo> <mi>β</mi> </mrow> </semantics></math> on Urban-3, (<b>e</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>/</mo> <mi>β</mi> </mrow> </semantics></math> on Urban-4, and (<b>f</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>/</mo> <mi>β</mi> </mrow> </semantics></math> on Urban-5.</p>
Full article ">Figure 5
<p>The spectral dimension selection on <math display="inline"><semantics> <mrow> <mo>(</mo> <mi mathvariant="bold">a</mi> <mo>)</mo> </mrow> </semantics></math> the Airport scenes, <math display="inline"><semantics> <mrow> <mo>(</mo> <mi mathvariant="bold">b</mi> <mo>)</mo> </mrow> </semantics></math> the Beach scenes, <math display="inline"><semantics> <mrow> <mo>(</mo> <mi mathvariant="bold">c</mi> <mo>)</mo> </mrow> </semantics></math> the Urban scenes.</p>
Full article ">Figure 6
<p>The ROC curves on the Airport-1 dataset. (<b>a</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>), (<b>b</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), (<b>c</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), and (<b>d</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>).</p>
Full article ">Figure 7
<p>The ROC curves on the Airport-2 dataset. (<b>a</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>), (<b>b</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), (<b>c</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), and (<b>d</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>).</p>
Full article ">Figure 8
<p>The ROC curves on the Beach-2 dataset. (<b>a</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>), (<b>b</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), (<b>c</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), and (<b>d</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>).</p>
Full article ">Figure 9
<p>The ROC curves on the Beach-3 dataset. (<b>a</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>), (<b>b</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), (<b>c</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), and (<b>d</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>).</p>
Full article ">Figure 10
<p>The ROC curves on the Urban-1 dataset. (<b>a</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>), (<b>b</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), (<b>c</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), and (<b>d</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>).</p>
Full article ">Figure 11
<p>The ROC curves on the Urban-2 dataset. (<b>a</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>), (<b>b</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), (<b>c</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), and (<b>d</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>).</p>
Full article ">Figure 12
<p>The ROC curves on the Urban-3 dataset. (<b>a</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>), (<b>b</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), (<b>c</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), and (<b>d</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>).</p>
Full article ">Figure 13
<p>The ROC curves on the Urban-4 dataset. (<b>a</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>), (<b>b</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), (<b>c</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), and (<b>d</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>).</p>
Full article ">Figure 14
<p>The ROC curves on the Urban-5 dataset. (<b>a</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>), (<b>b</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), (<b>c</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>), and (<b>d</b>) ROC curves of (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>D</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mi>τ</mi> </semantics></math>, <math display="inline"><semantics> <msub> <mi>P</mi> <mi>F</mi> </msub> </semantics></math>).</p>
Full article ">Figure 15
<p>The boxplots of all compared methods. <b>a</b>: RX; <b>b</b>: LRASR; <b>c</b>: GTVLRR; <b>d</b>: AUTO-AD; <b>e</b>: RGAE; <b>f</b>: DeCNN-AD; <b>g</b>: PTA; <b>h</b>: PCA-TLRSR; <b>i</b>: LARTVAD; <b>j</b>: WMS-LRTR.</p>
Full article ">Figure 16
<p>The detection maps on (<b>a</b>) Noisy Beach-3 and (<b>b</b>) Noisy Urban-3.</p>
Full article ">Figure 17
<p>The relative errors on <math display="inline"><semantics> <mrow> <mo>(</mo> <mi mathvariant="bold">a</mi> <mo>)</mo> </mrow> </semantics></math> the Airport scenes, <math display="inline"><semantics> <mrow> <mo>(</mo> <mi mathvariant="bold">b</mi> <mo>)</mo> </mrow> </semantics></math> the Beach scenes, <math display="inline"><semantics> <mrow> <mo>(</mo> <mi mathvariant="bold">c</mi> <mo>)</mo> </mrow> </semantics></math> the Urban scenes.</p>
Full article ">
13 pages, 716 KiB  
Article
Lightweight Denoising Diffusion Implicit Model for Medical Segmentation
by Rina Oh and Tad Gonsalves
Electronics 2025, 14(4), 676; https://doi.org/10.3390/electronics14040676 - 10 Feb 2025
Abstract
Automatic medical segmentation is crucial for assisting doctors in identifying disease regions effectively. As a state-of-the-art (SOTA) approach, generative AI models, particularly diffusion models, have surpassed GANs in generating high-quality images for tasks like segmentation. However, most diffusion-based architectures rely on U-Net designs [...] Read more.
Automatic medical segmentation is crucial for assisting doctors in identifying disease regions effectively. As a state-of-the-art (SOTA) approach, generative AI models, particularly diffusion models, have surpassed GANs in generating high-quality images for tasks like segmentation. However, most diffusion-based architectures rely on U-Net designs with multiple residual blocks and convolutional layers, resulting in high computational costs and limited applicability on general-purpose devices. To solve this issue, we propose an enhanced denoising diffusion implicit model (DDIM) that incorporates lightweight depthwise convolution layers within residual networks and self-attention layers. This approach significantly reduces computational overhead while maintaining segmentation performance. We evaluated the proposed DDIM on two distinct medical imaging datasets: X-ray and skin lesion and polyp segmentation. Experimental results demonstrate that our model achieves, with reduced resource requirements, accuracy comparable to standard DDIMs in both visual representation and region-based scoring. The proposed lightweight DDIM offers a promising solution for medical segmentation tasks, enabling easier implementation on general-purpose devices without the need for expensive high-performance computing resources. Full article
(This article belongs to the Special Issue Artificial Intelligence in Image and Video Processing)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Overview of the proposed DDIM model (while in the training phase).</p>
Full article ">Figure 2
<p>Overview of the reverse process using the proposed denoising model.</p>
Full article ">Figure 3
<p>Architecture of (<b>a</b>) residual blocks and (<b>b</b>) self-attention layers.</p>
Full article ">Figure 4
<p>Visualizing the predicted segmentation among trained models: (<b>a</b>) segmentation results using RSUA, (<b>b</b>) segmentation results using ISBI 2016, (<b>c</b>) and segmentation results using PolypDB.</p>
Full article ">Figure 5
<p>Comparison of the segmentation outputs from the PolypDB dataset using DDIM and Lightweight DDIM. (<b>a</b>,<b>b</b>): Lightweight DDIM output unstable segmentations with noticeable deviations from the ground truth, while DDIM demonstrated more consistent results. (<b>c</b>): Both DDIM and Lightweight DDIM failed to accurately segment the target region.</p>
Full article ">
21 pages, 4293 KiB  
Article
A Highly Robust Encoder–Decoder Network with Multi-Scale Feature Enhancement and Attention Gate for the Reduction of Mixed Gaussian and Salt-and-Pepper Noise in Digital Images
by Milan Tripathi, Waree Kongprawechnon and Toshiaki Kondo
J. Imaging 2025, 11(2), 51; https://doi.org/10.3390/jimaging11020051 - 10 Feb 2025
Abstract
Image denoising is crucial for correcting distortions caused by environmental factors and technical limitations. We propose a novel and highly robust encoder–decoder network (HREDN) for effectively removing mixed salt-and-pepper and Gaussian noise from digital images. HREDN integrates a multi-scale feature enhancement block in [...] Read more.
Image denoising is crucial for correcting distortions caused by environmental factors and technical limitations. We propose a novel and highly robust encoder–decoder network (HREDN) for effectively removing mixed salt-and-pepper and Gaussian noise from digital images. HREDN integrates a multi-scale feature enhancement block in the encoder, allowing the network to capture features at various scales and handle complex noise patterns more effectively. To mitigate information loss during encoding, skip connections transfer essential feature maps from the encoder to the decoder, preserving structural details. However, skip connections can also propagate redundant information. To address this, we incorporate attention gates within the skip connections, ensuring that only relevant features are passed to the decoding layers. We evaluate the robustness of the proposed method across facial, medical, and remote sensing domains. The experimental results demonstrate that HREDN excels in preserving edge details and structural features in denoised images, outperforming state-of-the-art techniques in both qualitative and quantitative measures. Statistical analysis further highlights the model’s ability to effectively remove noise in diverse, complex scenarios with images of varying resolutions across multiple domains. Full article
(This article belongs to the Special Issue Celebrating the 10th Anniversary of the Journal of Imaging)
Show Figures

Figure 1

Figure 1
<p>Images corrupted by mixed Gaussian (G) and random salt-and-pepper (RSP) noise: (<b>a</b>) G:10 + RSP, (<b>b</b>) G:30 + RSP, (<b>c</b>) G:50 + RSP, (<b>d</b>) G:70 + RSP, (<b>e</b>) G:90 + RSP.</p>
Full article ">Figure 2
<p>Architecture of the proposed HREDN.</p>
Full article ">Figure 3
<p>Architecture of the multi-scale feature enhancement block.</p>
Full article ">Figure 4
<p>In-depth architecture of proposed attention block.</p>
Full article ">Figure 5
<p>Visual comparison among eight facial image denoising methods on a single testing image from the FER2013 dataset with noise level (G:30 + RSP).</p>
Full article ">Figure 6
<p>Visual comparison results for eight facial image denoising methods on single testing image from CKPLUS dataset with noise level (G:30 + RSP).</p>
Full article ">Figure 7
<p>Visual comparison results for eight CT scan image denoising methods on single testing image from Curated COVID CT dataset with noise level (G:30 + RSP).</p>
Full article ">Figure 8
<p>Visual comparison results for eight remote sensing image denoising methods on single testing image from NWPU-RESISC45 dataset with noise level (G:30 + RSP).</p>
Full article ">
24 pages, 687 KiB  
Article
MtAD-Net: Multi-Threshold Adaptive Decision Net for Unsupervised Synthetic Aperture Radar Ship Instance Segmentation
by Junfan Xue, Junjun Yin and Jian Yang
Remote Sens. 2025, 17(4), 593; https://doi.org/10.3390/rs17040593 - 9 Feb 2025
Abstract
In synthetic aperture radar (SAR) images, pixel-level Ground Truth (GT) is a scarce resource compared to Bounding Box (BBox) annotations. Therefore, exploring the use of unsupervised instance segmentation methods to convert BBox-level annotations into pixel-level GT holds great significance in the SAR field. [...] Read more.
In synthetic aperture radar (SAR) images, pixel-level Ground Truth (GT) is a scarce resource compared to Bounding Box (BBox) annotations. Therefore, exploring the use of unsupervised instance segmentation methods to convert BBox-level annotations into pixel-level GT holds great significance in the SAR field. However, previous unsupervised segmentation methods fail to perform well on SAR images due to the presence of speckle noise, low imaging accuracy, and gradual pixel transitions at the boundaries between targets and background, resulting in unclear edges. In this paper, we propose a Multi-threshold Adaptive Decision Network (MtAD-Net), which is capable of segmenting SAR ship images under unsupervised conditions and demonstrates good performance. Specifically, we design a Multiple CFAR Threshold-extraction Module (MCTM) to obtain a threshold vector by a false alarm rate vector. A Local U-shape Feature Extractor (LUFE) is designed to project each pixel of SAR images into a high-dimensional feature space, and a Global Vision Transformer Encoder (GVTE) is designed to obtain global features, and then, we use the global features to obtain a probability vector, which is the probability of each CFAR threshold. We further propose a PLC-Loss to adaptively reduce the feature distance of pixels of the same category and increase the feature distance of pixels of different categories. Moreover, we designed a label smoothing module to denoise the result of MtAD-Net. Experimental results on the dataset show that our MtAD-Net outperforms traditional and existing deep learning-based unsupervised segmentation methods in terms of pixel accuracy, kappa coefficient, mean intersection over union, frequency weighted intersection over union, and F1-Score. Full article
Show Figures

Figure 1

Figure 1
<p>The overall architecture. In MCTM, thresholds corresponding to different false alarm rates are extracted. The LUFE module maps each pixel in the image to a high-dimensional feature space. The GVTE module employs a Vision Transformer structure to extract global features and maps these global features to the probabilities corresponding to the MCTM output thresholds. We use the designed loss function to update the weights of LUFE and GVTE.The inner product of the output vectors from MTCM and GVTE modules serves as the segmentation threshold for the input image. After segmentation, the result undergoes smoothing through the label smoothing module to obtain the final segmentation result.</p>
Full article ">Figure 2
<p>The image is divided into target region and noise region.</p>
Full article ">Figure 3
<p>Illustration of the Local U-shape Feature Extractor. (<b>a</b>) The overall structure of the LUFE module. After the input image passes through three Down modules and three Up modules, the feature map of the image is obtained. (<b>b</b>) The specific structure of the Down module. (<b>c</b>) The specific structure of the Up module.</p>
Full article ">Figure 4
<p>Illustration of the Global Vision Transformer Encoder. (<b>a</b>) The overall structure of the GVTE module. After Patch and Position Embedding, we feed it into Transformer Encoder, the global features of the image can be obtained. After MLP and SoftMax, global features can be mapped to the probabilities of each CFAR threshold. (<b>b</b>) The specific structure of Transformer Encoder.</p>
Full article ">Figure 5
<p>Schematic diagram of the role of the PLC-Loss function. (<b>a</b>) Distribution of different categories of pixels in the feature space before training. (<b>b</b>) Distribution of different categories of pixels in feature space after training.</p>
Full article ">Figure 6
<p>Qualitative analysis achieved by different unsupervised segmentation methods on small-size ship in the SSDD dataset. (<b>a</b>) Input image, (<b>b</b>) GT, (<b>c</b>) CFAR, (<b>d</b>) OTSU, (<b>e</b>) Kim-Net, (<b>f</b>) PiCIE, (<b>g</b>) CDA-SAR, (<b>h</b>) IDUDL, (<b>i</b>) Ours. The false negative areas and false positive areas are highlighted by green and red.</p>
Full article ">Figure 7
<p>Qualitative analysis achieved by different unsupervised segmentation methods on small-size ship in the HRSID dataset. (<b>a</b>) Input image (<b>b</b>) GT, (<b>c</b>) CFAR, (<b>d</b>) OTSU, (<b>e</b>) Kim-Net, (<b>f</b>) PiCIE, (<b>g</b>) CDA-SAR, (<b>h</b>) IDUDL, (<b>i</b>) Ours. The false negative areas and false positive areas are highlighted by green and red.</p>
Full article ">Figure 8
<p>Qualitative analysis achieved by different unsupervised segmentation methods on large ship in the SSDD dataset. (<b>a</b>) Input image, (<b>b</b>) GT, (<b>c</b>) CFAR, (<b>d</b>) OTSU, (<b>e</b>) Kim-Net, (<b>f</b>) PiCIE, (<b>g</b>) CDA-SAR, (<b>h</b>) IDUDL, (<b>i</b>) Ours. The false negative areas and false positive areas are highlighted by green and red.</p>
Full article ">Figure 9
<p>Qualitative analysis achieved by different unsupervised segmentation methods on large ship in the HRSID dataset. (<b>a</b>) Input image, (<b>b</b>) GT, (<b>c</b>) CFAR, (<b>d</b>) OTSU, (<b>e</b>) Kim-Net, (<b>f</b>) PiCIE, (<b>g</b>) CDA-SAR, (<b>h</b>) IDUDL, (<b>i</b>) Ours. The false negative areas and false positive areas are highlighted by green and red.</p>
Full article ">
Back to TopTop