Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (77)

Search Parameters:
Keywords = Wasserstein GAN

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 3915 KiB  
Article
Boosting EEG and ECG Classification with Synthetic Biophysical Data Generated via Generative Adversarial Networks
by Archana Venugopal and Diego Resende Faria
Appl. Sci. 2024, 14(23), 10818; https://doi.org/10.3390/app142310818 - 22 Nov 2024
Abstract
This study presents a novel approach using Wasserstein Generative Adversarial Networks with Gradient Penalty (WGAN-GP) to generate synthetic electroencephalography (EEG) and electrocardiogram (ECG) waveforms. The synthetic EEG data represent concentration and relaxation mental states, while the synthetic ECG data correspond to normal and [...] Read more.
This study presents a novel approach using Wasserstein Generative Adversarial Networks with Gradient Penalty (WGAN-GP) to generate synthetic electroencephalography (EEG) and electrocardiogram (ECG) waveforms. The synthetic EEG data represent concentration and relaxation mental states, while the synthetic ECG data correspond to normal and abnormal states. By addressing the challenges of limited biophysical data, including privacy concerns and restricted volunteer availability, our model generates realistic synthetic waveforms learned from real data. Combining real and synthetic datasets improved classification accuracy from 92% to 98.45%, highlighting the benefits of dataset augmentation for machine learning performance. The WGAN-GP model achieved 96.84% classification accuracy for synthetic EEG data representing relaxation states and optimal accuracy for concentration states when classified using a fusion of convolutional neural networks (CNNs). A 50% combination of synthetic and real EEG data yielded the highest accuracy of 98.48%. For EEG signals, the real dataset consisted of 60-s recordings across four channels (TP9, AF7, AF8, and TP10) from four individuals, providing approximately 15,000 data points per subject per state. For ECG signals, the dataset contained 1200 real samples, each comprising 140 data points, representing normal and abnormal states. WGAN-GP outperformed a basic generative adversarial network (GAN) in generating reliable synthetic data. For ECG data, a support vector machine (SVM) classifier achieved an accuracy of 98% with real data and 95.8% with synthetic data. Synthetic ECG data improved the random forest (RF) classifier’s accuracy from 97% with real data alone to 98.40% when combined with synthetic data. Statistical significance was assessed using the Wilcoxon signed-rank test, demonstrating the robustness of the WGAN-GP model. Techniques such as discrete wavelet transform, downsampling, and upsampling were employed to enhance data quality. This method shows significant potential in addressing biophysical data scarcity and advancing applications in assistive technologies, human-robot interaction, and mental health monitoring, among other medical applications. Full article
Show Figures

Figure 1

Figure 1
<p>Five decomposition of EEG waves using discrete wavelet transform (DWT) into approximation coefficient (<math display="inline"><semantics> <mrow> <mi>c</mi> <mi>A</mi> </mrow> </semantics></math>) and detailed coefficients (<math display="inline"><semantics> <mrow> <mi>c</mi> <msub> <mi>D</mi> <mrow> <mn>1</mn> <mo>−</mo> <mn>5</mn> </mrow> </msub> </mrow> </semantics></math>).</p>
Full article ">Figure 2
<p>WGAN-GP architecture.</p>
Full article ">Figure 3
<p>Workflow of synthetic EEG wave generation using WGAN-GP model.</p>
Full article ">Figure 4
<p>Two-dimensional CNN for EEG classification.</p>
Full article ">Figure 5
<p>Interface of the synthetic EEG generator, visualization, and CNN classification.</p>
Full article ">Figure 6
<p>EEG plot of TP9 channel for Subject A in concentration and relaxation states using WGAN-GP.</p>
Full article ">Figure 7
<p>PSD plot of TP9 channel for Subject A in EEG concentration and relaxation states.</p>
Full article ">Figure 8
<p>Real and synthetic normal ECG samples.</p>
Full article ">Figure 9
<p>Real and synthetic abnormal ECG samples.</p>
Full article ">Figure 10
<p>Bar chart of model accuracies with significance annotations. The label “<math display="inline"><semantics> <mrow> <mi>n</mi> <mi>s</mi> </mrow> </semantics></math>” stands for no statistical significance and the label “*” presents comparisons with statistical significance.</p>
Full article ">Figure 11
<p>Heatmap of pairwise statistical significance.</p>
Full article ">
32 pages, 8354 KiB  
Article
Estimation of Fractal Dimension and Detection of Fake Finger-Vein Images for Finger-Vein Recognition
by Seung Gu Kim, Jin Seong Hong, Jung Soo Kim and Kang Ryoung Park
Fractal Fract. 2024, 8(11), 646; https://doi.org/10.3390/fractalfract8110646 - 31 Oct 2024
Viewed by 584
Abstract
With recent advancements in deep learning, spoofing techniques have developed and generative adversarial networks (GANs) have become an emerging threat to finger-vein recognition systems. Therefore, previous research has been performed to generate finger-vein images for training spoof detectors. However, these are limited and [...] Read more.
With recent advancements in deep learning, spoofing techniques have developed and generative adversarial networks (GANs) have become an emerging threat to finger-vein recognition systems. Therefore, previous research has been performed to generate finger-vein images for training spoof detectors. However, these are limited and researchers still cannot generate elaborate fake finger-vein images. Therefore, we develop a new densely updated contrastive learning-based self-attention generative adversarial network (DCS-GAN) to create elaborate fake finger-vein images, enabling the training of corresponding spoof detectors. Additionally, we propose an enhanced convolutional network for a next-dimension (ConvNeXt)-Small model with a large kernel attention module as a new spoof detector capable of distinguishing the generated fake finger-vein images. To improve the spoof detection performance of the proposed method, we introduce fractal dimension estimation to analyze the complexity and irregularity of class activation maps from real and fake finger-vein images, enabling the generation of more realistic and sophisticated fake finger-vein images. Experimental results obtained using two open databases showed that the fake images by the DCS-GAN exhibited Frechet inception distances (FID) of 7.601 and 23.351, with Wasserstein distances (WD) of 18.158 and 10.123, respectively, confirming the possibility of spoof attacks when using existing state-of-the-art (SOTA) frameworks of spoof detection. Furthermore, experiments conducted with the proposed spoof detector yielded average classification error rates of 0.4% and 0.12% on the two aforementioned open databases, respectively, outperforming existing SOTA methods for spoof detection. Full article
Show Figures

Figure 1

Figure 1
<p>Overall flowchart of proposed method.</p>
Full article ">Figure 2
<p>Architecture of DCS-GAN.</p>
Full article ">Figure 3
<p>Samples for the selection of input and target image for training the generator and discriminator of DCS-GAN. * denotes one image randomly chosen in the intra-class of the input image, excluding the input image.</p>
Full article ">Figure 4
<p>Architecture of enhanced ConvNeXt-Small.</p>
Full article ">Figure 5
<p>Sample images of real finger-veins in the databases. (<b>a</b>) Examples from the ISPR database and (<b>b</b>) examples from the Idiap database.</p>
Full article ">Figure 6
<p>Examples of data augmentation on the Idiap database. (<b>a</b>) Original image, (<b>b</b>) image shifted upward, (<b>c</b>) image shifted downward, (<b>d</b>) image shifted to the left, (<b>e</b>) image shifted to the right.</p>
Full article ">Figure 7
<p>Graphs for the training and validation loss of DCS-GAN. (<b>a</b>) Training loss graph of the generator and the discriminator. (<b>b</b>) Validation loss graph of the generator and the discriminator.</p>
Full article ">Figure 8
<p>Training and validation accuracy (Acc) and loss (Loss) graphs of the enhanced ConvNeXt-Small. (<b>a</b>) Training accuracy and loss graphs. (<b>b</b>) Validation accuracy and loss graphs.</p>
Full article ">Figure 9
<p>Sample images of fake finger-vein images generated by DCS-GAN and other SOTA methods. Examples of (<b>a</b>) original image and images generated by (<b>b</b>) Pix2Pix, (<b>c</b>) Pix2PixHD, (<b>d</b>) CycleGAN, (<b>e</b>) CUT, and (<b>f</b>) DCS-GAN.</p>
Full article ">Figure 10
<p>FD estimation analysis for comparison between real and fake vein images: the first to the fourth images, from the left, in (<b>a</b>–<b>h</b>) mean finger vein image, CAM, BCAM, and FD graph, respectively. (<b>a</b>,<b>c</b>,<b>e</b>,<b>g</b>) show the real finger-vein images whereas (<b>b</b>,<b>d</b>,<b>f</b>,<b>h</b>) present the corresponding fake finger-vein images.</p>
Full article ">Figure 10 Cont.
<p>FD estimation analysis for comparison between real and fake vein images: the first to the fourth images, from the left, in (<b>a</b>–<b>h</b>) mean finger vein image, CAM, BCAM, and FD graph, respectively. (<b>a</b>,<b>c</b>,<b>e</b>,<b>g</b>) show the real finger-vein images whereas (<b>b</b>,<b>d</b>,<b>f</b>,<b>h</b>) present the corresponding fake finger-vein images.</p>
Full article ">Figure 10 Cont.
<p>FD estimation analysis for comparison between real and fake vein images: the first to the fourth images, from the left, in (<b>a</b>–<b>h</b>) mean finger vein image, CAM, BCAM, and FD graph, respectively. (<b>a</b>,<b>c</b>,<b>e</b>,<b>g</b>) show the real finger-vein images whereas (<b>b</b>,<b>d</b>,<b>f</b>,<b>h</b>) present the corresponding fake finger-vein images.</p>
Full article ">Figure 11
<p>ROC curves of TPR according to FPR by the proposed and the SOTA methods on (<b>a</b>) ISPR database and (<b>b</b>) Idiap database.</p>
Full article ">Figure 12
<p>Jetson TX2 board.</p>
Full article ">Figure 13
<p>Examples of correct spoof detection by the proposed method. (<b>a</b>) and (<b>c</b>) are examples of real images from the ISPR and Idiap databases, respectively, and (<b>b</b>) and (<b>d</b>) are corresponding examples of fake images.</p>
Full article ">Figure 14
<p>Examples of incorrect spoof detection by the proposed method. (<b>a</b>) and (<b>c</b>) are examples of real images from the ISPR and Idiap databases, respectively, and (<b>b</b>) and (<b>d</b>) are corresponding examples of fake images. In the proposed method, (<b>b</b>) and (<b>d</b>) are incorrectly identified as real images.</p>
Full article ">Figure 15
<p>Grad-CAM images. (<b>a</b>) shows Grad-CAM images for real images, while (<b>b</b>) shows Grad-CAM images for fake images generated from the real images in (<b>a</b>). In both (<b>a</b>,<b>b</b>), the first row is from the ISPR database, and the second row is from the Idiap database. Each row starts with the input image on the far left, followed by Grad-CAM images acquired from the first ConvNeXt Block, second ConvNeXt Block, third ConvNeXt Block, fourth ConvNeXt Block, and LKA attention of <a href="#fractalfract-08-00646-t004" class="html-table">Table 4</a>, respectively.</p>
Full article ">
12 pages, 1581 KiB  
Article
Airfoil Shape Generation and Feature Extraction Using the Conditional VAE-WGAN-gp
by Kazuo Yonekura, Yuki Tomori and Katsuyuki Suzuki
AI 2024, 5(4), 2092-2103; https://doi.org/10.3390/ai5040102 - 28 Oct 2024
Viewed by 692
Abstract
A machine learning method was applied to solve an inverse airfoil design problem. A conditional VAE-WGAN-gp model, which couples the conditional variational autoencoder (VAE) and Wasserstein generative adversarial network with gradient penalty (WGAN-gp), is proposed for an airfoil generation method, and then, it [...] Read more.
A machine learning method was applied to solve an inverse airfoil design problem. A conditional VAE-WGAN-gp model, which couples the conditional variational autoencoder (VAE) and Wasserstein generative adversarial network with gradient penalty (WGAN-gp), is proposed for an airfoil generation method, and then, it is compared with the WGAN-gp and VAE models. The VAEGAN model couples the VAE and GAN models, which enables feature extraction in the GAN models. In airfoil generation tasks, to generate airfoil shapes that satisfy lift coefficient requirements, it is known that VAE outperforms WGAN-gp with respect to the accuracy of the reproduction of the lift coefficient, whereas GAN outperforms VAE with respect to the smoothness and variations of generated shapes. In this study, VAE-WGAN-gp demonstrated a good performance in all three aspects. Latent distribution was also studied to compare the feature extraction ability of the proposed method. Full article
Show Figures

Figure 1

Figure 1
<p>Conditional GAN.</p>
Full article ">Figure 2
<p>Conditional VAE.</p>
Full article ">Figure 3
<p>Conditional VAEGAN.</p>
Full article ">Figure 4
<p>Network architectures of the encoder, decoder, and discriminator.</p>
Full article ">Figure 5
<p>Shape discretization.</p>
Full article ">Figure 6
<p>Histogram of <math display="inline"><semantics> <msub> <mi>C</mi> <mi mathvariant="normal">L</mi> </msub> </semantics></math>.</p>
Full article ">Figure 7
<p>Learning curve.</p>
Full article ">Figure 8
<p>Generated shapes. Numbers on top of each shape represent re-calculated <math display="inline"><semantics> <msub> <mi>C</mi> <mi mathvariant="normal">L</mi> </msub> </semantics></math>. Red figure implies <math display="inline"><semantics> <msub> <mi>C</mi> <mi mathvariant="normal">L</mi> </msub> </semantics></math> calculation did not converge.</p>
Full article ">Figure 9
<p>Generated shapes <math display="inline"><semantics> <mrow> <msub> <mi>C</mi> <mi mathvariant="normal">L</mi> </msub> <mo>=</mo> <mn>0.5</mn> </mrow> </semantics></math>. Different color represents different shapes.</p>
Full article ">Figure 10
<p>Latent distribution.</p>
Full article ">
17 pages, 18662 KiB  
Article
Symmetric Connected U-Net with Multi-Head Self Attention (MHSA) and WGAN for Image Inpainting
by Yanyang Hou, Xiaopeng Ma, Junjun Zhang and Chenxian Guo
Symmetry 2024, 16(11), 1423; https://doi.org/10.3390/sym16111423 - 25 Oct 2024
Viewed by 601
Abstract
This study presents a new image inpainting model based on U-Net and incorporating the Wasserstein Generative Adversarial Network (WGAN). The model uses skip connections to connect every encoder block to the corresponding decoder block, resulting in a strictly symmetrical architecture referred to as [...] Read more.
This study presents a new image inpainting model based on U-Net and incorporating the Wasserstein Generative Adversarial Network (WGAN). The model uses skip connections to connect every encoder block to the corresponding decoder block, resulting in a strictly symmetrical architecture referred to as Symmetric Connected U-Net (SC-Unet). By combining SC-Unet with a GAN, the study aims to reconstruct images more effectively and seamlessly. The traditional discriminators only differentiate the entire image as true or false. In this study, the discriminator calculated the probability of each pixel belonging to the hole and non-hole regions, which provided the generator with more gradient loss information for image inpainting. Additionally, every block of SC-Unet incorporated a Dilated Convolutional Neural Network (DCNN) to increase the receptive field of the convolutional layers. Our model also integrated Multi-Head Self-Attention (MHSA) into selected blocks to enable it to efficiently search the entire image for suitable content to fill the missing areas. This study adopts the publicly available datasets CelebA-HQ and ImageNet for evaluation. Our proposed algorithm demonstrates a 10% improvement in PSNR and a 2.94% improvement in SSIM compared to existing representative image inpainting methods in the experiment. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

Figure 1
<p>SC-Unet architecture, with a 3 × 256 × 256 image as input.</p>
Full article ">Figure 2
<p>(<b>left</b>) The original convolutional block of U-Net; (<b>right</b>) the Dilated Convolutional Neural Network in the convolutional block of SC-Unet.</p>
Full article ">Figure 3
<p>Multi-Head Self-Attention in the convolutional block of SC-Unet.</p>
Full article ">Figure 4
<p>The main framework of our image inpainting model.</p>
Full article ">Figure 5
<p>Sample of results with CelebA-HQ dataset. (<b>a</b>) Ground truth image. (<b>b</b>) Input image. (<b>c</b>) Reconstructed image. (<b>d</b>) Mask image. (<b>e</b>) Predicted image.</p>
Full article ">Figure 6
<p>Sample of results with ImageNet dataset. (<b>a</b>) Ground truth image. (<b>b</b>) Input image. (<b>c</b>) Reconstructed image. (<b>d</b>) Mask image. (<b>e</b>) Predicted image.</p>
Full article ">Figure 7
<p>Comparison with GLCIC and CA models on ImageNet.</p>
Full article ">Figure 8
<p>Comparison with GLCIC and CA models on CelebA-HQ.</p>
Full article ">Figure 9
<p>Ablation study with different model modules on ImageNet.</p>
Full article ">Figure 10
<p>Ablation study with different model modules on CelebA-HQ.</p>
Full article ">Figure A1
<p>Ablation study with different model modules on ImageNet.</p>
Full article ">Figure A2
<p>Ablation study with different model modules on ImageNet.</p>
Full article ">Figure A3
<p>Ablation study with different model modules on CelebA-HQ.</p>
Full article ">Figure A4
<p>Ablation study with different model modules on CelebA-HQ.</p>
Full article ">
13 pages, 7413 KiB  
Article
A Study on Enhancing the Visual Fidelity of Aviation Simulators Using WGAN-GP for Remote Sensing Image Color Correction
by Chanho Lee, Hyukjin Kwon, Hanseon Choi, Jonggeun Choi, Ilkyun Lee, Byungkyoo Kim, Jisoo Jang and Dongkyoo Shin
Appl. Sci. 2024, 14(20), 9227; https://doi.org/10.3390/app14209227 - 11 Oct 2024
Viewed by 570
Abstract
When implementing outside-the-window (OTW) visuals in aviation tactical simulators, maintaining terrain image color consistency is critical for enhancing pilot immersion and focus. However, due to various environmental factors, inconsistent image colors in terrain can cause visual confusion and diminish realism. To address these [...] Read more.
When implementing outside-the-window (OTW) visuals in aviation tactical simulators, maintaining terrain image color consistency is critical for enhancing pilot immersion and focus. However, due to various environmental factors, inconsistent image colors in terrain can cause visual confusion and diminish realism. To address these issues, a color correction technique based on a Wasserstein Generative Adversarial Network with Gradient Penalty (WGAN-GP) is proposed. The proposed WGAN-GP model utilizes multi-scale feature extraction and Wasserstein distance to effectively measure and adjust the color distribution difference between the input image and the reference image. This approach can preserve the texture and structural characteristics of the image while maintaining color consistency. In particular, by converting Bands 2, 3, and 4 of the BigEarthNet-S2 dataset into RGB images as the reference image and preprocessing the reference image to serve as the input image, it is demonstrated that the proposed WGAN-GP model can handle large-scale remote sensing images containing various lighting conditions and color differences. The experimental results showed that the proposed WGAN-GP model outperformed traditional methods, such as histogram matching and color transfer, and was effective in reflecting the style of the reference image to the target image while maintaining the structural elements of the target image during the training process. Quantitative analysis demonstrated that the mid-stage model achieved a PSNR of 28.93 dB and an SSIM of 0.7116, which significantly outperforms traditional methods. Furthermore, the LPIPS score was reduced to 0.3978, indicating improved perceptual similarity. This approach can contribute to improving the visual elements of the simulator to enhance pilot immersion and has the potential to significantly reduce time and costs compared to the manual methods currently used by the Republic of Korea Air Force. Full article
(This article belongs to the Special Issue Applications of Machine Learning Algorithms in Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>An overview of the architecture of the WGAN-GP model.</p>
Full article ">Figure 2
<p>Architecture of the Generator and Critic in the WGAN-GP model. (<b>a</b>) Architecture of the Generator (<b>b</b>) Architecture of the Critic.</p>
Full article ">Figure 3
<p>BigEarthNet-S2 RGB images (ref images).</p>
Full article ">Figure 4
<p>BigEarthNet-S2 preprocessing images (target images).</p>
Full article ">Figure 5
<p>Color matching results of BigEarthNet-S2 RGB images and BigEarthNet-S2 preprocessing images. (<b>a</b>) Generated image by model at the early stage of training, (<b>b</b>) generated image by model at the mid-stage of training, and (<b>c</b>) generated image by fully trained model.</p>
Full article ">Figure 6
<p>Precise texture reproduction results. (<b>a</b>) Generated image by model at the early stage of training, (<b>b</b>) generated image by model at the mid-stage of training, and (<b>c</b>) generated image by fully trained model.</p>
Full article ">Figure 7
<p>Comparison with other method’s results. (<b>a</b>) Image processed using histogram matching, showing limitations in maintaining color consistency and significant information loss in texture and detail; (<b>b</b>) image processed using the color transfer technique, which also shows limitations in maintaining the ground truth’s color consistency and lacks texture reproduction; (<b>c</b>) image generated by the early stage of the WGAN-GP-based model, where color distribution is irregular and texture representation is still underdeveloped; (<b>d</b>) image generated by the mid-stage model, demonstrating improved color matching and texture reproduction, with textures becoming more similar to the ground truth; and (<b>e</b>) image generated by the fully trained WGAN-GP model, showing a slight decrease in color consistency compared to the mid-stage model but offering superior texture reproduction compared to the other methods.</p>
Full article ">
16 pages, 5561 KiB  
Article
A Hybrid GAN-Inception Deep Learning Approach for Enhanced Coordinate-Based Acoustic Emission Source Localization
by Xuhui Huang, Ming Han and Yiming Deng
Appl. Sci. 2024, 14(19), 8811; https://doi.org/10.3390/app14198811 - 30 Sep 2024
Viewed by 983
Abstract
In this paper, we propose a novel approach to coordinate-based acoustic emission (AE) source localization to address the challenges of limited and imbalanced datasets from fiber-optic AE sensors used for structural health monitoring (SHM). We have developed a hybrid deep learning model combining [...] Read more.
In this paper, we propose a novel approach to coordinate-based acoustic emission (AE) source localization to address the challenges of limited and imbalanced datasets from fiber-optic AE sensors used for structural health monitoring (SHM). We have developed a hybrid deep learning model combining four generative adversarial network (GAN) variants for data augmentation with an adapted inception neural network for regression-based prediction. The experimental setup features a single fiber-optic AE sensor based on a tightly coiled fiber-optic Fabry-Perot interferometer formed by two identical fiber Bragg gratings. AE signals were generated using the Hsu-Nielsen pencil lead break test on a grid-marked thin aluminum plate with 35 distinct locations, simulating real-world structural monitoring conditions in bounded isotropic plate-like structures. It is demonstrated that the single-sensor configuration can achieve precise localization, avoiding the need for a multiple sensor array. The GAN-based signal augmentation expanded the dataset from 900 to 4500 samples, with the Wasserstein distance between the original and synthetic datasets decreasing by 83% after 2000 training epochs, demonstrating the high fidelity of the synthetic data. Among the GAN variants, the standard GAN architecture proved the most effective, outperforming other variants in this specific application. The hybrid model exhibits superior performance compared to non-augmented deep learning approaches, with the median error distribution comparisons revealing a significant 50% reduction in prediction errors, accompanied by substantially improved consistency across various AE source locations. Overall, this developed hybrid approach offers a promising solution for enhancing AE-based SHM in complex infrastructures, improving damage detection accuracy and reliability for more efficient predictive maintenance strategies. Full article
(This article belongs to the Special Issue Advanced Optical-Fiber-Related Technologies)
Show Figures

Figure 1

Figure 1
<p>Schematic of the fiber-optic coil-based acoustic emission sensing system. Inset: Close-up image of the sensor, showing the flexible mounting and dimensions (8 mm outer, 6 mm inner diameter).</p>
Full article ">Figure 2
<p>(<b>a</b>) Aluminum plate with the grid and fiber-optic sensor for AE testing (<b>b</b>) Schematic representation of the aluminum plate detailing the grid layout and test points.</p>
Full article ">Figure 3
<p>Time series augmentation showing the original data (orange) and generated data (green) to ensure each label has a balanced and sufficient number of samples for improved deep learning model performance.</p>
Full article ">Figure 4
<p>(<b>a</b>) Workflow of the hybrid network for AE source localization (<b>b</b>) Architecture of the Inception network for regression.</p>
Full article ">Figure 5
<p>Architecture of the generator and discriminator networks in the GAN for AE signal augmentation.</p>
Full article ">Figure 6
<p>The t-SNE visualization of synthetic and original datasets (<b>a</b>) The training epoch of 1 for GAN (<b>b</b>) The training epoch of 2000 for GAN (<b>c</b>) The training epoch of 2000 for WGAN (<b>d</b>) The training epoch of 2000 for DCGAN (<b>e</b>) The training epoch of 2000 for TSAGAN (<b>f</b>) Augmentation via addition of noise.</p>
Full article ">Figure 7
<p>The comparison of Wasserstein distance convergence across epochs for the four GAN variants (GAN, TSAGAN, WGAN, and DCGAN).</p>
Full article ">Figure 8
<p>The comparison of acoustic emission (AE) source localization performance. (<b>a</b>) Results from the hybrid deep learning model with GAN-based data augmentation and Inception network. (<b>b</b>) Results from the Inception network alone without GAN-based augmentation. Square markers represent actual source locations, star markers show predicted locations, and the large circular marker indicates the sensor position. The x and y axes represent dimensions in inches.</p>
Full article ">Figure 9
<p>The comparison of errors for the different methods.</p>
Full article ">
12 pages, 2079 KiB  
Article
Research on Default Classification of Unbalanced Credit Data Based on PixelCNN-WGAN
by Yutong Sun, Yanting Ji and Xiangxing Tao
Electronics 2024, 13(17), 3419; https://doi.org/10.3390/electronics13173419 - 28 Aug 2024
Viewed by 728
Abstract
Personal credit assessment plays a crucial role in the financial system, which not only relates to the financial activities of individuals but also affects the overall credit system and economic health of society. However, the current problem of data imbalance affecting classification results [...] Read more.
Personal credit assessment plays a crucial role in the financial system, which not only relates to the financial activities of individuals but also affects the overall credit system and economic health of society. However, the current problem of data imbalance affecting classification results in the field of personal credit assessment has not been fully solved. In order to solve this problem better, we propose a data-enhanced classification algorithm based on a Pixel Convolutional Neural Network (PixelCNN) and a Generative Adversarial Network (Wasserstein GAN, WGAN). Firstly, the historical data containing borrowers’ borrowing information are transformed into grayscale maps; then, data enhancement of default images is performed using the improved PixelCNN-WGAN model; and finally, the expanded image dataset is inputted into the CNN, AlexNet, SqueezeNet, and MobileNetV2 for classification. The results on the real dataset LendingClub show that the data enhancement algorithm designed in this paper improves the accuracy of the four algorithms by 1.548–3.568% compared with the original dataset, which can effectively improve the classification effect of the credit data, and to a certain extent, it provides a new idea for the classification task in the field of personal credit assessment. Full article
Show Figures

Figure 1

Figure 1
<p>GAN structure diagram.</p>
Full article ">Figure 2
<p>PixelCNN-WGAN model structure diagram.</p>
Full article ">Figure 3
<p>Structure of PixelCNN.</p>
Full article ">Figure 4
<p>Structure of discriminator.</p>
Full article ">Figure 5
<p>Pixel-WGAN default prediction flowchart.</p>
Full article ">Figure 6
<p>A grayscale map of credit data. There are obvious differences in the images corresponding to the normal samples and the default samples in some areas.</p>
Full article ">
16 pages, 4446 KiB  
Article
Method for Recognition of Communication Interference Signals under Small-Sample Conditions
by Rong Ge, Yusheng Li, Yonggang Zhu, Xiuzai Zhang, Kai Zhang and Minghu Chen
Appl. Sci. 2024, 14(13), 5869; https://doi.org/10.3390/app14135869 - 4 Jul 2024
Viewed by 680
Abstract
To address the difficulty in obtaining a large number of labeled jamming signals in complex electromagnetic environments, this paper proposes a small-sample communication jamming signal recognition method based on WDCGAN-SA (Wasserstein Deep Convolution Generative Adversarial Network–Self Attention) and C-ResNet (Convolution Block Attention Module–Residual [...] Read more.
To address the difficulty in obtaining a large number of labeled jamming signals in complex electromagnetic environments, this paper proposes a small-sample communication jamming signal recognition method based on WDCGAN-SA (Wasserstein Deep Convolution Generative Adversarial Network–Self Attention) and C-ResNet (Convolution Block Attention Module–Residual Network). Firstly, leveraging the DCGAN architecture, we integrate the Wasserstein distance measurement and gradient penalty mechanism to design the jamming signal generation model WDCGAN for data augmentation. Secondly, we introduce a self-attention mechanism to make the generation model focus on global correlation features in time–frequency maps while optimizing training strategies to enhance the quality of generated samples. Finally, real samples are mixed with generated samples and fed into the classification network, incorporating cross-channel and spatial information in the classification network to improve jamming signal recognition rates. The simulation results demonstrate that under small-sample conditions with a Jamming-to-Noise Ratio (JNR) ranging from −10 dB to 10 dB, the proposed algorithm significantly outperforms GAN, WGAN and DCGAN comparative algorithms in recognizing six types of communication jamming signals. Full article
Show Figures

Figure 1

Figure 1
<p>The overall architecture of WDCGAN-SA and C-ResNet.</p>
Full article ">Figure 2
<p>Time–frequency plots of six types of jamming signals.</p>
Full article ">Figure 3
<p>Self-Attention Module.</p>
Full article ">Figure 4
<p>Self-Attention Embedded Generative Model WDCGAN-SA.</p>
Full article ">Figure 5
<p>C-ResNet Network Structure.</p>
Full article ">Figure 6
<p>Comparative of Real Samples and Samples Generated by Two Different Network Models.</p>
Full article ">Figure 7
<p>The recognition accuracies of five different classification networks.</p>
Full article ">Figure 8
<p>The confusion matrix for (<b>a</b>) the CNN classification network; (<b>b</b>) our proposed method.</p>
Full article ">Figure 9
<p>Recognition Accuracy with Different Quantities of Generated Samples.</p>
Full article ">Figure 10
<p>The recognition rates of four different generative networks were evaluated under various conditions of sample size enhancement. (<b>a</b>) Enhancement with 30 images; (<b>b</b>) enhancement with 60 images; (<b>c</b>) enhancement with 90 images.</p>
Full article ">
23 pages, 5171 KiB  
Article
Image Enhancement Based on Dual-Branch Generative Adversarial Network Combining Spatial and Frequency Domain Information for Imbalanced Fault Diagnosis of Rolling Bearing
by Yuguang Huang, Bin Wen, Weiqing Liao, Yahui Shan, Wenlong Fu and Renming Wang
Symmetry 2024, 16(5), 512; https://doi.org/10.3390/sym16050512 - 24 Apr 2024
Cited by 2 | Viewed by 968
Abstract
To address the problems of existing 2D image-based imbalanced fault diagnosis methods for rolling bearings, which generate images with inadequate texture details and color degradation, this paper proposes a novel image enhancement model based on a dual-branch generative adversarial network (GAN) combining spatial [...] Read more.
To address the problems of existing 2D image-based imbalanced fault diagnosis methods for rolling bearings, which generate images with inadequate texture details and color degradation, this paper proposes a novel image enhancement model based on a dual-branch generative adversarial network (GAN) combining spatial and frequency domain information for an imbalanced fault diagnosis of rolling bearing. Firstly, the original vibration signals are converted into 2D time–frequency (TF) images by a continuous wavelet transform, and a dual-branch GAN model with a symmetric structure is constructed. One branch utilizes an auxiliary classification GAN (ACGAN) to process the spatial information of the TF images, while the other employs a GAN with a frequency generator and a frequency discriminator to handle the frequency information of the input images after a fast Fourier transform. Then, a shuffle attention (SA) module based on an attention mechanism is integrated into the proposed model to improve the network’s expression ability and reduce the computational burden. Simultaneously, mean square error (MSE) is integrated into the loss functions of both generators to enhance the consistency of frequency information for the generated images. Additionally, a Wasserstein distance and gradient penalty are also incorporated into the losses of the two discriminators to prevent gradient vanishing and mode collapse. Under the supervision of the frequency WGAN-GP branch, an ACWGAN-GP can generate high-quality fault samples to balance the dataset. Finally, the balanced dataset is utilized to train the auxiliary classifier to achieve fault diagnosis. The effectiveness of the proposed method is validated by two rolling bearing datasets. When the imbalanced ratios of the four datasets are 0.5, 0.2, 0.1, and 0.05, respectively, their average classification accuracy reaches 99.35% on the CWRU bearing dataset. Meanwhile, the average classification accuracy reaches 96.62% on the MFS bearing dataset. Full article
(This article belongs to the Section Engineering and Materials)
Show Figures

Figure 1

Figure 1
<p>The structure of the ACGAN.</p>
Full article ">Figure 2
<p>Framework of the proposed model.</p>
Full article ">Figure 3
<p>Structure of the SA: GAP represents the global average pooling, GN stands for group norm, <span class="html-italic">F</span>(<span class="html-italic">x</span>) = <span class="html-italic">ωx</span> + <span class="html-italic">b</span>, σ(⸱) represents the activation function, ⊗ represents the element-wise product, and C and S stand for the concat and channel shuffle operators, respectively.</p>
Full article ">Figure 4
<p>General structure of the proposed fault diagnosis method.</p>
Full article ">Figure 5
<p>The comparison of image generation ability among different models. (<b>a</b>) Normal; (<b>b</b>) BF7; (<b>c</b>) ORF14; (<b>d</b>) IRF14.</p>
Full article ">Figure 6
<p>Confusion matrix of classification results for the CWRU datasets with different imbalance ratios.To further illustrate the feature learning performance of the proposed diagnostic model with various unbalanced datasets, we utilized the t-distributed stochastic neighbor embedding (t-SNE) algorithm [<a href="#B41-symmetry-16-00512" class="html-bibr">41</a>] to visualize the classification results of the model on the test set. As shown in <a href="#symmetry-16-00512-f007" class="html-fig">Figure 7</a>, it is evident that the feature distributions of the test set samples with different health states in the four datasets exhibit significant differences. Although there are a few samples with conflated feature distributions, this impact on the model’s diagnosis results can be largely disregarded. This further validates the confusion matrices’ classification results and underscores the proposed model’s excellent feature learning and fault diagnosis capability.</p>
Full article ">Figure 7
<p>The visualization of t-SNE for classification results on the CWRU datasets with different imbalance ratios.</p>
Full article ">Figure 8
<p>The comparison of test accuracies among different models across four unbalanced datasets.</p>
Full article ">Figure 9
<p>Mechanical fault comprehensive simulation platform.</p>
Full article ">Figure 10
<p>Bearings with different fault types. (<b>a</b>) IRF; (<b>b</b>); ORF; (<b>c</b>) BF; (<b>d</b>); CF.</p>
Full article ">Figure 11
<p>Confusion matrix of classification results for the MFS datasets with different imbalance ratios. To better evaluate the diagnostic performance of the proposed model in a more intuitive way, the t-SNE dimensionality reduction visualization results are shown in <a href="#symmetry-16-00512-f012" class="html-fig">Figure 12</a>. It is evident that each fault category still exhibits a distinct classification boundary, further indicating the excellent data generation and fault diagnosis performance of the proposed model, along with its strong generalization ability.</p>
Full article ">Figure 12
<p>The visualization of t-SNE for classification results on the MFS datasets with different imbalance ratios.</p>
Full article ">Figure 13
<p>The comparison of test accuracies among different models across four imbalanced datasets.</p>
Full article ">
17 pages, 4399 KiB  
Article
Electroencephalographic Signal Data Augmentation Based on Improved Generative Adversarial Network
by Xiuli Du, Xinyue Wang, Luyao Zhu, Xiaohui Ding, Yana Lv, Shaoming Qiu and Qingli Liu
Brain Sci. 2024, 14(4), 367; https://doi.org/10.3390/brainsci14040367 - 9 Apr 2024
Cited by 2 | Viewed by 1439
Abstract
EEG signals combined with deep learning play an important role in the study of human–computer interaction. However, the limited dataset makes it challenging to study EEG signals using deep learning methods. Inspired by the GAN network in image generation, this paper presents an [...] Read more.
EEG signals combined with deep learning play an important role in the study of human–computer interaction. However, the limited dataset makes it challenging to study EEG signals using deep learning methods. Inspired by the GAN network in image generation, this paper presents an improved generative adversarial network model L-C-WGAN-GP to generate artificial EEG data to augment training sets and improve the application of BCI in various fields. The generator consists of a long short-term memory (LSTM) network and the discriminator consists of a convolutional neural network (CNN) which uses the gradient penalty-based Wasserstein distance as the loss function in model training. The model can learn the statistical features of EEG signals and generate EEG data that approximate real samples. In addition, the performance of the compressed sensing reconstruction model can be improved by using augmented datasets. Experiments show that, compared with the existing advanced data amplification techniques, the proposed model produces EEG signals closer to the real EEG signals as measured by RMSE, FD and WTD indicators. In addition, in the compressed reconstruction of EEG signals, adding the new data reduces the loss by about 15% compared with the original data, which greatly improves the reconstruction accuracy of the EEG signals’ compressed sensing. Full article
(This article belongs to the Section Neurotechnology and Neuroimaging)
Show Figures

Figure 1

Figure 1
<p>Basic GAN structure.</p>
Full article ">Figure 2
<p>LSTM basic building blocks.</p>
Full article ">Figure 3
<p>CNN basic structure.</p>
Full article ">Figure 4
<p>Overall structure of the L–C–WGAN–GP model.</p>
Full article ">Figure 5
<p>Generator structure of the L–C–WGAN–GP model.</p>
Full article ">Figure 6
<p>Discriminator structure of L–C–WGAN–GP model.</p>
Full article ">Figure 7
<p>Loss curve of the L–C–WGAN–GP discriminator.</p>
Full article ">Figure 8
<p>Time domain plots of generated data for five generative adversarial models.</p>
Full article ">Figure 9
<p>Time-frequency plots of generated data for five generative adversarial models.</p>
Full article ">Figure 10
<p>CS-ResNet compressed sensing reconstruction model.</p>
Full article ">
28 pages, 12037 KiB  
Article
Improved Generative Adversarial Network for Bearing Fault Diagnosis with a Small Number of Data and Unbalanced Data
by Zhaohui Qin, Faguo Huang, Jiafang Pan, Junlin Niu and Haihua Qin
Symmetry 2024, 16(3), 358; https://doi.org/10.3390/sym16030358 - 15 Mar 2024
Cited by 2 | Viewed by 1578
Abstract
Fault data under real operating conditions are often difficult to collect, making the number of trained fault data small and out of proportion to normal data. Thus, fault diagnosis symmetry (balance) is compromised. This will result in less effective fault diagnosis methods for [...] Read more.
Fault data under real operating conditions are often difficult to collect, making the number of trained fault data small and out of proportion to normal data. Thus, fault diagnosis symmetry (balance) is compromised. This will result in less effective fault diagnosis methods for cases with a small number of data and data imbalances (S&I). We present an innovative solution to overcome this problem, which is composed of two components: data augmentation and fault diagnosis. In the data augmentation section, the S&I dataset is supplemented with a deep convolutional generative adversarial network based on a gradient penalty and Wasserstein distance (WDCGAN-GP), which solve the problems of the generative adversarial network (GAN) being prone to model collapse and the gradient vanishing during the training time. The addition of self-attention allows for a better identification and generation of sample features. Finally, the addition of spectral normalization can stabilize the training of the model. In the fault diagnosis section, fault diagnosis is performed through a convolutional neural network with coordinate attention (CNN-CA). Our experiments conducted on two bearing fault datasets for comparison demonstrate that the proposed method surpasses other comparative approaches in terms of the quality of data augmentation and the accuracy of fault diagnosis. It effectively addresses S&I fault diagnosis challenges. Full article
Show Figures

Figure 1

Figure 1
<p>GAN framework diagram.</p>
Full article ">Figure 2
<p>Self-attention framework map.</p>
Full article ">Figure 3
<p>Coordinate attention structure.</p>
Full article ">Figure 4
<p>Method process diagram.</p>
Full article ">Figure 5
<p>Technology roadmap.</p>
Full article ">Figure 6
<p>Diagram of generator’s structure and parameters.</p>
Full article ">Figure 7
<p>Discriminator’s structure and parameters.</p>
Full article ">Figure 8
<p>WDCGAN-GP training flowchart.</p>
Full article ">Figure 9
<p>CNN-CA structure and parameter diagram.</p>
Full article ">Figure 10
<p>CWRU test platform.</p>
Full article ">Figure 11
<p>Different time domain images and corresponding TF images of bearings in the CWRU dataset (0 HP).</p>
Full article ">Figure 12
<p>Comparison of original TF images and the generated TF images for the CWRU dataset.</p>
Full article ">Figure 13
<p>Comparison of fault diagnosis results using different methods in CWRU dataset.</p>
Full article ">Figure 14
<p>Comparison of confusion matrices for different methods of fault diagnosis in Task 1.</p>
Full article ">Figure 15
<p>Comparison of t-SNE for different methods of fault diagnosis in Task 1.</p>
Full article ">Figure 16
<p>XJTU-SQ bearing test platform.</p>
Full article ">Figure 17
<p>Different time domain images and corresponding TF images of bearings from XJTU-SQ dataset.</p>
Full article ">Figure 18
<p>XJTU-SQ original samples and generated samples.</p>
Full article ">Figure 19
<p>Comparison of fault diagnosis results using different methods for XJTU-SQ dataset.</p>
Full article ">Figure 20
<p>Comparison of confusion matrices for different methods of fault diagnosis in Task 1.</p>
Full article ">Figure 21
<p>Comparison of t-SNE for different methods of fault diagnosis in Task 1.</p>
Full article ">Figure 21 Cont.
<p>Comparison of t-SNE for different methods of fault diagnosis in Task 1.</p>
Full article ">Figure 22
<p>Comparison of different models’ generated samples.</p>
Full article ">
18 pages, 6533 KiB  
Article
Rotating Machinery Fault Diagnosis with Limited Multisensor Fusion Samples by Fused Attention-Guided Wasserstein GAN
by Wenlong Fu, Ke Yang, Bin Wen, Yahui Shan, Shuai Li and Bo Zheng
Symmetry 2024, 16(3), 285; https://doi.org/10.3390/sym16030285 - 1 Mar 2024
Cited by 9 | Viewed by 1250
Abstract
As vital equipment in modern industry, the health state of rotating machinery influences the production process and equipment safety. However, rotating machinery generally operates in a normal state most of the time, which results in limited fault data, thus greatly constraining the performance [...] Read more.
As vital equipment in modern industry, the health state of rotating machinery influences the production process and equipment safety. However, rotating machinery generally operates in a normal state most of the time, which results in limited fault data, thus greatly constraining the performance of intelligent fault diagnosis methods. To solve this problem, this paper proposes a novel fault diagnosis method for rotating machinery with limited multisensor fusion samples based on the fused attention-guided Wasserstein generative adversarial network (WGAN). Firstly, the dimensionality of collected multisensor data is reduced to three channels by principal component analysis, and then the one-dimensional data of each channel are converted into a two-dimensional pixel matrix, of which the RGB images are obtained by fusing the three-channel two-dimensional images. Subsequently, the limited RGB samples are augmented to obtain sufficient samples utilizing the fused attention-guided WGAN combined with the gradient penalty (FAWGAN-GP) method. Lastly, the augmented samples are applied to train a residual convolutional neural network for fault diagnosis. The effectiveness of the proposed method is demonstrated by two case studies. When training samples per class are 50, 35, 25, and 15 on the KAT-bearing dataset, the average classification accuracy is 99.9%, 99.65%, 99.6%, and 98.7%, respectively. Meanwhile, the methods of multisensor fusion and the fused attention mechanism have an average improvement of 1.51% and 1.09%, respectively, by ablation experiments on the WT gearbox dataset. Full article
Show Figures

Figure 1

Figure 1
<p>The basic structure of GAN.</p>
Full article ">Figure 2
<p>The structure of the proposed FA mechanism.</p>
Full article ">Figure 3
<p>The structure of the proposed FAWGAN-GP.</p>
Full article ">Figure 4
<p>The proposed overall framework diagram.</p>
Full article ">Figure 5
<p>KAT test platform.</p>
Full article ">Figure 6
<p>Generated samples by training dataset A.</p>
Full article ">Figure 7
<p>Confusion matrix for four datasets in the KAT.</p>
Full article ">Figure 8
<p>T-SNE for four datasets in the KAT.</p>
Full article ">Figure 9
<p>Comparison of the number of generated samples.</p>
Full article ">Figure 10
<p>WT test platform.</p>
Full article ">Figure 11
<p>(<b>a</b>) Internal structure of planetary gearbox. (<b>b</b>) Healthy (NO), (<b>c</b>) broken tooth (BT), (<b>d</b>) worn gear (WR), (<b>e</b>) root crack (RC), (<b>f</b>) missing tooth (MT).</p>
Full article ">Figure 12
<p>Hyperparameters of learning rates and GP term coefficient experiments.</p>
Full article ">Figure 13
<p>Ablation experiment of four datasets.</p>
Full article ">Figure 14
<p>Comparative results of different classifiers.</p>
Full article ">
23 pages, 6027 KiB  
Article
An Image Classification Method of Unbalanced Ship Coating Defects Based on DCCVAE-ACWGAN-GP
by Henan Bu, Teng Yang, Changzhou Hu, Xianpeng Zhu, Zikang Ge and Honggen Zhou
Coatings 2024, 14(3), 288; https://doi.org/10.3390/coatings14030288 - 27 Feb 2024
Cited by 2 | Viewed by 1384
Abstract
Affected by the improper operation of the workers, environmental changes during drying and curing or the quality of the paint itself, diverse defects are produced during the process of ship painting. The traditional defect recognition method relies on expert knowledge or experience to [...] Read more.
Affected by the improper operation of the workers, environmental changes during drying and curing or the quality of the paint itself, diverse defects are produced during the process of ship painting. The traditional defect recognition method relies on expert knowledge or experience to detect defects, which is not conducive to ensuring the effectiveness of defect recognition. Therefore, this paper proposes an image generation and recognition model which is suitable for small samples. Based on a deep convolutional neural network (DCNN), the model combines a conditional variational autoencoder (DCCVAE) and auxiliary conditional Wasserstein GAN with gradient penalty (ACWGAN-GP) to gradually expand and generate various coating defect images for solving the overfitting problem due to unbalanced data. The DCNN model is trained based on newly generated image data and original image data so as to build a coating defect image classification model suitable for small samples, which is conducive to improving classification performance. The experimental results showed that our proposed model can achieve up to 92.54% accuracy, an F-score of 88.33%, and a G mean value of 91.93%. Compared with traditional data enhancement methods and classification algorithms, our proposed model can identify various defects in the ship painting process more accurately and consistently, which can provide effective theoretical and technical support for ship painting defect detection and has significant engineering research value and application prospects. Full article
(This article belongs to the Special Issue Recent Progress on Functional Films and Surface Science)
Show Figures

Figure 1

Figure 1
<p>The framework of the proposed image classification method for ship coating defects.</p>
Full article ">Figure 2
<p>Specific recognition process of ship coating defect images based on DCCVAE-ACWGAN-GP.</p>
Full article ">Figure 3
<p>The detailed description of the proposed DCCVAE network model structure.</p>
Full article ">Figure 4
<p>The detailed description of the proposed ACWGANGP network model structure.</p>
Full article ">Figure 5
<p>The following is the sagging image generated by the ACWGAN-GP model.</p>
Full article ">Figure 6
<p>The detailed description of the proposed DCNN model structure.</p>
Full article ">Figure 7
<p>Overall category-wise confusion matrix comparison between OURS and traditional data augmentation method: (<b>a</b>) Confusion matrix for the original imbalanced dataset. (<b>b</b>) Confusion matrix for balanced datasets augmented by traditional data augmentation methods. (<b>c</b>) The confusion matrix for balanced dataset augmented by our proposed method.</p>
Full article ">Figure 8
<p>Comparison of AUC performance on overall categories for OURS and current advanced data enhancement methods.</p>
Full article ">Figure 9
<p>The iterative curve of DCCVAE means square error with the iteration number.</p>
Full article ">Figure 10
<p>Iteration curves of accuracy and loss rate of training and test sets with the number of iterations.</p>
Full article ">Figure 11
<p>Effectiveness analysis of the weight coefficients of different loss functions on the performance of the proposed model.</p>
Full article ">Figure 12
<p>Comparison of different classification models: (<b>a</b>) real samples, (<b>b</b>) artificially generated samples.</p>
Full article ">Figure 13
<p>Visualization of distributions of features comparisons by t-SNE: (<b>a</b>) Original unbalanced data. (<b>b</b>) Data balanced by CVAE. (<b>c</b>) Data balanced by DCCVAE-ACWGAN-GP.</p>
Full article ">
30 pages, 32451 KiB  
Article
A Framework for Reconstructing Super-Resolution Magnetic Resonance Images from Sparse Raw Data Using Multilevel Generative Methods
by Krzysztof Malczewski
Appl. Sci. 2024, 14(4), 1351; https://doi.org/10.3390/app14041351 - 6 Feb 2024
Cited by 1 | Viewed by 1243
Abstract
Super-resolution magnetic resonance (MR) scans give anatomical data for quantitative analysis and treatment. The use of convolutional neural networks (CNNs) in image processing and deep learning research have led to super-resolution reconstruction methods based on deep learning. The study offers a G-guided generative [...] Read more.
Super-resolution magnetic resonance (MR) scans give anatomical data for quantitative analysis and treatment. The use of convolutional neural networks (CNNs) in image processing and deep learning research have led to super-resolution reconstruction methods based on deep learning. The study offers a G-guided generative multilevel network for training 3D neural networks with poorly sampled MR input data. The author suggest using super-resolution reconstruction (SRR) and modified sparse sampling to address these issues. Image-based Wasserstein GANs retain k-space data sparsity. Wasserstein Generative Adversarial Networks (WGANs) store and represent picture space knowledge. The method obtains null-valued k-space data and repairs fill gaps in the dataset to preserve data integrity. The proposed reconstruction method processes raw data samples and is able to perform subspace synchronization, deblurring, denoising, motion estimation, and super-resolution image production. The suggested algorithm uses different preprocessing methods to deblur and denoise datasets. Preliminary trials contextualize and speed up assessments. Results indicate that reconstructed pictures have better high-frequency features than sophisticated multi-frame techniques. This is supported by rising PSNR, MAE, and IEM measurements. A k-space correction block improves GAN network refinement learning in the suggested method. This block improves the network’s ability to avoid unnecessary data, speeding reconstruction. A k-space correction module can limit the generator’s output to critical lines, allowing the reconstruction of only missing lines. This improves convergence and speeds rebuilding. This study shows that this strategy reduces aliasing artifacts better than contemporaneous and noniterative methods. Full article
Show Figures

Figure 1

Figure 1
<p>Single k-space blade-based low-resolution MR images and their corresponding registration layer.</p>
Full article ">Figure 2
<p>A diagram illustrating the structure and components of the generation net.</p>
Full article ">Figure 3
<p>A flowchart depicting the architecture and functionality of the discriminator network.</p>
Full article ">Figure 4
<p>The algorithm for reconstructing super-resolution magnetic resonance images (MRI).</p>
Full article ">Figure 5
<p>The estimation of motion fields.</p>
Full article ">Figure 6
<p>Deblurring net.</p>
Full article ">Figure 7
<p>Denoising net.</p>
Full article ">Figure 8
<p>The subsequent depiction presents a clinical trial. The current study focuses on the reconstruction of an image using various techniques. In the first row, the regular sampling scheme is utilized without motion correction and SRR applied (1) (PSNR = 20.16 dB). The reconstruction process involves the use of a B-spline curve (2) (PSNR = 23.21 dB), Yang’s method [<a href="#B50-applsci-14-01351" class="html-bibr">50</a>] (3) (PSNR = 26.36 dB), Lim’s method (4) (PSNR = 29.11 dB) as referenced in [<a href="#B20-applsci-14-01351" class="html-bibr">20</a>], Zhang’s procedure (5) (PSNR = 28.71 dB) as referenced in [<a href="#B51-applsci-14-01351" class="html-bibr">51</a>], and Zhang’s second algorithm (6) (PSNR = 29.75 dB) as referenced in [<a href="#B43-applsci-14-01351" class="html-bibr">43</a>]. In the second row, Mahapatra’s method [<a href="#B52-applsci-14-01351" class="html-bibr">52</a>] (7) (PSNR = 30.03 dB) and Wang et al.’s procedure (8) (PSNR = 30.00 dB) are employed [<a href="#B53-applsci-14-01351" class="html-bibr">53</a>]. Furthermore, the reconstruction procedure utilizes Guerreiro’s et al.’s approach [<a href="#B54-applsci-14-01351" class="html-bibr">54</a>] (9) (PSNR = 30.03 dB), Pham et al.’s method [<a href="#B55-applsci-14-01351" class="html-bibr">55</a>] (10) (PSNR = 31.41 dB), Shi’s method [<a href="#B17-applsci-14-01351" class="html-bibr">17</a>] (11) (PSNR = 31.66 dB), as well as the author’s method (12) (PSNR = 32.99 dB) [<a href="#B20-applsci-14-01351" class="html-bibr">20</a>]. In addition, a sampling strategy and motion correction techniques are employed in order to achieve the super-resolution objectives. The aforementioned procedures are implemented without the inclusion of supplementary data. The compression ratio is 50%.</p>
Full article ">Figure 9
<p>The effectiveness of the algorithm being examined was assessed across different compression ratios of the k-space related to the figure mentioned in question (<a href="#applsci-14-01351-f008" class="html-fig">Figure 8</a>).</p>
Full article ">Figure 10
<p>The results of the second phase of the clinical brain imaging trial. The current study focuses on the reconstruction of an image using various techniques. In the first row, the regular sampling scheme is utilized without motion correction and SRR applied (1) (PSNR = 21.26 dB). The reconstruction process involves the use of the B-spline curve (2) (PSNR = 23.29 dB), Yang’s method [<a href="#B50-applsci-14-01351" class="html-bibr">50</a>] (3) (PSNR = 26.41 dB), Lim’s method (4) (PSNR = 29.22 dB) as referenced in [<a href="#B20-applsci-14-01351" class="html-bibr">20</a>], Zhang’s procedure (5) (PSNR = 28.71 dB) as referenced in [<a href="#B51-applsci-14-01351" class="html-bibr">51</a>], and Zhang’s second algorithm (6) as referenced in [<a href="#B43-applsci-14-01351" class="html-bibr">43</a>] (PSNR = 29.89 dB). In the second row, Mahapatra’s method [<a href="#B52-applsci-14-01351" class="html-bibr">52</a>] (7) (PSNR = 29.14 dB) and Wang et al.’s [<a href="#B53-applsci-14-01351" class="html-bibr">53</a>] procedure (8) (PSNR = 30.11 dB) are employed. Furthermore, the reconstruction procedure utilized Guerreiro’s approach [<a href="#B54-applsci-14-01351" class="html-bibr">54</a>] (9) (PSNR = 29.77 dB), Pham et al.’s method [<a href="#B55-applsci-14-01351" class="html-bibr">55</a>] (10) (PSNR = 23.48 dB), Shi’s method [<a href="#B17-applsci-14-01351" class="html-bibr">17</a>] (11) (12) (PSNR = 30.01 dB), as well as the author’s method (12) (PSNR = 32.61 dB). Furthermore, in order to achieve super-resolution, a suggested sampling strategy and motion correction techniques were employed. The aforementioned procedures were implemented without the inclusion of supplementary data. The compression ratio is 50%.</p>
Full article ">Figure 11
<p>The results of the second phase of the clinical brain imaging trial. This study utilizes the fastMRI test dataset [<a href="#B57-applsci-14-01351" class="html-bibr">57</a>]. The current study focuses on the reconstruction of an image using various techniques. In the first row, the regular sampling scheme is utilized without motion correction and with SRR applied (1) (PSNR = 20.11 dB). The reconstruction process involves the use of the B-spline curve (2) (PSNR = 23.31 dB), Yang’s method [<a href="#B50-applsci-14-01351" class="html-bibr">50</a>] (3) (PSNR = 27.01 dB), Lim’s method (4) (PSNR = 29.12 dB) as referenced in [<a href="#B20-applsci-14-01351" class="html-bibr">20</a>], Zhang’s procedure (5) (PSNR = 28.66 dB) as referenced in [<a href="#B51-applsci-14-01351" class="html-bibr">51</a>], and Zhang’s second algorithm (6) (PSNR = 29.71 dB) as referenced in [<a href="#B43-applsci-14-01351" class="html-bibr">43</a>]. In the second row, Mahapatra’s method [<a href="#B52-applsci-14-01351" class="html-bibr">52</a>] (7) (PSNR = 29.32 dB) and Wang et al.’s [<a href="#B53-applsci-14-01351" class="html-bibr">53</a>] procedure (8) (PSNR = 30.62 dB) are employed. Furthermore, the reconstruction procedure utilizes Guerreiro’s approach [<a href="#B54-applsci-14-01351" class="html-bibr">54</a>] (9) (PSNR = 28.82 dB), Pham et al.’s method [<a href="#B55-applsci-14-01351" class="html-bibr">55</a>] (10) (PSNR = 26.58 dB), Shi’s method [<a href="#B17-applsci-14-01351" class="html-bibr">17</a>] (11) (PSNR = 31.22 dB), and the author’s method (12) [<a href="#B20-applsci-14-01351" class="html-bibr">20</a>]. Furthermore, in order to achieve super-resolution, a suggested sampling strategy and motion correction techniques are employed. The aforementioned procedures are implemented without the inclusion of supplementary data. The compression ratio is 50%.</p>
Full article ">Figure 12
<p>The results of the second phase of the clinical brain imaging trial. This study utilizes the fastMRI test dataset [<a href="#B57-applsci-14-01351" class="html-bibr">57</a>]. The current study focuses on the reconstruction of an image using various techniques. In the first row, the regular sampling scheme is utilized without motion correction and with SRR applied (1) (PSNR = 22.89 dB). The reconstruction process involves the use of the B-spline curve (2) (PSNR = 23.22 dB), Yang’s method [<a href="#B50-applsci-14-01351" class="html-bibr">50</a>] (3) (PSNR = 26.30 dB), Lim’s method (4) (PSNR = 29.71 dB) as referenced in [<a href="#B20-applsci-14-01351" class="html-bibr">20</a>], Zhang’s procedure (5) (PSNR = 28.54 dB) as referenced in [<a href="#B51-applsci-14-01351" class="html-bibr">51</a>], and Zhang’s second algorithm (6) (PSNR = 29.77 dB) as referenced in [<a href="#B43-applsci-14-01351" class="html-bibr">43</a>]. In the second row, Mahapatra’s method [<a href="#B52-applsci-14-01351" class="html-bibr">52</a>] (7) (PSNR = 29.23 dB) and Wang et al.’s [<a href="#B53-applsci-14-01351" class="html-bibr">53</a>] procedure (8) (PSNR = 30.21 dB) are employed. Furthermore, the reconstruction procedure utilizes Guerreiro’s approach [<a href="#B54-applsci-14-01351" class="html-bibr">54</a>] (9) (PSNR = 30.77 dB), Pham et al.’s method [<a href="#B55-applsci-14-01351" class="html-bibr">55</a>] (10) (PSNR = 23.65 dB), Shi’s method [<a href="#B17-applsci-14-01351" class="html-bibr">17</a>] (11) (PSNR = 30.80 dB), and the author’s method (12) (PSNR = 34.02 dB) [<a href="#B20-applsci-14-01351" class="html-bibr">20</a>]. Furthermore, in order to achieve super-resolution, a suggested sampling strategy and motion correction techniques are employed. The aforementioned procedures are implemented without the inclusion of supplementary data. The compression ratio is 50%.</p>
Full article ">
17 pages, 7910 KiB  
Article
Fault Diagnosis in Hydroelectric Units in Small-Sample State Based on Wasserstein Generative Adversarial Network
by Wenhao Sun, Yidong Zou, Yunhe Wang, Boyi Xiao, Haichuan Zhang and Zhihuai Xiao
Water 2024, 16(3), 454; https://doi.org/10.3390/w16030454 - 30 Jan 2024
Cited by 2 | Viewed by 1247
Abstract
In the practical production environment, the complexity and variability of hydroelectric units often result in a need for more fault data, leading to inadequate accuracy in fault identification for data-driven intelligent diagnostic models. To address this issue, this paper introduces a novel fault [...] Read more.
In the practical production environment, the complexity and variability of hydroelectric units often result in a need for more fault data, leading to inadequate accuracy in fault identification for data-driven intelligent diagnostic models. To address this issue, this paper introduces a novel fault diagnosis method tailored for unbalanced small-sample states in hydroelectric units based on the Wasserstein generative adversarial network (W-GAN). Firstly, the fast Fourier transform is used to convert the signal from the time domain to the frequency domain to obtain the spectral data, and the W-GAN is trained to generate false spectral data with the same probability distribution as the real fault data, which are combined with the actual data and inputted into the 1D-CNN for feature extraction and fault diagnosis. In order to assess the effectiveness of the proposed model, a case study was conducted using actual data from a domestic hydropower plant, and the experimental results show that the sample features can be effectively enriched via data enhancement performed on small-sample data to improve the accuracy of fault diagnosis, which verifies the effectiveness of the method proposed in this paper. Full article
Show Figures

Figure 1

Figure 1
<p>Generating adversarial network structure.</p>
Full article ">Figure 2
<p>Generator structure.</p>
Full article ">Figure 3
<p>One-dimensional convolutional neural network structure.</p>
Full article ">Figure 4
<p>Flow diagram of the proposed method.</p>
Full article ">Figure 5
<p>Failure of Unit 3 of SK Power Station.</p>
Full article ">Figure 6
<p>Time domain waveform in different states. (<b>a</b>) Time domain waveform of Class 1 data. (<b>b</b>) Time domain waveform of Class 2 data. (<b>c</b>) Time domain waveform of Class 3 data.</p>
Full article ">Figure 7
<p>Spectrum in different states. (<b>a</b>) Spectrum of Class 1 data. (<b>b</b>) Spectrum of Class 2 data. (<b>c</b>) Spectrum of Class 3 data.</p>
Full article ">Figure 8
<p>Data generated via W-GAN at different epochs. (<b>a</b>) Class 2 data generated via W-GAN at different epochs. (<b>b</b>) Class 3 data generated via W-GAN at different epochs.</p>
Full article ">Figure 9
<p>Accuracy of different group data in the test set in 1D-CNN.</p>
Full article ">Figure 10
<p>Confusion matrix of 1D-CNN performance on different group test sets. (<b>a</b>) Group 1. (<b>b</b>) Group 2. (<b>c</b>) Group 3. (<b>d</b>) Group 4.</p>
Full article ">Figure 10 Cont.
<p>Confusion matrix of 1D-CNN performance on different group test sets. (<b>a</b>) Group 1. (<b>b</b>) Group 2. (<b>c</b>) Group 3. (<b>d</b>) Group 4.</p>
Full article ">Figure 11
<p>Visualization of T-sne downscaling of generated data with real data. (<b>a</b>) Group 2 data visualization results of dimensionality reduction before network classification. (<b>b</b>) Group 4 data visualization results of dimensionality reduction before network classification.</p>
Full article ">
Back to TopTop