Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Customized Trajectory Optimization and Compliant Tracking Control for Passive Upper Limb Rehabilitation
Previous Article in Journal
ADSTGCN: A Dynamic Adaptive Deeper Spatio-Temporal Graph Convolutional Network for Multi-Step Traffic Forecasting
Previous Article in Special Issue
GRU-Based Denoising Autoencoder for Detection and Clustering of Unknown Single and Concurrent Faults during System Integration Testing of Automotive Software Systems
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Metric Learning-Guided Semi-Supervised Path-Interaction Fault Diagnosis Method for Extremely Limited Labeled Samples under Variable Working Conditions

1
School of Mechanical and Aerospace Engineering, Jilin University, Changchun 130025, China
2
Sino-German College of Intelligent Manufacturing, Shenzhen Technology University, Shenzhen 518118, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(15), 6951; https://doi.org/10.3390/s23156951
Submission received: 24 June 2023 / Revised: 2 August 2023 / Accepted: 3 August 2023 / Published: 4 August 2023
Figure 1
<p>A conceptual illustration of the vanilla LAN.</p> ">
Figure 2
<p>A conceptual illustration of the triplet loss.</p> ">
Figure 3
<p>The architecture of the proposed method.</p> ">
Figure 4
<p>Demonstration of the 1D convolution operation.</p> ">
Figure 5
<p>A flowchart of the proposed method.</p> ">
Figure 6
<p>The experimental test rig of CWRU.</p> ">
Figure 7
<p>Confusion matrices for experiment 1 results on C1: (<b>a</b>) A<sub>1</sub>: five labels; (<b>b</b>) B<sub>1</sub>: two labels; (<b>c</b>) C<sub>1</sub>: one label.</p> ">
Figure 8
<p>Visualization features for experiment 1 results on C1: (<b>a</b>) five labels; (<b>b</b>) two labels; (<b>c</b>) one label.</p> ">
Figure 9
<p>Accuracy curves for experiment 1.</p> ">
Figure 10
<p>Ablation experimental results for experiment 1.</p> ">
Figure 11
<p>Accuracy curves of ablation experiment with 1 labeled sample.</p> ">
Figure 12
<p>Heatmaps of experiment 2 results: (<b>a</b>) five label; (<b>b</b>) two label; (<b>c</b>) one label.</p> ">
Figure 13
<p>Confusion matrices of experiment 2 results with one label: (<b>a</b>) C1–C4; (<b>b</b>) C4–C1.</p> ">
Figure 14
<p>Visualization features of experiment 2 results with one label: (<b>a</b>) C1–C4; (<b>b</b>) C4–C1.</p> ">
Figure 15
<p>Three-dimensional histograms of experiment 2 results for comparison: (<b>a</b>) is the comparison with Tri-LAN; (<b>b</b>) is the comparison with Tri-CNN.</p> ">
Figure 16
<p>Rotating machinery fault test system.</p> ">
Figure 17
<p>Heatmaps of motor experimental results with one label.</p> ">
Figure 18
<p>Confusion matrices for cross-working condition task: (<b>a</b>) C3–C4; (<b>b</b>) C4–C3.</p> ">
Figure 19
<p>The laboratory test rig of the motorized spindle.</p> ">
Figure 20
<p>Heatmaps of experimental results with 1 label.</p> ">
Figure 21
<p>Confusion matrices for cross-working condition task: (<b>a</b>) C1–C2; (<b>b</b>) C2–C1.</p> ">
Versions Notes

Abstract

:
The lack of labeled data and variable working conditions brings challenges to the application of intelligent fault diagnosis. Given this, extracting labeled information and learning distribution-invariant representation provides a feasible and promising way. Enlightened by metric learning and semi-supervised architecture, a triplet-guided path-interaction ladder network (Tri-CLAN) is proposed based on the aspects of algorithm structure and feature space. An encoder–decoder structure with path interaction is built to utilize the unlabeled data with fewer parameters, and the network structure is simplified by CNN and an element additive combination activation function. Metric learning is introduced to the feature space of the established algorithm structure, which enables the mining of hard samples from extremely limited labeled data and the learning of working condition-independent representations. The generalization and applicability of Tri-CLAN are proved by experiments, and the contribution of the algorithm structure and the metric learning in the feature space are discussed.

1. Introduction

As one of the most important components of intelligent manufacturing equipment, the health status of rotating machinery may affect the overall operation status of the equipment. For instance, the faults of bearings are prone to reducing the processing quality of the workpiece, and even result in considerable economic losses and potential safety hazards. Each fault of rotating machinery will eventually be embodied in the external excitation caused by mechanical structure defects, which produces mechanical vibration signals which differ from the healthy state. Due to the exceptional performance in solving the nonlinear feature extraction for machine vibration data, deep learning has witnessed remarkable success in the field of rotating machinery fault diagnosis [1]. In practical engineering, however, the lack of labeled data and the variable working conditions will restrict the profound study of prognostic and health management for rotating machinery [2].
Recent advances in supervised learning methods have been widely employed to overcome the challenge of variable working conditions. Xing et al. [3] proposed a distribution-invariant deep belief network (DBN) to learn distribution-invariant features by a locally connected structure. Zhao et al. [4] converted the one-dimensional signal to a three-dimensional image and applied a multiscale inverted residual convolution neural network (CNN) to learn different representations of variable load bearings. The gate units of a long short-term memory (LSTM) network were also utilized to store and transfer the classification information [5], and thus the working condition information could be ignored while the health condition was emphasized. The attention mechanism [6,7] combined with transfer learning enabled the model to retain invariant fault representation related to the faults during the training process. Although the aforementioned methods show superiority and outstanding stability in dealing with the inconsistent distributions within data under variable working conditions, these implementations have limitations in practical industrial scenarios. Ordinarily, the training of a decision-making model is based on the assumption of abundant labeled data, but it is unrealistic to label massive data in industrial applications.
Researchers have mainly made great efforts to alleviate the problem of insufficient labeled data from these three aspects: the feature learning-based strategy, the algorithm structure-based strategy and the data augmentation-based strategy [8]. From the perspective of the feature learning-based strategy, feature transfer based on transfer learning attained satisfactory diagnostic results. He et al. [9] designed a deep multi-wavelet autoencoder to select high-quality auxiliary samples for parameter knowledge transfer. Li et al. [10] constructed a multi-layer CNN to extract transferable features from the limited labeled data of the source domain and reduced the discrepancy of the marginal and the conditional probability distribution for limited labeled tasks. From the perspective of data resources, feature transfer based on transfer learning cannot encompass the entire fault dataset and mine useful information of unlabeled data, which causes a certain waste of the available information resources.
Taking the considerable fault information of unlabeled data into account, which is the most inexpensive data available in industrial scenarios, designing a semi-supervised algorithm structure appears to be a viable solution to address the issue above. The graph-based semi-supervised learning method [11,12,13] constructed a graph structure by regarding samples as vertices and regarding the similarity between points as edges, and thus the attribution of labeled samples could be propagated to unlabeled samples due to the hierarchy structure. To fully use the more abundant unlabeled data, Wu et al. [14] designed a hybrid classification autoencoder as a one-input two-output configuration consisting of the reconstruction of the input and the prediction of the health condition. Analogously, encoder–decoder network architectures based on CNNs [15] and LSTM [16] are established to distinguish the abnormal regime from the normal operating regimes by the magnitude of the reconstruction loss. As is common practice, a skipped connection was introduced in the encoder–decoder architectures, which was known as a vanilla ladder network (LAN) [17]. The vanilla LAN constantly varied in the backbone based on a typical deep learning algorithm to obtain higher training efficiency [15,18]. Zhang et al. [19] established two independent variational autoencoder (VAE)-based deep generative models to obtain the low-dimension latent features for labeled and unlabeled data, respectively. Accordingly, the multi-channel structure enabled the semi-supervised network to learn the fault representation of both labeled and unlabeled data.
Regarding aspects of the data augmentation-based strategy, some researchers attempt to extract more sensitive fault features based on signal processing. Zhang et al. [20] input the time-frequency wavelet coefficients into a multiple association layers network combining LAN and a variational autoencoder with less-labeled samples. Roozbeh et al. [21] fused the information of the raw sensory measurements in three different domains, and Yu et al. [22] employed seven data augmentation strategies. However, the tremendous data preprocessing procedure ignores the end-to-end feature extraction ability of deep learning. Furthermore, to alleviate the limited labeled problem, generating data with the same distribution of labeled data is regarded as an intuitive solution [8]. Ding et al. [23] utilized the probabilistic mixture model and the Markov Chain Monte Carlo algorithm to expand the fault dataset, which could provide large amounts of fake data. Tao et al. [24] generated pseudo-cluster labels for labeled and unlabeled data by adopting density peak clustering strategies. In addition, deep generative models were often utilized to generate new samples for labeled minority fault samples, such as GAN [25,26,27,28] and VAE [29,30]. Difficulties arise, however, when the quality of generated samples should be ensured to implement the data augmentation-based strategies.
Taken together, the research described above has the following shortcomings when facing the lack of labeled data and the variable working conditions:
  • These two challenges are usually overcome individually, and few works in the literature have studied these two issues simultaneously.
  • Closer attention is paid to expanding labeled data for supervised learning, while considerable fault information contained in unlabeled data is ignored and wasted.
  • More than ten labeled training samples are chiefly required; however, the available labeled samples are fewer in real industrial scenarios.
Recent advances in face recognition are attributed to the rise of metric learning. Unlike generative networks, which need to pay attention to each detail of the labeled data distribution, metric learning shows its promising potential to learn discriminative embeddings that can distinguish from other samples. Typically, contrastive loss [31] and triplet loss [32] could group intra-class samples closely while pushing inter-class samples distantly in the embedding space of pairwise samples. The contrastive loss could be introduced as a regularization [33,34,35] to learn working condition-independent features. Rombach et al. [36] considered triplets of training samples and learned invariance representation in the context of changing operations. Customarily, the hard example mining strategy [37,38] is often integrated with triplet loss to enhance the representation learning ability for the later network training stage. As a result, it provides the possibility of mining limited labeled data, which lays emphasis on the similarity among pairwise samples in the embedding space, and is able to learn fault-related rather than working condition-related representation.
Given the shortcomings of the above methods and enlightened by metric learning, both algorithm structure-level and feature-level aspects are considered in this paper. In terms of the algorithm’s structure-level, a CNN-based ladder network (CLAN) with path interaction is established to extract features from the most readily available unlabeled data and the limited labeled data. From the aspects of the feature-level, the similarity among anchor, positive and negative samples are calculated in the embedding space based on metric learning, in which extremely limited labeled samples can be regarded as hard samples to mine fault-related information and eliminate the working condition shifting effect. Therefore, the acquired classification error, reconstruction error and triplet loss are jointly defined as the objective function for the proposed method. The main contributions of this study, as well as the acquisition of the objective function, are listed as follows:
  • CLAN, a novel CNN-based ladder network, replaces the vanilla ladder network (LAN) backbone with a CNN and integrates the structure of the vanilla ladder network. Thus, the classification error of labeled samples and the reconstruction error of unlabeled samples can be obtained, and the parameters of the training process can be reduced by a simplified combination activation function and a path-interaction strategy.
  • To further alleviate the feature distribution shifting problem under variable working conditions, the triplet loss with the hard sample mining strategy is utilized to enlarge the margin among the embeddings of the limited labeled samples under different working conditions, which enables the proposed method to emphasize the fault-related features.
  • The proposed method is evaluated on two datasets: the first is the public bearing dataset from Case Western Reserve University (CWRU) for comparison with other state-of-the-art algorithms and the second is the experimental bearing dataset from our laboratory test rig of the motorized spindle to illustrate its extensive applicability. A few labeled data are selected randomly to verify the effectiveness of the proposed method. Moreover, variable working conditions are able to prove the ability of the learning distribution-invariant features.
The remaining part of the paper is organized as follows. In Section 2, the theoretical background is expounded. Section 3 concentrates on introducing the details of the proposed method. In Section 4, three case studies are given to illustrate the accuracy and robustness of the proposed method for extremely limited labeled samples under variable working conditions. Finally, Section 5 concludes this work and gives direction for future work.

2. Primary Theoretical Background of the Proposed Method

2.1. Semi-Supervised LAN

In the field of unsupervised learning, a two-stage strategy of “unsupervised pre-training + supervised fine-tuning” [39] is adopted; however, the two stages are independent from each other. Semi-supervised learning is an amalgamation of supervised and unsupervised learning. The unsupervised learning part can retain the original data information to the greatest extent through data reconstruction, while the supervised learning part attempts to keep the task-related information. Given this, to make the supervised learning compatible with the unsupervised learning, a specific structure is designed for the semi-supervised LAN which provides fault information extraction paths for both labeled and unlabeled data, respectively.
The structure of the vanilla LAN was proposed by Rasmus et al. [17], which consisted of a corrupted encoder path inspired by the denoising autoencoder [40], a decoder path for data reconstruction and a clean encoder path, as shown in Figure 1. In particular, the skipped connections between the corrupted encoder path and the decoder path enable the model to put emphasis on classification-relevant features at higher layers, while the original data information is ensured to transmit to the decoder path for data reconstruction. A cardinal principle of the vanilla LAN is described as follows.
Given a sequence of sample set X = x i i = 1 M 1 × D , let M denote the number of samples and D denote the length of a sample. According to whether there is a label, the sample set is further divided into a labeled dataset X L = x i , y i i = 1 N 1 × D and an unlabeled dataset X U = x i i = N + 1 M 1 × D , where y i { 1 , 2 , 3 } is the label for the labeled dataset, and N denotes the number of labeled samples, N M .
Generally, the vanilla LAN is based on a fully connected autoencoder network, and batch normalization (BN) and rectified linear units (ReLU) are applied to each layer, including the top-level layer. The corrupted noise ε l l = 1 L obeying the Gaussian distribution is implemented for the labeled samples, which is an auxiliary task to denoise representations at every level. Thus, the supervised classification cost C c of the noisy output y ˜ and y i can be obtained as Equation (1).
C c = 1 N i = 1 N log P y ˜ = y i x i ,
As for the unlabeled samples, a ladder-shaped encoder–decoder architecture provides an additional target with a data reconstruction error. More specifically, fault representations of each layer can be obtained through the corrupted encoder path, and then the top-level fault representation is fed into the decoder path in reverse order. Formally, the operation can be expressed as follows:
x ˜ , z ˜ ( 1 ) , , z ˜ ( L ) = Encoder c o r r u p t e d x   ,
x ^ , z ^ ( 1 ) , , z ^ ( L 1 ) = D e c o d e r z ˜ ( 1 ) , z ˜ ( 2 ) , , z ˜ ( L ) ,
where Encoder c o r r u p t e d   and D e c o d e r are the fully connected autoencoder network, and x , x ˜ , x ^ , z ˜ L are the input, the corrupted input, the reconstructed input and the top-level fault representation, respectively. The variables z ˜ l and z ^ l are the corrupted and the reconstructed fault representation in hidden layer l .
To incorporate the information of the upper layer and the skipped connection in the decoder path, a combinator activation function g , is designed in an element-wise manner, as expressed in Equation (4).
z ^ ( l ) = g z ˜ ( l ) , u ( l + 1 ) ,
where u l + 1 is the vertical fault representation learned from the upper layer. To provide the encoder–decoder architecture with clean reconstruction targets, the unlabeled samples are fed in the clean encoder path, which has a similar operation with E n c o d e r c o r r u p t e d but without the Gaussian noise:
x , z ( 1 ) , , z ( L ) = E n c o d e r c l e a n x ,
where z l represents the fault representation learned in hidden layer l . Thus, the unsupervised reconstruction costs can be obtained:
C R = R e c o n s C o s t z ( l ) , z ^ ( l ) ,
where R e c o n s C o s t is in terms of square error.
The final objective function is a weighted sum of C c and C U , as expressed in Equation (7), which is trained by backpropagation to assist the supervised learning by adding unsupervised tasks.
  C o s t = - λ C n = 1 N log P y ˜ = y n x n + λ R n = n + 1 M l = 1 L λ l R e c o n s C o s t z ( l ) n , z ^ ( l ) n ,
where λ C and λ R are the weight for the supervised classification cost and the unsupervised reconstruction costs. λ l denotes a layer-wise hyperparameter to determine the importance of the denoising cost in each layer.

2.2. Triplet Loss

Triplet loss is initially proposed in face recognition tasks [32], which introduces the concept of positive and negative samples to learn representations in the embedding space. Superior to the predetermined categories of the SoftMax function, triplet loss pays close attention to learning a mapping of a Euclidean space where distances directly correspond to a measure of the pairwise samples. As shown in Figure 2, the triplet samples consist of an anchor sample x i a , a positive sample x i p and a negative sample x i n . The triplet loss attempts to enforce a margin between different categories and to group samples from the same category according to the distribution discrepancy. Thus, the following condition needs to be met:
f x i a f x i p 2 2 + α < f x i a f x i n 2 2 , f x i a , f x i p , f x i n Γ ,
where f is a mapping function for Euclidean space, α is the margin maintained between classes and Γ includes all possible triplets of the samples. CNN, a deep network architecture, is widely used to minimize the loss L t r i :
L T r i = a , p , n y a = y p y n max f x i a f x i p 2 2 f x i a f x i n 2 2 + α , 0 ,

3. The Proposed Method

3.1. An Overview of the Proposed Method

In this work, we innovatively proposed a triplet-guided path-interaction CNN-based ladder network (Tri-CLAN) as a semi-supervised model for extremely limited labeled samples under variable working conditions. Compared to conventional semi-supervised deep learning methods, the innovation of Tri-CLAN is mainly reflected in two aspects: on the one hand, a CNN is utilized to substitute for the fully connected (FC) layer in the vanilla LAN, and reasonable model simplification is implemented to prevent over-fitting; on the other hand, triplet loss is introduced to further mine hard samples from limited labeled data, which forces the model to learn distinctive and working condition-independent embeddings among samples. The architecture of Tri-CLAN is displayed in Figure 3, and the three acquisition paths of the major losses are described in the following subsections.

3.2. Reconstruction Loss for Unlabeled Data

As depicted in Path 1 of Figure 3, the acquisition path of reconstruction loss for unlabeled data is composed of a corrupted encoder and a decoder. In the corrupted encoder, the Gaussian noise ε with mean 0 and variance are injected into the unlabeled samples X U = x i i = N + 1 M to learn internal stable representation and resist the noise perturbation, which originates from the vanilla ladder network. Specifically, standard Gaussian noise with mean 0 and variance 1 is selected because of its symmetrical characteristic and the consistence with the data distribution after batch normalization. The corrupted representation z ˜ ( l ) can be obtained by the following operation:
z ˜ ( 0 ) , z ˜ ( 1 ) , , z ˜ ( L ) = Encoder c o r r u p t e d X U ,
where Encoder c o r r u p t e d denotes the 1D convolution operation Conv(∙) followed by the batch normalization operation BN(∙) and the non-linear activation operation ReLU(∙) as follows:
Z ˜ ( l ) = h ˜ ( l ) = X U + ε , l = 0 ,
z ˜ pre ( l ) = C o n v h ˜ ( l 1 ) , 1 l L ,
z ˜ ( l ) = B N z ˜ p r e l ,
h ˜ ( l ) = R e L U γ ( l ) × z ˜ ( l ) + β ( l ) , 1 l L ,
where z ˜ p r e l , h ˜ ( l ) stand for the intermediate variables at layer l and γ ( l ) and β ( l ) are the scaling and offset coefficients of the batch normalization operation, respectively. To be specific, Conv(∙) multiplies the local input data by the shared weight of the convolution kernel, and thus the feature matrix can be obtained and the corresponding output of Conv(∙) at layer l can be described as:
  Conv o u t p u t = b i ( l ) + W i ( l ) X l j = k = 1 K W i l k X l j + k ,
where is the valid cross-correlation operator, X l j denotes the j th local data to be convoluted, W i ( l ) and b i ( l ) are the weight and bias of the i th convolution kernel, K is the width of the convolution kernel and the detailed operation process of the 1D convolution operation is shown in Figure 4. Additionally, BN(∙) is represented by B N ( x ) = x m e a n ( x ) var ( x ) , and ReLU(∙) is formulated by R e L U ( x ) = max ( 0 , x ) .
In the decoder, the corrupted representation z ˜ ( L ) at each level and the top-level representation h ˜ L of the corrupted encoder are taken as the input by introducing the skipped connection; then, sufficient fault-related information can be reserved for classification and the original low-level data can still flow into the decoder path. Thus, the contradictions between supervised fault classification and unsupervised data reconstruction can be resolved. With this strategy, the denoising representation z ^ ( L ) can be calculated by combining the corrupted representation z ˜ ( L ) from the corresponding layer of the corrupted encoder and z ^ ( l + 1 ) from the upper layer, as expressed by:
  z ^ ( 0 ) , z ^ ( 1 ) , , z ^ ( L ) = D e c o d e r z ˜ ( 0 ) , z ^ ( 1 ) , z ˜ ( 1 ) , z ^ ( 2 ) , , z ˜ ( L ) , h ˜ ( L ) ,
where D e c o d e r denotes the 1D transposed convolution operation T r a n s C o n v followed by the batch normalization operation BN(∙) and an element-wise activation operation g , as follows:
u ( l ) pre = h ˜ ( L ) , if   l = L T r a n s C o n v z ^ ( l + 1 ) , 1 l < L ,
u ( l ) = B N u ( l ) pre ,
z ^ ( l ) = g z ˜ ( l ) , u ( l ) ,
where u ( l ) pre and u ( l ) stand for the intermediate variables for the denoising representation z ^ ( l ) at layer l, and g ( , ) is a simplified combination activation function z ^ ( l ) = z ˜ ( l ) + u l from which internal elements participate in the operation.
So far, the corrupted representation z ˜ ( l ) and the denoising representation z ^ ( l ) at each layer are obtained with the skipped connection to reconstruct the unlabeled data X U , and the reconstruction loss for unlabeled data can be expressed as:
  C R e c o n = R e c o n s C o s t z 0 , z ^ ( 0 ) = z ( 0 ) B N z ^ ( 0 ) 2 ,
where R e c o n s C o s t stands for the l 2 distance between z 0 and z ^ 0 . Obviously, the reconstruction errors at each layer of the vanilla LAN are reducible to an error between the input corrupted data and the reconstructed data, and thus the computation burden of the acquisition path for reconstruction error can be reduced.

3.3. The Classification Error for Labeled Data

The acquisition path of the classification error consists of a corrupted encoder sharing parameters with the corrupted encoder of the acquisition path for reconstruction loss, which is a branch followed by fully connected layers for the prediction of fault types, as depicted in Path 2 of Figure 3. In the corrupted encoder for labeled samples X L = x i , y i i = 1 N , the embedding e m b e d l of fully connected layers and the classification output y ˜ can be computed as follows:
  e m b e d l = B N W l e m b e d l 1 , if   layer   l   is   a   fully   connected   layer ,
  y ˜ = S o f t M a x e m b e d i ( L ) ,
where W l is the weight matrix of fully connected layer l , L is a classification layer and S o f t M a x is represented by S o f t M a x x i = e x i k = 1 K e x k . Thus, the classification error for labeled data can be expressed as the cross-entropy:
  C C l a s s = 1 N i = 1 N   log   P   y   ˜ = y i x i ,

3.4. Triplet Loss for Labeled Data

The acquisition path of triplet loss, another branch of labeled data, is composed of pairwise distance calculation, hard labeled sample mining and hardest triplet training. From the aspect of mapping space, the fault information is mined to the greatest extent according to the similarity among triple samples. Intuitively, the large amount of possible generating triplets may result in uninformative representation, rendering a slow convergence and an overlong training duration. With this realization, selecting the hardest positive and negative samples provides an organizational modification for the triplet loss, which dramatically helps learning essential fault-related representation of limited labeled data. As depicted in Path 3 of Figure 3, all pairwise distances among each anchor sample e m b e d a L , randomly sampling positive samples e m b e d p L and negative samples e m b e d n L are first calculated, and then triplets that violate the constraint in Equation (8) are selected. Eventually, the hardest triplet is trained according to Equation (24), which is the loss of learning:
C T r i = i = 1 P a = 1 K [ m a x p = 1 K f e m b e d a i L f e m b e d p i L 2 2 m i n j = 1 P n = 1 K j i f e m b e d a i L f e m b e d n j L 2 2 + a + ,
where P is the number of the fault classes, K is the sample number of each fault class and P K is the quantity of all labeled samples. At this point, it is essential to note that the first term represents the distance metric of the hardest positive sample, and the second term represents the distance metric of the hardest negative samples in the Euclidean space.

3.5. The Final Objective Function

Profiting from the joining of semi-supervised learning and metric learning, the proposed method can not only excavate discriminative features from unlabeled and labeled samples, but it can also optimize the embedding space of limited labeled samples to further obtain semantically meaningful fault features. As the illustration of the information flow shown in Figure 5 shows, the final objective function is composed of the reconstruction loss, the cross-entropy loss and the triplet loss, which can be defined as:
  C T o t a l = C R e c o n + C C l a s s + C T r i = 1 N + M i = 1 N + M R e c o n s C o s t z ( i ) ( 0 ) , z ^ ( i ) ( 0 ) 1 N i = 1 N log P y ˜ = y i x i + C T r i ,
Unlike the fixed “pre-trained + fine-tuning” mode, C R e c o n and C T r i can both be regarded as a regularized term into the supervised classification costs, which is able to optimize the multi-path costs in a one-phase training simultaneously. Here, historical data of mechanical equipment under different health conditions is preprocessed to obtain the three losses in the offline training stage, and then the final objective function is minimized by introducing the Adam algorithm [41]. When Tri-CLAN has been well trained, the second path with the classifier is used for detecting fault types of online real-time data.

4. Experimental Studies

Several experiments on two case studies are conducted to evaluate the effectiveness and applicability of the proposed method. One is the public bearing dataset from CWRU, which is regarded as the standard dataset for objective comparison with the state-of-the-art algorithms, and the other is the laboratory test rig of the motorized spindle.

4.1. Implementation Details

In experimental studies, the raw data are usually 1D, and fast Fourier transform (FFT) is used to highlight more frequency-domain information. In order to facilitate the subsequent description of the architecture of the proposed method and the corresponding comparison methods, several basic modules are built, including a corrupted encoder module based on 1D CNN (CE-CNN), a decoder module based on 1D CNN (DE-CNN), a corrupted encoder module based on FC (CE-FC), a decoder module based on FC (DE-FC), a predicting module and a metric learning module. In the CE-CNN and DE-CNN, three 1D convolutional layers and three 1D transposed convolutional layers are separately constructed followed by BN and ReLU operation. In the CE-FC and DE-FC, a fully connected form with three hidden layers is adopted for a fair comparison. Likewise, the predicting module with FC is used for dimension reduction and fault classification. Put slightly differently, the predicting #1 requires an additional flattening operation to integrate the features of several channels into a 1D form. The detailed kernel size and neuron number of each module are listed in Table 1.
This study applies comparison to conduct a systematic and intensive study, and several deep learning methods based on basic modules are employed to verify the effectiveness of the proposed method, which contains a conventional CNN and vanilla LAN. Additionally, methods without the metric learning module are employed to discuss the constructive contribution for limited labels and variable working conditions. The detailed architecture of these methods can be shown in Table 2, and the parameter setup for experiments in case studies are listed in Table 3.
The experiments are carried out on a computer with NVIDIA Quadro RTX 6000 GPU, and PyTorch platform is implemented as the backend for programming. For the experimental error reduction, each case study experiment is repeated with ten trials to avoid contingency.

4.2. Case Study 1: Public Bearing Dataset of CWRU

4.2.1. Fault Dataset Description

As is well known, the public bearing dataset of CWRU is most commonly used to verify the effectiveness of diagnostic algorithms. As shown in Figure 6, the experimental test rig of CWRU mainly consists of a 2 HP motor, a dynamometer and a connecting part with sensors, including a torque transducer and an encoder. In this study, the drive end bearing supporting the motor shaft is selected as the research object, and its vibration acceleration signal is collected using the acceleration sensor placed on the bearing pedestal at the drive end. Considering that it is often difficult to maintain a stable working condition and reach a high sampling frequency in practical industrial applications, the vibration data collected at the sampling frequency of 12 kHz under different working conditions, listed in Table 4, are selected as training data and testing data. At the same time, in order to meet the industrial requirements, it is not only necessary to identify the fault location of the research object, but also to distinguish its damage degree to provide a basis for subsequent maintenance strategies. Therefore, both training data and testing data are divided into ten categories according to the fault location and damage degree, as shown in Table 5.
Before beginning the experiments, the datasets under different working conditions need to be preprocessed uniformly, and a sample consisting of 2048 points is intercepted from the original vibration data of each fault label. Then, FFT and zero-mean normalization processing are introduced into each sample to obtain samples with a length of 1024. Hereafter, 3000 and 1000 samples are generated for training and test datasets, respectively. According to different experimental purposes, the training data will be subsequently divided into labeled training data and unlabeled training data in different proportions.

4.2.2. Experiments Setup for Fault Diagnosis

Two experiments were conducted to verify the superiority of the proposed Tri-CLAN. Firstly, to preliminarily demonstrate the diagnostic performance, three datasets (A1, B1, C1) with different labeled sample numbers are established by randomly selecting from working condition C1; the exact number of training samples and testing samples for each category is listed in Table 6. Then, the ability of learning distribution-invariant features under variable working conditions is ulteriorly proved by building four datasets (A2, B2, C2, D2) on the foundation of experiment 1, and the number of training samples and testing samples is listed in Table 7.

4.2.3. Results Analysis for Experiment 1

From a more specific perspective, the matching matrix between the predicted diagnosis results and the actual labels is displayed in a visual form to reflect the detailed classification, and the clustering effects of features are shown by t-distributed stochastic neighbor embedding (t-SNE). It is worth adding that the confusion matrix diagram shown in Figure 7 and the visualization features shown in Figure 8 are the diagnosis results which are closest to the average accuracy rate of the ten trials. It can be seen that the fault location and damage degree of the bearing are all correctly classified when the labeled sample number is five and the accuracy is 100%. Samples belonging to different fault categories in the feature space are distinguished clearly, and there is no overlap in the feature space. Compared with the experiment taking five labeled samples, experiments with two and one labeled samples can be regarded as extremely severe conditions, and the accuracy is 99.10% and 91.10%, respectively. Specifically, only nine test sample are wrongly classified when the labeled sample number is two, which belongs to the misclassed damage degree for bearing OR. It can also be observed from the feature space that sample features are misclassified into other categories. When the labeled sample number is one, only one sample of ball faults is predicted as OR faults, and the accuracy of fault location classification can reach 99.90%. The remaining 88 misclassed test samples are all damage degree identification errors, which are not identified as normal and which affected routine repair and maintenance of the equipment.
As shown in Figure 9, the convergence during training iterations is analyzed for experiment 1. When the labeled sample numbers are five and two, the accuracy curve gradually trends towards smoothing after 20 iterations. By comparison, the accuracy curve with one labeled sample appears to unstably change with sharp fluctuations at the beginning, and it trends towards smoothing after 40 iterations. It is relatively hard to provide sufficient fault information due to the lack of labeled fault samples, which results in a smaller slope of the convergence curve and a lower value of the last convergence accuracy.
On the premise of preliminarily proving the effectiveness of the proposed algorithm with extremely limited labeled samples, we further discuss the contribution of the ladder-shaped semi-supervised network structure and triplet loss for the proposed Tri-CLAN. Therefore, the methods listed in Table 2, classic baseline (CNN), methods with ladder-shaped architecture in the form of encoder–decoder (CLAN, Tri-LAN, Vanilla LAN) and methods adding metric-learning (Tri-CNN, Tri-LAN), are established to provide a reference for analysis. In order to intuitively compare ablation experimental results, accuracy bar charts with error bands for five, two and one labeled samples are presented in Figure 10. All methods in Table 2 are ranked according to the average accuracy with standard deviation. It is evident that the accuracy of the proposed Tri-CLAN algorithm is significantly higher than that of CLAN, CNN and Vanilla LAN under various labeled sample numbers. Therefore, we only discuss the comparison among the proposed Tri-CLAN, Tri CNN and Tri LAN. When the labeled sample number is five, the accuracy of Tri-CLAN, Tri-CNN and Tri-LAN is higher than 95%, and the accuracy of Tri-CLAN is 3.97% and 4.22% higher than Tri-CNN and Tri-LAN, respectively. Moreover, the standard deviation of Tri-CLAN is close to zero; in other words, the proposed method has good robustness and its accuracy is basically stable at 100%. When the labeled sample number is two, the accuracy of Tri-CLAN can still stay above 99%, while the accuracy of Tri-CNN and Tri-LAN are reduced to less than 95%. As the case with the least labeled data, one labeled sample is a tremendous test for diagnosis algorithms to learn fault representation; however, the average accuracy of Tri-CLAN is surprisingly 92.45%. As shown in the confusion matrix above, the accuracy of fault location classification can reach 99.70%. From the perspective of standard deviation, the accuracy of the proposed Tri-CLAN basically fluctuates within 1% in the three cases of various labeled samples, while the standard deviation of the Tri-CNN and Tri-LAN is greater than 2% even in the case of five labeled samples.
According to the ranking in Figure 10 and the accuracy curves in Figure 11, methods adding a metric learning module generally achieve relatively high accuracy and early convergence, such as Tri-CLAN, Tri-CNN and Tri-LAN. It follows that the triplet loss plays a decisive role in the contribution to diagnosis accuracy. Subsequently, the method based on the CNN has higher accuracy and smaller fluctuation than that based on the autoencoder, and we believe that this phenomenon is related to the excellent generalization capability of the shared convolutional kernel. From the perspective of algorithm structure, the ladder-shaped encoder–decoder architecture is able to exploit an enormous quantity of unlabeled data which is usually ignored, and thus the actual distribution of each fault can be obtained. As a comprehensive combination of the triplet loss, the ladder-shaped encoder–decoder architecture and the CNN-based backbone, the proposed Tri-CLAN can deservedly learn effective fault representations and improve the accuracy with extremely limited labeled samples.
Based on the comparisons of architectures with different modules, we systematically select some excellent references from recent years with the same diagnostic purpose and compared the experimental results with the proposed Tri-CLAN in this paper. The classification category, the number of training samples and the test accuracy of the comparative experiment are all listed in Table 8. Obviously, these methods have achieved good fault identification results with few labels; for instance, the experimental setup in reference [42] is basically consistent with the experiment with five labeled samples in this paper, with an average accuracy of 98.40%. However, it can be found that the misclassed test samples are wrongly divided to other fault locations by observing the confusion matrix results in reference [42]. References [43,44] only classified fault locations and not the damage degree with an average accuracy about 90%, and we analyzed that it is bound up with the failure to utilize the unlabeled data as auxiliary.

4.2.4. Results Analysis for Experiment 2

On the whole, experiment 1 focuses on the performance comparison of limited labeled samples, and from the above discussion of the experimental results, it can be proved that the proposed Tri-CLAN can effectively utilize the unlabeled samples to assist extremely limited labeled samples and improve the accuracy of the method. In a follow-up experiment, we decided to increase the difficulty of experiment 1 by training with extremely limited labeled samples under one working condition and testing under other working conditions. Herein, the actual manufacturing operation is simulated, in which it is difficult to obtain data under the same working condition.
To better understand the working condition shifting effect with extremely limited samples, Figure 12 shows the results of experiment 2 in the form of heatmaps. It clearly shows the accuracy distribution by displaying different colors; broadly speaking, the accuracy of the diagnosis results is relatively reduced with the decrease of labeled samples. Secondly, a specific example under different working conditions with five labeled samples is taken for detailed discussion, as shown in Figure 12a. As a matter of course, the diagnosis results of the same working condition are higher than that of different working conditions, which can reach more than 99.50% universally. As the difference between training and testing conditions becomes larger, the accuracy rate appears to have a significant downtrend because of the difference between the probability distributions in the representation space. Another interesting phenomenon appears when tasks between different working conditions have the same difference degree, such as C1 → C4 and C4 → C1, and the diagnosis task from low-speed to high-speed shows relatively higher accuracy. It is noted that the load under different working conditions only affects the speed of the motor, and there is no mechanism to convert the torque load into the radial load on the bearing [45]. As the rotation speed decreases, the amplitude of the vibration signal becomes larger, and thus the vibration caused by faults can be highlighted and more fault representations can be learned relatively. Therefore, the fault representations learned from the low-speed C4 are sufficient to support the robust diagnostic accuracy for test under the high-speed C1. Instead, the model trained under C1 with a high rotating speed cannot afford to provide sufficient fault representations for C4 and achieve an ideal diagnosis. Similarly, this interesting phenomenon also appears in tasks under different working conditions with two and one labeled samples.
Probing into the experimental group with the most significant difference degree and the least labeled sample is meaningful to prove the effectiveness and ability for learning distribution-invariant features of the proposed Tri-CLAN. As can be seen from the above heatmap, the average accuracy of C1 → C4 and C4 → C1 is 83.05% and 86.58%, respectively, which are the worst diagnosis test results in experiment 2. To further observe the specific situation of the classification, confusion matrices and visualization features which are closest to the average accuracy are shown in Figure 13 and Figure 14, respectively. It can be seen that 15 samples are misclassified as other fault locations in the C1 → C4 experiment, and the remaining samples belong to the damage degree identification error. The features of different fault locations appear in a clear clustering state, and only a few parts among samples of different damage degrees overlap in the feature space. All misclassified samples belong to the damage degree identification error in the C4 → C1 experiment, and all samples with different fault locations can be well distinguished in the feature space. Therefore, the proposed Tri-CLAN has achieved satisfactory diagnostic results with extremely limited labeled samples under variable working conditions.
We review experiment 1 and prove that triplet loss plays a crucial role when the labeled samples are extremely few. To reduce unnecessary comparisons, we directly compare the proposed Tri-CLAN with the top two algorithms in experiment 1, which are the methods adding metric-learning (Tri-CNN, Tri-LAN). The comparative experimental results are presented in the form of three-dimensional histograms in Figure 15. It is clear that the proposed Tri-CLAN performs the best in experiment 2, which primarily depends on the distinctive structure. Making a concrete analysis, the ladder-shaped encoder–decoder module can obtain more unsupervised data auxiliary information and the CNN-backbone prevents overfitting, which remedies the feature distribution shifting problem and improves the overall test accuracy.

4.3. Case Study 2: Motor Fault Dataset of SZTU

4.3.1. Fault Dataset Description

The motor fault dataset is collected and organized by Shenzhen Technology University (SZTU). The test system is shown in Figure 16, which mainly consists of a three-phase asynchronous motor, two bearing seats, a rotating disc, a planetary gear box, a frequency converter and a magnetic powder brake. The rotating components are connected through couplings, and the vibration of the test system is obtained through a sequence of vibration acceleration sensors, a data acquisition system and a host computer. The motor speed is controlled by a frequency converter and can be adjusted within the range of 0–1750 rpm. The load of the motor is adjusted through a magnetic particle brake, which can provide a torque load of 0–50 N∙m.
In order to verify the diagnostic performance of the proposed Tri-CLAN under different working conditions, four working conditions are listed in Table 9. Especially, C1 and C3 working conditions are set to the rated speed of the motor, and C1 and C4 working conditions are set to the highest torque load. The motor fault dataset includes six fault categories and one health category, for which the corresponding labels are listed in Table 10. The vibration data at the 12 o’clock direction of the motor fan end is selected, and each 2048 sampling points are set as one sample. The sample number of the training dataset and the testing dataset are 300 and 100 for each category. Then FFT and zero-mean normalization processing are introduced into each sample. The number of training and testing samples for each category under different operating conditions is listed in Table 11, which includes labeled samples and unlabeled samples.

4.3.2. Results Analysis

We directly select one labeled sample for experimental verification under variable conditions, which is the most extreme case. The heatmaps of experimental results for one labeled sample are displayed in Figure 17. It can be seen that the result of the classification task can maintain around 95% under the same working condition. Faced with variable working conditions with one labeled sample, the classification task of fault location can still reach more than 90%. C1 and C3 working conditions have the greatest difference in both speed and load, and Figure 18 shows the confusion matrices of the cross-working condition task. It can be observed that BRM and BBM are easily misclassified among the seven fault categories. It is worth noting that none of the fault samples are misclassified as healthy samples, which can ensure the safe operation of the motor.
Compared to the publicly available CWRU dataset, the motor fault dataset has more fault categories and more complex working conditions, which can sufficiently demonstrate the applicability of the proposed method in more complex practical scenarios.

4.4. Case Study 3: Laboratory Bearing Dataset of Motorized Spindle Test Rig

4.4.1. Fault Dataset Description and Experiment Setup

Here, to further illustrate the extensive applicability of the proposed method, a laboratory bearing of the motorized spindle test rig (MSTR) [46] is separately conducted to acquire data and analysis. The overall test rig is illustrated in Figure 19, which consists of a supporting system, a loading system, a force convert system and an accessory system. The bearing required for experimental verification is the core part of the force convert system, which needs to bear dynamic loads during the high-speed rotation test of the motorized spindle, such as radial load, axial load and torque load. Therefore, the bearing under dynamic load and high-speed rotation is the basis for ensuring the safe and stable operation of the reliability test of the motorized spindle. In order to ensure the accurate application of load and the safety of the experimenter, it is necessary to identify the health state of the bearing. As the research object of this case study, the vibration acceleration signal of the bearing is collected by the acceleration sensor placed on the force convert system sleeve.
To align with the experiment of case study 1, the sampling frequency of the vibration data is set to 12 kHz. The rotation speed of the bearing is consistent with that of the motorized spindle, and two working conditions with a large difference are selected for experimental verification according to the reliability test conditions in the literature [46], which are 5000 rpm (C1) and 1000 rpm (C2), respectively. Four kinds of health conditions are carried out for 10 s, which include Normal, IF, OF and B. Furthermore, the same data preprocessing as case study 1 is executed and the experimental settings are listed in Table 12.

4.4.2. Results Analysis

We carried out two typical working conditions including four kinds of identification tasks, respectively, as the most demanding data condition, and the heatmaps of the experimental results for one labeled sample are displayed in Figure 20. It can be seen that the result of the classification task can reach 100% under the same working conditions, which is significantly higher than the experimental results of the CWRU bearing dataset in case study 1. The classification target focusing on fault location is considered a prime reason for this phenomenon, and we can see to some extent that the neglecting of the damage degree reduces the training difficulty for the proposed method. Faced with variable working conditions with one labeled sample, the classification task of the fault location can still reach more than 95%. Observing the confusion matrices in Figure 21, which is the closest to the average accuracy, the experimental results are analyzed separately under two different working conditions. Specifically, only 18 and 5 samples belonging to OR faults are wrongly classified as IR faults, respectively, and none of the faults’ health statuses are misidentified, which can ensure the safe operation of the MSTR long-term reliability test. From the experiments of this case study, we further prove that the proposed Tri-CLAN has extensive applicability and universality. In different application scenarios, it can effectively mine extremely limited labeled data and utilize easily available unlabeled data to achieve fault diagnosis under variable working conditions.

5. Conclusions

Given the actual industrial data, it is urgent and necessary to carry out end-to-end intelligent fault diagnosis. This paper innovatively proposes a triplet-guided path-interaction CNN-based ladder network, which realizes the intelligent fault diagnosis with extremely limited labeled samples under variable working conditions. To accommodate the data requirements, the proposed method can be elaborated from the aspects of algorithm structure and feature space. First, to integrate the advantage of the convolution operation and achieve information fusion with fewer parameters, this paper replaces the vanilla LAN backbone with a CNN and constructs a path-interaction semi-supervised architecture with a simplified combination activation function. Furthermore, benefiting from the contribution to the triplet loss with a hard sample mining strategy, the feature distribution shifting problem between variable working conditions is alleviated by learning fault-related representation at the feature level. The public CWRU bearing dataset is utilized to verify the effectiveness of the proposed method, and two experimental datasets are applied to illustrate its extensive applicability. In future research, we will collect data from more engineering application scenarios to verify the proposed algorithm. Furthermore, we will further study the intelligent fault diagnosis method for the purpose of its application to other extreme circumstances in the industry.

Author Contributions

Conceptualization, Z.Y. and F.C.; data curation, X.Z.; formal analysis, Z.Q.; funding acquisition, F.C. and B.X.; investigation, Z.Q.; methodology, Z.Y.; project administration, F.C.; resources, X.Z.; supervision, F.C.; validation, B.M. and Z.Q.; visualization, B.M.; writing—original draft, Z.Y.; writing—review and editing, Z.Y. and F.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Guangdong Province Key Construction Discipline Scientific Research Capacity Improvement Project, grant number 2022ZDJS114; the Key Laboratory of Robot Intelligence Technology for 3C Machinery Industry, grant number 2022SA-07-18; the Self-made Experimental Instruments and Equipment Project of Shenzhen Technology University, grant number JSZZ202301013; and the Sichuan Science and Technology Planning Project, grant number 2021YJ0371.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request due to restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
  2. Yang, Z.; Xu, B.B.; Luo, W.; Chen, F. Autoencoder-based representation learning and its application in intelligent fault diagnosis: A review. Measurement 2022, 189, 110460. [Google Scholar] [CrossRef]
  3. Xing, S.; Lei, Y.; Wang, S.; Jia, F. Distribution-invariant deep belief network for intelligent fault diagnosis of machines under new working conditions. IEEE Trans. Ind. Electron. 2021, 68, 2617–2625. [Google Scholar] [CrossRef]
  4. Zhao, W.; Wang, Z.; Cai, W.; Zhang, Q.; Wang, J.; Du, W.; Yang, N.; He, X. Multiscale inverted residual convolutional neural network for intelligent diagnosis of bearings under variable load condition. Measurement 2022, 188, 110511. [Google Scholar] [CrossRef]
  5. An, Z.; Li, S.; Wang, J.; Jiang, X. A novel bearing intelligent fault diagnosis framework under time-varying working conditions using recurrent neural network. ISA Trans. 2020, 100, 155–170. [Google Scholar] [CrossRef]
  6. Yang, S.; Kong, X.; Wang, Q.; Li, Z.; Cheng, H.; Xu, K. Deep multiple auto-encoder with attention mechanism network: A dynamic domain adaptation method for rotary machine fault diagnosis under different working conditions. Knowl. Based Syst. 2022, 249, 108639. [Google Scholar] [CrossRef]
  7. Wu, H.; Li, J.; Zhang, Q.; Tao, J.; Meng, Z. Intelligent fault diagnosis of rolling bearings under varying operating conditions based on domain-adversarial neural network and attention mechanism. ISA Trans. 2022, 130, 477–489. [Google Scholar] [CrossRef]
  8. Zhang, T.; Chen, J.; Li, F.; Zhang, K.; Lv, H.; He, S.; Xu, E. Intelligent fault diagnosis of machines with small & imbalanced data: A state-of-the-art review and possible extensions. ISA Trans. 2022, 119, 152–171. [Google Scholar] [CrossRef]
  9. He, Z.; Shao, H.; Wang, P.; Lin, J.; Cheng, J.; Yang, Y. Deep transfer multi-wavelet auto-encoder for intelligent fault diagnosis of gearbox with few target training samples. Knowl. Based Syst. 2020, 191, 105313. [Google Scholar] [CrossRef]
  10. Li, Q.; Tang, B.; Deng, L.; Wu, Y.; Wang, Y. Deep balanced domain adaptation neural networks for fault diagnosis of planetary gearboxes with limited labeled data. Measurement 2020, 156, 107570. [Google Scholar] [CrossRef]
  11. Zaman, S.M.K.; Liang, X. An effective induction motor fault diagnosis approach using graph-based semi-supervised learning. IEEE Access 2021, 9, 7471–7482. [Google Scholar] [CrossRef]
  12. Wang, X.; Feng, H.; Fan, Y. Fault detection and classification for complex processes using semi-supervised learning algorithm. Chemom. Intell. Lab. Syst. 2015, 149, 24–32. [Google Scholar] [CrossRef]
  13. Wang, X.; Wang, T.; Ming, A.; Zhang, W.; Li, A.; Chu, F. Semi-supervised hierarchical attribute representation learning via multi-layer matrix factorization for machinery fault diagnosis. Mech. Mach. Theory 2022, 167, 104445. [Google Scholar] [CrossRef]
  14. Wu, X.; Zhang, Y.; Cheng, C.; Peng, Z. A hybrid classification autoencoder for semi-supervised fault diagnosis in rotating machinery. Mech. Syst. Signal Process. 2021, 149, 107327. [Google Scholar] [CrossRef]
  15. Li, S.; Luo, J.; Hu, Y. Semi-supervised process fault classification based on convolutional ladder network with local and global feature fusion. Comput. Chem. Eng. 2020, 140, 106843. [Google Scholar] [CrossRef]
  16. Pandiyan, V.; Akeddar, M.; Prost, J.; Vorlaufer, G.; Varga, M.; Wasmer, K. Long short-term memory based semi-supervised encoder-decoder for early prediction of failures in self-lubricating bearings. Friction 2022, 11, 109–124. [Google Scholar] [CrossRef]
  17. Rasmus, A.; Valpola, H.; Honkala, M.; Berglund, M.; Raiko, T. Semi-supervised learning with ladder networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; MIT Press: Cambridge, MA, USA, 2015; pp. 3546–3554. [Google Scholar]
  18. Zhang, S.; Qiu, T. Semi-supervised LSTM ladder autoencoder for chemical process fault diagnosis and localization. Chem. Eng. Sci. 2022, 251, 117467. [Google Scholar] [CrossRef]
  19. Zhang, S.; Ye, F.; Wang, B.; Habetler, T.G. Semi-supervised bearing fault diagnosis and classification using variational autoencoder-based deep generative models. IEEE Sens. J. 2021, 21, 6476–6486. [Google Scholar] [CrossRef]
  20. Zhang, K.; Tang, B.; Qin, Y.; Deng, L. Fault diagnosis of planetary gearbox using a novel semi-supervised method of multiple association layers networks. Mech. Syst. Signal Process. 2019, 131, 243–260. [Google Scholar] [CrossRef]
  21. Razavi-Far, R.; Hallaji, E.; Farajzadeh-Zanjani, M.; Saif, M.; Kia, S.H.; Henao, H.; Capolino, G.-A. Information fusion and semi-supervised deep learning scheme for diagnosing gear faults in induction machine systems. IEEE Trans. Ind. Electron. 2019, 66, 6331–6342. [Google Scholar] [CrossRef]
  22. Yu, K.; Lin, T.R.; Ma, H.; Li, X.; Li, X. A multi-stage semi-supervised learning approach for intelligent fault diagnosis of rolling bearing using data augmentation and metric learning. Mech. Syst. Signal Process. 2021, 146, 107043. [Google Scholar] [CrossRef]
  23. Ding, X.; Lu, X.; Wang, D.; Lv, Q.; Zhai, H. Rolling bearing fault diagnosis based on probabilistic mixture model and semi-supervised ladder network. Adv. Mech. Eng. 2020, 12, 168781402097774. [Google Scholar] [CrossRef]
  24. Tao, X.; Ren, C.; Li, Q.; Guo, W.; Liu, R.; He, Q.; Zou, J. Bearing defect diagnosis based on semi-supervised kernel local fisher discriminant analysis using pseudo labels. ISA Trans. 2021, 110, 394–412. [Google Scholar] [CrossRef] [PubMed]
  25. Xu, M.; Wang, Y. An imbalanced fault diagnosis method for rolling bearing based on semi-supervised conditional generative adversarial network with spectral normalization. IEEE Access 2021, 9, 27736–27747. [Google Scholar] [CrossRef]
  26. Wang, Z.; Xuan, J.; Shi, T. A novel semi-supervised generative adversarial network based on the actor-critic algorithm for compound fault recognition. Neural Comput. Appl. 2022, 34, 10787–10805. [Google Scholar] [CrossRef]
  27. Pan, T.; Chen, J.; Zhang, T.; Liu, S.; He, S.; Lv, H. Generative adversarial network in mechanical fault diagnosis under small sample: A systematic review on applications and future perspectives. ISA Trans. 2021, 128, 1–10. [Google Scholar] [CrossRef]
  28. Zhang, T.; He, S.; Chen, J.; Pan, T.; Zhou, Z. Toward small sample challenge in intelligent fault diagnosis: Attention-weighted multidepth feature fusion net with signals augmentation. IEEE Trans. Instrum. Meas. 2022, 71, 1–13. [Google Scholar] [CrossRef]
  29. Wang, Y.; Sun, G.; Jin, Q. Imbalanced sample fault diagnosis of rotating machinery using conditional variational auto-encoder generative adversarial network. Appl. Soft Comput. J. 2020, 92, 106333. [Google Scholar] [CrossRef]
  30. Liu, S.; Jiang, H.; Wu, Z.; Li, X. Rolling bearing fault diagnosis using variational autoencoding generative adversarial networks with deep regret analysis. Measurement 2021, 168, 108371. [Google Scholar] [CrossRef]
  31. Yi, S.; Wang, X.; Tang, X. Deep learning face representation by joint identification-verification. arXiv 2014, arXiv:1406.4773. [Google Scholar]
  32. Schroff, F.; Kalenichenko, D.; Philbin, J. FaceNet: A unified embedding for face recognition and clustering. arXiv 2015, arXiv:1503.03832. pp. 815–823. [Google Scholar]
  33. Zhou, H.; Huang, X.; Wen, G.; Dong, S.; Lei, Z.; Zhang, P.; Chen, X. Convolution enabled transformer via random contrastive regularization for rotating machinery diagnosis under time-varying working conditions. Mech. Syst. Signal Process. 2022, 173, 109050. [Google Scholar] [CrossRef]
  34. Ragab, M.; Chen, Z.; Zhang, W.; Eldele, E.; Wu, M.; Kwoh, C.-K.; Li, X. Conditional contrastive domain generalization for fault diagnosis. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  35. Peng, P.; Lu, J.; Xie, T.; Tao, S.; Wang, H.; Zhang, H. Open-set fault diagnosis via supervised contrastive learning with negative out-of-distribution data augmentation. IEEE Trans. Ind. Inf. 2022, 19, 2463–2473. [Google Scholar] [CrossRef]
  36. Rombach, K.; Michau, G.; Fink, O. Contrastive learning for fault detection and diagnostics in the context of changing operating conditions and novel fault types. Sensors 2021, 21, 3550. [Google Scholar] [CrossRef]
  37. Qu, F.; Liu, J.; Liu, X.; Jiang, L. A multi-fault detection method with improved triplet loss based on hard sample mining. IEEE Trans. Sustain. Energy 2021, 12, 127–137. [Google Scholar] [CrossRef]
  38. Wang, X.; Liu, F. Triplet loss guided adversarial domain adaptation for bearing fault diagnosis. Sensors 2020, 20, 320. [Google Scholar] [CrossRef] [Green Version]
  39. Hinton, G.E.; Osindero, S.; Teh, Y. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  40. Vincent, P.; Larochelle, H.; Bengio, Y.; Manzagol, P.-A. Extracting and composing robust features with denoising autoencoders. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; Association for Computing Machinery: New York, NY, USA, 2008; pp. 1096–1103. [Google Scholar]
  41. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  42. Gao, Y.; Chen, M.; Yu, D. Semi-supervised graph convolutional network and its application in intelligent fault diagnosis of rotating machinery. Measurement 2021, 186, 110084. [Google Scholar] [CrossRef]
  43. Wang, D.; Zhang, M.; Xu, Y.; Lu, W.; Yang, J.; Zhang, T. Metric-based meta-learning model for few-shot fault diagnosis under multiple limited data conditions. Mech. Syst. Signal Process. 2021, 155, 107510. [Google Scholar] [CrossRef]
  44. Li, S.; Yang, W.; Zhang, A.; Liu, H.; Huang, J.; Li, C.; Hu, J. A novel method of bearing fault diagnosis in time-frequency graphs using InceptionResnet and deformable convolution networks. IEEE Access 2020, 8, 92743–92753. [Google Scholar] [CrossRef]
  45. Smith, W.A.; Randall, R.B. Rolling element bearing diagnostics using the Case Western Reserve University data: A benchmark study. Mech. Syst. Signal Process. 2015, 64–65, 100–131. [Google Scholar] [CrossRef]
  46. Yang, Z.; Chen, F.; Luo, W.; Xu, B.; Wang, S. Reliability test rig of the motorized spindle and improvements on its ability for high-speed and long-term tests. Shock Vib. 2021, 2021, 6637335. [Google Scholar] [CrossRef]
Figure 1. A conceptual illustration of the vanilla LAN.
Figure 1. A conceptual illustration of the vanilla LAN.
Sensors 23 06951 g001
Figure 2. A conceptual illustration of the triplet loss.
Figure 2. A conceptual illustration of the triplet loss.
Sensors 23 06951 g002
Figure 3. The architecture of the proposed method.
Figure 3. The architecture of the proposed method.
Sensors 23 06951 g003
Figure 4. Demonstration of the 1D convolution operation.
Figure 4. Demonstration of the 1D convolution operation.
Sensors 23 06951 g004
Figure 5. A flowchart of the proposed method.
Figure 5. A flowchart of the proposed method.
Sensors 23 06951 g005
Figure 6. The experimental test rig of CWRU.
Figure 6. The experimental test rig of CWRU.
Sensors 23 06951 g006
Figure 7. Confusion matrices for experiment 1 results on C1: (a) A1: five labels; (b) B1: two labels; (c) C1: one label.
Figure 7. Confusion matrices for experiment 1 results on C1: (a) A1: five labels; (b) B1: two labels; (c) C1: one label.
Sensors 23 06951 g007
Figure 8. Visualization features for experiment 1 results on C1: (a) five labels; (b) two labels; (c) one label.
Figure 8. Visualization features for experiment 1 results on C1: (a) five labels; (b) two labels; (c) one label.
Sensors 23 06951 g008
Figure 9. Accuracy curves for experiment 1.
Figure 9. Accuracy curves for experiment 1.
Sensors 23 06951 g009
Figure 10. Ablation experimental results for experiment 1.
Figure 10. Ablation experimental results for experiment 1.
Sensors 23 06951 g010
Figure 11. Accuracy curves of ablation experiment with 1 labeled sample.
Figure 11. Accuracy curves of ablation experiment with 1 labeled sample.
Sensors 23 06951 g011
Figure 12. Heatmaps of experiment 2 results: (a) five label; (b) two label; (c) one label.
Figure 12. Heatmaps of experiment 2 results: (a) five label; (b) two label; (c) one label.
Sensors 23 06951 g012
Figure 13. Confusion matrices of experiment 2 results with one label: (a) C1–C4; (b) C4–C1.
Figure 13. Confusion matrices of experiment 2 results with one label: (a) C1–C4; (b) C4–C1.
Sensors 23 06951 g013
Figure 14. Visualization features of experiment 2 results with one label: (a) C1–C4; (b) C4–C1.
Figure 14. Visualization features of experiment 2 results with one label: (a) C1–C4; (b) C4–C1.
Sensors 23 06951 g014
Figure 15. Three-dimensional histograms of experiment 2 results for comparison: (a) is the comparison with Tri-LAN; (b) is the comparison with Tri-CNN.
Figure 15. Three-dimensional histograms of experiment 2 results for comparison: (a) is the comparison with Tri-LAN; (b) is the comparison with Tri-CNN.
Sensors 23 06951 g015
Figure 16. Rotating machinery fault test system.
Figure 16. Rotating machinery fault test system.
Sensors 23 06951 g016
Figure 17. Heatmaps of motor experimental results with one label.
Figure 17. Heatmaps of motor experimental results with one label.
Sensors 23 06951 g017
Figure 18. Confusion matrices for cross-working condition task: (a) C3–C4; (b) C4–C3.
Figure 18. Confusion matrices for cross-working condition task: (a) C3–C4; (b) C4–C3.
Sensors 23 06951 g018
Figure 19. The laboratory test rig of the motorized spindle.
Figure 19. The laboratory test rig of the motorized spindle.
Sensors 23 06951 g019
Figure 20. Heatmaps of experimental results with 1 label.
Figure 20. Heatmaps of experimental results with 1 label.
Sensors 23 06951 g020
Figure 21. Confusion matrices for cross-working condition task: (a) C1–C2; (b) C2–C1.
Figure 21. Confusion matrices for cross-working condition task: (a) C1–C2; (b) C2–C1.
Sensors 23 06951 g021
Table 1. Basic modules.
Table 1. Basic modules.
ModuleNetwork LayerKernel SizeModuleNeurons Number
CE-CNNConv1d #116 × 15CE-FC896-512-256
Conv1d #232 × 15DE-FC512-896-1024
Conv1d #364 × 5Predicting#1Embedding-1024
DE-CNNTransConv1d #132 × 5#21024-64
TransConv1d #216 × 15#364-fault classes
TransConv1d #31 × 15Metric learning ~
Table 2. Architectures of the methods in case studies.
Table 2. Architectures of the methods in case studies.
MethodsCE-CNNDE-CNNCE-FCDE-FCPredictingMetric
Learning
Tri-CLAN
CLAN
Tri-CNN
CNN
Tri-LAN
Vanilla LAN
Table 3. Parameter setup for experiments.
Table 3. Parameter setup for experiments.
ParametersValues
Learning rate0.001
Training epochs100
Batch size of labeled datafault classes
Batch size of unlabeled data200
Gaussian noise ε (0,1)
Table 4. Working conditions of CWRU.
Table 4. Working conditions of CWRU.
Working ConditionMotor Load (hp)Motor Speed (rpm)
C101797
C211772
C321750
C431730
Table 5. Labels of fault location with different damage degree.
Table 5. Labels of fault location with different damage degree.
Fault LocationFault Diameter (Inch)Fault Labels
None (Normal)00
Inner Raceway (IR)0.071
0.142
0.213
Outer Raceway (OR)0.074
0.145
0.216
Ball (B)0.077
0.148
0.219
Table 6. Experiment 1: settings of datasets under C1.
Table 6. Experiment 1: settings of datasets under C1.
NameTraining Samples (Labeled/Unlabeled)Testing Samples
A15/100100
B12/100100
C11/100100
Table 7. Experiment 2: settings of datasets under variable working conditions.
Table 7. Experiment 2: settings of datasets under variable working conditions.
NameTraining Samples (Labeled/Unlabeled)Testing Samples
C1C2C3C4C1C2C3C4
A21/100000100100100100
2/100
5/100
B201/10000
2/100
5/100
C2001/100
2/100
5/100
D20001/100
2/100
5/100
Table 8. Comparisons with state-of-the-art methods for experiment 1 on CWRU.
Table 8. Comparisons with state-of-the-art methods for experiment 1 on CWRU.
ReferencesFault
Location
Damage
Degree
Training Samples
(Labeled/Unlabeled)
Accuracy (%)
[42]50/95098.40
[43]-10/-90.93
[44]-900/-88.54
[19]300/12,90087.63
This work50/100099.98
20/100099.43
10/100092.45
Table 9. Working conditions of the motor fault dataset.
Table 9. Working conditions of the motor fault dataset.
Working ConditionSetting Speed (rpm)Actual Speed (rpm)Load
(N·m)
C11750172233
C21500149017
C31750174017
C490087533
Table 10. Labels of motor fault location.
Table 10. Labels of motor fault location.
Fault LocationFault Labels
Normal0
Rotor unbalanced motor (RUM)1
Bending rotor motor (BRM)2
Faulty bearing motor (FBM)3
Broken bar motor (BBM)4
Stator winding fault motor (WFM)5
Single phase fault motor (SPM)6
Table 11. Experimental settings of motor datasets under variable working conditions.
Table 11. Experimental settings of motor datasets under variable working conditions.
NameTraining Samples (Labeled/Unlabeled)Testing Samples
C1C2C3C4C1C2C3C4
A1/100000100100100100
B01/10000
C001/1000
D0001/100
Table 12. Experimental settings of MSTR under variable working conditions.
Table 12. Experimental settings of MSTR under variable working conditions.
NameTraining Samples
(Labeled/Unlabeled)
Testing Samples
C1C2C1C2
A1/1000100100
B01/100
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Z.; Chen, F.; Xu, B.; Ma, B.; Qu, Z.; Zhou, X. Metric Learning-Guided Semi-Supervised Path-Interaction Fault Diagnosis Method for Extremely Limited Labeled Samples under Variable Working Conditions. Sensors 2023, 23, 6951. https://doi.org/10.3390/s23156951

AMA Style

Yang Z, Chen F, Xu B, Ma B, Qu Z, Zhou X. Metric Learning-Guided Semi-Supervised Path-Interaction Fault Diagnosis Method for Extremely Limited Labeled Samples under Variable Working Conditions. Sensors. 2023; 23(15):6951. https://doi.org/10.3390/s23156951

Chicago/Turabian Style

Yang, Zheng, Fei Chen, Binbin Xu, Boquan Ma, Zege Qu, and Xin Zhou. 2023. "Metric Learning-Guided Semi-Supervised Path-Interaction Fault Diagnosis Method for Extremely Limited Labeled Samples under Variable Working Conditions" Sensors 23, no. 15: 6951. https://doi.org/10.3390/s23156951

APA Style

Yang, Z., Chen, F., Xu, B., Ma, B., Qu, Z., & Zhou, X. (2023). Metric Learning-Guided Semi-Supervised Path-Interaction Fault Diagnosis Method for Extremely Limited Labeled Samples under Variable Working Conditions. Sensors, 23(15), 6951. https://doi.org/10.3390/s23156951

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop