Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Three-Dimensional Digital Zooming of Integral Imaging under Photon-Starved Conditions
Next Article in Special Issue
Long Short-Term Memory Neural Networks for Modeling Dynamical Processes and Predictive Control: A Hybrid Physics-Informed Approach
Previous Article in Journal
In-Orbit Calibration of Phase Retardance for Channeled Spectropolarimeter
Previous Article in Special Issue
Achieving Reliability in Cloud Computing by a Novel Hybrid Approach
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Empowered Digital Twin Using Acoustic Signal for Welding Quality Inspection

School of Mechanical Engineering, Universiti Sains Malaysia, Nibong Tebal 14300, Pulau Pinang, Malaysia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(5), 2643; https://doi.org/10.3390/s23052643
Submission received: 5 February 2023 / Revised: 23 February 2023 / Accepted: 24 February 2023 / Published: 28 February 2023
(This article belongs to the Special Issue Fuzzy Systems and Neural Networks for Engineering Applications)
Figure 1
<p>Digital twin system.</p> ">
Figure 2
<p>Welding defects: (<b>a</b>) incomplete penetration; (<b>b</b>) excessive penetration; (<b>c</b>) dent; (<b>d</b>) burn-through.</p> ">
Figure 3
<p>Welding acoustic signals: (<b>a</b>) signals with obvious welding defects; (<b>b</b>) signals without obvious welding defects.</p> ">
Figure 4
<p>Weld images: (<b>a</b>) Front face of weld with obvious welding defects; (<b>b</b>) back side of weld with obvious welding defects; (<b>c</b>) welds without obvious welding defects. (<b>1</b>) Excessive penetration, (<b>2</b>) burn-through, and (<b>3</b>) dent.</p> ">
Figure 5
<p>Time- and frequency-domain diagrams of environmental noise.</p> ">
Figure 6
<p>Time domain and frequency domain of two randomly selected welding signals (First row: Time-domain diagram; Second row: Normalised frequency spectrum; Third row: Normalised persistent frequency spectrum.</p> ">
Figure 7
<p>Comparison of wavelet denoising results for three random samples with different signal-to-noise ratios (Row 1: original signal; Row 2: soft threshold wavelet filtering; Row 3: Hard threshold wavelet filtering; Row 4: Improved valence wavelet filtering).</p> ">
Figure 8
<p>Improved wavelet denoising results for two actual welding signals.</p> ">
Figure 9
<p>SeCNN-LSTM structure diagram (<b>N</b>): Number of sampling points per group (<b>M</b>): Number of data groups.</p> ">
Figure 10
<p>Weight initializer comparison.</p> ">
Figure 11
<p>Test confusion matrix.</p> ">
Figure 12
<p>Test accuracy curve.</p> ">
Figure 13
<p>ROC curve comparison.</p> ">
Figure 14
<p>P–R curve comparison.</p> ">
Versions Notes

Abstract

:
Weld site inspection is a research area of interest in the manufacturing industry. In this study, a digital twin system for welding robots to examine various weld flaws that might happen during welding using the acoustics of the weld site is presented. Additionally, a wavelet filtering technique is implemented to remove the acoustic signal originating from machine noise. Then, an SeCNN-LSTM model is applied to recognize and categorize weld acoustic signals according to the traits of strong acoustic signal time sequences. The model verification accuracy was found to be 91%. In addition, using numerous indicators, the model was compared with seven other models, namely, CNN-SVM, CNN-LSTM, CNN-GRU, BiLSTM, GRU, CNN-BiLSTM, and LSTM. A deep learning model, and acoustic signal filtering and preprocessing techniques are integrated into the proposed digital twin system. The goal of this work was to propose a systematic on-site weld flaw detection approach encompassing data processing, system modeling, and identification methods. In addition, our proposed method could serve as a resource for pertinent research.

1. Introduction

Several countries have selected Industry 4.0 as a strategic goal for their industrial development because of its huge potential and exciting possibilities. In industrial production, welding is the most common method for joining materials. Various welding processes can be accomplished very efficiently using automatic equipment, such as industrial robots [1]. Providing the robot with only expected instructions allows it to begin working immediately.
Although industrial robots have a high level of intelligence, the whole welding process does not have the same level of intelligence. Online intelligent weld defect detection still has a troublesome track record, at least when it comes to detecting weld defects [2]. Online weld testing not only makes weld quality more reliable and stable, but also improves production efficiency and reduces production costs [3]. GMAW (Gas Metal Arc Welding) is one of the most commonly used welding processes. However, high temperature, spatter, high-strength arc light, and complex welding environments make GMAW online weld detection difficult [4].
The most popular method for finding weld defects online is to use visual images to obtain more understandable weld and weld pool information. By gathering an image of the weld center and by examining the geometric and spatial distribution properties of the weld, Ma suggested a weld defect detection method based on active visual sensing and machine learning [5].
Zhang developed an image collection system to acquire image data in all welding directions. After that, a CNN (Convolutional Neural Network) is used to identify it. It is important to note that the system incorporates welding arc light as complementary information rather than weakening or eliminating arc light interference [6]. Arc changes can be reflected in welding current and voltage. In order to determine the weld penetration, Cao used the arc voltage in the peak current period and the average arc voltage during this time and proposed a bilinear model to design a gain-scheduled model predictive controller [7]. Ultrasonic testing has also been used for online testing as a conventional, non-destructive procedure. Bonikila established a non-destructive weld detection method based on ultrasonics utilizing a machine learning algorithm [8].
The analysis of arc acoustic signals has always been a crucial factor in the online assessment of weld quality. An acoustic signal may perfectly match the detection criteria, especially in the online detection of weld fields. However, acoustic signals are easily interrupted, and welding is not a linear change process. A reason for this is the fact that acoustic signal extraction and analysis are highly challenging.
Through experimentation, Horvat examined the acoustic signal of GMAW welding, identified the two primary noise sources, and proposed a classification algorithm [9]. Gao developed the notion that professional welders can detect welding process flaws through their auditory experience. By gathering significant crowd input, a subjective assessment model based on human hearing was presented [10]. Following the extraction of acoustic information using LPCC (Licensed Professional Clinical Counselors), Lv suggested a BPANN (Back Propagation Artificial Neural Network) pie-change algorithm and constructed a controller using this algorithm. The controller accuracy in the experiment ranged from 80% to 90% [11]. Cui studied the arc bias blowing phenomenon in K-TIG (Keyhole Tungsten Inert Gas) welding and subsequently suggested an ECOC-SVM-GSCV technique based on machine learning to categorize weld penetration flaws [3]. Liu focused on cold metal transfer lap welding (CMT) and emphasized how air intake affects the caliber of the weld. An RNN-based BiLSTM-CTC method was suggested to determine the signal features by tracking the sound signal of air intake [12]. Acoustic experimental hardware was upgraded by Chen, who also set up an array of several microphones and used the FastICA (Fast algorithm for Independent Component Analysis) to separate weld noise [13]. Wang developed a central audit perception model, tested the model effectiveness in online weld detection, and used the model to examine the human hearing principle to identify defects in GMAW welding [14].
The acoustic signal is a key parameter in welding site inspection. The real-time and accurate acoustic diagnosis of weld defects has long been a challenging subject because of the complicated on-site environment and highly non-linear welding process. In order to identify weld defects, this study suggests a system based on digital twin acoustic signals. Real-time, accurate defect identification was considered when defining the fundamental needs of the digital twin system.
This paper presents a digital twin system for weld sound detection that includes the setup of hardware for experimental purposes, creating a virtual environment for experimentation. In Section 2, the communication network of the virtual reality system and other components are introduced. In Section 3, the preprocessing approach for sound signals is introduced. By examining the time-domain and frequency-domain properties of the signal, we then suggest an adaptive wavelet threshold denoising method. To extract signal features and to detect and classify signals based on the timing properties of acoustic signals and the requirements of real-time systems, an improved SeCNN-LSTM deep learning model is proposed in Section 4. In this work, seven other models were tested and compared with the SeCNN-LSTM model, and a variety of metrics, such as precision and sample balance, were used for validation. This paper is summarized in Section 5, which also examines its flaws and suggests a future line of inquiry.

2. Digital Twin System

2.1. Digital Twin of Industrial Robot

Digital twin (DT) is currently acknowledged as one of the key avenues for advancement in achieving Industry 4.0 and industrial intelligence [15]. The earliest application of DTs was in the aerospace industry to build a fictitious digital object for tracking and forecasting spacecraft performance [16]. Digital twins have started to show their potential in recent years as a result of the significant advancement in and richness of the communication sector as well as the precision and variety of sensors. DTs have swiftly spread across numerous sectors and research fields in just a few years [17]. DTs represent a potent tool for producing real-time analysis systems, because they are virtual replicas of real systems and can communicate in two directions with the physical copy [18].
Lu proposed a straightforward DT model in their article [19]. The model makes it clear that the inclusion of physical objects, digital objects, and communication is the fundamental characteristic of DTs. Information models and data processing operations make up the majority of digital objects. Real-time, bidirectional communication is the fundamental criterion. Ren developed a digital twin of a coaxial single-sided resistance spot weld for real-time prediction of transient temperature fields, based on an experimentally validated finite element model [20].
Numerous reports have been published so far on the use of DTs in the realm of industrial robotics. Xu presented an industrial cloud robot based on DT that combines cloud computing and industrial robots. They utilized it for sorting logistics [21]. With the use of a DT system, Zhuang assembled and predicted spaceships using a model of a workshop complete with industrial robots [22]. Wang focused on the picture data of the weld pool during the welding operation in order to construct a DT system [23]. Tiparey suggested a way for flexible industrial robots to pick and place workpieces based on the idea of DT [24].
The real-time reflection of the condition of physical things and bidirectional communication are the fundamental requirements of DTs. The approach to on-site weld inspection is entirely consistent with this. Tao, F proposes a digital twin modelling approach for five-dimensional models [25], as shown in Formula (1).The DT system used in this research utilizes the five-dimensional model creation method.
M D T = ( P E , D E , S s , D D , C N )
where P E is the physical object, D E is the digital object, S s is the service target or service standard, D D is data processing, and C N is the communication mode.

2.1.1. Physical Entity

The physical entity is the physical component of the activity or the component that needs to be digitized. The physical parts of industrial robots usually include the robot body, the control cabinet, and FlexPendant. Welding robots also need to include gas cylinders and welding control cabinets. There are numerous different structures in the robot body. The robot model used in this paper was KUKA KR 10 R1420 HP, which is a classic six-axis serial robot. The robot controller model was KUKA KR C4. In addition, there were supporting wire feeding mechanisms, welding gun, and workpieces. A TBi Industries RoboMIG RM2 system was adopted for welding gun and wire feeder. The microphone and acoustic signal acquisition tools employed is KY-037, which consists of three main components: microphone; audio amplifier and comparator. An electret condenser microphone (ECM) as an audio sensor to detect and measure the area physically sounds and produce an analog signal. The audio amplifier receives the analogue signal from the ECM and amplifies it. The comparator compares the amplified signal with the reference value and change output level. The physical entity also includes the workspace where the robot is located, which was an enclosed space of about 10 m2 in our case. Computing resources are I7 CPU and RTX 3050Ti GPU.

2.1.2. Digital Entity

The digital entity is a highly integrated and quantified model in the DT system. It accurately simulates a wide range of physical entity features. The digital entity characteristics are displayed in Formula (2).
D E = ( G m , P m , D m , R m )
where G m is a geometric model. The geometric model of an industrial robot is usually represented by its 3D model, which was here built using CAD software. It should be noted that the geometric model of the workpiece also had to be built, because the focus of this paper is the robot welding procedure.
In the above equation, P m is the physical property model. Through the physical property model, the material and physical properties of the entity are expressed in the physical property model. Welding wire materials and their physical qualities, workpiece materials and their welding characteristics, and welding shielding gas are all important considerations in welding process experiments. The characteristics of the materials used here are shown in Table 1.
In the above equation, D m is a dynamic model that mostly refers to the robot trajectory, drive mode, and other kinematic and power information. The geometry of the workpiece and the operator, in general, determine the motion information of an industrial robot.
In the above equation, R m is the rule model. The secret to unlocking the intelligence of the digital twin system is the rule model. It refers to the guidelines for robot evaluation, prediction, and judgment. A sort of SeCNN-LSTM deep learning algorithm was employed in this experiment to develop the rule model for detecting welding errors in robots, which are crucial to locating weld flaws. In Section 3 and Section 4, this section is thoroughly explained.

2.1.3. Service Project

S s refers to the goal of common services of digital entities and physical entities and is expressed by Formula (3).
S s = ( F u n , I n p u t , O u t p u t , Q u a , S t )
where F u n is a functional term. This experiment’s primary purpose was to identify welding process defects by means of the on-the-job monitoring of weld defects. Robot motion and position data made up the majority of I n p u t and O u t p u t , with the sound signals captured by sensors also serving as inputs. Q u a refers to quality, which is understood as weld quality and recognition efficiency in this paper. Presenting comparative tests, Section 5 of this paper thoroughly introduces the topic of recognition efficiency. S t is the robot running condition.

2.1.4. Data

Since the DT system is essentially a virtual picture product, it must receive data from every system participant.
D D = ( D P e , D D e , D S s , D e )
where D P e are data from robots and sensors, D D e are the feedback data of virtual entities, D S s are server data, and D e are the expert data in the knowledge base.

2.1.5. Communication

The DT system uses a variety of communication channels because of its stringent requirements for information transmission. B i D E _ D D , B i D E _ S s , and B i D D _ S s stand for digital entities and databases, database and server interaction, and two-way communication between the digital entity and the server, respectively. The communication between the physical entity and the server is represented by P E _ D D and P E _ S s , and by P E _ D D and P E _ S s , respectively. It is important to note that the industrial robot is highly integrated equipment. Typically, the industrial robot control cabinet or FlexPendant has a few external connectors. Some degree of permission to control the motion status of each robot joint can be obtained through the external interface. B i I N r o b refers to two-way external interface communication with the robot. On the KUKA robot control cabinet and FlexPendant we employed, there were Ethernet external ports. Since E X s e n was used to connect external sound sensors, only one-way communication was possible in this section.
C N = ( P E _ D D , B i D E _ D D , P E _ S s , B i D E _ S s , B i D D _ S s )
P E _ D D = P E _ S s = ( B i I N r o b , E X s e n )

2.2. Building a DT

As seen in Figure 1, the digital twin system in the experiment was built using the 5D digital twin modeling technique. This paper includes a detailed description of the essential tools and setup for this experiment in the physical entity section. The workspace information was also reflected in the system, because the digital twin system needs to take into account more thorough physical entities. The operator gives the first instructions as the service terminal. Three categories of data—workspace data (environmental data), robot data, and weld status data—are used to organize the data in the database. Robot data mostly refer to the data that the robot itself naturally possesses, such as motion status, geometric details, etc. Weld status data, which primarily record weld quality, weld flaws, and weld types, represent the weld field status fed back through the rule model. To improve data accuracy, the sensor data must go through several filtering processes, which are fully discussed in Section 3. The three primary kinds of digital entity models are geometric models of robots, dynamic models, and physical property models, each of which is independently modeled and returns the necessary data. Regarding the rule model, Section 4. provides a detailed introduction to weld status recognition and weld defect classification.

3. Signal Preprocessing

3.1. Acoustic Signal Analysis

In welding research, there are many different sorts of defects, such as cold cracks, hot cracks, incomplete fusion, incomplete penetration, burn-through, blowholes, excessive penetration, dents, and so forth. Some defects may occur after welding, for example, cold cracks may occur in the cooling stage after welding. Some defects occur during welding, such as incomplete penetration, burn-through, collapse, and pits. Figure 2a shows a root that has not completely entered, which is characteristic of incomplete penetration. The workpiece break angle, low current, excessive speed, and inappropriate welding wire angle could all be contributing factors. Incomplete penetration can easily lead to a buildup of stress at the root, which can result in fissures and structural harm. Excessive weld penetration is depicted in Figure 2b, in which the front of the weld is collapsed and the back is lifted. This is the result of excessive workpiece clearance or current, or excessive metal penetration through the rear. Local low-lying areas that occur on the weld surface define the pit shape. The working section of the weld becomes smaller because of pits. The most significant occurrence is burn-through. Molten metal pours out of the groove to create the perforation, as depicted in Figure 2d. This happens because the workpiece is typically overheated, the welding current is too high, and the welding pace is too sluggish. Generally speaking, welding current, the production of these faults, voltage, and speed are connected.
The welding robot workspace in this experiment was an enclosed space. Possible sources of noise in the workspace included the two fans that were there. Second, there was noise generated by the robot movement, as well as some silent noise makers in order to lessen how noise affected the acoustic signals. Both time-domain and frequency-domain analyses of two groups of acoustic sounds were performed in this study. In Figure 3, there are two groups of acoustic signals: those with apparent weld defects and those with good weld quality. Comparable pictures of real welds are shown in Figure 4.
Figure 3a matches Figure 4a,b, while Figure 3b matches Figure 4c. The acoustic signals in Figure 3 have sections of signal with reduced amplitude at the front and back, which represent the situations before and after welding, respectively. At this time, machine noise and outside noise make up the majority of the signal.
It should be noticed that the amplitude abruptly reduces in the middle of Figure 3a, which is the result of the arc extinguishing phenomenon caused by the arc burning through the workpiece and the arc touching the base plate. Indicated by mark (2) in Figure 4, the burnt-through hole is visible. The front and back images of the weld are shown in Figure 4 as (a) and (b), respectively. Combining the two images reveals a range of different weld flaws.
Figure 3 makes it abundantly evident that there are some irregular peaks and relatively small amplitude variations in the signal time-domain characteristics of the pre-welding and post-welding stages. The machine internal noise and the fan background noise were the main sources of noise in the enclosed environment, as already mentioned. The portion of background noise was collected separately in the shutdown stage, as shown in Figure 5.
The amplitude did not exceed 0.1, and the energy was primarily concentrated below the frequency of 200 Hz, characteristic of environmental and fan noise, respectively. In order to separate the machine noise, we extracted 0.5 S data from the signals at both ends of Figure 2 and obtained the time-domain diagram, frequency spectrum, and persistent frequency spectrum of the signal, as shown in Figure 6. The persistent frequency spectrum clearly shows the changes in signal components. In the persistent frequency spectrum, it can be observed that there is an obvious disturbance between the normalized frequencies of 0.4 and 0.6.

3.2. Improved Wavelet Denoising

Figure 6 shows that machine noise, which was the major noise source, strongly correlated with weld acoustic signals and could only be seen in certain frequency bands. If only this frequency band were filtered, it would cause serious signal loss. Therefore, this research proposes a wavelet denoising method with an improved threshold function to lessen the interference of machine noise.
C j + 1 ( n ) = k C j ( k ) h ( k 2 n ) , W j + 1 ( n ) = k W j ( k ) g ( k 2 n ) ,
The above is the wavelet decomposition of the signal, where C j n are scale factors and W j n are wavelet coefficients.
The traditional threshold functions have a hard threshold and soft threshold, as shown in Formula (8) and Formula (9), respectively.
W t = W W t 0 W < t
W t = s i g n ( W ) ( W t ) W t 0 W < t
where W t is the wavelet coefficient after the application of the threshold.
The threshold is set to
t = σ 2 ln N
where σ is the standard variance of noise, N is the length of the signal. However, hard thresholds and soft thresholds have their own shortcomings. The hard threshold is discontinuous at t = W . The soft threshold may be over-compressed when t W .
In fact, the energy of arc signals is relatively high and concentrated, and excessive compression affects signal recovery. Therefore, based on the threshold value, we set coefficient A as shown in Formula (11).
A = 1 Z ( t W )
W t = s i g n ( W ) ( W A t ) W t 0 W < t
where 0 < Z < 1 when W t , A < 1 . Therefore, the improved threshold is always greater than the soft threshold, which means that the value of Z can be appropriately adjusted according to the signal.
Reconstruction is the reverse process of decomposition. The formula is as follows:
C j n = k [ C j + 1 k h k 2 n + W j + 1 ( k ) g ( k 2 ) ]
This study compared three groups of various sample signals to demonstrate the efficacy of the wavelet technique, as shown in Figure 7. The test sample signal information was taken from the website SC.chinaz.com. Noise signals of varying decibel levels were added to the three groups of sample signals. The sample signals were subjected to soft threshold-, hard threshold-, and improved threshold-based filtering using the Haar wavelet basis function. The method described in this paper performed well in terms of the noise reduction effect compared with the signal-to-noise ratio (SNR) of the processing results, as shown in Table 2.
The measured signal was decomposed using a three-layer wavelet, and the signal was denoised as shown in Figure 8.

4. Identification and Classification

4.1. Classification Model

The usual sequential signal, the acoustic signal, has a very high degree of continuity. In order to identify and categorize the information reflected by acoustic signals, in this work, we developed an SeCNN-LSTM depth learning model.
The discipline of speech recognition has extensively explored acoustic signal processing. However, although they can be used as a guide, related voice recognition technologies are not entirely relevant, because signals from human speech are the primary focus of speech recognition. For instance, the widely used MFCC (mel-frequency cepstral coefficient) feature classification approach enhances the human auditory frequency range while suppressing the high-frequency band [26,27]. Suppressing high-frequency signals results in a significant loss of information when industrial machinery is used as the research object.
There are currently three established techniques for identifying auditory signals. First, the MFCC approach is represented by cepstral feature extraction and classification [28]. For example, a signal of 16 KHZ usually intercepts a frame of the cepstrum, with a 39D channel output. Cepstral feature extraction also makes use of LPCCs (linear prediction cepstral coefficients), LSFs (line spectral frequencies), PLP (perceptual linear prediction), and other algorithms [29,30]. Additionally, filter banks that have an output of up to 80D channels are used to extract the features. The alternative is to sample and categorize signals directly. The benefit of this is that the output channel can be altered to prevent catastrophes caused by high and small numbers of dimensions.
Figure 9 displays the deep learning model created in this paper. The experimental requirements for the deep learning model include a lightweight design in addition to an effective recognition rate, because a certain level of portability is necessary for the creation of digital twin systems. As a result, we built the one-dimensional convolution layer feature extraction process as a three-layer parallel extended structure. Although it resembles the traditional Squeeze network, the parallel convolution kernel size is the same. Data are folded for one-dimensional convolution when they are input and then expanded, and their time series are regressed. The output timing signal group has 256 channels. Finally, the two layers are LSTM (Long Short-Term Memory) structures of 128 and 32 hidden units.

4.2. Model Training and Parameters

Dataset: The sampling frequency of the acoustic signal was 4.8 KHz, so 0.1 S signals were intercepted as a group, where each group contained 480 sampling points. A total of 90 groups were intercepted upon the occurrence of each defect type. Because the burn-through signal had a certain degree of instantaneity and abruptness, the burn-through signal was 40 groups. The data for this experiment is 310 timing signals and the amount of data is 148,800. 80% is the training set and 20% is the testing set.
Loss function: Softmax is often applied to multi-category tasks, also known as multi-category categorical cross entropy loss.
S ( x , y ) = i = 1 N q ( x i ) log q ( x i )
q ( x i ) = y i c i i [ 1 , N ]
where c i is x i and corresponds to the target class.
Activation function: In the one-dimensional convolution part of the model, Formula (16) is used as the activation function.
R u e l ( x ) = m a x ( 0 , x ) = 0 , x < 0 x , x > = 0
LSTM performs well in current time series data classification and prediction tasks. The saturated activation function is more suitable for the requirements of LSTM. Formula (17) is used as the gate activation function for the forgetting gate, input gate, and output gate. When generating candidate memory, the state activation function is Formula (18).
S i g m o i d ( x ) = 1 1 + e x
T a n h ( x ) = e x e x e x + e x
Weight initialization: In the training process, weight has a significant impact on the effect of network training. The optimal scheme is obtained by comparing different weight initialization functions. The ”Glorot” initializer independently samples from a uniform distribution with 0 and weights W g .
W g U 1 n , 1 n
where U 1 n , 1 n is the uniform distribution in interval 1 n , 1 n and n is the size of the previous layer. The "He" initializer independently samples from a uniform distribution with 0 and weights W h .
W h U 0 , 2 V n i
where V n i indicates the current input size. The “Orthogonal” initializer refers to the orthogonal matrix decomposed from the random matrix sampled from the normal distribution. “Narrow normal” refers to sampling from 0 to 0.01 standard deviation.
Figure 10 shows the training results under the influence of different initialization weight functions. From the figure, it can be seen that the training result of the "HE" initialization function was the best one.
The maximum iteration was 100, and the learning rate was 0.005.
The SeCNN-LSTM network was trained as shown in Figure 11. Target class refers to the prediction classification item, while output class denotes the true classification item. The sample was unbalanced because of the experimental circumstances, as was indicated in the preceding article. Additionally, since a random sample among all the samples was used to create the test set data, each sample in the confusion matrix was likewise out of balance.
A c c = t r ( c o n ) S u m ( c o n )
where A c c refers to accuracy; t r ( c o n ) refers to the trace of the confusion matrix, that is, the sum of the main diagonal elements; and S u m ( c o n ) is the sum of all elements in the matrix. The test set confusion matrix verification accuracy reached 91.0%.

4.3. Model Comparison

In order to verify the superiority of the model described in this paper, the SeCNN-LSTM model was compared with other seven models using multiple indicators. Models LSTM, BiLSTM(Bidirectional Long Short-Term Memory), CNN-LSTM, CNN-BiLSTM, GRU(Gated Recurrent Units), CNN-GRU, and CNN-SVM were compared. In the area of sequence classification, the widely applied and developed deep learning models LSTM and BiLSTM were used. BiLSTM is composed of two LSTMs, with one receiving inputs forwards and the other backwards. Another type of RNN is GRU. GRU is better in line with the demands of processing power and time cost than LSTM and has fewer parameters. The goal of this paper was to demonstrate the benefits of the deep learning model described in this paper in one-dimensional sequential classification networks. The deep learning model in this paper is a CNN-RNN structure, and the majority of the models involved in the comparison are also CNN-RNN structures.
The test set accuracy of each model is shown in Figure 12. The test accuracy of the SeCNN-LSTM model used in this study was 90.99% after 100 iterations, while those of CNN-SVM and CNN-BiLSTM were 83.78% and 78.28%, respectively. The test accuracy values of other models were under 80%.
In binary classification models, precision and recall are commonly employed as important metrics to check the model classification outcomes. This idea is expanded in the multi-classification model to cover the classification of each sample separately. A metric used to describe the outcomes of predictions is precision. It speaks of the likelihood of a real positive sample among all anticipated positive samples. The initial sample is described by the recall rate. It is the likelihood that the expected positive sample will be positive.
As can be seen in Table 3, the SeCNN-LSTM model classification precision was significantly higher than those of other classification models, with the first three samples being classified correctly with more than 90% precision and the fourth sample being classified correctly with over 85% precision.
Each sample in the SeCNN-LSTM model had a recall rate of over 88%, which is significantly greater than those of conventional classification models.
In actual classification, sample imbalance is inevitable. ROC (receiver operating characteristic) and AUC (area under the curve) indicators can ignore sample imbalance to assess the classification and prediction abilities of a model.
ROC refers to the relationship between sensitivity (recall) and specificity under different classification thresholds. ROC is a concept based on a binary classification model. In a multivariate classification model, each sample needs to be considered separately. As shown in Figure 13, the ROC curves of the eight classification models are compared. It could be observed that the ROC curve of the SeCNN-LSTM model is smoother and more concentrated in the ( 0 , 1 ) coordinates.
AUC refers to the area values enclosed by ROC curve and coordinates. The closer the AUC value is to 1, the better the model is. When the AUC value is less than 0.5, the model is invalid. It can be seen from Table 4 that the AUC values of all samples of the SECNN-LSTM model are above 0.9, much higher than other models.
As shown in Figure 14, the P–R curve refers to the relationship between precision and recall under different classification thresholds. Precision and recall are a pair of mutually exclusive indicators. In an actual sample, the highest accuracy rate and recall rate cannot be obtained at the same time, so only an ideal balance value can be obtained. It can be observed from the P–R curves of the eight models that the curves of SeCNN-LSTM model are smoother and more concentrated in the ( 1 , 1 ) coordinates.
The area values enclosed by the P–R curves (P–R AUCs) are shown in Table 5. The value of each sample of the SeCNN-LSTM model is above 0.83.
The F1-score indicates the best balance between precision and recall. As shown in Table 6, comparing the F1-score of each sample of the eight models, the model described in this paper showed the best performance.
In summary, we compared the eight different models’ accuracy, precision, recall, AUC, ROC, F1 score and other measures. The findings demonstrate that the model presented in this research outperforms the one-dimensional sequence data classification model in terms of performance, stability, and sample balance.

5. Conclusions

This research established a digital twin system for robot weld field detection based on acoustic signal processing and acoustic signal analysis. The identification and classification of auditory signals was performed using an enhanced SeCNN-LSTM deep learning model. The digital twin system incorporates both a deep learning model and a signal processing technique.
The digital twin system modeled in this paper is insufficient. First, the modeling of rule models was covered at length in this study, but the modeling methods of the physical material model and the dynamic model in the digital twin system were not discussed in detail. Second, the deep learning model discussed in this research was found to have 91% verification accuracy, which is not the best possible outcome, meaning that there is still room for development.
Weld field inspection is both a research focus and a challenge in the welding industry, due to the complicated on-site environment, influencing factors, non-linear welding conditions, and other considerations. Despite having solid timing, the acoustic signal is susceptible to interference. Therefore, future research could focus on multi-sensor or multi-modal weld information detection to reflect the welding conditions using several types of weld information.

Author Contributions

Conceptualization, T.J.; methodology, T.J.; software, T.J. and N.M.N.; validation, T.J. and N.M.N.; formal analysis, T.J.; investigation, T.J.; resources, N.M.N.; data curation, T.J.; writing—original draft preparation, T.J.; writing—review and editing, N.M.N.; visualization, N.M.N.; supervision, N.M.N.; project administration, N.M.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the authors upon reasonable request.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could appear to have influenced the work reported in this paper.

References

  1. Zhang, Z.; Wen, G.; Chen, S. On-Line Monitoring and Defects Detection of Robotic Arc Welding: A Review and Future Challenges. In Transactions on Intelligent Welding Manufacturing; Springer: Berlin/Heidelberg, Germany, 2019; pp. 3–28. [Google Scholar] [CrossRef]
  2. Xu, J.; Chen, S. The Review of Spectrum Detection and Ultrasonic Vibration Control of Porosity Defects in Aluminum Alloy Welding. In Transactions on Intelligent Welding Manufacturing; Springer: Berlin/Heidelberg, Germany, 2020; pp. 3–24. [Google Scholar] [CrossRef]
  3. Cui, Y.; Shi, Y.; Zhu, T.; Cui, S. Welding penetration recognition based on arc sound and electrical signals in K-TIG welding. Measurement 2020, 163, 107966. [Google Scholar] [CrossRef]
  4. Chen, S.B.; Lv, N. Research evolution on intelligentized technologies for arc welding process. J. Manuf. Process. 2014, 16, 109–122. [Google Scholar] [CrossRef]
  5. Ma, G.; Yuan, H.; Yu, L.; He, Y. Monitoring of weld defects of visual sensing assisted GMAW process with galvanized steel. Mater. Manuf. Process. 2021, 36, 1178–1188. [Google Scholar] [CrossRef]
  6. Zhang, Z.; Wen, G.; Chen, S. Weld image deep learning-based on-line defects detection using convolutional neural networks for Al alloy in robotic arc welding. J. Manuf. Process. 2019, 45, 208–216. [Google Scholar] [CrossRef]
  7. Cao, Y.; Wang, Z.; Hu, S.; Wang, W. Modeling of weld penetration control system in GMAW-P using NARMAX methods. J. Manuf. Process. 2021, 65, 512–524. [Google Scholar] [CrossRef]
  8. Bonikila, P.R.; Indimath, S.S.; Shajan, N. Failure assessment of Mash Seam Weld breakage and development of online weld inspection system for early detection of weld failure. Eng. Fail. Anal. 2022, 133, 105967. [Google Scholar] [CrossRef]
  9. Horvat, J.; Prezelj, J.; Polajnar, I.; Čudina, M. Monitoring Gas Metal Arc Welding Process by Using Audible Sound Signal. Stroj. Vestnik– J. Mech. Eng. 2011, 2011, 267–278. [Google Scholar] [CrossRef]
  10. Gao, Y.; Zhao, J.; Wang, Q.; Xiao, J.; Zhang, H. Weld bead penetration identification based on human-welder subjective assessment on welding arc sound. Measurement 2020, 154, 107475. [Google Scholar] [CrossRef]
  11. Lv, N.; Xu, Y.; Li, S.; Yu, X.; Chen, S. Automated control of welding penetration based on audio sensing technology. J. Mater. Process. Technol. 2017, 250, 81–98. [Google Scholar] [CrossRef]
  12. Liu, L.; Chen, H.; Chen, S. Quality analysis of CMT lap welding based on welding electronic parameters and welding sound. J. Manuf. Process. 2022, 74, 1–13. [Google Scholar] [CrossRef]
  13. Na, L.; Chen, S.j.; Chen, Q.h.; Tao, W.; Zhao, H.; Chen, S.b. Dynamic welding process monitoring based on microphone array technology. J. Manuf. Process. 2021, 64, 481–492. [Google Scholar] [CrossRef]
  14. Wang, Q.; Gao, Y.; Huang, L.; Gong, Y.; Xiao, J. Weld bead penetration state recognition in GMAW process based on a central auditory perception model. Measurement 2019, 147, 106901. [Google Scholar] [CrossRef]
  15. Boschert, S.; Rosen, R. Digital Twin—The Simulation Aspect. In Mechatronic Futures; Springer: Berlin/Heidelberg, Germany, 2016; Book Section Chapter 5; pp. 59–74. [Google Scholar] [CrossRef]
  16. Tao, F. Make more digital twins. Nature 2019, 573, 490–491. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Qi, Q.; Tao, F.; Hu, T.; Anwer, N.; Liu, A.; Wei, Y.; Wang, L.; Nee, A.Y.C. Enabling technologies and tools for digital twin. J. Manuf. Syst. 2021, 58, 3–21. [Google Scholar] [CrossRef]
  18. Cimino, C.; Negri, E.; Fumagalli, L. Review of digital twin applications in manufacturing. Comput. Ind. 2019, 113, 103130. [Google Scholar] [CrossRef]
  19. Lu, Y.; Liu, C.; Wang, K.I.K.; Huang, H.; Xu, X. Digital Twin-driven smart manufacturing: Connotation, reference model, applications and research issues. Robot. Comput.-Integr. Manuf. 2020, 61, 101837. [Google Scholar] [CrossRef]
  20. Ren, S.; Ma, Y.; Ma, N.; Chen, Q.; Wu, H. Digital twin for the transient temperature prediction during coaxial one-side resistance spot welding of Al5052/CFRP. J. Manuf. Sci. Eng. 2022, 144, 031015. [Google Scholar] [CrossRef]
  21. Xu, W.; Cui, J.; Li, L.; Yao, B.; Tian, S.; Zhou, Z. Digital twin-based industrial cloud robotics: Framework, control approach and implementation. J. Manuf. Syst. 2021, 58, 196–209. [Google Scholar] [CrossRef]
  22. Zhuang, C.; Miao, T.; Liu, J.; Xiong, H. The connotation of digital twin, and the construction and application method of shop-floor digital twin. Robot. Comput.-Integr. Manuf. 2021, 68, 102075. [Google Scholar] [CrossRef]
  23. Wang, Q.; Jiao, W.; Zhang, Y. Deep learning-empowered digital twin for visualized weld joint growth monitoring and penetration control. J. Manuf. Syst. 2020, 57, 429–439. [Google Scholar] [CrossRef]
  24. Tipary, B.; Erdős, G. Generic development methodology for flexible robotic pick-and-place workcells based on Digital Twin. Robot. Comput.-Integr. Manuf. 2021, 71, 102140. [Google Scholar] [CrossRef]
  25. Tao, F.; Zhang, M.; Liu, Y.; Nee, A.Y.C. Digital twin driven prognostics and health management for complex equipment. CIRP Ann. 2018, 67, 169–172. [Google Scholar] [CrossRef]
  26. Imran, M.S.; Rahman, A.F.; Tanvir, S.; Kadir, H.H.; Iqbal, J.; Mostakim, M. An analysis of audio classification techniques using deep learning architectures. In Proceedings of the 2021 6th International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 20–22 January 2021; pp. 805–812. [Google Scholar]
  27. Tran, V.T.; Tsai, W.H. Acoustic-Based Train Arrival Detection Using Convolutional Neural Networks with Attention. IEEE Access 2022, 10, 72120–72131. [Google Scholar] [CrossRef]
  28. Boulmaiz, A.; Messadeg, D.; Doghmane, N.; Taleb-Ahmed, A. Design and implementation of a robust acoustic recognition system for waterbird species using TMS320C6713 DSK. Int. J. Ambient Comput. Intell. (IJACI) 2017, 8, 98–118. [Google Scholar] [CrossRef]
  29. Maria, A.; Jeyaseelan, A.S. Development of optimal feature selection and deep learning toward hungry stomach detection using audio signals. J. Control Autom. Electr. Syst. 2021, 32, 853–874. [Google Scholar] [CrossRef]
  30. Saddam, S.A.W. Wind Sounds Classification Using Different Audio Feature Extraction Techniques. Informatica 2022, 45. [Google Scholar] [CrossRef]
Figure 1. Digital twin system.
Figure 1. Digital twin system.
Sensors 23 02643 g001
Figure 2. Welding defects: (a) incomplete penetration; (b) excessive penetration; (c) dent; (d) burn-through.
Figure 2. Welding defects: (a) incomplete penetration; (b) excessive penetration; (c) dent; (d) burn-through.
Sensors 23 02643 g002
Figure 3. Welding acoustic signals: (a) signals with obvious welding defects; (b) signals without obvious welding defects.
Figure 3. Welding acoustic signals: (a) signals with obvious welding defects; (b) signals without obvious welding defects.
Sensors 23 02643 g003
Figure 4. Weld images: (a) Front face of weld with obvious welding defects; (b) back side of weld with obvious welding defects; (c) welds without obvious welding defects. (1) Excessive penetration, (2) burn-through, and (3) dent.
Figure 4. Weld images: (a) Front face of weld with obvious welding defects; (b) back side of weld with obvious welding defects; (c) welds without obvious welding defects. (1) Excessive penetration, (2) burn-through, and (3) dent.
Sensors 23 02643 g004
Figure 5. Time- and frequency-domain diagrams of environmental noise.
Figure 5. Time- and frequency-domain diagrams of environmental noise.
Sensors 23 02643 g005
Figure 6. Time domain and frequency domain of two randomly selected welding signals (First row: Time-domain diagram; Second row: Normalised frequency spectrum; Third row: Normalised persistent frequency spectrum.
Figure 6. Time domain and frequency domain of two randomly selected welding signals (First row: Time-domain diagram; Second row: Normalised frequency spectrum; Third row: Normalised persistent frequency spectrum.
Sensors 23 02643 g006
Figure 7. Comparison of wavelet denoising results for three random samples with different signal-to-noise ratios (Row 1: original signal; Row 2: soft threshold wavelet filtering; Row 3: Hard threshold wavelet filtering; Row 4: Improved valence wavelet filtering).
Figure 7. Comparison of wavelet denoising results for three random samples with different signal-to-noise ratios (Row 1: original signal; Row 2: soft threshold wavelet filtering; Row 3: Hard threshold wavelet filtering; Row 4: Improved valence wavelet filtering).
Sensors 23 02643 g007
Figure 8. Improved wavelet denoising results for two actual welding signals.
Figure 8. Improved wavelet denoising results for two actual welding signals.
Sensors 23 02643 g008
Figure 9. SeCNN-LSTM structure diagram (N): Number of sampling points per group (M): Number of data groups.
Figure 9. SeCNN-LSTM structure diagram (N): Number of sampling points per group (M): Number of data groups.
Sensors 23 02643 g009
Figure 10. Weight initializer comparison.
Figure 10. Weight initializer comparison.
Sensors 23 02643 g010
Figure 11. Test confusion matrix.
Figure 11. Test confusion matrix.
Sensors 23 02643 g011
Figure 12. Test accuracy curve.
Figure 12. Test accuracy curve.
Sensors 23 02643 g012
Figure 13. ROC curve comparison.
Figure 13. ROC curve comparison.
Sensors 23 02643 g013
Figure 14. P–R curve comparison.
Figure 14. P–R curve comparison.
Sensors 23 02643 g014
Table 1. Characteristics of materials of physical equipment.
Table 1. Characteristics of materials of physical equipment.
Name/BrandMaterial
WorkpieceStainless steel S30400 C = 0.07 , M n = 2 , P = 0.045 , S = 0.03 , S i = 0.75 , C r = 17.5–19.5, N i = 8–10.5
Solder WireER70s-6 C 0.08 , S i 1.15 , M n = 1.4–1.85, P 0.025 , S 0.035 , C u 0.5
Shielding GasCO2
Table 2. Signal-to-noise ratio (SNR).
Table 2. Signal-to-noise ratio (SNR).
Samle 1Sample 2Sample 3
Soft shreshold3.65256.582812.2800
Hart shreshold3.56456.363812.1649
Improved shreshold4.18866.973912.3590
Table 3. Accuracy, precision and recall comparison.
Table 3. Accuracy, precision and recall comparison.
AccuracyPrecisionRecall
Incomplete PenetrationExcessive PentrationDentBurn-ThroughIncomplete PenetrationExcessive PentrationDentBurn-Through
SeCNN-LSTM90.99%93.93%92.59%90%85.71%88.57%89.28%90%100%
LSTM63.96%93.3%47.3%93.8%64%38.9%86.7%53.6%94.1%
GRU61.26%67.9%61.9%85.7%66.87%59.4%83.9%42.9%90 %
CNN-SVM83.78%71.4%92.3%73.1%0%74.1%88.9%70.4%0%
CNN-LSTM73.87%89.3%73.5%68.6%92.9%61%96.2%88.9%76.5%
CNN-GRU74.77%72.7%83.8%83.3%75%88.9%91.2%50%81.8%
CNN-BiLSTM79.28%64.7%84.4%88.9%72.2%81.5%84.4%63.2%92.9%
BiLSTM68.47%86.7%58.8%78.3%68.2%39.4%90.9%62.1%93.8%
Table 4. AUC comparison.
Table 4. AUC comparison.
Incomplete PenetrationExcessive PentrationDentBurn-ThroughAverage
SeCNN-LSTM0.92740.96870.90520.98950.9443
LSTM0.71720.78270.78090.92790.7952
GRU0.85730.83430.76640.92690.8320
CNN-SVM0.76870.81880.72410.94700.7634
CNN-LSTM0.84260.91590.83530.98680.8951
CNN-GRU0.86410.90120.84480.97950.8965
CNN-BiLSTM0.82380.90160.88420.93100.8872
BiLSTM0.86950.82030.73460.90970.8210
Table 5. P–R AUC comparison.
Table 5. P–R AUC comparison.
Incomplete PenetrationExcessive PentrationDentBurn-ThroughAverage
SeCNN-LSTM0.88380.94280.83910.93220.8938
LSTM0.62560.59910.71600.74610.6719
GRU0.79000.70610.71030.73230.7303
CNN-SVM0.61250.63170.466500.4501
CNN-LSTM0.81540.84010.84790.84920.8323
CNN-GRU0.79550.86120.77630.93290.8358
CNN-BiLSTM0.76960.83380.79510.80700.8061
BiLSTM0.75460.66780.66830.72840.7018
Table 6. F1 score comparison.
Table 6. F1 score comparison.
Incomplete PenetrationExcessive PentrationDentBurn-ThroughAverage
SeCNN-LSTM0.74060.77410.69840.75610.7436
LSTM0.56280.44540.59110.54140.5716
GRU0.63450.55560.58570.56720.6212
CNN-SVM0.50230.63210.538400.3984
CNN-LSTM0.68180.74880.71660.71990.7212
CNN-GRU0.69450.75730.66850.78710.7296
CNN-BiLSTM0.64410.77200.67660.73870.7127
BiLSTM0.56520.54880.53250.56310.5792
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ji, T.; Mohamad Nor, N. Deep Learning-Empowered Digital Twin Using Acoustic Signal for Welding Quality Inspection. Sensors 2023, 23, 2643. https://doi.org/10.3390/s23052643

AMA Style

Ji T, Mohamad Nor N. Deep Learning-Empowered Digital Twin Using Acoustic Signal for Welding Quality Inspection. Sensors. 2023; 23(5):2643. https://doi.org/10.3390/s23052643

Chicago/Turabian Style

Ji, Tao, and Norzalilah Mohamad Nor. 2023. "Deep Learning-Empowered Digital Twin Using Acoustic Signal for Welding Quality Inspection" Sensors 23, no. 5: 2643. https://doi.org/10.3390/s23052643

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop