Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Generalized Parking Occupancy Analysis Based on Dilated Convolutional Neural Network
Next Article in Special Issue
A Novel Surface Descriptor for Automated 3-D Object Recognition and Localization
Previous Article in Journal
Smart Water Management Platform: IoT-Based Precision Irrigation for Agriculture
Previous Article in Special Issue
Dressing Tool Condition Monitoring through Impedance-Based Sensors: Part 1—PZT Diaphragm Transducer Response and EMI Sensing Technique
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Self-Adaptive 1D Convolutional Neural Network for Flight-State Identification

1
School of Civil Aviation, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China
2
Shanghai Engineering Research Center of Civil Aircraft Health Monitoring, Shanghai Aircraft Customer Service Co., Ltd., Shanghai 200241, China
3
Department of Mechanical, Aerospace, and Nuclear Engineering, Rensselaer Polytechnic Institute, Troy, NY 12180, USA
4
School of Electronic, Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
5
Department of Aeronautics and Astronautics, Stanford University, Stanford, CA 94305, USA
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(2), 275; https://doi.org/10.3390/s19020275
Submission received: 26 November 2018 / Revised: 27 December 2018 / Accepted: 2 January 2019 / Published: 11 January 2019
(This article belongs to the Special Issue Sensor Applications for Smart Manufacturing Technology and Systems)

Abstract

:
The vibration of a wing structure in the air reflects coupled aerodynamic–mechanical responses under varying flight states that are defined by the angle of attack and airspeed. It is of great challenge to identify the flight state from the complex vibration signals. In this paper, a novel one-dimension convolutional neural network (CNN) is developed, which is able to automatically extract useful features from the structural vibration of a recently fabricated self-sensing wing through wind-tunnel experiments. The obtained signals are firstly decomposed into various subsignals with different frequency bands via dual-tree complex-wavelet packet transformation. Then, the reconstructed subsignals are selected to form the best combination for multichannel inputs of the CNN. A swarm-based evolutionary algorithm called grey-wolf optimizer is utilized to optimize a set of key parameters of the CNN, which saves considerable human efforts. Two case studies demonstrate the high identification accuracy and robustness of the proposed method over standard deep-learning methods in flight-state identification, thus providing new perspectives in self-awareness toward the next generation of intelligent air vehicles.

1. Introduction

A novel “fly-by-feel” (FBF) concept was recently proposed, inspired by the remarkable flying capabilities of birds in nature. This concept aims at the development of the next generation of intelligent air-vehicle structures that can “feel”, “think”, and “react” based on high-resolution state-sensing, awareness, and self-diagnostic capabilities [1,2]. This kind of bioinspired systems are able to sense the environment (e.g., temperature, pressure, and aerodynamic forces) at unprecedented length and time scales. In addition, it could think in real time and be aware of its current flight state and structural health condition. Further, such a system could react intelligently under various situations and achieve superior performance and agility with optimal flight control. Compared with current flying approaches, the FBF concept has the following advantages: (1) sensing the external environment and internal structural condition (stresses, strains, damage) using the same integrated intelligent structure [3], (2) being aware of the general flight state and some critical states (flutter, stall, etc.) through effective data interpretation from intelligent structure, (3) optimal decision-making and autonomous flight control based on self-awareness [2]. Toward this end, great challenges have been exerted on current structural design and data-processing techniques, with a departure from existing technologies.
Recent years have witnessed investigations into different sensing-network architectures and simulations [4,5,6,7]. A typical example is that a stretchable network made of polymer-based substrates was designed by the Structure and Composites Lab (SACL) at Stanford University. This network contains many micronodes with the ability to integrate microsensors, actuators, and electronics for multiple applications [8]. Based on the development of microfabrication techniques [9,10,11], a composite wing structure with the sensor network monolithically embedded was successfully fabricated [3], as shown in Figure 1. This multifunctional wing consists of four sensor networks, and each network is integrated with strain gauges, resistive temperature detectors (RTD), and piezoelectric lead–zirconate titanate (PZT) transducers. Specifically, the strain gauge measures wing-strain distribution and can used to identify the dangerous area. RTD detects the temperature distribution to provide temperature compensation [12]. PZT transducers can be used under both active and passive modes. In active mode, they are used for structural health monitoring for any damage detection, while in passive mode, wing structural vibrations during flying are measured to indicate the air dynamic characteristics [3].
The flight state is characterized by a set of critical variables, such as altitude, angle of attack, airspeed, and temperature, forming a flight-state vector that indicates a specific aerodynamic condition. For the self-sensing wing mentioned above, its flight-state vector is defined by a combination of airspeed and angle of attack, which can be reflected by the signals obtained from the embedded sensor network. Kopsaftopoulos and Chang proposed a stochastic global identification framework on the basis of a developed Vector-dependent Functionally Pooled (VFP) model [2,13,14]. It combines the state-space method and stochastic time-series analysis, and is able to capture and predict the structural dynamics and the aeroelastic response under multiple flight states. From another perspective, if some indicative features can be extracted from the continuous structural aerodynamic behavior, it is possible to identify the flight state directly using the limited feature information instead of detailed characterization of the structural responses. Therefore, effective feature extraction is crucial in flight-state identification from the sensing-signal space to the practical-state space.
Feature extraction is one of the key steps in data-analysis processes, which can be divided into two approaches, by manpower and by machine [15]. Features selected by human expertise are explainable and straightforward. For example, various statistical features can be calculated, such as the mean value, standard deviation, peak value, and kurtosis, from both time domain and frequency domain for time-series signals [16,17,18]. The authors have also proposed novel feature-extraction and selection methods to achieve highly important features with low colinearity in the previous research [19]. However, these approaches heavily rely on domain knowledge. There might be useful features that remain uncovered. Alternatively, converting raw data into a set of features can be complemented by automatic feature construction using deep learning, which belongs to the broad family of machine-learning algorithms using multiple layers for feature extraction and transformation [20]. The development of deep learning was boosted since 2006, when a multilayer feedforward neural network consisting of stacked restricted Boltzmann machines showed superior performance in data representation [21]. Many DL architectures for signal and information processing have been developed, including deep autoencoders (DAE) [22], deep belief networks (DBN) [23], convolutional neural networks (CNN) [24], recurrent neural networks (RNN) [25], and their modifications. Although various deep neural networks (DNN) can save considerable human efforts in feature extraction and achieve competitive results, there are certain number of parameters in the DNN structure that are difficult to adjust. To enable adaptive DNN configuration, many researchers have developed various metaheuristic methods for parameter optimization [26]. To list a few examples, in a modified DAE for vibrational signal analysis, an artificial fish-swarm algorithm was used to optimized the critical parameters of the DAE [27]. Classical algorithm particle-swarm optimization was employed for determination of the proper structure of a trained DBN and a CNN for rolling bearing-fault diagnosis [28,29]. Three popular algorithms, simulated annealing, differential evolution, and harmony search, were used to optimize a CNN on classifying the MNIST and CIFAR datasets [30]. The cuckoo search algorithm was implemented in conjunction with an RNN and two back propagation neural networks for fast convergence and local minima avoidance [31]. These optimized DNNs outperformed their original structures, and produced results comparable or even superior to human experts or classic machine-learning algorithms.
In this study, we intend to apply the deep-learning technique to address the flight-state identification of the self-sensing wing. The signals collected from the PZT sensors embedded in the wing structure reflect the coupled aerodynamic–mechanical responses under varying flight states, with each state characterized by a specific angle of attack (AoA) and airspeed, and kept constant during the data collection. Therefore, these noise-corrupted structural responses can represent different flight states, which can be treated as a multiclassification problem. The objective of this paper is the introduction and evaluation of a novel multilayer network method for accurate flight-state identification based on vibration signals. Specifically, a one-dimensional (1D) CNN structure is developed for automatic feature learning. Taking advantage of the multichannel property of the CNN, reconstructed subsignals at different frequency bands via dual-tree complex-wavelet packet analysis are obtained as network inputs. A gray-wolf optimizer (GWO) [32] is then employed to determine the CNN key parameters. The proposed method does not only have strong feature-learning capabilities, but can also self-adapt to signal features for better classification performance without human intervention. Finally, we compare the flight-state identification accuracy with the standard CNN and other machine-learning models, and visualize the hierarchical feature-learning process through t-distributed stochastic neighbor embedding (t-SNE) [33]. The framework from the data acquisition, methodology development, evaluation, and application is shown in Figure 2.
The rest of paper is organized as follows: Section 2 presents the problem statement of this study. Section 3 develops the adaptive multichannel 1D CNN method using decomposed signals with parameter optimization by GWO. Two case studies with regard to general flight-state identification and stall detection and alerting are presented in Section 4, followed by their results and discussions in Section 5. Concluding marks are finally made.

2. Problem Statement

The problem statement of this work is as follows. Based on the sensing signals collected from a series of wind-tunnel experiments under varying flight states, we aimed to make the self-sensing wing accurately and automatically identify its undergoing flight state from the vibration time series. This problem can be further divided into three successive subproblems: (1) whether useful information can be learnt in an automatic manner, which are good at distinguishing different flight states; (2) how to establish the mapping relationship from the signal space to the physical flying state space; (3) how to improve the identification accuracy and robustness.
Accordingly, the first two problems are addressed via the development of a 1D CNN that is capable of autonomous feature learning in a layer-wise approach. The mapping relationship is established after a certain number of network operations, in which useful features are obtained as data representation of the original raw data to indicate different states. The third problem is tackled through two approaches. The first approach is to decompose the preprocessed signals into various signal segments using dual-tree complex-wavelet packet transformation (DTCWPT) and the reconstructed signals at different levels with different combinations are used as multichannel inputs. The second approach is to optimize the CNN hyperparameters by applying a GWO. The above developed method is evaluated for multiple flight-state identification and then applied to the specific case of stall detection and alerting.
The main novel aspects of this study include:
(1)
A tailored 1D deep CNN structure with multiple input channels using DTCWPT is developed for automatic feature learning instead of feature extraction and selection by human experts.
(2)
A self-adaptive CNN is proposed by combining the 1D CNN with a swarm-based GWO for automatic parameter determination instead of relying on human experience.
(3)
The flight-state identification of the self-sensing wing is treated as a classification problem by directly establishing the mapping relationship from the raw data to the physical space characterized by varying angle of attack and airspeed through wind tunnel experiments.
(4)
The application on stall detection and alerting with high identification accuracy provides new perspectives for autonomous flight control towards the “fly-by-feel” air vehicles.

3. Methodology Development

In this section, a self-adaptive 1D CNN method is proposed for flight-state identification of the self-sensing wing including three parts. Firstly, the basic theory of CNN is introduced. Then, a 1D deep CNN structure with DTCWPT is developed for autonomous feature extraction from the reconstructed subsignals. Lastly, a GWO is used to optimize CNN hyperparameters. The realization of the methods is programmed using the Python language (v 3.6.2) with tensorflow 1.8.

3.1. Basic CNN Theory

CNN is one of the active models in deep learning and has been widely applied in fields such as computer vision [34], speech recognition [35], and fault diagnosis [24]. Different from a traditional neural network with full connection throughout each layer, a CNN significantly reduces the network parameters by local connectivity and weights sharing using convolutional layers. This core building block consists of a set of kernels (or filters) which have a small receptive field. Each kernel moves across the input volume in a specified manner performing the convolution operation. Meanwhile, the kernel parameters remain the same to control the total number of free parameters. For a convolutional layer in the lth layer, the computation is expressed as
X k ( l ) = f ( c W k ( l ) , c X ( l 1 ) , c + B k l )
where k denotes the kernel number, c represents the channel number of the input X ( l 1 ) . W k ( l ) , c is the kth convolutional kernel corresponding to the cth channel, and B k l is the learnable bias corresponding to the kth kernel, f ( · ) is the activation function and is the elementwise multiplication [28].
Another import building block is the pooling layer, which is a form of nonlinear subsampling. A pooling layer is commonly inserted between successive convolutional layers for parameters reduction with the intuition that the rough location relative to other features are more important than the exact location. Common subsampling functions are max pooling by selecting the maximum value in each kernel and mean pooling by averaging each kernel. After a number of convolutional and pooling layers, fully connected layers are attached for high-level reasoning, which are identical to the layers in the classic neural networks.
The CNN training mechanism is based on backpropagation. Between densely connected layers, an error term δ of the lth layer is defined as
δ ( l ) = ( ( W ( l ) ) T δ ( l + 1 ) ) · f ( z ( l ) )
where W ( l ) is the parameter matrix of the lth layer, z ( l ) denotes the total weighted sum of inputs in layer l including the bias term, e.g., z ( l ) = W ( l 1 ) a ( l 1 ) + b ( l 1 ) , a ( l 1 ) = f ( z ( l 1 ) ) , which is the activation value in layer l-1. For output layer n l , δ ( n l ) = J z ( n l ) , where J stands for the cost function, f ( · ) is the derivative of the activation function.
For pooling layers, the error term of the lth layer is computed as
δ k ( l ) = u ( ( W k ( l ) ) T δ k ( l + 1 ) ) · f ( z k ( l ) )
where k is the kernel number as Equation (1) and W k l is the kth kernel, u ( · ) is the upsample operation which propagates the error through the pooling layer by calculating the error in regard of the incoming to the pooling layer.
Finally, for convolutional layers, the error term of lth layer is propagated through as
δ k ( l ) = c o n v ( δ k ( l + 1 ) , rot 180 ( W k l + 1 ) ) · f ( z k ( l ) )
where c o n v ( · ) is the convolution operation and rot 180 ( · ) denotes the 180 degrees rotation to make the convolution function perform cross-correction.

3.2. 1D CNN with DTCWPT

For signals with multimode aliasing, decomposition is necessary to achieve higher quality subsignals with different frequency bands, some of which may be better for classification. Compared with the conventional discrete wavelet packet transformation, DTCWPT provides a more precise frequency-band partition with approximate shift invariance [36]. It achieves signal decomposition and reconstruction through two parallel transforms of discrete wavelet, both of the which meet the reconstruction condition and each of the trees contains a set of low-pass and high-pass filters [37].
Herein, a sample signal is firstly decomposed and the subnodes at the same level or different levels are reconstructed to form the subsignals with the same length as the original signal. The reconstructed subsignals are then used as the CNN input instead of the original sample signal. To enable nonexperts’ intervention in 1D signal analysis for pattern recognition, a tailored deep CNN structure is developed as shown in Figure 3. Each vertical bar represents a network layer with different numbers of neurons corresponding to the exact data points. The multichannel inputs receive the reconstructed subsignals via DTCWPT. Features are extracted through several convolutional and maxpooling operation with increasing feature maps. A much shorter but deeper feature set is obtained and flatten to form a fully connected layer followed by a Softmax layer for multiclassification.
The configuration of the 1D deep CNN model used in this paper consists of an input layer, a convolutional layer C1, a pooling layer P1, a convolutional layer C2, a pooling layer P2, a convolutional layer C3, a pooling layer P3, a fully connected layer FC, and an output layer. For C1 to C3, the kernel numbers are 64, 256, and 128. The first kernel size is set to 10 while the other two sizes are denoted as s. For P1 to P3, max pooling is used, and the subregions are nonoverlapping with a size of 4. The learning rate is η , dropout is used in the last layer, and the dropout rate is λ . Relu is used as the activation function, and adam is used as the network training algorithm.

3.3. Parameter Optimization by GWO

In terms of the determination of the CNN key parameters, there is no mature method in theory. In this paper, we employ a swarm-based method called Grey Wolf Optimizer (GWO), which is one of the latest additions to the group of nature inspired optimization heuristics, to determine the optimum parameters of the trained 1D CNN. The GWO is inspired by the leadership hierarchy and hunting mechanism of grey wolves in nature and have demonstrated competitive results against some well-known evolutionary algorithms such as particle swarm optimization, genetic algorithm, and differential evolution [32]. The population of GWO is divided into four hierarchies. The first three fittest solutions are alpha ( α ) , beta ( β ) , and delta ( δ ) , which guide other wolves omega ( ω ) for hunting. The main hunting phase include: encircling, hunting, attacking, and searching.
In encircling, the wolves update their positions according to the prey as follows:
D = | C · X p ( t ) X ( t ) |
X ( t + 1 ) = X p ( t ) A · D
where t is the current iteration, X p is the position vector of the prey, while X denotes the position vector of a grey wolf. A and C are coefficient vectors, A = 2 a · r 1 a , C = 2 r 2 , in which a is linearly decreased from 2 to 0 over the course of iterations, and r 1 , r 2 are random vectors in [0,1].
During hunting, all wolves are obliged to update their positions according to first three best solutions obtained from encircling as follows:
X 1 = X α A 1 · D α
X 2 = X β A 2 · D β
X 3 = X δ A 3 · D δ
X ( t + 1 ) = X 1 + X 2 + X 3 3
where X α , X β , and X δ are the positions of alpha, beta and delta. D α , D β and D δ are calculated using Equation (5) with different coefficient C .
Attacking occurs when |A| < 1, otherwise, wolves diverge from each other for searching, which emphasizes further global exploration.
The optimization procedure of 1D deep CNN by GWO is illustrated in the following steps:
Step 1: Prepare the trained 1D deep CNN and set the complementary of the classification accuracy to be the fitness of GWO.
Step 2: Initialize the grey-wolf population X i = [ x 1 i , x 2 i , x n i ] , where X denotes the vector of parameters from CNN to be optimized.
Step 3: Initialize the GWO parameters a, A, and C.
Step 4: Train the 1D deep CNN with the initialized parameters corresponding to each agent.
Step 5: Evaluate the fitness of each agent and obtain the top three agents X α , X β , and X δ .
Step 6: Update the position of each search agent by Equation (10).
Step 7: Update the GWO parameters a, A, and C and return to Step 5 to update the top three agents X α , X β , and X δ until the maximum iteration is reached.
Step 8: Return X α as the optimized 1D deep CNN parameter vector.

4. Case Study

In this section, data collection and preparation from a series of wind-tunnel experiments are introduced. Then two cases are presented including a general flight-state identification with 16 classes for methodology demonstration and a novel application on stall detection and alerting with 12 classes.

4.1. Wind-Tunnel Experiment and Data Preparation

A series of wind-tunnel experiments of the self-sensing composite wing were conducted under various angles of attack (AoAs) and freestream velocities at Stanford University. The AoAs range from 0 degree up to 18 degrees with an incremental step of 1 degree. At each degree, data were collected for all velocities ranging from 9–22 m/s (incremental step of 1 m/s). For experimental details, please see Reference [2].
A series of wind-tunnel experiments of the self-sensing composite wing were conducted under various angles of attack (AoAs) and freestream velocities at Stanford University. The AoAs range from 0 degree up to 18 degrees with an incremental step of 1 degree. At each degree, data were collected for all velocities in the range of 9–22 m/s (incremental step of 1 m/s). For experimental details, please see Reference [2].
PZT signals reflect the coupled airflow-structural dynamics through the wing structural vibration and each time series contains coupled behavior with repeated patterns of a certain flight state. In each experiment, the structural vibration responses (90,000 data points) were recorded from the PZT located near the wing root at 1000 Hz sampling frequency. For each flight state, data are prepared in two steps: (1) the entire signal of 90,000 data points is divided into 180 segments (500 data points for each segment) to ensure enough samples for training while each segment has sufficient data points for feature extraction; (2) all signal samples are subtracted by their mean values eliminate the influence of zero drift.
To evaluate the effectiveness of the proposed method and apply it for dangerous state prewarning, two sets of data are collected for general flight-state identification and stall detection and alerting.

4.2. General Flight-State Identification

The first dataset includes PZT signals covering a relative coarse resolution of 16 flight states corresponding to a combination of four AoAs (3, 5, 7, 9 degrees) and four airspeeds (12, 14, 16, 18 m/s). An example of signal segments representing 16 flight states is shown in Figure 4. The vertical axes are the signal amplitudes and the horizontal axes are the signal lengths.
It can be observed from Figure 4 that the amplitude of the vibration signal tends to be larger with the increasing AoAs and the airspeeds. However, most of the adjacent flight states can hardly be distinguished without detailed investigation.

4.3. Stall Detection and Alerting

The second dataset covers a dense resolution of flight states (AoAs: 11, 12, 13 degrees, airspeeds: 10, 11, 12, 13 m/s) for critical-state alerting. In aerodynamics, stall phenomenon is one of the dangerous conditions wherein a sudden reduction of the lift coefficient occurs as the angle of attack increases beyond a critical point. According to a previous analysis [2], the signal energy can be used as an indicator of the lift loss of the self-sensing wing. From the wind-tunnel experiments, the mean values of the signal energy for a series of AoAs (0–17 degrees) under four airspeeds (10, 11, 12, 13 m/s) are obtained and shown in Figure 5.
Signal-energy variation with respect to the angle of attack is similar under four different airspeeds. It is noticed that, for all velocities, energy starts to increase at approximately 12 degrees. For 10 m/s, a stall happens at 13 degrees with a large increase of energy, while for the other three velocities, a stall occurs one step late at 14 degrees. To be conservative, we defined the orange-shaded area starting from 13 degrees as the Stall region, which should be avoided. The initial increase region around 12 degrees is defined as the Alert region, in which warnings should be provided to the flight control for angle reduction when the self-sensing wing flies in this region. The left side before 12 degrees is the Safe region as shaded in light green.

5. Results and Discussion

5.1. General Flight-State Identification

The proposed method (DTCWPT-CNN-GWO) is evaluated through the general flight-state identification.

5.1.1. Signal Decomposition and Reconstructed-Signal Selection

Firstly, a three-layer DTCWPT is employed to decompose the signals into various components with different frequency bands as shown in Figure 6. Reconstructions are then conducted for all three level signals without combination and obtained totally 14 subsignals from ‘a’ at the first level to ‘ddd’ at the bottom level as shown in Figure 6 and Figure 7. The vertical axes are the reconstructed signal amplitudes and the horizontal axes are the signal lengths.
The signal samples are randomly divided into 80% training set and 20% test set with equal classification category. The training set are further split into four subsets for cross-validation. The undetermined CNN parameters (the kernel size in the last two convolutional layers, the learning rate, and the dropout rate) [ s , η , λ ] are set as 10, 0.01, and 0.5. The objective of the primary classification experiment is to evaluate the classification ability of each reconstructed subsignals. The average identification accuracies using different subsignals as single input through the proposed CNN structure are shown in Figure 8.
It can be observed that the subsignals reconstructed from the first two decomposition levels have higher classification ability, among which the high-frequency band signal ‘d’ at the first decomposition level has the best performance. The top four subsignals are selected, ‘a’, ‘d’, ‘ad’, and ‘dd’, for the following analysis.
The second classification experiment is to make use of the multichannel mechanism of the CNN input layer and find the best combination from the four subsignals. The identification accuracies of various combinations are shown in Figure 9.
The last orange bar in Figure 9 indicates the classification accuracy using the original signal after preprocessing without DTCWPT. It is found that some subsignal combinations as multichannel inputs outperformance the original signal as single channel input, among which the three-channel combination ‘a/d/ad’ achieves the highest accuracy as 82%, improving more than 6% compared with the original signal.

5.1.2. Parameter Optimization and Identification Accuracy Comparison

The third experiment was to employ GWO to optimize three key parameters, kernel size s , learning rate η , and dropout rate λ . The classification error was set as the fitness function. The CNN parameter setting with and without GWO is shown in Table 1.
To examine the effectiveness of the proposed method (DTCWPT-CNN-GWO), three other methods were used for comparison, a normal 1D deep CNN, a Deep Neural Network (DNN), and a classic Back-Propagation Network (BPN).
The structure and parameter setting of the other three methods are as follows: The determination of all parameters is by experience and repeated experiments.
The structure of DNN is 500-250-250-100-16, ‘relu’ was used as the activation function, ‘adam’ was used as the training algorithm, the learning rate is 0.001, and the maximum iteration is 100.
The structure of BPN was 500-250-16, ‘relu’ was used as the activation function, ‘adam’ was used as the training algorithm, the learning rate was 0.001, and the maximum iteration was 200. The determination of all parameters is determined by experience and repeated experiments.
The training and testing were conducted for 10 times; the identification accuracies of the four methods are shown in Figure 10 and the related information is summarized in Table 2.
The results show that: (1) the proposed method has the highest identification accuracy among the four methods, especially surpassing the standard 1D CNN by approximately four points. This proves that the proposed method can adaptively deal with aerodynamic–mechanical coupled structural vibration and achieve robust classification results. (2) Both CNN methods have much higher accuracies than the DNN and BPN. It can be explained that the local connection mechanism in CNN significantly improves the feature-learning performance compared to the conventional full connection networks, which have very limited classification ability using preprocessed signal data without feature extraction by manpower.

5.1.3. Visualization of the Learning Process

To examine the feature-learning effects in the hierarchical CNN structure of the proposed method, 3D visualization is conducted through dimension reduction using a relatively new method named t-Distributed Stochastic Neighbor Embedding (t-SNE). It is a manifold learning technique by mapping to probability distributions through affine transformation, which is particularly suitable for nonlinear and high-dimensional datasets [33]. The learning results for five selected layers of the CNN are visualized sequentially in Figure 11. Sixteen colors represent 16 flight states, and a single point in each 3D plot indicates a sample.
It can be observed that the 16 flight states in the input layer are in a chaos. After the first three layers, increasing flight states are gradually distinguished but it is still difficult for all-states identification. From the forth layer, most of the flight states tend to be separated and the best classification effect is obtained by the last output layer. The results indicate a layer-wise improvement in classification performance by transforming the low-level signal data into high-level features.

5.2. Stall Detection and Alerting

The second case study applies the proposed method for pre-warning of one of the typical dangerous conditions in aerodynamics. Similarly, the parameters in CNN are optimized by GWO as shown in Table 3.
Ten trials of each identification method (the proposed method, 1D CNN, DNN, BPN) are conducted and the comparison between the four methods is shown in Figure 12 while the mean values and the standard deviations are summarized in Table 4.
It can be seen that the proposed method has the best performance in the classification of 12 dense-resolution flight states with fewer network parameters. To further examine the stall detection and alerting effects, a classification report of one the ten trials is obtained in Table 5 with three criteria: Precision, Recall, and F1 score. Precision is the ratio of correctly predicted positive observations to the total predicted positive observations, while Recall is the ratio of correctly predicted positive observations to the all observations in the actual class. F1 Score is the weighted average of Precision and Recall: F1 Score = 2 * (Recall * Precision)/(Recall + Precision) [38]. Safe, Alert, and Stall regions are divided with corresponding flight states. The overall identification accuracy (recall) is 93%.
To facilitate detailed analysis, a normalized confusion matrix is presented in Figure 13. Each row of the matrix represents the test samples in a true class label while each column indicates the samples in a predicted class label [39].
Herein, the recall value and precision value can be interpreted as the miss-alarm ratio and false alarm ratio. As shown in Figure 11, the lowest classification accuracy occurs at State 8 (Alert region), in which 19% samples belonging to State 8 are wrongly predicted as State 4 (Safe region) as can be found in the eighth row of the confusion matrix. This is denoted as the miss-alarm ratio. From another perspective, 11% samples of the predicted samples actually belong to State 4 (Safe region) as can be found in the eighth column of the confusion matrix, meaning that the wing flying in Safe region yet receives a false alarm. This partial precision value is regarded as false alarm ratio. Overall, the confusion matrix provides a detailed insight of the identification accuracy on each flight state for further improvement actions.

6. Conclusions

In this paper, a novel adaptive CNN is proposed for flight-state identification from structural vibration signals obtained from a self-sensing composite wing through wind tunnel experiments. A 1D deep CNN structure is established to facilitate the automatic feature extraction from the lengthy vibration signal. To improve the identification accuracy, the subsignals are decomposed and reconstructed via the DTCWPT, which is able to obtain the most differentiated subsignal components and take advantage of the multichannel mechanism of the CNN input. To avoid subjective human experience in the key parameter determination, a GWO algorithm is further attached to the CNN, which enables the self-adaptive parameter setting to extract more informative features.
The proposed method is evaluated through the general flight state identification and applied in dangerous flight states detection. Results from both cases demonstrate the superior performance of the proposed method over the conventional CNN method and general deep-learning network without modification. In all, the proposed method has the advantage of saving considerable efforts in feature extraction and parameter setting of the deep neural network by human experts and is promised to be used in broader applications for intelligent structural vibration analysis.

Author Contributions

X.C. analyzed the data and developed the 1D deep CNN method; F.K. and F.-K.C. designed the self-sensing wing and performed the wind-tunnel experiments; Q.W. provided the parameter optimization algorithm; H.R. and F.-K.C. co-ordinated the research and revised the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No. 51705242), Shanghai Sailing Program (Grant No. 16YF1404900), and the U.S. Air Force Office of Scientific Research (AFOSR) program ‘‘Avian-Inspired Multi-functional Morphing Vehicles” under grant FA9550-16-1-0087 with Program Manager Byung-Lip (‘‘Les”) Lee.

Acknowledgments

The authors would like to thank Pengchuan Wang, Ravi Gondaliya, Jun Wu, and Shaobo Liu for their help during the wind-tunnel experiments. Finally, the authors would like to acknowledge the support of Lester Su and John Eaton in the wind-tunnel facility at Stanford University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. NASA. Fly-By-Feel Systems Represent The Next Revolution In Aircraft Controls; NASA: Washington, DC, USA, 2011.
  2. Kopsaftopoulos, F.; Nardari, R.; Li, Y.H.; Chang, F.K. A stochastic global identification framework for aerospace structures operating under varying flight states. Mech. Syst. Signal Process. 2018, 98, 425–447. [Google Scholar] [CrossRef]
  3. Kopsaftopoulos, F.; Nardari, R.; Li, Y.H.; Wang, P.; Ye, B.; Chang, F.K. Experimental identification of structural dynamics and aeroelastic properties of a self-sensing smart composite wing. In Proceedings of the 10th International Workshop on Structural Health Monitoring, Stanford, CA, USA, 1–3 September 2015. [Google Scholar]
  4. Brenner, M.J. Controller Performance Evaluation of Fluy-by-feel (FBF) Technology; NASA: Washington, DC, USA, 2012.
  5. Mangalam, A.S.; Brenner, M.J. Fly-by-Feel Sensing and Control: Aeroservoelasticity. In Proceedings of the AIAA Atmospheric Flight Mechanics Conference, National Harbor, MD, USA, 13–17 January 2014. [Google Scholar]
  6. Suh, P.; Chin, A.; Mavris, D. Virtual Deformation Control of the X-56A Model with Simulated Fiber Optic Sensors. In Proceedings of the AIAA Atmospheric Flight Mechanics Conference, Boston, MA, USA, 19–22 August 2013. [Google Scholar]
  7. Suh, P.M.; Chin, A.; Mavris, D.N. Robust Modal Filtering and Control of the X-56A Model with Simulated Fiber Optic Sensor Failures. In Proceedings of the AIAA Atmospheric Flight Mechanics Conference, National Harbor, MD, USA, 13–17 January 2014. [Google Scholar]
  8. Lanzara, G.; Feng, J.; Chang, F.-K. Design of Micro-Scaled Highly Expandable Networks of Polymer Based Substrates for Macro-Scale Applications. Smart Mater. Struct. 2010, 19, 45013. [Google Scholar] [CrossRef]
  9. Salowitz, N.; Guo, Z.; Li, Y.H.; Kim, K.; Lanzara, G.; Chang, F.K. Bio-inspired stretchable network-based intelligent composites. J. Compos. Mater. 2013, 47, 97–105. [Google Scholar] [CrossRef]
  10. Salowitz, N.; Guo, Z.; Roy, S.; Nardari, R.; Li, Y.H.; Kim, S.J.; Kopsaftopoulos, F.; Chang, F.K. Recent advancements and vision toward stretchable bio-inspired networks for intelligent structures. Struct. Heal. Monit. 2014, 13, 609–620. [Google Scholar] [CrossRef]
  11. Guo, Z.; Aboudi, U.; Peumans, P.; Howe, R.T.; Chang, F.K. A Super Stretchable Organic Thin-Film Diodes Network That Can Be Embedded into Carbon Fiber Composite Materials for Sensor Network Applications. J. Microelectromec. Syst. 2016, 25, 524–532. [Google Scholar] [CrossRef]
  12. Roy, S.; Lonkar, K.; Janapati, V.; Chang, F.-K. A novel physics-based temperature compensation model for structural health monitoring using ultrasonic guided waves. Struct. Heal. Monit. Int. J. 2014, 13, 321–342. [Google Scholar] [CrossRef]
  13. Kopsaftopoulos, F.P.; Fassois, S.D. Vector-dependent functionally pooled ARX models for the identification of systems under multiple operating conditions. IFAC Proc. Vol. 2012, 16, 310–315. [Google Scholar] [CrossRef]
  14. Kopsaftopoulos, F.; Nardari, R.; Li, Y.-H.; Wang, P.; Chang, F.-K. Stochastic global identification of a bio-inspired self-sensing composite UAV wing via wind tunnel experiments. Health Monit. Struct. Biol. Syst. 2016, 9805. [Google Scholar] [CrossRef]
  15. Guyon, I. Feature Extraction Foundations and Applications; Springer: Berlin, Germany, 2006; Volume 207, ISBN 9783540354871. [Google Scholar]
  16. Samanta, B. Gear fault detection using artificial neural networks and support vector machines with genetic algorithms. Mech. Syst. Signal Process. 2004, 18, 625–644. [Google Scholar] [CrossRef]
  17. Shen, Z.; Chen, X.; Zhang, X.; He, Z. A novel intelligent gear fault diagnosis model based on EMD and multi-class TSVM. Measurement 2012, 45, 30–40. [Google Scholar] [CrossRef]
  18. Xi, Y.L.; Xi, Z.H.; Xi, Y.Z. Fault Diagnosis of Rotating Machinery Based on Multiple ANFIS Combination with Gas Fault diagnosis of rotating machinery based on multiple ANFIS combination with GAs. Mech. Syst. Signal Process. 2007, 21, 2280–2294. [Google Scholar] [CrossRef]
  19. Chen, X.; Kopsaftopoulos, F.; Wu, Q.; Ren, H.; Chang, F.K. Flight state identification of a self-sensingwing via an improved feature selection method and machine learning approaches. Sensors 2018, 18, 1379. [Google Scholar] [CrossRef]
  20. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  21. Hinton, G.E.; Osindero, S.; Teh, Y.W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  22. Gehring, J.; Miao, Y.; Metze, F.; Waibel, A. Extracting deep bottleneck features using stacked auto-encoders. In Proceedings of the ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013. [Google Scholar]
  23. Yu, D.; Deng, L. Deep Learning and Its Applications to Signal and Information Processing. IEEE Signal Process. Mag. 2011, 28, 145–150. [Google Scholar] [CrossRef]
  24. Janssens, O.; Slavkovikj, V.; Vervisch, B.; Stockman, K.; Loccufier, M.; Verstockt, S.; Van de Walle, R.; Van Hoecke, S. Convolutional Neural Network Based Fault Detection for Rotating Machinery. J. Sound Vib. 2016, 377, 331–345. [Google Scholar] [CrossRef]
  25. Graves, A.; Mohamed, A.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013. [Google Scholar]
  26. Madić, M.; Marković, D.; Radovanović, M. Comparison of meta-heuristic algorithms for solving machining optimization problems. Mech. Eng. 2013, 11, 29–44. [Google Scholar]
  27. Shao, H.; Jiang, H.; Zhao, H.; Wang, F. A novel deep autoencoder feature learning method for rotating machinery fault diagnosis. Mech. Syst. Signal Process. 2017, 95, 187–204. [Google Scholar] [CrossRef]
  28. Fuan, W.; Hongkai, J.; Haidong, S.; Wenjing, D.; Shuaipeng, W. An adaptive deep convolutional neural network for rolling bearing fault diagnosis. Meas. Sci. Technol. 2017, 28. [Google Scholar] [CrossRef]
  29. Shao, H.; Jiang, H.; Zhang, X.; Niu, M. Rolling bearing fault diagnosis using an optimization deep belief network. Meas. Sci. Technol. 2015, 26. [Google Scholar] [CrossRef]
  30. Rasdi Rere, L.M.; Fanany, M.I.; Arymurthy, A.M. Metaheuristic Algorithms for Convolution Neural Network. Comput. Intell. Neurosci. 2016, 2016. [Google Scholar] [CrossRef]
  31. Nawi, N.M.; Khan, A.; Rehman, M.Z.; Herawan, T.; Deris, M.M. Comparing Performances of Cuckoo Search Based Neural Networks; Springer: Cham, Switzerland, 2014; Volume 287, ISBN 9783319076911. [Google Scholar]
  32. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  33. Van Der Maaten, L.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 620, 267–284. [Google Scholar] [CrossRef]
  34. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 1–9. [Google Scholar] [CrossRef]
  35. Abdel-Hamid, O.; Mohamed, A.; Jiang, H.; Deng, L.; Penn, G.; Yu, D. Convolutional Neural Networks for Speech Recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 1533–1545. [Google Scholar] [CrossRef]
  36. Bayram, I.; Selesnick, I.W. On the dual-tree complex wavelet packet and M-band transforms. IEEE Trans. Signal Process. 2008, 56, 2298–2310. [Google Scholar] [CrossRef]
  37. Wang, Y.; He, Z.; Zi, Y. Enhancement of signal denoising and multiple fault signatures detecting in rotating machinery using dual-tree complex wavelet transform. Mech. Syst. Signal Process. 2010, 24, 119–137. [Google Scholar] [CrossRef]
  38. Davis, J.; Goadrich, M. The relationship between Precision-Recall and ROC curves. In Proceedings of the 23rd International Conference on Machine Learning—ICML’06, Pittsburgh, PA, USA, 25–29 June 2006. [Google Scholar]
  39. Powers, D.M.W. Evaluation: From Precision, Recall and F-Measure To Roc, Informedness, Markedness and Correlation. J. Mach. Learn. Technol. 2011, 2, 37–63. [Google Scholar]
Figure 1. Self-sensing composite wing design [2].
Figure 1. Self-sensing composite wing design [2].
Sensors 19 00275 g001
Figure 2. Framework of the proposed methodology.
Figure 2. Framework of the proposed methodology.
Sensors 19 00275 g002
Figure 3. 1D deep convolutional neural network (CNN) structure.
Figure 3. 1D deep convolutional neural network (CNN) structure.
Sensors 19 00275 g003
Figure 4. Piezoelectric lead–zirconate titanate (PZT) signal segments under 16 flight states after data preprocessing.
Figure 4. Piezoelectric lead–zirconate titanate (PZT) signal segments under 16 flight states after data preprocessing.
Sensors 19 00275 g004
Figure 5. Signal energy under various flight states.
Figure 5. Signal energy under various flight states.
Sensors 19 00275 g005
Figure 6. Three-level decomposition structure.
Figure 6. Three-level decomposition structure.
Sensors 19 00275 g006
Figure 7. Reconstructed subsignals at different decomposition levels.
Figure 7. Reconstructed subsignals at different decomposition levels.
Sensors 19 00275 g007
Figure 8. Identification accuracy under various reconstructed subsignals.
Figure 8. Identification accuracy under various reconstructed subsignals.
Sensors 19 00275 g008
Figure 9. Identification accuracy under various reconstructed subsignal combinations.
Figure 9. Identification accuracy under various reconstructed subsignal combinations.
Sensors 19 00275 g009
Figure 10. Identification accuracy of four methods.
Figure 10. Identification accuracy of four methods.
Sensors 19 00275 g010
Figure 11. 3D visualization of the hierarchical learning process.
Figure 11. 3D visualization of the hierarchical learning process.
Sensors 19 00275 g011
Figure 12. Identification accuracy of four methods.
Figure 12. Identification accuracy of four methods.
Sensors 19 00275 g012
Figure 13. Confusion matrix of flight-state identification.
Figure 13. Confusion matrix of flight-state identification.
Sensors 19 00275 g013
Table 1. Features in time domain.
Table 1. Features in time domain.
Parameter DescriptionValue after GWOValue Before (Used in Normal CNN in the Following Comparison)
Population of GWO20-
Iteration number20-
Dimensionality of particles3-
Kernel size in C2 and C3 layers510
Learning rate 0.00120.001
Dropout rate0.40.5
Table 2. Features in frequency domain.
Table 2. Features in frequency domain.
MethodsInput DimensionSize of Training/Testing SampleAverage Testing AccuracyStandard DeviationTotal Parameters
Proposed method5002304/57685.15%2.07%260,432
1D CNN5002304/57677.53%2.20%500,816
DNN5002304/57615.45%1.22%214,100
BPN5002304/5761.37%0.27%129,000
Table 3. Parameters used with and without GWO.
Table 3. Parameters used with and without GWO.
Parameter DescriptionValue after GWOValue before (Used in Normal CNN in the Following Comparison)
Population of GWO20-
Iteration number20-
Dimensionality of particles3-
Kernel size in C2 and C3 layers310
Learning rate0.00110.001
Dropout rate0.40.5
Table 4. Average testing accuracy and the parameter numbers of four methods.
Table 4. Average testing accuracy and the parameter numbers of four methods.
MethodsInput DimensionSize of Training/Testing SampleAverage Testing AccuracyStandard DeviationTotal Parameters
Proposed method5001728/43292.43%1.48%160,588
1D CNN5001728/43277.11%2.27%498,764
DNN5001728/43226.41%1.24%213,700
BPN5001728/4327.82%0.94%128,000
Table 5. Classification report.
Table 5. Classification report.
States IDAoA degSpeed m/sPrecisionRecallF1 Score
Safe111100.970.940.96
211110.900.970.93
311120.940.940.94
411130.820.890.85
Alert512101.000.970.99
612110.950.970.96
712121.000.920.96
812130.780.810.79
Stall913100.940.940.94
1013110.940.920.93
1113120.970.970.97
1213131.000.940.97

Share and Cite

MDPI and ACS Style

Chen, X.; Kopsaftopoulos, F.; Wu, Q.; Ren, H.; Chang, F.-K. A Self-Adaptive 1D Convolutional Neural Network for Flight-State Identification. Sensors 2019, 19, 275. https://doi.org/10.3390/s19020275

AMA Style

Chen X, Kopsaftopoulos F, Wu Q, Ren H, Chang F-K. A Self-Adaptive 1D Convolutional Neural Network for Flight-State Identification. Sensors. 2019; 19(2):275. https://doi.org/10.3390/s19020275

Chicago/Turabian Style

Chen, Xi, Fotis Kopsaftopoulos, Qi Wu, He Ren, and Fu-Kuo Chang. 2019. "A Self-Adaptive 1D Convolutional Neural Network for Flight-State Identification" Sensors 19, no. 2: 275. https://doi.org/10.3390/s19020275

APA Style

Chen, X., Kopsaftopoulos, F., Wu, Q., Ren, H., & Chang, F. -K. (2019). A Self-Adaptive 1D Convolutional Neural Network for Flight-State Identification. Sensors, 19(2), 275. https://doi.org/10.3390/s19020275

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop