Next Article in Journal
Multiscale Analysis of Runoff Complexity in the Yanhe Watershed
Next Article in Special Issue
Combine Harvester Bearing Fault-Diagnosis Method Based on SDAE-RCmvMSE
Previous Article in Journal
Towards Generalizing the Information Theory for Neural Communication
Previous Article in Special Issue
Exposing Deep Representations to a Recurrent Expansion with Multiple Repeats for Fuel Cells Time Series Prognosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Deep Transfer Learning Method for Intelligent Fault Diagnosis Based on Variational Mode Decomposition and Efficient Channel Attention

1
School of Electronic Information, Hangzhou Dianzi University, Hangzhou 310018, China
2
Zhejiang Provincial Key Lab of Equipment Electronics, Hangzhou 310018, China
3
Tianneng Battery Group Co., Ltd, Changxing 313100, China
*
Author to whom correspondence should be addressed.
Submission received: 4 July 2022 / Revised: 2 August 2022 / Accepted: 5 August 2022 / Published: 6 August 2022

Abstract

:
In recent years, deep learning has been applied to intelligent fault diagnosis and has achieved great success. However, the fault diagnosis method of deep learning assumes that the training dataset and the test dataset are obtained under the same operating conditions. This condition can hardly be met in real application scenarios. Additionally, signal preprocessing technology also has an important influence on intelligent fault diagnosis. How to effectively relate signal preprocessing to a transfer diagnostic model is a challenge. To solve the above problems, we propose a novel deep transfer learning method for intelligent fault diagnosis based on Variational Mode Decomposition (VMD) and Efficient Channel Attention (ECA). In the proposed method, the VMD adaptively matches the optimal center frequency and finite bandwidth of each mode to achieve effective separation of signals. To fuse the mode features more effectively after VMD decomposition, ECA is used to learn channel attention. The experimental results show that the proposed signal preprocessing and feature fusion module can increase the accuracy and generality of the transfer diagnostic model. Moreover, we comprehensively analyze and compare our method with state-of-the-art methods at different noise levels, and the results show that our proposed method has better robustness and generalization performance.

1. Introduction

Mechanical equipment is one of the most critical components for the normal operation and production of intelligent factories. Due to the long-term use of the equipment and some human factors, the equipment may malfunction, causing major economic losses and personnel injuries. Therefore, fault location and maintenance of mechanical equipment are particularly significant. In practical industrial applications, most equipment fault location and maintenance rely on previous experience or expert knowledge, which requires a lot of manpower and time costs. To solve this problem, intelligent fault diagnosis combines artificial intelligence technology with fault diagnosis technology, which has become an important branch of intelligent health management. Traditional machine learning methods have been widely used in the field of equipment fault diagnosis, such as support vector machine (SVM) [1], random forest (RF) [2], k-nearest neighbor (KNN) [3], and artificial neural network (ANN) [4]. However, these machine learning methods for fault diagnosis rely heavily on industry experts to extract artificial features from raw signals [5]. Moreover, due to the explosive growth of available manufacturing data and the increasing complexity of equipment, traditional machine learning can no longer meet the current requirements for intelligent fault diagnosis.
The application of deep learning has developed rapidly in recent years [6,7]. To solve the above problems, deep learning has been widely used in the field of intelligent fault diagnosis and has achieved fruitful results [8]. Janssens et al. [9] used a convolutional neural network (CNN) to automatically extract the fault features of the original signals, avoiding the trouble of manual feature extraction. Zhao et al. [10] embedded a soft threshold denoising module into a deep residual network. The deep learning diagnostic model has improved feature learning capability for strong noise vibration signals. Nevertheless, the improvement of the soft thresholding function embedded in the network needs further discussion. Lei et al. [11] constructed the long-term dependence of time series captured by long- and short-term memory networks. It is applied to wind turbine condition monitoring and fault diagnosis and has achieved great results. Although the deep learning diagnosis method has a strong feature extraction ability, the model diagnosis performance will be greatly reduced when the training data and test data do not meet the independent and identical distribution conditions. This problem is known as the domain shift problem.
Transfer learning [12,13] improves the diagnostic performance of the diagnostic model in the target domain by learning relevant knowledge from similar scenarios. It provides a new way to solve the domain shift problem. When the data distribution of the source domain and the target domain is different, but the learning task is the same, the transfer learning in this scenario is called domain adaptation. Deep domain adaptation [14,15] combines the powerful feature extraction capability of deep learning. It can effectively learn the domain-invariant representation features of data and enhance the performance of diagnostic models in the target domain. Lei et al. [16] proposed a deep convolutional transfer learning network to explore intelligent fault diagnoses for machines with unlabeled data. The maximum mean difference (MMD) [17] domain discriminator was used to optimize the model and improve accuracy. Jiao et al. [18] proposed a residual joint domain adaptive adversarial network model. Joint maximum mean difference (JMMD) and adversarial domain adaptive discriminator are introduced to align the source and target domain feature spaces. However, the performance of these methods in noisy scenarios needs to be verified. Based on this problem, signal preprocessing in the feedforward network is one of the options for denoising. Qian et al. [19] proposed an unsupervised transfer learning model for a convolutional autoencoder. The Autoencoder is used to denoise the signal. However, the output of the autoencoder is lossy. The removal of noise will also remove some useful weak signals. Li et al. [20] thought from the perspective of a frequency domain. A frequency-domain fusion convolutional neural network for domain adaptive fault diagnosis is proposed. It provides an example adaptation network design for the unified domain but extracting frequency characteristics directly from the CNN still cannot be fully interpreted. Ben Ali et al. [21] proposed an automatic bearing fault diagnosis method based on empirical mode decomposition (EMD) to successfully detect the severity of defects online. However, determining the intrinsic mode functions (IMFs) best suited for bearing fault diagnosis requires expert experience. DASENet achieves high accuracy with few input data points [22]. The inputs to this model are the frequency domain signal and the time-frequency graph signal. In addition, it fuses features through the squeeze-and-excitation attention (SEA) module, which improves the generality of the diagnostic model. Nevertheless, this method incorporates the features of duplication and is not a transfer diagnostic model. Wu et al. [23] introduced ensemble empirical mode decomposition (EEMD) signal preprocessing and attention mechanisms into the transfer diagnosis model to extract domain-invariant features. In their study, the signal preprocessing part is just a function of denoising. In addition, the diagnosis results are effective, indicating that signal preprocessing is necessary. Although the above methods integrate signal preprocessing technology into deep learning or transfer learning methods to improve the performance of diagnostic models, there are still several open questions, and the following situations should be considered.
  • Due to differences in the distribution of acquired vibration data, the effectiveness of the deep learning model diagnosis method is limited. Therefore, it is necessary to introduce signal preprocessing into the transfer learning diagnostic model.
  • The original vibration signal is a typical non-stationary signal [24]. However, the commonly used Fourier transform method has insufficient processing ability for non-stationary signals (signals whose frequencies change with time). Although the EMD method eliminates the limitation of Fourier transform, it also has a serious mode aliasing problem.
  • The signal preprocessing method and the intelligent diagnosis model based on transfer learning are not effectively related. An effective connection module is required between them instead of being directly preprocessed into new data for input.
To bridge these gaps, this paper proposes a novel deep transfer learning method for intelligent fault diagnosis based on Variational Mode Decomposition and Efficient Channel Attention. The corresponding framework of this method is divided into three parts: signal preprocessing using VMD, mode feature fusion using ECA, and a deep transfer network for intelligent fault diagnosis. Since different modes represent signals with different center frequencies, it makes sense to re-weight the channel features. Moreover, joint domain adaptation is used to align the joint distribution of input features and output labels in the source domain and the target domain to achieve an efficient intelligent fault diagnosis. We comprehensively analyze and compare our method with state-of-the-art methods, and the results show that our proposed method has better robustness and generalization performance. The novelty and main contributions of this paper can be summarized as follows:
  • The signal preprocessing of the VMD is applied for the transfer fault diagnosis. The VMD algorithm is used to decompose original fault vibration signals into a series of intrinsic mode functions (IMFs) with specific bandwidth so that the transfer learning model can learn fault features better.
  • To fuse the mode features after VMD decomposition more effectively, ECA is used to learn channel attention. This module avoids dimensionality reduction and effectively captures cross-channel interactions. Since different modes represent signals with different center frequencies, it makes sense to re-weight the channel features.
  • A novel deep transfer learning method is proposed for intelligent fault diagnosis based on Variational Mode Decomposition and Efficient Channel Attention (VMD-ECA-DTN). Considering that it can accurately fit the domain-invariant representation of fault features, our proposed method has better performance (robustness and generalization) in transfer diagnosis tasks compared with the state-of-the-art methods.
The remainder of this paper begins with the related preliminaries in Section 2. Section 3 introduces the construction of the VMD-ECA-DTN and its optimization objective. In Section 4, the experiments and analyses of the study are given in Section 4. Finally, the conclusions are drawn in Section 5.

2. Preliminaries

2.1. Domain Adaptation Problem

Domain adaptation refers to transfer learning in a scenario where the source domain and the target domain have different data distribution and the same learning task. The source domain is defined as 𝒟s = {( x 1 s , y 1 s ), …, ( x n s s , y n s s )}, and the target domain is defined as 𝒟t = {( x 1 t , y 1 t ), …, ( x n t t , y n t t )}, Where ns is the sample number in the source domain, nt is the sample number in the target domain. The target domain under unsupervised domain adaptation has no label, and the target domain is de fined as 𝒟t = {( x 1 t ), …, ( x n t t )}. In this paper, the problem of unsupervised domain adaptation fault diagnosis is mainly studied. As shown in Figure 1, the main goal of domain adaptation is to align the feature space of the source. Domain and the target domain to realize cross-domain intelligent fault diagnosis.

2.2. Variational Mode Decomposition

The fault vibration signal of equipment is a typical non-stationary signal. To better learn the fault features of the original signal, we use the VMD method to decompose fault vibration signals during signal preprocessing. The VMD algorithm is a completely non-recursive signal decomposition method that decomposes time series to obtain relatively stable sub-signals containing multiple different frequency scales [25,26,27]. This algorithm assumes that all components are narrowband signals concentrated near their respective central frequencies. Therefore, VMD establishes constrained optimization problems according to component narrowband conditions. Then, the center frequency of the signal component is estimated, and the corresponding component is reconstructed. The constrained variational problem is expressed as follows:
min ( u k , w k ) { k = 1 K t [ ( δ ( t ) + j π t ) u k ( t ) ] e j w k t 2 2 } s . t . x ( t ) = k = 1 K u k
where K indicates the number of modes. {uk}: = {u1, …, uK} represents the set of all modes. {wk}: = {w1, …, wK} represents the set of their center frequencies. δ(t) is the Dirac distribution, and (*) defines the convolution operation. x(t) represents the time series of the fault vibration signal. To solve the above mathematical problem, the equation can be equivalent to an unconstrained optimization problem by means of augmented Lagrange functions, as shown in:
L ( u k , w k , λ ) = α k = 1 K t [ ( δ ( t ) + j π t ) u k ( t ) ] e j w k t 2 2 + x ( t ) k = 1 K u k ( t ) 2 2 + λ ( t ) , x ( t ) k = 1 K u k ( t )
where α is the quadratic penalty parameter, and λ is Lagrangian multipliers.
The minimization scheme of the original problem is solved by the alternate direction method of multipliers (ADMM), which keeps the other two variables constant and updates one of them. Therefore, the update formulas for uk and wk are expressed as follows:
u ^ k n + 1 = x ^ ( w ) i > k u ^ i ( w ) i < k u ^ i ( w ) + λ ^ ( w ) 2 1 + 2 α ( w w k ) 2
w k n + 1 = 0 w | u ^ k n + 1 ( w ) | 2 d w 0 | u ^ k n + 1 ( w ) | 2 d w
where n is the number of iterations. (^) denotes the Fourier transform. w ≥ 0. The result of solving uk is a Wiener filter, and the result of solving wk is a barycentric method of solving signal bandwidth. Given precision ε, iteration stops when the following convergence criterion is reached:
k = 1 K u ^ k n + 1 u ^ k n 2 2 / u ^ k n 2 2 < ε
Finally, all modes of signal decomposition are obtained through the above iterative updating formula. The VMD algorithm is used to decompose original fault vibration signals into a series of intrinsic mode functions (IMFs) with specific bandwidth so that the transfer learning model can learn fault features better.

2.3. Efficient Channel Attention

The channel attention mechanism has proven to have great potential for improving the performance of deep convolutional neural networks (CNN). Efficient Channel Attention (ECA) [28] is improved based on Squeeze-and-Excitation Attention (SEA) [29]. First, the ECA module aggregates convolution features using global average pooling (GAP) without dimensionality reduction. Second, kernel size k is determined adaptively. Then, a one-dimensional convolution operation is performed, and finally Sigmoid function was performed to learn channel attention. The diagram of the ECA is shown in Figure 2:
Let the input be X, and the output X′ of the ECA is expressed as:
X = σ ( C 1 D k ( g ( X ) ) ) X
where C1D indicates one-dimensional convolution, and k is the kernel size. g(X) represents the global average pooling function.
The ECA module uses the 1D convolution layer directly after the global average pooling layer, removing the full connection layer. This module avoids dimensionality reduction and effectively captures cross-channel interactions. In addition, ECA involves only a few parameters to achieve good results. Different from general networks, ECA’s main role in the study is to fuse the features of different modes after VMD decomposition.

3. The Proposed VMD-ECA-DTN Method

In this section, an advanced and novel deep transfer learning method for intelligent fault diagnosis based on Variational Mode Decomposition and Efficient Channel Attention is proposed. As shown in Figure 3, the corresponding framework of this method is divided into three parts: signal preprocessing using VMD, mode feature fusion using ECA, and a deep transfer network for intelligent fault diagnosis. Signal preprocessing is a significant step in fault analysis. The VMD method adaptively matches the optimal center frequency and finite bandwidth of each mode. The effective separation of IMFs and the frequency domain division of signals are realized. Finally, the effective decomposition components of a given signal are obtained. To fuse the mode features more effectively after VMD decomposition, ECA is used to learn channel attention. Since different modes represent signals with different center frequencies, it makes sense to re-weight the channel features. The last deep transfer network includes a convolution feature extractor, a fault classifier for prediction, and distribution discrepancy metrics with joint domain adaptation to reduce the difference in the joint distribution of input features and output tags between the source domain and the target domain. The goal of the proposed approach is to achieve a better cross-domain device intelligent fault diagnosis by enhancing mode features, which are more favorable for the diagnosis task. It should be noted that the source of training data has two parts: one is the source domain data with labels, and the other is the target domain data without labels. The processing operations of these two parts of data in the signal preprocessing and mode feature fusion stages are similar. As detailed in the following subsections.

3.1. Architecture of the Proposed VMD-ECA-DTN

In this subsection, the specific architecture of the proposed method will be explained in detail. Inspired by a deep transfer network with joint distribution adaptation, the proposed deep transfer network architecture is designed based on VMD and ECA. The detailed architecture of the proposed method is shown in Figure 4. In a general way, the nonlinear correction unit function Parametric Rectified Linear Unit (PRELU) [30] is adopted for the nonlinear activation layer. Given a signal length of 1024 for a single sample, the input shape of the original vibration signal is (64, 1, 1024). 64 indicates the batch size, and 1 indicates the channel size. Set the number of modes of the VMD algorithm as 5, because the length of the mode component obtained after VMD decomposition is consistent with the original signal, the shape of the IMF signal is (64, 5, 1024). To fuse mode features more effectively, the ECA is used to learn channel attention. The ECA block consists of an adaptive global average pooling layer (Ada-Avg-Pool), a convolution layer (Conv), and a sigmoid layer. In addition, there is a weight multiplier at the end. Notably, it does not reduce dimension and effectively captures information about cross-channel interactions. The shape of the signal does not change as it passes through the ECA module, which simply re-weights the mode’s features. Next is the CNN feature extractor, which consists of four convolution layers containing batch normalization (BN). The convolution kernel size of these convolution layers is 15, 3, 3, and 3, respectively. In particular, the reason for using large convolution kernels in the first layer is to increase the receptive field. Inspired by Alexnet, max pooling (Max-Pool) is adopted for the second convolution layer, and adaptive global max pooling (Ada-Max-Pool) is adopted for the fourth convolution layer. To prevent overfitting, we use the dropout trick in the later fully connected layers (FC) as an alternative form of regularization. Set the dropout ratio to p, where each output node is set to zero with probability p. In addition, the architecture of the fault classifier and the distribution discrepancy metrics is similar. The specific parameters of the proposed network architecture are shown in Table 1.

3.2. Optimization Objective of VMD-ECA-DTN

Given the mode’s signal set X = {u1, …, uK} after VMD decomposition, the output X′ = σ(C1Dk(g(X)))X of ECA is obtained by modal feature fusion based on the attentional mechanism module. The size of the feature map does not change after the ECA module, but the importance of each channel feature has changed. Let xl be the input of the lth convolution layer, and then the feature map after the convolution operation and pooling operation can be expressed as Formula (7):
x c o n v l = p o o l ( K l x l + b l )
where l is the lth convolution layer; xl is the input; Kl denotes the convolution kernel; bl denotes the bias. The feature extractor has four convolution layers, and the convolution calculation method is consistent with Formula (7).
According to the structure of the universal classifier, the cross-entropy loss function is used for the fault classifier. We are given the source sample labels. yi ∈ {1, 2, 3, …, m}, m denotes the total number of sample categories (including normal categories and fault categories). Therefore, the classification loss can be expressed as follows:
c l s = 1 n ( i = 1 n j = 1 m I ( y i = j ) log ( p j ) )
where N is the sample batch size of the source domain during training; pj denotes the probability that the sample is predicted to be jth class. I is a judgment function that outputs 1 if the input is true and 0 otherwise.
To align the joint distribution of input features and output labels in the source and target domains, the joint maximum mean distance (JMMD) [31] is used for distribution discrepancy metrics. The calculation of this loss function is expressed by:
d = E x s ~ P ( x ) [ l = 1 m ϕ l ( x s l ) ] E x t ~ Q ( x ) [ l = 1 m ϕ l ( x t l ) ] l = 1 m H l 2
where P(x) denotes the feature space distribution of the source domain samples. Q(x) represents the feature spatial distribution of the target domain samples. l = 1 m ϕl( x s l ) = ϕ1( x s 1 )⊗ϕ2( x s 2 )⊗···⊗ϕl( x s l ) is the feature mapping of the source domain in the tensor product Hilbert space. l is the set of higher network layers; l = 1 m ϕl( x t l ) is the feature mapping of the source domain in the tensor product Hilbert space; l = 1 m H l denotes an mth order tensor product feature space.
Therefore, the total loss function can be composed of fault classification loss and domain alignment loss, as shown below:
( θ E , θ F , θ C , θ D ) = c l s + λ d
where θE, θF, θC, and θD represent the parameters of the ECA module, CNN feature extractor, fault classifier, and distribution discrepancy metrics, respectively. λ is the weighting parameter. The training process of VMD-ECA-DTN consists of two processes: pre-training and cross-domain adaptation. During the source domain pre-training, fault classification loss is used to optimize the goal. In domain adaptation training, fault classification loss and domain confusion loss are used to optimize the target. Based on Formula (10), the parameters θE, θF, θC, and θD are updated in the following order:
θ E θ E μ ( c l s θ E + λ d θ E )
θ F θ F μ ( c l s θ F + λ d θ F )
θ C θ C μ c l s θ C
θ D θ D μ d θ D
where μ is the learning rate.

4. Experimental Results and Analysis

4.1. Dataset Introduction and Experiment Setup

To demonstrate the effectiveness of the proposed method, in this section, we select the CWRU bearing fault dataset for experiments, which is a widely used dataset [32]. It was provided by the Case Western Reserve University laboratory and is currently the most used open-source dataset. We selected the driver end fault data with a sampling frequency of 12 kHz and normal sample data. The faults are divided into inner ring faults (IF), outer ring faults (OF), and rolling body faults (RF). The diameter of the fault damage is 7 mils, 14 mils, and 21 mils (1 mil = 0.001 inches). There are 10 categories of fault samples plus normal samples, and 260 samples are selected for each category. According to load and speed, it is divided into four diagnostic tasks, as shown in Table 2. Cross-domain diagnosis tasks 0-1 indicate that the slave source domain is the condition where task code 0 resides, and the target domain is the condition where task code 1 resides. There are 6 cross-domain diagnosis tasks in pairs.
The data is augmented by partial resampling techniques. Each sample of the training set and testing set is a continuous time series with a length of 1024. The dataset for the whole experiment includes data collected under three different working conditions. According to different working conditions, the data are divided into source domain data and target domain data. The source domain data contain 1000 samples, and the target domain data contan 1600 samples. Source domain data have labels, but target domain data have no labels. The details of the number of samples in the training dataset, validation dataset, and testing dataset are given in Table 3. x − y indicates that the source domain is x, and the target domain is y.
During the following comparative experiment, all calculation was carried out on a computer with Intel(R) Core(TM) i7-6700K (4.00 GHz), NVIDIA GTX 1080Ti graphics cards, 64 GB RAM and Pytorch framework. To reduce the impact of accidental behavior in the experiment, all classification processes were repeated 10 times, and the trimmed mean value was taken as the final result (average after removing the highest and lowest scores). The max training epoch is set to 200, with a batch size of 64. In practical applications, the original vibration data collected by sensors are often mixed with noise, so the intelligent fault diagnosis models is required to have a certain anti-noise ability. To demonstrate the anti-noise capability of the proposed method, various Gaussian noises were added to the original vibration signal, and SNR [33] was used to evaluate the noise level, which was defined as:
SNR ( d B ) = 10 · lg ( P s i g na l / P n o i s e )
where Psignal and Pnoise represent the effective power of the raw vibration signal and noise signal, respectively.

4.2. Comparison of Signal Preprocessing Methods

Signal preprocessing techniques play an important role in intelligent fault diagnosis. The effective signal preprocessing method is beneficial to the intelligent diagnosis model for better learning the fault features in the original signal. To illustrate the effectiveness and advantages of the proposed VMD preprocessing method in the whole intelligent fault diagnosis framework, popular preprocessing methods are compared. Notably, all comparison methods are consistent with the rest of the framework, except for signal preprocessing. All details of the contrastive methods are illustrated as follows:
  • FFT [34]: The basic idea of FFT (Fast Fourier Transform) is to decompose the original sequence of N points into a series of short sequences.
  • EMD [35]: EMD (Empirical Mode Decomposition) is a signal processing method in the time-frequency domain, which is based on the time-scale characteristics of the data itself, without setting any basis function in advance. EMD has obvious advantages in processing non-stationary and nonlinear data and is suitable for analyzing nonlinear non-stationary signal sequences with a high signal-to-noise ratio.
  • VMD: The signal decomposition method we use in the signal preprocessing stage. It shifts the acquisition of signal components into a variational framework. A non-recursive processing strategy is used to decompose the original signal by constructing and solving a constrained variational problem, which can effectively avoid problems such as modal aliasing, over-envelope, under-envelope, and boundary effects.
In addition, Gaussian noises of 4 dB and 0 dB were added to the original vibration signal, respectively. The purpose of adding noise is to verify the anti-noise and robustness of the signal preprocessing method. The comparison results of the signal preprocessing methods under different noise levels are illustrated in Figure 5 and Table 4.
From the comparison results, we can draw the following conclusions:
  • When the original vibration signal is not added with Gaussian noise, the average accuracy of all signal preprocessing methods for the six transfer diagnosis tasks is more than 97.00%. After Gaussian noise is added to the original vibration signal, the diagnostic accuracy of all signal preprocessing methods is almost reduced. This shows that the addition of noise seriously affects the performance of the model. In addition, the higher the noise level, the more serious the performance of the intelligent fault diagnosis model decreases.
  • At the same noise level, the FFT preprocessing method has the worst performance, while the VMD preprocessing method has the best performance. Although the VMD preprocessing method is not optimal in individual transfer diagnosis tasks (such as 0-2 transfer diagnosis task without adding noise), the VMD preprocessing method has the highest cross-domain diagnosis accuracy in most transfer diagnosis tasks as a whole. The results show that VMD preprocessing can provide better anti-noise and robustness for intelligent fault diagnosis models.
Based on the above two points, it can be concluded that the VMD preprocessing method has a beneficial effect on neural network learning fault features. The original signal, decomposition signal, and signal center frequency are shown in Figure 6. This shows that the different modes of decomposition represent the different central frequency characteristics of the fault signals. Decomposition of the signal will be more conducive to intelligent model learning and will play a role in noise reduction. Notably, the number of modes K is a hyperparameter. The effect of VMD decomposition is mainly affected by the selected value of the modal number. When the selected value of the mode is small, since the VMD algorithm is equivalent to an adaptive filter bank, some important information in the original signal will be filtered, affecting the subsequent feature extraction. When the selected value of the mode is large, the center frequencies of the adjacent mode components will be close to each other, resulting in mode repetition or additional noise. The main difference between the different modes is the difference in center frequency. Therefore, the appropriate mode value is selected by observing the distribution of the center frequency under different mode numbers. We set K equal to 5 by this method. In addition, we conducted some experiments on different hyperparameters K, as shown in Figure 7. This indicates that the VMD algorithm needs an appropriate K setting.

4.3. Comparison of Feature Fusion Methods

How to fuse the decomposed mode signal features is a difficult problem. After the original signal decomposition, each mode represents the fault features of different central frequencies. The aim of adding channel attention mechanisms to the intelligent fault diagnosis model is to reweight and fuse different mode features. This means that the transfer learning fault diagnosis network automatically learns which modes are more important to the final objective optimization. To illustrate the effectiveness and advantages of the proposed ECA feature fusion method in the whole intelligent fault diagnosis framework, popular feature fusion methods are compared. Notably, all comparison methods are consistent with the rest of the framework, except for feature fusion. We used Concatenate, Add, SEA, and ECA (the attentional mechanism we use in the proposed transfer learning framework) to conduct experiments for comparison. The experimental results are shown in Figure 8 and Table 5.
The experimental results show that different feature fusion methods have different effects on transfer fault diagnosis. However, the effect of feature fusion is not as important as signal preprocessing. In addition, the feature fusion method needs to be combined with the pretreatment method. In this way, the decomposed modes of VMD can be better utilized. As can be seen from Table 5, the ECA feature fusion method achieves the highest average diagnostic accuracy. In addition, the accuracy standard deviation of the ECA method is relatively lower.

4.4. Comparison between the Proposed Method and State-Of-The-Art Methods

To verify the universality and effectiveness of the proposed VMD-ECA-DTN, we compared it with state-of-the-art methods. The methods to be compared are WDCNN (Wide Deep Convolutional Neural Networks) [36], DDC (Deep Domain Confusion Network) [37], DANN (Deep Adversarial Neural Network) [38], and DTN (Deep Transfer Network with joint distribution adaptation) [39]. In particular, the WDCNN model is trained only by the labeled samples in the source domain, while the other models are trained by a combination of labeled samples in the source domain and unlabeled samples in the target domain. To facilitate comparison, the structural design of these models adopts a CNN feature extractor. The number of network layers inside CNN is consistent with the proposed VMD-ECA-DTN, and the hyperparameters of these models are also consistent with the VMD-ECA-DTN. Moreover, to comprehensively analyze and compare all methods objectively, all classification processes were repeated 10 times, and the trimmed mean value was taken as the result. The comparison results of the proposed method with other state-of-the-art methods are shown in Table 6.
From the comparison results above, we can draw the following conclusions:
  • Among the five methods, only WDCNN is not a method of domain adaptation, and its average classification accuracy is lower than that of the other four methods, indicating the effectiveness of the domain adaptive method for cross-domain diagnosis. After adding 0 dB noise to the original data, the average diagnostic accuracy of WDCNN is only 93.31%.
  • Adding Gaussian noise to the original signal affects the ability of the intelligent fault diagnosis model to learn fault features. The proposed method achieves the best average performance in different noise environments. Even under 0 dB noise, the average diagnostic accuracy of the proposed method is 95.24%. This shows that the proposed method has better robustness and universality.
  • In a specific transfer diagnostic task, the performance of the proposed method is not the best. For example, in the 0-1 diagnostic task with 0 dB noise, the diagnostic accuracy of the proposed method is only 90.73%. This may indicate that different levels of noise will affect VMD signal preprocessing, and thus affect neural network learning fault features. However, in most transfer diagnosis tasks, our proposed method is superior to other state-of-the-art methods. This again shows the effectiveness and robustness of our proposed method.

5. Conclusions

In this study, we propose a novel deep transfer learning method for intelligent fault diagnosis based on Variational Mode Decomposition and Efficient Channel Attention. VMD signal preprocessing is used to decompose the original vibration signals into mode signals of different center frequencies. The effective separation of IMFs and the frequency domain division of signals are realized. Finally, the effective decomposition components of a given signal are obtained. To fuse the mode features more effectively after VMD decomposition, ECA is used to learn channel attention. It is used to adaptively learn the importance of different mode features for transfer diagnostic tasks, thereby improving model performance. We study the effects of signal preprocessing, feature fusion, and synthesis on the diagnosis results in the task of transfer diagnosis. The conclusions are as follows:
  • An appropriate signal preprocessing method is beneficial to the transfer diagnosis model. The purpose of preprocessing is to denoise the signal and extract the frequency signal, which is useful for diagnostic tasks.
  • Feature fusion is an important step in learning the main mode features. Compared with other fusion methods, the ECA module integrates the features of different modes, which facilitates the further learning of vibration signals by the diagnostic model.
  • The combination of signal preprocessing and attention mechanisms can be used to extract meaningful features. Our proposed method has better performance (robustness and generalization) in transfer diagnosis tasks compared with state-of-the-art methods.
It has certain application potential in cross-domain intelligent fault diagnosis. However, there are also some shortcomings. First, it is difficult to obtain balanced datasets in practical applications. How to improve the method for imbalanced datasets remains to be discussed. Second, the hyperparameters of VMD preprocessing in this method are still determined by expert experience. Getting rid of additional interventions remains difficult.
In the future, it is still a challenge to popularize intelligent fault diagnosis models in battery manufacturing, aerospace, and other industrial fields. For single application scenarios, such as battery manufacturing equipment, an intelligent fault diagnosis model based on unsupervised transfer learning is feasible. However, it is very difficult to solve the problem of fault diagnosis from laboratory data to actual equipment data. The actual data collected in the industry may be unlabeled and almost glitch-free. The data without any labels are useless for an intelligent model. It takes a long way to get there.

Author Contributions

Conceptualization, methodology and validation, C.L. and X.Z.; formal analysis and investigation, Z.H. and Z.B.; writing—original draft preparation, C.L. and X.Z.; writing—review and editing, Z.H., C.L. and Z.B.; supervision, M.G. and W.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China [Grant No. 2020YFB1710600], the Key Research Development Program of Zhejiang Province [Grant No. 2021C01111], and the National Natural Science Foundation of China [Grant No. 62171170].

Data Availability Statement

All data used in the experiments can be downloaded from the following links: http://csegroups.case.edu/bearingdatacenter/pages/download-data-file (accessed on 26 July 2021) (CWRU dataset).

Acknowledgments

The authors would like to thank all anonymous reviewers and editors for their helpful suggestions for the improvement of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Parmar, U.; Pandya, D.H. Experimental Investigation of Cylindrical Bearing Fault Diagnosis with SVM. Mater. Today Proc. 2021, 44, 1286–1290. [Google Scholar] [CrossRef]
  2. Cerrada, M.; Zurita, G.; Cabrera, D.; Sánchez, R.-V.; Artés, M.; Li, C. Fault Diagnosis in Spur Gears Based on Genetic Algorithm and Random Forest. Mech. Syst. Signal Process. 2016, 70–71, 87–103. [Google Scholar] [CrossRef]
  3. Song, B.; Tan, S.; Shi, H.; Zhao, B. Fault Detection and Diagnosis via Standardized k Nearest Neighbor for Multimode Process. J. Taiwan Inst. Chem. Eng. 2020, 106, 1–8. [Google Scholar] [CrossRef]
  4. Amiruddin, A.; Zabiri, H.; Taqvi, S.A.A.; Tufa, L.D. Neural Network Applications in Fault Diagnosis and Detection: An Overview of Implementations in Engineering-Related Systems. NEURAL Comput. Appl. 2020, 32, 447–472. [Google Scholar] [CrossRef]
  5. Sun, T.D.; Yu, G.; Gao, M.; Zhao, L.L.; Bai, C.; Yang, W.Q. Fault Diagnosis Methods Based on Machine Learning and Its Applications for Wind Turbines: A Review. IEEE Access 2021, 9, 147481–147511. [Google Scholar] [CrossRef]
  6. Dong, Z.; Xiaoyue, J.; Zhou, G.; Gao, M.; Donglian, Q. Multimodal Neuromorphic Sensory-Processing System with Memristor Circuits for Smart Home Applications. IEEE Trans. Ind. Appl. 2022, in press. [Google Scholar] [CrossRef]
  7. Ji, X.Y.; Dong, Z.K.; Lai, C.S.; Qi, D.L. A Brain-Inspired In-Memory Computing System for Neuronal Communication via Memristive Circuits. IEEE Commun. Mag. 2022, 60, 100–106. [Google Scholar] [CrossRef]
  8. Zhang, S.; Zhang, S.; Wang, B.; Habetler, T.G. Deep Learning Algorithms for Bearing Fault Diagnostics—A Comprehensive Review. IEEE Access 2020, 8, 29857–29881. [Google Scholar] [CrossRef]
  9. Janssens, O.; Slavkovikj, V.; Vervisch, B.; Stockman, K.; Loccufier, M.; Verstockt, S.; Van de Walle, R.; Van Hoecke, S. Convolutional Neural Network Based Fault Detection for Rotating Machinery. J. Sound Vib. 2016, 377, 331–345. [Google Scholar] [CrossRef]
  10. Zhao, M.H.; Zhong, S.S.; Fu, X.Y.; Tang, B.P.; Pecht, M. Deep Residual Shrinkage Networks for Fault Diagnosis. IEEE Trans. Ind. Inform. 2020, 16, 4681–4690. [Google Scholar] [CrossRef]
  11. Lei, J.H.; Liu, C.; Jiang, D.X. Fault Diagnosis of Wind Turbine Based on Long Short-Term Memory Networks. Renew. Energy 2019, 133, 422–432. [Google Scholar] [CrossRef]
  12. Li, C.; Zhang, S.H.; Qin, Y.; Estupinan, E. A Systematic Review of Deep Transfer Learning for Machinery Fault Diagnosis. Neurocomputing 2020, 407, 121–135. [Google Scholar] [CrossRef]
  13. Li, J.; Lin, M.; Li, Y.; Wang, X. Transfer Learning with Limited Labeled Data for Fault Diagnosis in Nuclear Power Plants. Nucl. Eng. Des. 2022, 390, 111690. [Google Scholar] [CrossRef]
  14. Wang, M.; Deng, W.H. Deep Visual Domain Adaptation: A Survey. Neurocomputing 2018, 312, 135–153. [Google Scholar] [CrossRef]
  15. Zhang, S.; Su, L.; Gu, J.; Li, K.; Zhou, L.; Pecht, M. Rotating Machinery Fault Detection and Diagnosis Based on Deep Domain Adaptation: A Survey. Chinese J. Aeronaut. 2021, in press. [Google Scholar] [CrossRef]
  16. Guo, L.; Lei, Y.G.; Xing, S.B.; Yan, T.; Li, N.P. Deep Convolutional Transfer Learning Network: A New Method for Intelligent Fault Diagnosis of Machines With Unlabeled Data. IEEE Trans. Ind. Electron. 2019, 66, 7316–7325. [Google Scholar] [CrossRef]
  17. Sejdinovic, D.; Sriperumbudur, B.; Gretton, A.; Fukumizu, K. Equivalence of Distance-Based and RKHS-Based Statistics in Hypothesis Testing. Ann. Stats 2013, 41, 2263–2291. [Google Scholar] [CrossRef]
  18. Jiao, J.; Zhao, M.; Lin, J.; Liang, K. Residual Joint Adaptation Adversarial Network for Intelligent Transfer Fault Diagnosis. Mech. Syst. Signal Process. 2020, 145, 106962. [Google Scholar] [CrossRef]
  19. Qian, Q.; Qin, Y.; Wang, Y.; Liu, F.Q. A New Deep Transfer Learning Network Based on Convolutional Auto-Encoder for Mechanical Fault Diagnosis. Measurement 2021, 178, 109352. [Google Scholar] [CrossRef]
  20. Li, X.D.; Zheng, J.H.; Li, M.T.; Ma, W.Z.; Hu, Y. Frequency-Domain Fusing Convolutional Neural Network: A Unified Architecture Improving Effect of Domain Adaptation for Fault Diagnosis. Sensors 2021, 21, 450. [Google Scholar] [CrossRef]
  21. Ben Ali, J.; Fnaiech, N.; Saidi, L.; Chebel-Morello, B.; Fnaiech, F. Application of Empirical Mode Decomposition and Artificial Neural Network for Automatic Bearing Fault Diagnosis Based on Vibration Signals. Appl. Acoust. 2015, 89, 16–27. [Google Scholar] [CrossRef]
  22. Wang, Y.; Yang, M.M.; Zhang, Y.P.; Xu, Z.D.; Huang, J.G.; Fang, X. A Bearing Fault Diagnosis Model Based on Deformable Atrous Convolution and Squeeze-and-Excitation Aggregation. IEEE Trans. Instrum. Meas. 2021, 70, 3524410. [Google Scholar] [CrossRef]
  23. Wu, H.; Li, J.; Zhang, Q.; Tao, J.; Meng, Z. Intelligent Fault Diagnosis of Rolling Bearings under Varying Operating Conditions Based on Domain-Adversarial Neural Network and Attention Mechanism. ISA Trans. 2022, in press. [Google Scholar] [CrossRef]
  24. Liu, R.N.; Yang, B.Y.; Zio, E.; Chen, X.F. Artificial Intelligence for Fault Diagnosis of Rotating Machinery: A Review. Mech. Syst. Signal Process. 2018, 108, 33–47. [Google Scholar] [CrossRef]
  25. Dragomiretskiy, K.; Zosso, D. Variational Mode Decomposition. IEEE Trans. Signal Process. 2014, 62, 531–544. [Google Scholar] [CrossRef]
  26. Zhang, X.; Miao, Q.; Zhang, H.; Wang, L. A Parameter-Adaptive VMD Method Based on Grasshopper Optimization Algorithm to Analyze Vibration Signals from Rotating Machinery. Mech. Syst. Signal Process. 2018, 108, 58–72. [Google Scholar] [CrossRef]
  27. Zhang, J.L.; Wu, J.M.; Hu, B.B.; Tang, J.H. Intelligent Fault Diagnosis of Rolling Bearings Using Variational Mode Decomposition and Self-Organizing Feature Map. J. Vib. Control 2020, 26, 1886–1897. [Google Scholar] [CrossRef]
  28. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 11531–11539. [Google Scholar]
  29. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E.H. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef]
  30. Niu, G.X.; Wang, X.; Golda, M.; Mastro, S.; Zhang, B. An Optimized Adaptive PReLU-DBN for Rolling Element Bearing Fault Diagnosis. Neurocomputing 2021, 445, 26–34. [Google Scholar] [CrossRef]
  31. Long, M.; Zhu, H.; Wang, J.; Jordan, M.I. Deep Transfer Learning with Joint Adaptation Networks. arXiv 2016, arXiv:1605.06636. [Google Scholar]
  32. Smith, W.A.; Randall, R.B. Rolling Element Bearing Diagnostics Using the Case Western Reserve University Data: A Benchmark Study. Mech. Syst. Signal Process. 2015, 64–65, 100–131. [Google Scholar] [CrossRef]
  33. Naylor, G.; Johannesson, R.B. Long-Term Signal-to-Noise Ratio at the Input and Output of Amplitude-Compression Systems. J. Am. Acad. Audiol. 2009, 20, 161–171. [Google Scholar] [CrossRef] [PubMed]
  34. Wang, T.Z.; Qi, J.; Xu, H.; Wang, Y.D.; Liu, L.; Gao, D.J. Fault Diagnosis Method Based on FFT-RPCA-SVM for Cascaded-Multilevel Inverter. ISA Trans. 2016, 60, 156–163. [Google Scholar] [CrossRef]
  35. Vashishtha, G.; Chauhan, S.; Kumar, A.; Kumar, R. An Ameliorated African Vulture Optimization Algorithm to Diagnose the Rolling Bearing Defects. Meas. Sci. Technol. 2022, 33, 075013. [Google Scholar] [CrossRef]
  36. Zhang, W.; Peng, G.; Li, C.; Chen, Y.; Zhang, Z. A New Deep Learning Model for Fault Diagnosis with Good Anti-Noise and Domain Adaptation Ability on Raw Vibration Signals. Sensors 2017, 17, 425. [Google Scholar] [CrossRef]
  37. Fu, S.; Zhang, Y.; Lin, L.; Zhao, M.; Zhong, S. Deep Residual LSTM with Domain-Invariance for Remaining Useful Life Prediction across Domains. Reliab. Eng. Syst. Saf. 2021, 216, 108012. [Google Scholar] [CrossRef]
  38. Mao, W.; Ding, L.; Liu, Y.; Afshari, S.S.; Liang, X. A New Deep Domain Adaptation Method with Joint Adversarial Training for Online Detection of Bearing Early Fault. ISA Trans. 2021, 122, 444–458. [Google Scholar] [CrossRef]
  39. Han, T.; Liu, C.; Yang, W.G.; Jiang, D.X. Deep Transfer Network with Joint Distribution Adaptation: A New Intelligent Fault Diagnosis Framework for Industry Application. ISA Trans. 2020, 97, 269–281. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Domain adaptation. The target classifier on the source domain can also be applied to the target domain after the alignment of the domain adaptation feature space.
Figure 1. Domain adaptation. The target classifier on the source domain can also be applied to the target domain after the alignment of the domain adaptation feature space.
Entropy 24 01087 g001
Figure 2. Diagram of efficient channel attention.
Figure 2. Diagram of efficient channel attention.
Entropy 24 01087 g002
Figure 3. The architecture of VMD-ECA-DTN for intelligent fault diagnosis.
Figure 3. The architecture of VMD-ECA-DTN for intelligent fault diagnosis.
Entropy 24 01087 g003
Figure 4. Details of the designed VMD-ECA-DTN architecture.
Figure 4. Details of the designed VMD-ECA-DTN architecture.
Entropy 24 01087 g004
Figure 5. Comparison results of signal preprocessing methods under different transfer diagnosis tasks.
Figure 5. Comparison results of signal preprocessing methods under different transfer diagnosis tasks.
Entropy 24 01087 g005
Figure 6. Diagram of signal decomposition. (a) Original signal; (b) Decomposition signal; (c) Spectral decomposition.
Figure 6. Diagram of signal decomposition. (a) Original signal; (b) Decomposition signal; (c) Spectral decomposition.
Entropy 24 01087 g006aEntropy 24 01087 g006b
Figure 7. Experimental results with different hyperparameters K.
Figure 7. Experimental results with different hyperparameters K.
Entropy 24 01087 g007
Figure 8. Comparison results of different feature fusion methods.
Figure 8. Comparison results of different feature fusion methods.
Entropy 24 01087 g008
Table 1. Specific parameters of the proposed network architecture.
Table 1. Specific parameters of the proposed network architecture.
TypeLayerKernel SizeStrideChannel SizeInput SizeOutput Size
InputReshape////(64, 1, 1024)
VMDVMD output////(64, 5, 1024)
ECAAda-Avg-Pool///(64, 5, 1024)(64, 5, 1)
Reshape///(64, 5, 1)(64, 1, 5)
Conv111(64, 1, 5)(64, 1, 5)
Sigmoid///(64, 1, 5)(64, 1, 5)
Reshape///(64, 1, 5)(64, 5, 1)
Multiplier///(64, 5, 1)
(64, 1, 1024)
(64, 1, 1024)
CNN feature extractorConv15116(64, 1, 1024)(64, 16, 1010)
BN///(64, 16, 1010)(64, 16, 1010)
Conv3132(64, 16, 1010)(64, 32, 1008)
BN///(64, 32, 1008)(64, 32, 1008)
Max-Pool2232(64, 32, 1008)(64, 32, 504)
Conv3164(64, 32, 504)(64, 64, 502)
BN///(64, 64, 502)(64, 64, 502)
Conv31128(64, 64, 502)(64, 128, 500)
BN///(64, 128, 500)(64, 128, 500)
Ada-Max-Pool///(64, 128, 500)(64, 128, 4)
Reshape///(64, 128, 4)(64, 512)
FC///(64, 512)(64, 256)
Dropout///(64, 256)(64, 256)
Fault classifier (Output)FC///(64, 256)(64, 128)
Dropout///(64, 128)(64, 128)
FC///(64, 128)(64, 10)
Distribution discrepancy metricsFC///(64, 256)(64, 128)
Dropout///(64, 128)(64, 128)
FC///(64, 128)(64, 10)
Table 2. The diagnostic tasks of CWRU.
Table 2. The diagnostic tasks of CWRU.
Task CodeSpeed (rpm)Load (HP)
017300
117501
217722
Table 3. Sample division of six transfer diagnosis tasks.
Table 3. Sample division of six transfer diagnosis tasks.
Transfer Diagnosis TaskTraining DatasetValidation Dataset (Target Domain Data: B)Testing
Dataset (Target Domain Data: C)
Source Domain DataTarget Domain Data: A
x − y (x, y ∈ [0, 2]; x, y ∈ N; x ≠ y)Labeled
samples: 1000
Unlabeled samples: 1000Samples: 300Samples: 300
Table 4. Comparison results of signal preprocessing methods under different transfer diagnosis tasks.
Table 4. Comparison results of signal preprocessing methods under different transfer diagnosis tasks.
0-10-21-01-22-02-1Average
FFT99.68 ± 0.49100.00 ± 0.0098.98 ± 5.45100.00 ± 0.0087.74 ± 6.9096.39 ± 2.0797.13
EMD97.88 ± 3.8298.20 ± 2.1894.86 ± 6.66100.00 ± 0.0095.17 ± 6.7298.84 ± 0.7397.49
VMD99.71 ± 0.8899.28 ± 0.5198.89 ± 0.70100.00 ± 0.0096.60 ± 1.6899.71 ± 0.2599.03
FFT(4 dB)97.08 ± 1.9597.62 ± 1.2994.38 ± 2.4499.46 ± 0.3771.90 ± 5.2598.05 ± 0.3693.08
EMD(4 dB)96.46 ± 0.7597.11 ± 2.2594.90 ± 4.0499.28 ± 0.4295.32 ± 1.9197.40 ± 0.3696.75
VMD(4 dB)97.32 ± 0.8897.22 ± 0.9095.77 ± 1.4599.28 ± 0.3996.39 ± 2.1698.41 ± 0.2897.40
FFT(0 dB)92.42 ± 1.7983.98 ± 2.3989.14 ± 1.3196.00 ± 1.5975.10 ± 5.2595.56 ± 2.4688.70
EMD(0 dB)93.07 ± 1.9892.35 ± 1.3394.47 ± 0.3296.17 ± 1.2891.92 ± 2.0694.23 ± 1.3693.70
VMD(0 dB)94.30 ± 1.2497.04 ± 1.4890.73 ± 3.7797.91 ± 0.2594.81 ± 2.6796.61 ± 1.0395.24
Table 5. Comparison results of different feature fusion methods.
Table 5. Comparison results of different feature fusion methods.
0-10-21-01-22-02-1Average
Concatenate99.85 ± 0.1898.84 ± 0.8597.96 ± 4.9899.35 ± 0.1096.94 ± 0.8999.57 ± 0.5098.75
Add99.93 ± 0.2898.98 ± 5.2889.20 ± 4.8999.93 ± 0.1994.47 ± 7.6098.92 ± 1.4696.90
SEA99.64 ± 3.8698.41 ± 0.6497.96 ± 7.4299.28 ± 0.4795.58 ± 4.8599.64 ± 0.5998.42
ECA99.71 ± 0.8899.28 ± 0.5198.89 ± 0.70100.00 ± 0.0096.60 ± 1.6899.71 ± 0.2599.03
Table 6. Comparison results of the proposed method with other state-of-the-art methods.
Table 6. Comparison results of the proposed method with other state-of-the-art methods.
0-10-21-01-22-02-1Average
WDCNN96.62 ± 2.3494.61 ± 3.3598.36 ± 0.8199.98 ± 0.0697.50 ± 2.2596.36 ± 1.5997.24
DDC98.15 ± 2.0698.77 ± 1.9397.79 ± 1.50100.00 ± 0.0098.89 ± 3.7498.48 ± 0.5798.68
DANN99.42 ± 1.0399.20 ± 1.4499.23 ± 0.1899.71 ± 0.7797.79 ± 3.1295.24 ± 1.6498.43
DTN97.45 ± 0.8095.24 ± 2.3697.70 ± 1.38100.00 ± 0.0098.72 ± 0.5699.49 ± 0.6898.10
Proposed99.71 ± 0.8899.28 ± 0.5198.89 ± 0.70100.00 ± 0.0096.60 ± 1.6899.71 ± 0.2599.03
WDCNN(4 dB)92.33 ± 3.6795.41 ± 2.5396.25 ± 1.4397.16 ± 0.9795.02 ± 2.3995.41 ± 0.9695.26
DDC(4 dB)95.17 ± 2.1098.13 ± 3.7596.43 ± 0.7198.56 ± 0.7495.66 ± 1.5996.68 ± 2.4396.77
DANN(4 dB)93.72 ± 4.1197.47 ± 4.4296.60 ± 3.7799.28 ± 0.5995.92 ± 1.2396.68 ± 1.2196.61
DTN(4 dB)97.50 ± 0.3297.46 ± 1.0297.11 ± 0.9398.63 ± 2.0795.49 ± 2.3596.83 ± 1.1697.17
Proposed(4 dB)97.32 ± 0.8897.22 ± 0.9095.77 ± 1.4599.28 ± 0.3996.39 ± 2.1698.41 ± 0.2897.40
WDCNN(0 dB)93.33 ± 1.6493.09 ± 2.0391.90 ± 2.9195.28 ± 0.6591.08 ± 3.9395.19 ± 1.2493.31
DDC(0 dB)93.65 ± 2.6794.08 ± 1.5494.05 ± 2.5296.32 ± 1.3393.03 ± 1.4793.80 ± 2.7094.15
DANN(0 dB)94.44 ± 1.7895.17 ± 1.8092.93 ± 2.5897.55 ± 1.1392.35 ± 2.2295.45 ± 1.3494.65
DTN(0 dB)92.71 ± 0.8593.80 ± 2.1393.11 ± 1.3896.32 ± 1.0192.94 ± 1.8595.02 ± 1.6793.98
Proposed(0 dB)94.30 ± 1.2497.04 ± 1.4890.73 ± 3.7797.91 ± 0.2594.81 ± 2.6796.61 ± 1.0395.24
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, C.; Zheng, X.; Bao, Z.; He, Z.; Gao, M.; Song, W. A Novel Deep Transfer Learning Method for Intelligent Fault Diagnosis Based on Variational Mode Decomposition and Efficient Channel Attention. Entropy 2022, 24, 1087. https://0-doi-org.brum.beds.ac.uk/10.3390/e24081087

AMA Style

Liu C, Zheng X, Bao Z, He Z, Gao M, Song W. A Novel Deep Transfer Learning Method for Intelligent Fault Diagnosis Based on Variational Mode Decomposition and Efficient Channel Attention. Entropy. 2022; 24(8):1087. https://0-doi-org.brum.beds.ac.uk/10.3390/e24081087

Chicago/Turabian Style

Liu, Caiming, Xiaorong Zheng, Zhengyi Bao, Zhiwei He, Mingyu Gao, and Wenlong Song. 2022. "A Novel Deep Transfer Learning Method for Intelligent Fault Diagnosis Based on Variational Mode Decomposition and Efficient Channel Attention" Entropy 24, no. 8: 1087. https://0-doi-org.brum.beds.ac.uk/10.3390/e24081087

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop