Next Article in Journal
Learning in Convolutional Neural Networks Accelerated by Transfer Entropy
Previous Article in Journal
An Efficient Partition-Based Approach to Identify and Scatter Multiple Relevant Spreaders in Complex Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fault Feature Extraction for Reciprocating Compressors Based on Underdetermined Blind Source Separation

1
Mechanical Science and Engineering Institute, Northeast Petroleum University, Daqing 163318, China
2
Heilongjiang Key Laboratory of Petroleum Machinery Engineering, Daqing 163318, China
*
Author to whom correspondence should be addressed.
Submission received: 30 July 2021 / Revised: 11 September 2021 / Accepted: 13 September 2021 / Published: 15 September 2021

Abstract

:
In practical engineering applications, the vibration signals collected by sensors often contain outliers, resulting in the separation accuracy of source signals from the observed signals being seriously affected. The mixing matrix estimation is crucial to the underdetermined blind source separation (UBSS), determining the accuracy level of the source signals recovery. Therefore, a two-stage clustering method is proposed by combining hierarchical clustering and K-means to improve the reliability of the estimated mixing matrix in this paper. The proposed method is used to solve the two major problems in the K-means algorithm: the random selection of initial cluster centers and the sensitivity of the algorithm to outliers. Firstly, the observed signals are clustered by hierarchical clustering to get the cluster centers. Secondly, the cosine distance is used to eliminate the outliers deviating from cluster centers. Then, the initial cluster centers are obtained by calculating the mean value of each remaining cluster. Finally, the mixing matrix is estimated with the improved K-means, and the sources are recovered using the least square method. Simulation and the reciprocating compressor fault experiments demonstrate the effectiveness of the proposed method.

1. Introduction

With the rapid development of modern industry and technology, mechanical equipment is growing larger and more precise with high-speed [1]. Any fault of the equipment can possibly cause a breakdown of the entire mechanical system, which may cause the interruption of production, huge economic losses, and even casualties. Many mechanical fault diagnosis methods have been proposed to ensure the normal and stable operation of equipment, such as oil analysis, acoustic emission detection, temperature detection, and vibration analysis. The advantage of vibration analysis is its ability to retain abundant fault information, and thus is widely used in condition monitoring and feature extraction of mechanical equipment [2].
Reciprocating compressor are commonly used in the petroleum, chemical, and transportation industry with the advantages of wide pressure distribution and high compression efficiency. Therefore, how to maintain the normal operation of reciprocating compressor has attracted considerable attention in recent years [3,4,5]. Due to the complex structure of the reciprocating compressor, numerous internal excitation sources, and the mutual coupling of each source signal, the collected vibration signals mostly present non-stationarity and nonlinearity characteristics. Traditional signal analysis methods such as empirical mode decomposition (EMD) [6,7], local mean decomposition (LMD) [8,9], variational mode decomposition (VMD) [10,11] have been successfully applied in the field of rotating machinery fault diagnosis. However, these methods may fail to extract effective features from the vibration signals in the fault diagnosis procedures of reciprocating compressor [1].
Blind source separation (BSS) is an advanced signal processing technique that can accurately separate the source signals from their mixtures without relying on prior knowledge, which has wide applications in various areas such as biomedical engineering [12,13], speech recognition [14,15], and fault diagnosis [16,17]. Recently, BSS algorithms have been rapidly developed and have been successfully applied in real engineering applications. For example, Zhou et al. [18] combined independent component analysis (ICA) and principal component analysis for blast furnace ironmaking condition monitoring and fault diagnosis. To extract the weak fault feature of rotating machinery from the observed signals, Miao et al. [19] proposed a BSS method based on fast independent component analysis (Fast ICA) and wavelet packet analysis. He et al. [20] studied the weak characteristic determination of centrifugal compressors with underdetermined blind source separation (UBSS) based on sparse component analysis (SCA) and successfully detected the cracked blade information from the pressure pulsation. However, the above studies mainly focus on the positive or overdetermined cases, i.e., the number of sensors is greater than or equal to the number of sources, but this constraint is difficult to be met in practical engineering applications. Moreover, much research methods based on BSS or UBSS have focused on rotating machinery and seldom literature has considered the fault diagnosis on the reciprocating compressor. Therefore, we proposed a new UBSS method and apply it to the fault feature extraction of the reciprocating compressor.
Sparse component analysis, as a common UBSS method, can separate the source signals by exploiting the sparsity characteristics of sources in the transform domain [21,22]. Generally, the SCA algorithm consists of two steps: mixing matrix estimation and source recovery [23]. Among the two-step approach, the mixing matrix estimation has been widely considered to be the most important step [15,24,25]. The potential function and clustering are the two most common methods used to estimate the mixing matrix. The potential function can estimate the mixing matrix without knowing the number of the sources. For example, Bofill et al. [26] firstly proposed the potential function method to estimate the mixing matrix to separate the flute signals from the mixing signals. Hao et al. [27] further extended the potential function to three-dimensional space and achieved good results in separating bearing compound fault signals and extracting the fault feature. However, in summary, these methods based on potential function are limited in how many sensors they can use. In contrast, the clustering methods, such as hierarchical clustering [28,29], Fuzzy C-means (FCM) clustering [30,31], K-means clustering [32,33], can get rid of this limitation by allowing an arbitrary number of sensors. The clustering methods mentioned above have drawn tremendous interest and are popularly used for the mixing matrix estimation in the SCA. As an illustration, Lu et al. [34] applied hierarchical clustering to estimate the mixing matrix of the cylindrical structure fault signals, and thus identified the major vibration of the mechanical systems. Hu et al. [35] applied the FCM and L1 norm to extract the fault feature of the wind turbine gearbox, which successfully detected the fault location. He et al. [36] proposed an improved K-means by combining the Laplace potential function to estimate the mixing matrix of bearing vibration signals.
In summary, although these methods can all estimate the mixing matrix, there are still some accuracy level limitations in the source recovery as reported in [28,36]. In contrast to hierarchical clustering and FCM clustering, the K-means method has a fast computation speed and relatively higher estimation accuracy. However, the clustering accuracy is usually unstable because its initial clustering centers are generated randomly. To overcome the drawback that K-means is sensitive to the initial clustering centers, Xie et al. [37] employed the firefly optimization algorithm (FOA) to optimize the initial clustering centers, and thus achieved enhanced centers with fast convergence. Arthur et al. [38] proposed a novel K-means++ algorithm based on K-means to solve the above-mentioned problem. In addition, He et al. [39] applied the K-means to estimate the mixing matrix by combining the affinity propagation (AP) clustering to preset the initial clustering of K-means, which can avoid the random selection. The above various improved K-means algorithms have achieved great results in the mixing matrix estimation. However, another challenge in the K-means is that the outliers existing in the observed signals may have a significant influence on clustering, which has been ignored. To alleviate the effect of these outliers, Reju et al. [40] firstly proposed a quadratic hierarchical clustering method to remove the outliers existing in the clusters. More specifically, the initial clustering centers are obtained by the first hierarchical clustering, the cosine distance is used to eliminate the outliers deviating from the initial clustering centers, and then the remaining clusters are clustered again by the second hierarchical clustering.
Inspired by the idea of quadratic clustering and integrating the advantages of K-means, a two-stage clustering method is proposed by combining hierarchical clustering and K-means. The purpose of this study is to solve the two main problems in the K-means algorithm: the random selection of initial cluster centers and the sensitivity of the algorithm to outliers. Firstly, the hierarchical clustering method is used to cluster the mixed signals to obtain the clustering centers, and some outliers deviating from the clustering center are eliminated. Then, the new clustering centers obtained by calculating the mean value of the remaining clusters are set as the initial clustering centers of K-means. Finally, the source signals are separated with the least square method. The simulation and reciprocating compressor experiments prove the effectiveness of the proposed method.
The rest part of the paper is organized as follows: Section 2 introduces the UBSS method based on SCA. Section 3 states the problem existing in the K-means and presents the proposed mixing matrix estimation method by combining hierarchical clustering and K-means clustering. Section 5 presents a simulation analysis, a comparison with the traditional K-means, and the application in reciprocating compressor fault diagnosis and draws conclusions.

2. Underdetermined Blind Source Separation Based on SCA

2.1. Basic Theory of UBSS

The linear instantaneous mixed model of UBSS is shown as follows:
x ( t ) = A s ( t )
where x ( t ) = [ x 1 ( t ) , x 2 ( t ) , , x m ( t ) ] T is an m-dimensional observed signal, xi is the observation value of the i-th sensor at time point t. A = [ a 1 , a 2 , , a n ] m × n is a mixing matrix, a i is the i-th column vector of the matrix A . s ( t ) = [ s 1 ( t ) , s 2 ( t ) , , s n ( t ) ] T is an n-dimensional source signal, and si(t) is the sample point of the i-th source at time point t. m and n are the number of the observed signal and source signals, respectively. However, this study focuses on the UBSS under the condition that n > m 2 , which is more suitable for engineering applications.
To obtain the sparse signals, the mixed model in Equation (1) is transformed to the time-frequency (TF) domain with the short-time Fourier transform (STFT) [41,42]. The process is described as:
x ( t , f ) = A s ( t , f ) = i = 1 n a i s i ( t , f )
where x(t) and s(t) are the TF coefficients of x(t,f) and s(t,f), respectively.
After obtaining the sparse signals x(t,f), the UBSS based SCA algorithm can be employed to recover the source signals. The detailed steps of the UBSS-SCA are depicted in Figure 1.

2.2. Single Source Points Detection

The single source points (SSPs) detection is used as a preprocessing before the mixing matrix estimation for enhancing the sparse representation of the source signals in the TF domain and thus improving the accuracy of the estimated mixing matrix. In the literature [40], an SSPs identification method was proposed by comparing the directions of the absolute values of the real and imaginary parts of the observed signals by:
R { x ( t , f ) } T I { x ( t , f ) } R { x ( t , f ) } 2 I { x ( t , f ) } 2   < cos ( Δ θ )
where · represents the absolute value and Δ θ is a threshold usually set to be close to 0, the literature [26] suggested θ = 0.2 0.8 .
Since noise may interfere with the sparse representation of observed signals in the TF domain, so it is necessary to effectively reduce the strength of interference by eliminating the low-energy mixed TF vectors if:
x ( t , f ) 2 < ε max x ( 1 , 1 ) 2 , , x ( T , K ) 2
Note that, in (4), T and K represent the number of time and frequency points, respectively, and ε represents a fraction number, commonly set to be 0.01.

2.3. Mixing Matrix Estimation Based on K-Means

K-means is an unsupervised learning clustering algorithm with the advantages of fast computation and high accuracy, which is widely used for the mixing matrix estimation in UBSS. The principle is described as follows [43].
Given a sample set X = { x 1 , x 2 , , x n } , randomly select an initial cluster center ( m 1 , m 2 , , m k ) , and find a partition C such that the objective function is minimized by:
min C l = 1 k C ( i ) = l x i m l 2
where · denotes the euclidean norm and k is the number of clusters.
Then, for a given division C, the center ( m 1 , m 2 , , m k ) of each cluster is found, such that the objective function is minimized by:
min m 1 , m 2 , , m k l = 1 k C ( i ) = l x i m l 2
For each class G l containing ni samples, update its mean value ml:
m l = 1 n l C ( i ) = l x i , l = 1 , 2 , , k
Repeat the above steps until the cluster center ml does not change anymore. The obtained clustering center ( m 1 , m 2 , , m k ) is the estimated mixing matrix A ^ .

2.4. Source Recovery

After obtaining the estimated mixing matrix A ^ , the sources are recovered by solving a series of least square problems [44]. Specifically, A is defined as m × ( m 1 ) sub-matrices of A ^ , which is represented by:
A = { A i | A i = [ a θ 1 , a θ 2 , , a θ m 1 ] }
Obviously, A obtains C n ( m 1 ) elements, i.e., i = 1 , 2 , , C n ( m 1 ) .
For any given TF point ( t , f ) , there is a matrix A = [ a ϕ 1 , a ϕ 2 , , a ϕ m 1 ] in the set A , which satisfies (9):
x ˜ ( t , k ) = A A x ˜ ( t , k )
where A is the pseudo-inverse of A .
Then, the source signals can be recovered through the following formula:
s ^ j ( t , k ) =   e i ,   if   j = ϕ i 0 ,   otherwise  
where e = [ e 1 , e 2 , , e m 1 ] T = A x ^ ( t , f ) , and A can be obtained by:
A = arg min A i A x ˜ ( t , k ) A i A i x ˜ ( t , k ) 2
Finally, the estimated TF source signal s ^ ( t , f ) is transformed back into the time domain by inverse short-time Fourier transform (ISTFT), which provides the time domain estimated signals s ^ ( t ) as required.

3. The Proposed Mixing Matrix Estimation Method

3.1. The Statement of K-Means Clustering

From the principle of K-means described in Section 2.3, we can find that there are two major problems in K-means. The one is that the K-means algorithm is sensitive to outliers. When some outliers, such as noise, exist in the original sample sets, the K-means will take these outliers as target data points, and divide them into the nearest clusters from the initial clustering center during the clustering process. Then, the initial clustering center will be updated based on the obtained clusters. Therefore, the clustering accuracy level will decrease in the iteration process because of the existence of outliers. The other is that the random selection issue of the initial clustering centers. From Equation (5), we can find that the initial clustering centers are randomly generated and thus the sets are divided into k clusters according to the given clustering centers. As we know, K-means is a partition-based clustering algorithm whose results greatly depend on the selection of the initial clustering centers. Once the clustering centers are set, the iteration will be executed according to the original clustering centers, which cannot guarantee that the solution is the global optimal.
To address the above challenging issues, a novel cluster algorithm is proposed by combining hierarchical clustering and K-means. Unlike FCM and K-means, hierarchical clustering does not have the problem of a random selection of the initial clustering centers and its result is relatively stable. Therefore, it is chosen as the first cluster in this study to get the original clustering centers, and then cosine distance was used to eliminate part outliers deviating from that clustering centers. Moreover, aiming at the random selection problem of the initial clustering centers in K-means, the revised clustering centers after removing the outliers are set as the initial clustering centers. Finally, the improved K-means algorithm is employed to estimate the mixing matrix.

3.2. Sources Number Estimation Based on EMD and BIC

Before the mixing matrix estimation, the number of clusters (the source number) needs to be predetermined. The Bayesian information criterion (BIC) is robust to non-Gaussian sources, and thus it can be used to estimate the number of mechanical vibration sources in this study.
The objective of the Minaka Bayesian selection model (MIBS) based on BIC is to find a number k = n   ( 1 k l ) that maximizes the cost function. The cost function of the MIBS model is as follows [45]:
M I B S ( k ) = p ( x | k ) p k j = 1 k λ j N / 2 σ ˜ k N ( l k ) A k 1 / 2 ( 2 π ) d k + k / 2 N k / 2
Ignoring all terms that do not grow with N, MIBS can be approximated by the BIC [40]:
B I C ( k ) = j = 1 k λ j N / 2 σ ˜ k N ( l k ) / 2 N d k + k / 2
where N is the length of the data, l is the number of non-zero eigenvalues.
The BIC requires the number of mixed signals m must be larger than the number of source signals n. To address this problem in UBSS, the EMD decomposition algorithm is introduced in this study to construct a virtual multi-channel sensor. EMD, as an adaptive nonlinear and nonstationary signal analysis method, can decompose a complex single-component signal into a finite number of Intrinsic Mode Function (IMF). For the detailed theory of EMD, please refer to the reference [6].
The steps based on EMD and BIC to estimate the source number are as follows:
Step 1:
EMD decomposition. A single channel signal x i in mixed signal x ( t ) = [ x 1 ( t ) , x 2 ( t ) , , x m ( t ) ] T is selected as the EMD decomposition object to obtain a multi-channel signal x I M F ( t ) = [ c 1 , c 2 , , c n 1 , c n ] T , n is the number of IMF.
Step 2:
Calculate the covariance matrix. The covariance matrix of x I M F is calculated by R I M F = E [ x I M F ( t ) x I M F H ( t ) ] .
Step 3:
Singular value decomposition. Compute the singular values of R I M F to obtain M non-zero singular values in descending order Λ = { λ 1 , λ 2 , , λ M } .
Step 4:
Sources number estimation. The BIC(k) (k = 1, 2, …, M) of each eigenvalue is calculated, and the k corresponding to the largest BIC value is used as the number of the estimated sources.

3.3. Mixing Matrix Estimation by Combining Hierarchical Clustering and K-Means

First, the mixed signals are normalized and mapped to the upper-right half hypersphere:
x ˜ ( t , f ) = x ( t , f ) x ( t , f ) 2 × s i g n ( x 1 ( t , f ) )
Then, the mixed signal x ˜ ( t , f ) is clustered by the hierarchical clustering algorithm for the first time to obtain the cluster center C f i r s t   = [ c 1 , c 2 , , c k ] and its corresponding cluster G f i r s t   = [ g 1 , g 2 , , g k ] . Note that c k represents the cluster center corresponding to the cluster g k , and k represents the number of clusters.
Due to the influence of outliers such as noise, the computed cluster may deviate from the true cluster center. Therefore, the pairwise cosine distances dc between all data points in each cluster and its corresponding cluster center are calculated, and the data points that satisfy the following condition in (15) are taken to be outliers and are thus removed:
d c = 1 c i g i c i 2 g i 2 > Δ ω , i = 1 , 2 , , k
where Δ ω is a threshold, typically set to be close to zero. Clearly, the smaller the threshold Δ ω is, the more data points will be taken to be outliers and removed.
The new cluster center C n e w   = [ c n e w 1 , c n e w 2 , , c n e w k ] is obtained by calculating the mean value of each cluster G n e w =   [ g n e w 1 ,   g n e w 2 ,   ,   g n e w k ] after removing the outliers:
c n e w i = m e a n ( g n e w i ) ,   i = 1 , 2 , , k
Finally, Cnew is used as the initial clustering center of K-means, and the second clustering is performed by the improved K-means to get the clustering center Csecond, i.e., the estimated mixing matrix A ^ . The procedure of the proposed method is summarized in Algorithm 1.
Algorithm 1 The procedure of the proposed mixing matrix estimation algorithm.
Input: The observed signals x ( t ) , the sources number k, the constraint parameter Δ ω .
1:  Transform the observed signals x ( t ) into the TF domain using STFT to obtain the sparse signal x ( t , f ) .
2:  Normalize x ( t , f ) by Equation (14).
3:  Hierarchical clustering is used to cluster the normalized data to obtain the cluster center C f i r s t and cluster G f i r s t .
4:  Remove the outliers deviating from the clustering center C f i r s t via Equation (15) to get the new center G n e w .
5:  The cluster center C s e c o n d is obtained by calculating the mean value of each cluster g n e w i via Equation (16).
6:  Cluster by setting C n e w as the initial clustering center of K-means.
7:  Calculate the cluster centers of k clusters as the estimated mixing matrix A ^ .
Output: The estimated mixing matrix A ^ .
Combining the description of UBSS-SCA in Section 2, the proposed method is shown by the flowing flowchart.

4. Application Cases

4.1. Simulation Analysis

The simulated source signal s = [ s 1 , s 2 , s 3 ] T is constructed to verify the effectiveness of the proposed method, as shown in Equation (17) with the sampling frequency of 1000 Hz and sampling points of 1024. In all the experiments, the window length is 512, the number of overlapped samples is 508 and the Hanning window is selected as the window function, the SSPs threshold θ = 0.6 and the threshold ω = 0.0001 . The time domain waveform and Fourier spectrums of the three source signals are shown in Figure 2.
s 1 = sin ( 2 π f 1 t ) s 2 = 0 . 6 cos ( 2 π f 2 t + 10 ) s 3 = sin ( 2 π f 3 t + 0 . 2 cos ( 2 π f r t )
We note that, f 1 = 100 Hz , f 2 = 220 Hz , f 3 = 300 Hz , and f r = 20 Hz .
A random mixing matrix A is generated to obtain the mixed signal x(t) as follows:
A = 0.6986 0.5575 0.9295 0.7155 0.8301 0.3688
The mixed signal x(t) is constructed according to Equation (1), and 10% root mean square (RMS) Gaussian white noise with mean 0 and standard deviation 0.1 generated by a random function is added to the mixed signals. The time domain waveforms and Fourier spectrums of the mixed signal are shown in Figure 3:
x ( t ) = A s ( t ) + N o i s e
As shown in Figure 3, the three source signals are completely mixed in the mixed signal x1 and x2, and the waveforms of source signals in Figure 2a cannot be identified. In the spectra of mixed signals, the frequency components interfere with each other. It is also obvious that the 280 Hz and 320 Hz modulation frequency components are completely submerged in the spectrum of the mixed signal x2.
Firstly, the BIC is used to estimate the number of source signals. The mixed signal x1 is selected as a single channel signal and input to EMD for decomposition to obtain 9 modal components, and the multi-channel signal is constructed. Secondly, the covariance matrix of ximf is calculated, and SVD is used to decompose the matrix to obtain the singular values in descending order, as shown in Table 1.
Then, the eigenvalues are input to Equation (13), and the results are shown in Figure 4. When k = 3, the corresponding BIC value is the largest. Therefore, the number of source signals can be determined to be 3. The result is consistent with the number of source signals in Figure 2.
The mixed signals are transformed into the TF domain using the STFT, and the scatter plot is shown in Figure 5. It can be observed from Figure 5 that the signals present strong linear characteristics in the TF domain, while the distribution of scatter plots in the time domain is chaotic. The mixed signals in Figure 5b are approximately clustered into three straight lines, indicating that the mixed signals have stronger sparsity in the TF domain than in the time domain.
The SSPs are selected from the TF points to enhance the scatter plot’s linear characteristics and improve the accuracy of the estimated mixing matrix. Figure 6b is the scatter plot after SSPs processing. The linear characteristics between the mixed signals are clearer in Figure 6b after identifying the SSPs than in Figure 6a. Besides, there is no overlap between the frequency points in Figure 6a,b, which is beneficial for clustering.
After identifying the SSPs, the TF points are mapped to the upper right half hypersphere by Equation (14), and the normalized TF scatter plot is shown in Figure 7a. The boundary of the three clusters is not obvious, causing a great interference to cluster. Therefore, the outliers in Figure 7a are removed by Equation (15), and the obtained scatter plot is shown in Figure 7b. The clusters in Figure 7b are more compact than Figure 7a, and the boundary between each is clear.
The final estimated mixing matrix is obtained as follows:
A ^ = 0.7032 0.5566 0.9298 0.7110 0.8308 0.3680
The difference between the absolute values of the mixing matrix A and the estimated mixing matrix A ^ is shown in Equation (21), indicating that the estimated mixing matrix is very close to the original mixing matrix and the clustering accuracy of the proposed method is very high:
A A ^   =   0.0046 0.0009 0.0003 0.0045 0.0007 0.0008
After the mixing matrix is accurately obtained, the source signal is recovered by the least square method, and time domain waveforms and spectrums of the estimated source signals are shown in Figure 8. The results are obtained by calculating the correlation coefficients of the time domain waveforms in Figure 8a and Figure 2a, and the values are 0.9981, 0.9960, and 0.9981, which indicates that the estimated source signal s ^ t is very similar to the source signal s(t). The Fourier spectrums in Figure 8b show that the main frequencies of the source signal s(t) can be recovered accurately, thus verifying the effectiveness of the proposed method.

4.2. Comparision with the Traditional K-Means

To confirm the proposed method has superiority in the mixing matrix estimation and source signals, the traditional K-means algorithm is used in this study. As a comparison, the mixed signals in Figure 3 are also analyzed by the K-means. The estimated mixing matrix is given as follows:
A ^ = 0.7043 0.5703 0.9332 0.7099 0.8214 0.3592
Normalized mean square error (NMSE) and the deviation angle are used to evaluate the accuracy of the mixing matrix estimation [40,46], and the calculation formulas are shown in Equations (23) and (24), respectively:
NMSE = 10 log 10 i = 1 m j = 1 n ( a ^ i j a i j ) 2 i = 1 m j = 1 n a i j 2
where m and n represent the number of columns and rows of the mixing matrix A, respectively, a ^ i j and aij represent the elements of the mixing matrix A and the estimated mixing matrix A ^ in the i-th row and j-th column, respectively. The smaller the NMSE value is, the more accurate the estimated mixing matrix is:
a n g ( a , a ^ ) = 180 π cos 1 a T a ^ a · a ^
where a is the column vector of the original mixing matrix A and a ^ is the column vector of the estimated mixing matrix A ^ . The smaller the value of ang is, the more accurate the estimated mixing matrix is.
The above two mixing matrix evaluation criteria are used to evaluate the estimation accuracy of the mixing matrix between the traditional K-means method and the proposed method. The results are listed as follows.
Table 2 shows that the proposed method has the smallest NMSE value and the smallest angular deviation in each column, indicating that the proposed method is superior to the traditional method in the mixing matrix estimation.
The mixing matrix estimation is the most important step in UBSS, determining the performance of source recovery. After obtaining the mixing matrix, the source signals are recovered in the same way as in Section 4.1. The estimated sources are shown in Figure 9 and their Fourier spectrums are given in Figure 10c.
Comparing the estimated source signals in Figure 2a, Figure 8a and Figure 9, there is no obvious difference between these signals in terms of time domain waveforms. However, when comparing with their Fourier spectrums shown in Figure 10, we can find that the characteristic frequency of s ^ 2 and s ^ 3 in Figure 10c are disturbed by 300 Hz and 220 Hz, respectively. On the contrary, the proposed method estimated Fourier spectrums shown in Figure 10b are more similar to those of source signals in Figure 10a. Therefore, the results showed that the proposed method could outperform the traditional K-means algorithm.
To qualitatively measure the performance of the estimated source signals by the proposed method, the correlation coefficient and mean squared error (MSE) are used in this study. The correlation coefficient between the source signals and the estimated source signal is calculated by:
R = i = 1 n [ s i ( t ) s ¯ i ( t ) ] [ s ^ i ( t ) s ^ ¯ i ( t ) ] T s i ( t ) s ¯ i ( t ) 2 s ^ i ( t ) s ^ ¯ i ( t ) 2
where s ¯ i and s ^ ¯ i denote the means of the signal si and s ^ i , respectively, and n is the number of the sources. The MSE is defined as follows:
MSE = 10 log 10 1 n i = 1 n min δ s i δ s ^ i 2 2 s i 2 2
where s ^ i denotes the estimation of the source signal si, and δ is a scalar reflecting the scalar ambiguity.
Table 3 shows the performance comparison of the source recovery with the evaluation criteria of correlation coefficients and MSE. As can be seen in Table 3, the correlation coefficient of each estimated source signal is larger than that of the traditional K-means. The MSE of the recovered sources by the K-means and the proposed method are −14.8696 dB and −22.8119 dB, which indicates that the recovered error by the proposed method is smaller. The above results validate that the improved K-means algorithm is superior to the traditional K-means. Moreover, as revealed in Table 2 and Table 3, the accuracy of the source signals recovery greatly depends on the estimation accuracy of the mixing matrix. Therefore, the mixing matrix estimation plays a more decisive role in the UBSS.

4.3. Experiment and Discussion

The reciprocating compressor has been widely applied in the petroleum and chemical industry, and their operating state and security are thought to be challenging research subjects. In this paper, a 2D12-70/0.1~13 double-acting reciprocating compressor was studied with the UBSS.
The shaft power of the reciprocating compressor is 500 kW, the exhaust volume is 70 m3/min, the exhaust pressure is 0.2746–0.2942 MPa, the piston stroke is 240 mm, the motor speed is 496 r/min, and the corresponding motor rotation frequency is 8.27 Hz. Photos of the test bench and the schematic diagram of the reciprocating compressor transmission mechanism are displayed in Figure 11 and Figure 12, respectively. As shown in Figure 12, the reciprocating compressor is driven by the machine to rotate the main shaft and transmit the power to the crankcase, which then drives the connecting rod, crosshead, and piston to make the reciprocating motion in the work engineering. However, due to manufacturing tolerances, defects, or after working for a while, the sliding bearings between the big end of the connecting rod and the crankshaft, and the small end of the connecting rod and the crosshead are susceptible to wear because of the long-term action of the reciprocating force, the bearing clearance is too large, resulting in equipment failure. Therefore, the fault diagnosis of the reciprocating compressor is carried out by combining UBSS and refined composite multiscale fuzzy entropy (RCMFE) [47].
Bearing bush wear is a long-term process, but if not detected and replaced early enough, it may lead to huge economic losses and even casualties. The early fault signal is weak and often drowned in noise, which makes it difficult to extract the fault characteristic information effectively. Therefore, this study takes the early weak fault signal of the bearing bush as the research object to verify the effectiveness of the proposed method, the connecting rod and its components are shown in Figure 13. After working for a while, the normal bearing bush in Figure 13d will wear out and then becomes the fault state in Figure 13c, resulting in a clearance between the journal and bearing bush, and the machine will produce abnormal sound and vibration. During the test, for the first-stage connecting rod big end fault state, two sensor measurement points were arranged on the compressor for data acquisition, located at the upper end of the crosshead slide and the top of the crankcase. The specific installation locations are shown in the red triangle in Figure 14. The vibration signals were collected with the Beijing Oriental Institute’s INV306U-6660 multi-channel intelligent data acquisition instrument. As shown in Figure 11b,c, the integrated circuit piezoelectric (ICP) acceleration sensors (CT1010LC) were installed in the red rectangular box with a sensitivity of 100 mv/g, a measurement range of ± 50 g, and a frequency range of 0.5~5 kHz. The sampling frequency is 50 kHz, and the test lasts for 4 s. The time domain waveforms and envelope spectrums of the acquired mixed signal x = [x1,x2]T are shown in Figure 15.
Figure 15a shows that the mixed signal xi waveform is more chaotic with no obvious characteristic information due to the mutual coupling of multiple excitation sources of the reciprocating compressor possibly. The waveform of mixed signal x2 shows an obvious periodic shock component caused by some fault possibly. In the envelope spectrums of Figure 15b, the double frequency of 16.78 Hz is obvious, which is determined by the mechanism of the reciprocating compressor. Because under the action of reciprocating inertial force, there will be two collisions in a cycle, and thus the double frequency of the envelope spectrum will have a peak.
According to the flowchart in Figure 1, the source signals are recovered using the proposed UBSS method. The results of the estimated source signal s ^ = [ s ^ 1 , s ^ 2 , s ^ 3 ] T are shown in Figure 16. From Figure 16a, we can see that the estimated source signal s ^ 1 is a periodic excitation source, the s ^ 2 has obvious periodic impact characteristics, and the s ^ 3 is chaotic. From the envelope spectrums in Figure 16b, the first two signals have distinct double frequency components of the rotation frequency, and that characteristic frequency can not be found in s ^ 3 .
As a comparison, the mixing matrix was estimated by the traditional K-means algorithm, and then the sources were recovered in the same way shown. The time waveforms and their envelope spectrums were shown in Figure 17 and Figure 18b. From Figure 17, we can see that there is no obvious difference between the two methods. To facilitate comparison and analysis of the effects of the proposed method and the K-means in the frequency domain, the envelope spectrums in Figure 16b were put together with that of K-means in Figure 18.
Comparing with Figure 18a,b, the first two signals estimated by the K-means have lost some amplitude information, resulting in their characteristic frequencies being interfered with by other components, especially the signal s ^ 2 is submerged in other frequency components, and the s ^ 3 is also chaotic. Therefore, this suggests that the proposed method could outperform the contrast method in feature extraction.
Different from the fault diagnosis of the rotating machinery, the type of the estimated source signals in the reciprocating compressor can not be identified from the envelope spectrums because of the complex structure of reciprocating compressors, various forms of component movements (e.g., reciprocating and rotating movements), and the multi-component coupled signals of vibration signals. Therefore, the fault diagnosis of the reciprocating compressor is performed by comparing the collected signals with those in a standard database, which is obtained by long-term monitoring in the laboratory, serving as a standard state for fault decisions.
As a nonlinear dynamical method, information entropy can measure the degree of uncertainty of things and portray the complexity of systems, which is widely used in machinery fault diagnosis [5,48,49]. Therefore, the RCMFE-based information entropy is introduced to explore the deep fault information hidden in the vibration waveform. Firstly, the RCMFE values of the normal state signal, the first-stage connecting rod big end fault signal, and the first-stage connecting rod small end fault signal in the standard database are calculated, as shown by the solid lines in Figure 19. From Figure 19, the curves of the three states are clearly distinguished. Under normal conditions, the entropy of the collected signal is large because of the great unpredictability of the system operating state and the small self-similarity of the vibration signal. The malfunction of the equipment component generates a periodic shock, a strong self-similarity of the vibration signal, a low degree of confusion, and a small entropy value. Since the failure mechanism of the connecting rod big end and the connecting rod small end are different, their entropy values are not the same in Figure 19. The dotted lines in Figure 19a represent the RCMFE values of the estimated source signals in Figure 16a, where the blue curve has the largest entropy value, indicating the greatest degree of confusion. s ^ 3 is determined as the environmental noise with the analysis of the source signal vibration waveform estimated in Figure 16a. The green entropy curve has a low entropy value at the first two scales, while its entropy value is close to that of the noisy signal s ^ 3 at most scale factors. Further analysis of the estimated source signal combined with the Figure 16a suggests that the estimated source signal s ^ 1 is a signal coupled with multiple excitation source components. The red entropy curve under different scale factors τ displays the same trend as the black curve in the standard database, indicating that both curves contain the same fault characteristic information and s ^ 2 is a connecting rod big end fault. However, the red curve has a lower entropy value than the black curve because the signal contains other excitation sources such as noise due to the influence of the transmission path and other factors in the acquisition process. Besides, the UBSS can filter out some noise and excitation sources in the source signals separation process, making the estimated source signal s ^ 2 purer so that its disorder degree is lower and entropy value is smaller than the big end fault signal in the standard database, but the overall trend is still consistent with the big end fault entropy curve. As a comparison, the results by the K-means method were also calculated shown in Figure 19b. From Figure 19a,b, we can find that the results of the two methods are basically consistent. However, the entropy values of the estimated signal s ^ 2 by the K-means are larger on all scales, and its disorder degree is greater than that of s ^ 2 by the proposed method, indicating that it is disturbed by noise, which is consistent with the previous analysis results in Figure 18b.

5. Conclusions

This paper proposed a two-stage clustering method by combining hierarchical clustering and K-means to estimate the mixing matrix from their instantaneous mixtures in underdetermined systems. Our method can remove some outliers existing in the observed signals and then obtain the accurate initial clustering centers when compared with the K-means whose initial clustering centers are randomly generated.
The simulation analysis showed that the NMSE value of the estimated mixing matrix by the proposed method was smaller than that of the traditional K-means and the correlation coefficient of the estimated source signal was larger, indicating the proposed method could outperform the traditional K-means. The application of reciprocating compressor showed that the improved UBSS method by combining RCMFE could accurately extract fault features and identify the fault location. Although the proposed method can improve the accuracy of the mixing matrix estimation, there are still some limitations. The two-stage clustering is more time-consuming than other single clustering algorithms. And the proposed method is only employed in the underdetermined mixture system, where at least two sensors are needed. In future work, we will focus on the sparse representation of sources and applying fewer samples to cluster and reduce the computing time.

Author Contributions

Conceptualization, J.W. and X.C.; software, X.C.; validation, J.W., H.Z. and Y.L.; formal analysis, H.Z.; investigation, Y.L.; writing—original draft preparation, J.W.; writing—review and editing, X.C. and Z.L.; supervision, H.Z.; project administration, J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Heilongjiang Natural Science Foundation (No. LH2021E021), in part by the Northeast Petroleum University Youth Foundation (No. 2018ANC-31), and in part by the National Natural Science Foundation of China (No. 51505079).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used or analyzed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Haiyang, Z.; Jindong, W.; Lee, J.; Ying, L. A Compound Interpolation Envelope Local Mean Decomposition and Its Application for Fault Diagnosis of Reciprocating Compressors. Mech. Syst. Signal Process. 2018, 110, 273–295. [Google Scholar] [CrossRef]
  2. Li, G.; Tang, G.; Luo, G.; Wang, H. Underdetermined Blind Separation of Bearing Faults in Hyperplane Space with Variational Mode Decomposition. Mech. Syst. Signal Process. 2019, 120, 83–97. [Google Scholar] [CrossRef]
  3. Cabrera, D.; Guamán, A.; Zhang, S.; Cerrada, M.; Sanchez, R.-V.; Cevallos, J.; Long, J.; Li, C. Bayesian Approach and Time Series Dimensionality Reduction to LSTM-Based Model-Building for Fault Diagnosis of a Reciprocating Compressor. Neurocomputing 2020, 380, 51–66. [Google Scholar] [CrossRef]
  4. Mondal, D.; Zhen, D.; Gu, F.; Ball, A.D. Fault diagnosis of reciprocating compressor using empirical mode decomposition-based Teager energy spectrum of airborne acoustic signal. In Advances in Asset Management and Condition Monitoring; Springer: Berlin/Heidelberg, Germany, 2020; pp. 939–952. [Google Scholar] [CrossRef]
  5. Sharma, V.; Parey, A. Performance Evaluation of Decomposition Methods to Diagnose Leakage in a Reciprocating Compressor under Limited Speed Variation. Mech. Syst. Signal Process. 2019, 125, 275–287. [Google Scholar] [CrossRef]
  6. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.-C.; Tung, C.C.; Liu, H.H. The Empirical Mode Decomposition and the Hilbert Spectrum for Nonlinear and Non-Stationary Time Series Analysis. Proc. R. Soc. Lond. Ser. Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  7. Hoseinzadeh, M.S.; Khadem, S.E.; Sadooghi, M.S. Quantitative Diagnosis for Bearing Faults by Improving Ensemble Empirical Mode Decomposition. ISA Trans. 2018, 83, 261–275. [Google Scholar] [CrossRef] [PubMed]
  8. Smith, J.S. The Local Mean Decomposition and Its Application to EEG Perception Data. J. R. Soc. Interface 2005, 2, 443–454. [Google Scholar] [CrossRef]
  9. Zhang, L.; Wang, Z.; Quan, L. Research on Weak Fault Extraction Method for Alleviating the Mode Mixing of LMD. Entropy 2018, 20, 387. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Dragomiretskiy, K.; Zosso, D. Variational Mode Decomposition. IEEE Trans. Signal Process. 2013, 62, 531–544. [Google Scholar] [CrossRef]
  11. Liang, T.; Lu, H.; Sun, H. Application of Parameter Optimized Variational Mode Decomposition Method in Fault Feature Extraction of Rolling Bearing. Entropy 2021, 23, 520. [Google Scholar] [CrossRef]
  12. Kaur, C.; Bisht, A.; Singh, P.; Joshi, G. EEG Signal Denoising Using Hybrid Approach of Variational Mode Decomposition and Wavelets for Depression. Biomed. Signal Process. Control 2021, 65, 102337. [Google Scholar] [CrossRef]
  13. Holobar, A.; Farina, D. Noninvasive Neural Interfacing With Wearable Muscle Sensors: Combining Convolutive Blind Source Separation Methods and Deep Learning Techniques for Neural Decoding. IEEE Signal Process. Mag. 2021, 38, 103–118. [Google Scholar] [CrossRef]
  14. Pezzoli, M.; Carabias-Orti, J.J.; Cobos, M.; Antonacci, F.; Sarti, A. Ray-Space-Based Multichannel Nonnegative Matrix Factorization for Audio Source Separation. IEEE Signal Process. Lett. 2021, 28, 369–373. [Google Scholar] [CrossRef]
  15. Zhen, L.; Peng, D.; Zhang, H.; Sang, Y.; Zhang, L. Underdetermined Mixing Matrix Estimation by Exploiting Sparsity of Sources. Measurement 2020, 152, 107268. [Google Scholar] [CrossRef]
  16. Zhao, X.; Qin, Y.; He, C.; Jia, L. Underdetermined Blind Source Extraction of Early Vehicle Bearing Faults Based on EMD and Kernelized Correlation Maximization. J. Intell. Manuf. 2020, 1–17. [Google Scholar] [CrossRef]
  17. Ho, S.K.; Nedunuri, H.C.; Balachandran, W.; Kanfoud, J.; Gan, T.-H. Monitoring of Industrial Machine Using a Novel Blind Feature Extraction Approach. Appl. Sci. 2021, 11, 5792. [Google Scholar] [CrossRef]
  18. Zhou, P.; Zhang, R.; Xie, J.; Liu, J.; Wang, H.; Chai, T. Data-Driven Monitoring and Diagnosing of Abnormal Furnace Conditions in Blast Furnace Ironmaking: An Integrated PCA-ICA Method. IEEE Trans. Ind. Electron. 2020, 68, 622–631. [Google Scholar] [CrossRef]
  19. Miao, F.; Zhao, R.; Jia, L.; Wang, X. Multisource Fault Signal Separation of Rotating Machinery Based on Wavelet Packet and Fast Independent Component Analysis. Int. J. Rotating Mach. 2021, 2021. [Google Scholar] [CrossRef]
  20. He, C.; Li, H.; Zhao, X. Weak Characteristic Determination for Blade Crack of Centrifugal Compressors Based on Underdetermined Blind Source Separation. Measurement 2018, 128, 545–557. [Google Scholar] [CrossRef]
  21. Babatas, E.; Erdogan, A.T. Time and Frequency Based Sparse Bounded Component Analysis Algorithms for Convolutive Mixtures. Signal Process. 2020, 173, 107590. [Google Scholar] [CrossRef]
  22. Xiao, Y.; Lu, W.; Yan, Q.; Zhang, H. Blind Separation of Coherent Multipath Signals with Impulsive Interference and Gaussian Noise in Time-Frequency Domain. Signal Process. 2021, 178, 107750. [Google Scholar] [CrossRef]
  23. Eqlimi, E.; Makkiabadi, B.; Samadzadehaghdam, N.; Khajehpour, H.; Mohagheghian, F.; Sanei, S. A Novel Underdetermined Source Recovery Algorithm Based on K-Sparse Component Analysis. Circuits Syst. Signal Process. 2019, 38, 1264–1286. [Google Scholar] [CrossRef] [Green Version]
  24. Xie, Y.; Xie, K.; Xie, S. Underdetermined Blind Separation of Source Using Lp-Norm Diversity Measures. Neurocomputing 2020, 411, 259–267. [Google Scholar] [CrossRef]
  25. Lu, J.; Cheng, W.; He, D.; Zi, Y. A Novel Underdetermined Blind Source Separation Method with Noise and Unknown Source Number. J. Sound Vib. 2019, 457, 67–91. [Google Scholar] [CrossRef]
  26. Bofill, P.; Zibulevsky, M. Underdetermined Blind Source Separation Using Sparse Representations. Signal Process. 2001, 81, 2353–2362. [Google Scholar] [CrossRef] [Green Version]
  27. Hao, Y.; Song, L.; Ren, B.; Wang, H.; Cui, L. Step-by-Step Compound Faults Diagnosis Method for Equipment Based on Majorization-Minimization and Constraint SCA. IEEE/ASME Trans. Mechatron. 2019, 24, 2477–2487. [Google Scholar] [CrossRef]
  28. Xu, Y.; Brownjohn, J.M.; Hester, D. Enhanced Sparse Component Analysis for Operational Modal Identification of Real-Life Bridge Structures. Mech. Syst. Signal Process. 2019, 116, 585–605. [Google Scholar] [CrossRef] [Green Version]
  29. Cheng, W.; Jia, Z.; Chen, X.; Han, L.; Zhou, G.; Gao, L. Underdetermined Convolutive Blind Source Separation in the Time–Frequency Domain Based on Single Source Points and Experimental Validation. Meas. Sci. Technol. 2020, 31, 095001. [Google Scholar] [CrossRef]
  30. Sgouros, T.; Mitianoudis, N. A Novel Directional Framework for Source Counting and Source Separation in Instantaneous Underdetermined Audio Mixtures. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 2025–2035. [Google Scholar] [CrossRef]
  31. Askari, S. Fuzzy C-Means Clustering Algorithm for Data with Unequal Cluster Sizes and Contaminated with Noise and Outliers: Review and Development. Expert Syst. Appl. 2021, 165, 113856. [Google Scholar] [CrossRef]
  32. Rezaee, M.J.; Eshkevari, M.; Saberi, M.; Hussain, O. GBK-Means Clustering Algorithm: An Improvement to the K-Means Algorithm Based on the Bargaining Game. Knowl.-Based Syst. 2021, 213, 106672. [Google Scholar] [CrossRef]
  33. Chowdhury, K.; Chaudhuri, D.; Pal, A.K. An Entropy-Based Initialization Method of K-Means Clustering on the Optimal Number of Clusters. Neural Comput. Appl. 2021, 33, 6965–6982. [Google Scholar] [CrossRef]
  34. Lu, J.; Cheng, W.; Zi, Y. A Novel Underdetermined Blind Source Separation Method and Its Application to Source Contribution Quantitative Estimation. Sensors 2019, 19, 1413. [Google Scholar] [CrossRef] [Green Version]
  35. Hu, C.; Yang, Q.; Huang, M.; Yan, W. Sparse Component Analysis-Based under-Determined Blind Source Separation for Bearing Fault Feature Extraction in Wind Turbine Gearbox. IET Renew. Power Gener. 2017, 11, 330–337. [Google Scholar] [CrossRef]
  36. Jun, H.; Chen, Y.; Zhang, Q.-H.; Sun, G.; Hu, Q. Blind Source Separation Method for Bearing Vibration Signals. IEEE Access 2017, 6, 658–664. [Google Scholar] [CrossRef]
  37. Xie, H.; Zhang, L.; Lim, C.P.; Yu, Y.; Liu, C.; Liu, H.; Walters, J. Improving K-Means Clustering with Enhanced Firefly Algorithms. Appl. Soft Comput. 2019, 84, 105763. [Google Scholar] [CrossRef]
  38. Arthur, D.; Vassilvitskii, S. K-Means++: The Advantages of Careful Seeding. In Proceedings of the 18th Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA, 7–9 January 2007; pp. 1027–1035. [Google Scholar]
  39. He, X.; He, F.; Cai, W. Underdetermined BSS Based on K-Means and AP Clustering. Circuits Syst. Signal Process. 2016, 35, 2881–2913. [Google Scholar] [CrossRef]
  40. Reju, V.G.; Koh, S.N.; Soon, Y. An Algorithm for Mixing Matrix Estimation in Instantaneous Blind Source Separation. Signal Process. 2009, 89, 1762–1773. [Google Scholar] [CrossRef]
  41. Amini, F.; Hedayati, Y. Underdetermined Blind Modal Identification of Structures by Earthquake and Ambient Vibration Measurements via Sparse Component Analysis. J. Sound Vib. 2016, 366, 117–132. [Google Scholar] [CrossRef]
  42. Belouchrani, A.; Amin, M.G. Blind Source Separation Based on Time-Frequency Signal Representations. IEEE Trans. Signal Process. 1998, 46, 2888–2897. [Google Scholar] [CrossRef]
  43. Hartigan, J.A.; Wong, M.A. Algorithm AS 136: A K-Means Clustering Algorithm. Appl. Stat. 1979, 28, 100. [Google Scholar] [CrossRef]
  44. Zhen, L.; Peng, D.; Yi, Z.; Xiang, Y.; Chen, P. Underdetermined Blind Source Separation Using Sparse Coding. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 3102–3108. [Google Scholar] [CrossRef] [PubMed]
  45. Minka, T. Automatic Choice of Dimensionality for PCA. Adv. Neural Inf. Process. Syst. 2000, 13, 598–604. [Google Scholar]
  46. Sun, J.; Li, Y.; Wen, J.; Yan, S. Novel Mixing Matrix Estimation Approach in Underdetermined Blind Source Separation. Neurocomputing 2016, 173, 623–632. [Google Scholar] [CrossRef]
  47. Azami, H.; Fernández, A.; Escudero, J. Refined Multiscale Fuzzy Entropy Based on Standard Deviation for Biomedical Signal Analysis. Med. Biol. Eng. Comput. 2017, 55, 2037–2052. [Google Scholar] [CrossRef]
  48. Rodriguez, N.; Alvarez, P.; Barba, L.; Cabrera-Guerrero, G. Combining Multi-Scale Wavelet Entropy and Kernelized Classification for Bearing Multi-Fault Diagnosis. Entropy 2019, 21, 152. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Minhas, A.S.; Singh, G.; Singh, J.; Kankar, P.K.; Singh, S. A Novel Method to Classify Bearing Faults by Integrating Standard Deviation to Refined Composite Multi-Scale Fuzzy Entropy. Measurement 2020, 154, 107441. [Google Scholar] [CrossRef]
Figure 1. The flow chart of fault feature extraction based on underdetermined blind source separation.
Figure 1. The flow chart of fault feature extraction based on underdetermined blind source separation.
Entropy 23 01217 g001
Figure 2. Source signals: (a) Waveforms; (b) Fourier spectrums.
Figure 2. Source signals: (a) Waveforms; (b) Fourier spectrums.
Entropy 23 01217 g002
Figure 3. Mixed signals: (a) Waveforms; (b) Fourier spectrums.
Figure 3. Mixed signals: (a) Waveforms; (b) Fourier spectrums.
Entropy 23 01217 g003
Figure 4. Source number estimation by BIC.
Figure 4. Source number estimation by BIC.
Entropy 23 01217 g004
Figure 5. The scatter plot of mixed signals: (a) Time domain; (b) TF domain by STFT.
Figure 5. The scatter plot of mixed signals: (a) Time domain; (b) TF domain by STFT.
Entropy 23 01217 g005
Figure 6. The scatter plot of mixed signals in TF domain: (a) Before SSPs; (b) After SSPs.
Figure 6. The scatter plot of mixed signals in TF domain: (a) Before SSPs; (b) After SSPs.
Entropy 23 01217 g006
Figure 7. Normalized TF scatter plot: (a) Scatter plot before processing; (b) Scatter plot after using cosine distance to remove the outliers based on (a).
Figure 7. Normalized TF scatter plot: (a) Scatter plot before processing; (b) Scatter plot after using cosine distance to remove the outliers based on (a).
Entropy 23 01217 g007
Figure 8. Estimated source signals: (a) Waveforms; (b) Fourier spectrums.
Figure 8. Estimated source signals: (a) Waveforms; (b) Fourier spectrums.
Entropy 23 01217 g008
Figure 9. Estimated source signals by the K-means.
Figure 9. Estimated source signals by the K-means.
Entropy 23 01217 g009
Figure 10. Comparison of the estimated source signals in the frequency domain: (a) Source signals; (b) The proposed method; (c) K-means.
Figure 10. Comparison of the estimated source signals in the frequency domain: (a) Source signals; (b) The proposed method; (c) K-means.
Entropy 23 01217 g010
Figure 11. A two-stage double-acting reciprocating compressor of type 2D12: (a) Photos of the test bench; (b) Sensor on the top of the crankcase; (c) Sensor on the top of the crosshead guide surface.
Figure 11. A two-stage double-acting reciprocating compressor of type 2D12: (a) Photos of the test bench; (b) Sensor on the top of the crankcase; (c) Sensor on the top of the crosshead guide surface.
Entropy 23 01217 g011
Figure 12. Schematic diagram of the reciprocating compressor transmission mechanism.
Figure 12. Schematic diagram of the reciprocating compressor transmission mechanism.
Entropy 23 01217 g012
Figure 13. Photos of connecting rod: (a) Connecting rod; (b) Normal state; (c) Big end fault bearing; (d) Bearing bush.
Figure 13. Photos of connecting rod: (a) Connecting rod; (b) Normal state; (c) Big end fault bearing; (d) Bearing bush.
Entropy 23 01217 g013
Figure 14. The structural draw of reciprocating compressor and measuring points.
Figure 14. The structural draw of reciprocating compressor and measuring points.
Entropy 23 01217 g014
Figure 15. Mixed signals: (a) Waveforms; (b) Envelope spectrums.
Figure 15. Mixed signals: (a) Waveforms; (b) Envelope spectrums.
Entropy 23 01217 g015
Figure 16. Estimated source signals: (a) Waveforms; (b) Envelope spectrums.
Figure 16. Estimated source signals: (a) Waveforms; (b) Envelope spectrums.
Entropy 23 01217 g016
Figure 17. Estimated source signals by the K-means.
Figure 17. Estimated source signals by the K-means.
Entropy 23 01217 g017
Figure 18. Comparison of the estimated source signals in the frequency domain: (a) The proposed method; (b) The K-means.
Figure 18. Comparison of the estimated source signals in the frequency domain: (a) The proposed method; (b) The K-means.
Entropy 23 01217 g018
Figure 19. RCMFE analysis of the estimated source signals by different method. (a) The proposed method; (b) The K-means.
Figure 19. RCMFE analysis of the estimated source signals by different method. (a) The proposed method; (b) The K-means.
Entropy 23 01217 g019
Table 1. Eigenvalue of the x i m f .
Table 1. Eigenvalue of the x i m f .
SVD λ 1 λ 2 λ 3 λ 4 λ 5 λ 6 λ 7 λ 8 λ 9
Value0.49850.30490.00170.000860.000140.000050.000040.000040.00002
Table 2. Angular differences and NMSE of K-means and the proposed method.
Table 2. Angular differences and NMSE of K-means and the proposed method.
MethodsAngular DifferenceNMSE (dB)
a n g ( a 1 , a ^ 1 ) a n g ( a 2 , a ^ 2 ) a n g ( a 3 , a ^ 3 )
K-means0.45780.88680.5895−38.6513
Proposed method0.37150.06200.0515−48.3349
Table 3. Performance comparison of the estimated source signals with correlation coefficients R and MSE.
Table 3. Performance comparison of the estimated source signals with correlation coefficients R and MSE.
MethodsCorrelation Coefficients RMSE (dB)
< s 1 ,   s ^ 1 > < s 2 ,   s ^ 2 > < s 3 ,   s ^ 3 >
K-means0.99560.98710.9927−17.8696
Proposed method0.99810.99600.9981−22.8119
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, J.; Chen, X.; Zhao, H.; Li, Y.; Liu, Z. Fault Feature Extraction for Reciprocating Compressors Based on Underdetermined Blind Source Separation. Entropy 2021, 23, 1217. https://0-doi-org.brum.beds.ac.uk/10.3390/e23091217

AMA Style

Wang J, Chen X, Zhao H, Li Y, Liu Z. Fault Feature Extraction for Reciprocating Compressors Based on Underdetermined Blind Source Separation. Entropy. 2021; 23(9):1217. https://0-doi-org.brum.beds.ac.uk/10.3390/e23091217

Chicago/Turabian Style

Wang, Jindong, Xin Chen, Haiyang Zhao, Yanyang Li, and Zujian Liu. 2021. "Fault Feature Extraction for Reciprocating Compressors Based on Underdetermined Blind Source Separation" Entropy 23, no. 9: 1217. https://0-doi-org.brum.beds.ac.uk/10.3390/e23091217

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop