Brought to you by:
Paper

Remaining useful life prediction for bearing based on automatic feature combination extraction and residual multi-Head attention GRU network

, , and

Published 22 December 2023 © 2023 IOP Publishing Ltd
, , Citation Jiawen He et al 2024 Meas. Sci. Technol. 35 036003 DOI 10.1088/1361-6501/ad1652

0957-0233/35/3/036003

Abstract

Rolling bearings are indispensable parts in mechanical equipment, and predicting their remaining useful life is critical to normal operation and keep equipment in good repair. However, the complex characteristics of bearings make it difficult to describe their degradation characteristics. To address this issue, a novel method that combines an automatic feature combination extraction mechanism with a gated recurrent unit (GRU) network that has a residual multi-head attention mechanism for rolling bearing life prediction is proposed. Firstly, the automatic feature combination extraction mechanism is used to learn the degradation representation of the bearing vibration signal in the time domain, frequency domain, and time–frequency joint domain, and automatically extract the optimal bearing degradation feature combination. Then, the GRU network with residual multi-head attention mechanism is developed to weight and distinguish the learned degradation features, thereby improving the network's attention to important bearing degradation features. In the end, the proposed method is validated on the prediction and the health management of systems dataset and compared to other advanced approaches. The experimental results show that the proposed method can effectively capture the complex and dynamic features of rolling bearings and has high accuracy and generalization ability in rolling bearing life prediction.

Export citation and abstract BibTeX RIS

1. Introduction

The Industrial Internet of Things and modern technology have made fault prediction and the health management of systems (PHM) common in industrial system health management [1]. As an important mechanical component, rolling bearings find wide applications in rotating machinery across various fields such as electrical industry, machine tools, and automobiles. The long-term use of rolling bearings may cause bearing performance degradation and noise increase due to various factors. If not detected and dealt with in time, the bearings may suddenly fail during use, causing equipment damage or even endangering personnel safety [2]. Rolling bearing remaining useful life (RUL) represents the remaining time or service life of a bearing that can work safely until failure under the current working condition. In order to diagnose possible faults and damages in advance, carry out maintenance and replacement work, and avoid accidents or production losses caused by equipment downtime during operation, rolling bearing RUL prediction comes into being [3, 4]. Therefore, RUL prediction of rolling bearings is a significant technology in modern intelligent manufacturing.

Regarding the RUL prediction of rolling bearings, the current research methods mainly include model-based methods, data-driven methods, and hybrid methods [5]. Model-based methods establish mathematical models of bearing operation to analyze the influencing factors of bearing damage and failure, facilitating accurate RUL prediction [6]. For example, the finite element method and system identification technology are utilized to establish bearing operation models, which are combined with experimental data for calibration and verification [7, 8]. In the early stages of industrial development, model-based methods demonstrated better results. However, as industrial systems become increasingly complex, this method is difficult to meet the needs of industrial applications. Data-driven methods use machine learning algorithms to comprehend the inherent relationship between bearing operational status and lifespan from massive bearing data, and use the learned model for RUL prediction [9]. For instance, support vector machines [10], ANN [11], and hidden Markov models [12] are commonly employed for feature extraction and modeling. Recent breakthroughs in communication technology and networks have significantly lowered the threshold for data-driven methods, and the required professional knowledge and expert experience are relatively less, thus making this method gradually receive attention in the research field. Hybrid methods combine the characteristics of these two methods, but its application on industrial systems still faces challenges [13, 14]. This paper chooses data-driven methods that are more suitable for the current era as the research method.

Currently, data-driven RUL prediction methods can be mainly divided into two-stage and end-to-end approaches. The two-stage rolling bearing RUL prediction method decomposes the RUL prediction task into two stages for processing [15, 16]. Firstly, the rolling bearing degradation state is modeled to obtain its corresponding feature representation; then, the RUL prediction model takes the feature representation as input and produces the RUL prediction output of the rolling bearing. End-to-end RUL prediction methods for rolling bearings directly use raw data as input to the prediction model, simultaneously achieving degradation state representation and RUL prediction tasks [17, 18]. End-to-end models are often intricate, requiring an extensive amount of time and data for training. Conversely, two-stage methods break down the task into two parts for processing, enabling the use of fewer data and computing resources at each stage to improve the model robustness and generalizability. Moreover, the flexibility of the two-stage methods lies in their ability to adopt different feature extraction and preprocessing methods to suit various data types and distributions. Therefore, a data-driven two-stage framework is adopted in this paper to research the RUL prediction of rolling bearings.

As one of the most classical data-driven methods, remarkable performance has been demonstrated through deep learning in numerous prediction tasks [19, 20]. The common deep neural networks for RUL prediction are mainly recurrent neural networks (RNNs) and convolutional neural networks (CNNs). Research on CNNs is extensive, including traditional convolutional networks and time sequence convolutional network [21]. Cao et al [22] proposed a time convolutional network with residual self-attention mechanism and verified it on multiple datasets. However, only marginal spectrum is extracted in this study, which contains relatively limited degradation information. Similarly, Song et al [23] proposed a time convolutional network based on distributed attention for RUL prediction of aircraft engines, which uses a single attention mechanism and has doubts about whether the information contained in multi-modal data can be fully exploited. On the other hand, Jiang et al [24] proposed a CNN for RUL prediction of rolling bearings based on dual attention mechanism. The feature combination used in this study includes both time domain and frequency domain features, but no screening process was applied to the feature combination. Yang et al [25] proposed an intelligent RUL prediction method based on dual CNN model architecture, which does not need to extract features in advance, but directly inputs the original vibration signal into the network, and uses the powerful feature extraction ability of CNN to learn degradation features. However, large-scale raw vibration signal data may lead to a heavy computational burden for network training. Another approach involves the use of RNN, which is a type of neural network with the capability to model and learn from time series data. As rolling bearing RUL prediction has strong temporal characteristics, the utilization of RNN in RUL prediction is widely prevalent [26, 27]. Guo et al [28] proposed a predictive model based on multi-scale gated CNN and Transformer to address the rolling bearing RUL prediction problem. This method employs CNN and gated recurrent unit (GRU) for feature extraction, followed by prediction using a Transformer model. However, it exhibits relatively slower training speeds. Another study, introduced by Shen et al [29], utilizes a combination of multi-head attention mechanisms (MHA) and Bi-long-short term memory (Bi-LSTM) for bearing RUL research. In terms of feature selection, degenerate feature combinations are selected by considering monotonicity and trend. Furthermore, Liu et al [30] introduced a multi-head neural network model with asymmetric constraints. This research combines Bi-GRU with self-attention mechanisms and employs an ensemble of multiple subnetworks for learning. While this method demonstrates significant effectiveness in focusing on the degradation features with diverse transformations, it has relatively long training times. Finally, Qin et al [31] proposed a GRU neural network with dual attention gates for rolling bearing RUL prediction. However, this study only considered the root mean square of the signals as the network's input, which may not be sufficient for extracting degradation features comprehensively. According to the above literature, the existing bearing RUL prediction methods still have some limitations.

  • (1)  
    Most of the current deep learning-based methods for predicting bearing RUL mainly concentrate on feature extraction and modeling, while ignoring the importance of feature combination. Feature combination can effectively integrate multiple types of features and improve the representation ability of features, thereby enhancing the accuracy of RUL prediction.
  • (2)  
    Some methods rely on one type of manually crafted features, which may not effectively capture the complex patterns in vibration signals. Other methods only consider single-type attention mechanisms, failing to fully utilize the information contained in multimodal data. These methods have limitations in capturing the complex and dynamic characteristics of rolling bearings.

A RUL prediction method is presented in this paper, which integrates an automatic feature combination extraction mechanism and a GRU network with residual MHA to mitigate the aforementioned issues. The main contributions of this paper are as follows:

  • (1)  
    A method for automatically extracting feature combinations is proposed in this paper. The method calculates a series of candidate features and then automatically selects a combination of features that better expresses bearing degradation according to a random forest regression-based method.
  • (2)  
    A GRU network model with residual MHA is proposed. The residual MHA weights the GRU network's output, emphasizing the contribution of important hidden features to RUL prediction. By employing a residual connection, the GRU network's deficiency in extracting long-term sequence features is addressed.
  • (3)  
    By integrating the above two approaches, a new method for RUL prediction in rolling bearings is proposed. To validate the proposed method's effectiveness and accuracy, experimental predictions are conducted on a dataset of rolling bearing degradation.

The rest of the paper is arranged as follows: section 2 introduces the theoretical basis required for the study, section 3 details the proposed methodology, section 4 validate the proposed method through a contrast experiment. Finally, a conclusion is drawn in section 5.

2. Theoretical basis

2.1. Hilbert–Huang transformation (HHT)

HHT has been extensively applied in bearing PHM research owing to its excellent analysis and processing capabilities for nonlinear and non-stationary signals [32]. By decomposing the original signal into various Intrinsic Mode Functions (IMFs), which indicate different vibration modes, this method accomplishes signal decomposition and analysis. Then the instantaneous frequency and amplitude of each IMF can be obtained by calculating the Hilbert transform. This allows for the energy distribution of the signal at different frequencies to be obtained. The specific calculation steps are as follows:

The first step involves employing empirical mode decomposition on the original signal a(t) to obtain several IMFs, it can be defined as:

Equation (1)

Where, ${{I_i}(t)}$ is the ith IMF and Re(t) is the residual term.

Subsequently, the Hilbert transform of each ${{I_i}(t)}$ is calculated, which is defined as:

Equation (2)

Thus, the analytic signal ${{s_i}(t)}$ of each ${{I_i}(t)}$ is defined as:

Equation (3)

where, ${B_i}(t)$ is the instantaneous amplitude and ${\phi _i}(t)$ is the instantaneous phase, which are defined as:

Equation (4)

Equation (5)

The instantaneous frequency of each analytic signal may be derived by differentiating the instantaneous phase with respect to time, which can be calculated by:

Equation (6)

According to equations (1), (3) and (6), the original signal a(t) can be rephrased as:

Equation (7)

The marginal spectrum is an application of HHT, which is employed to analyze the instantaneous frequency and amplitude of time–frequency signals [33]. It can be calculated by:

Equation (8)

where, $H(f,t)$ denotes the HHT transform of the signal at frequency f and time t, T is the upper bound of integration and represents the time horizon of integration.

2.2. Gate recurrent unit network

In the field of time series modeling and prediction, GRU has become a popular choice due to its effectiveness. It is a type of recurrent neural network that is derived from LSTM [34]. Compared with LSTM, GRU uses only an update gate and a reset gate to control the selection and retention of memory, thus significantly reducing the number of parameters and computational complexity. The structure of a single GRU cell is shown in figure 1.

Figure 1.

Figure 1. GRU cell structure.

Standard image High-resolution image

GRU network can predict future states by learning the historical information of sequential data. The process of computing a GRU neuron can be described as: at time step t, given an input sequence xt , the GRU cell computes the reset gate rt and the update gate zt . The reset gate controls the forgetting of historical information, while the update gate controls the fusion of new information. Next, the candidate memory cell ${\tilde h_t}$ is calculated and interpolated with the previous memory cell ${h_{t - 1}}$ to obtain the current memory cell ${h_t}$. Finally, the predicted value or hidden state ht at time step t is outputted and used for prediction and computation at the next time step. The specific calculation formulas are as follows:

Equation (9)

where, ${{\mathcal H}^G}$ is the non-linear transformation function of the GRU neuron. The weights Vz , Vr and Vh are the weights of the current input xt and the recurrent input $h_{t-1}$. The biases pz , pr , and ph are the bias weights of the corresponding nodes. σ is the sigmoid function, $\textrm{tanh}$ is the hyperbolic tangent function, and $\odot$ is element-wise multiplication. zt and rt are the outputs of the update gate and reset gate of the GRU cell unit. ${\tilde h_t}$ is the candidate hidden state at the current time step, ht is the hidden state at the current time step, and ${h_{t - 1}}$ is the hidden state at the previous time step.

2.3. MHA

The attention mechanism is a structure specifically designed to handle the interdependence between input and output data, which can automatically learn and calculate the weight of each input data's contribution to the output data [35]. In the field of deep learning, attention mechanisms have become an important technology widely used in various tasks, especially in natural language processing. The MHA is an algorithm derived from the attention mechanism, which was first applied to the transformer model and proposed by Vaswani et al [36]. It captures key information in different subspaces through parallel computation and combines them into the final output. The state of rolling bearings is usually affected by multiple factors, such as vibration, temperature, pressure, etc. Compared with the common attention mechanism, the use of MHA mechanism allows the model to focus on multiple different key information at the same time, and thus improve the accuracy and reliability of predictions [37]. The MHA can be calculated as follows:

Equation (10)

Equation (11)

Equation (12)

Where, Q, K, and V denote the query, key, and value, respectively. n is the number of heads of MHA. W is the learnable parameter. αi is the score of the ith attention head. Qi , Ki , Vi are the query subspace, key subspace, and value subspace respectively, which can be obtained by multiplying Q, K, V with the learnable weight parameters $W_i^{\,Q}$, $W_i^K$, $W_i^V$ respectively. $\sqrt {{\textrm{d}_k}} $ is the scaling factor.

2.4. Residual connection

Deep neural networks have powerful representation capabilities, but they often encounter the issue of gradient vanishing during training. To mitigate this problem, the residual connection can be used, which introduces direct connections across layers in the network to ensure that the output contains information from the input, thus allowing gradients to propagate to deeper layers [38]. The residual connection is defined as follows:

Equation (13)

where a is the input vector, θl is the parameter of the lth layer, $f(a,\theta_l)$ is the transformation of the input a.

Residual connections can be used to build deep neural networks by stacking multiple residual blocks. During training, residual connections allow gradients to flow without losing information. Compared to traditional network structures, residual connections can better mitigate the issue of vanishing gradients and expedite model convergence.

3. Methodology

This section introduces the proposed framework for estimating the RUL of rolling bearings, depicted in its entirety in figure 2. The approach comprises two primary modules: automatic feature combination extraction and residual multi-Head attention gated recurrent unit (RMAGRU). Automatic feature combination extraction (AFCE) includes two parts: feature extraction and automatic feature combination selection. The feature extraction module uses feature parameters to define the computation of features. Next, in the extracted features, the random forest algorithm was applied and enhanced to allow for the automatic selection of feature combinations from the extracted features. These selected combinations meet a specific threshold in terms of their scores and represent features with more pronounced degradation. Finally, we have developed RMAGRU and applied it to predict the RUL of rolling bearings. The following sections will provide a detailed description of each module.

Figure 2.

Figure 2. The proposed framework for RUL prediction.

Standard image High-resolution image

3.1. Automatic feature combination extraction

3.1.1. Feature extraction.

Although deep learning methods such as GRU possess formidable nonlinear modeling capabilities and the capacity for autonomous feature acquisition, the high sampling frequency of sensors results in an excessive volume of raw bearing data. If one were to directly employ a GRU network for feature extraction from the raw data and subsequent RUL prediction, it would engender an inflation in network parameters, thereby significantly augmenting the computational burden during network training [39]. In the field of rolling bearing prognostics, engineers typically harbor a profound understanding of bearing operational principles and features. Consequently, they are adept at the curation and delineation of feature parameters germane to the problem at hand. Through the judicious selection of appropriate feature parameters, it is possible to ameliorate the computational onus on the network whilst simultaneously diminishing the model's susceptibility to noise and extraneous information, rendering it more adept at accommodating diverse datasets.

Based on the above considerations, we calculated and extracted 23 types of rolling bearing features except f24 according to [40], because the calculation of f24 may appear complex numbers, which is inconvenient to calculate on the computer. These features can represent various aspects of the signal in both the frequency and time domains and are of paramount importance for predicting RUL. In addition, the signal envelope R feature and the sum of edge spectra have been proved to effectively characterize signal performance [41]. Consequently, we computed the signal envelope R feature and the sum of edge spectra, designating them as f24 and f25. The specific calculation process is as follows:

Initially, the Hilbert transform is computed for the raw signal ${{I_i}(t)}$ by equation (2).

Next, the envelope of the Hilbert transform is extracted by equation (4).

Then, the envelope feature R of the signal is calculated as f24 by:

Equation (14)

where, $E[{B^4}(t)]$ is the fourth moment, and ${E^{\,2}}[{B^2}(t)]$ is the square of the second moment.

According to equation (8), the sum of marginal spectra of the HHT transform is calculated as f25 by:

Equation (15)

Where, K is the number of spectral lines, $f_{r_k}$ is the frequency value of the kth spectral line.

The final 25 features are shown in table 1, where $f_1-f_{11}$ are time domain features, $f_{12}-f_{23}$ are frequency domain features, and $f_{24}-f_{25}$ are time–frequency domain features.

Table 1. The feature parameters.

FeatureequationFeatureequationFeatureequation
f1 $\frac{\sum_{i = 1}^Na_i}N $ f10 $\frac{f_4}{\frac1N\sum_{i = 1}^N\left|a_i\right|}$ f19 $\sqrt{\frac{\sum_{k = 1}^Kf_{r_k}^{\,4}s_k}{\sum_{k = 1}^Kf_{r_k}^{\,2}s_k}}$
f2 $\sqrt{\frac{\sum_{i = 1}^{N}(a_i-f_1)^2}{N-1}}$ f11 $\frac{f_5}{\frac1N\sum_{i = 1}^N\left|a_i\right|}$ f20 $\frac{\sum_{k = 1}^{K}f_{r_k}^{\,2}s_k}{\sqrt{\sum_{k = 1}^{K}s_k\sum_{k = 1}^{K}f_{r_k}^{\,4}s_k}}$
f3 $\left(\frac{\sum_{i = 1}^N\sqrt{\left|a_i\right|}}{N}\right)^2$ f12 $\frac{\sum_{k = 1}^{K}s_k}{K}$ f21 $\frac{f_{17}}{f_{16}}$
f4 $\sqrt{\frac{\sum_{i = 1}^N(a_i)^2}N}$ f13 $\frac{\sum_{k = 1}^K(s_k-f_{12})^2}{K-1}$ f22 $\frac{\sum_{k = 1}^K(f_{r_k}-f_{16})^3s_k}{Kf_{17}^{\,3}}$
f5 $\max|a_i|$ f14 $\frac{\sum_{k = 1}^K(s_k-f_{12})^3}{K(\sqrt{f_{13}})^3}$ f23 $\frac{\sum_{k = 1}^K(f_{r_k}-f_{16})^4s_k}{Kf_{17}^{\,4}}$
f6 $\frac{\sum_{i = 1}^N(a_i-f_1)^3}{(N-1)f_2^3}$ f15 $\frac{\sum_{k = 1}^{K}(s_k-f_{12})^{4}}{Kf_{13}^{\,2}}$ f24 $\left| {\frac{{E[{B^4}(t)] - {E^{\,2}}[{B^2}(t)]}}{{{E^{\,2}}[{B^2}(t)]}}} \right|$
f7 $\frac{\sum_{i = 1}^{N}(a_i-f_{1})^{4}}{(N-1)f_{2}^{4}}$ f16 $\frac{\sum_{k = 1}^Kf_{r_k}s_k}{\sum_{k = 1}^Ks_k}$ f25 $\sum_{k = 1}^K {M(f_{r_k})} $
f8 $\frac{f_5}{f_4}$ f17 $\sqrt{\frac{\sum_{k = 1}^{K}(f_{r_k}-P_{16})^{2}s_k}{K}}$   
f9 $\frac{f_5}{f_3}$ f18 $\sqrt{\frac{\sum_{k = 1}^{K}f_{r_k}^{\,2}s_k}{\sum_{k = 1}^{K}s_k}}$   

Where ai is a sequence of signals for $i = 1,2,\ldots ,N$. N is the number of data points. sk is a spectrum for $k = 1,2,\ldots ,K$ and K is the number of spectral lines. $F_{r_k}$ is the frequency value of the kth spectral line.

3.1.2. Feature combination automatic selection.

The raw vibration signal monitored by the sensor includes horizontal and vertical vibration signals. These distinct vibration signals convey differential information pertaining to the degradation of the bearing, respectively. Extracting features from the horizontal vibration signal yields a total of 25 distinctive features denoted as $F_{h_1}$ to $F_{h_{25}}$. Similarly, the extraction of features from the vertical vibration signal results in another 25 unique features designated as $F_{v_1}$ to $F_{v_{25}}$. Consequently, a grand total of 50 distinct features are obtained. However, some of these features have similar properties, while some degenerate properties may not be significant enough. If all these features are fed into the neural network for training, it will increase the computational burden and may even lead to information loss of degenerate features, thus reducing the prediction accuracy of RUL. Moreover, selecting an optimal combination of features is a challenging task, and random or subjective selection may not necessarily achieve the best results. To address this issue, this paper proposes a feature combination algorithm based on random forest automatic selection. This algorithm evaluates the importance of each feature combination and selects the best feature combination with a score that meets the preset threshold.

Random forest obtains the final classification or regression result through the voting of multiple decision trees. In terms of feature selection, random forest has been shown to have good performance [42]. Random forest algorithms typically use classification and regression trees (CART) as decision trees. The traditional CART tree uses the gini impurity function to measure the criterion of splitting nodes, but it may not be sensitive enough to the degradation characteristics of bearings. In order to better adapt to the bearing degradation characteristics, we propose a weighted gini impurity function. This modified gini impurity function utilizes the bearing degradation correlation to assist the node splitting optimization of the CART tree. The calculation formula is given in equation (16),

Equation (16)

Where x is a splitting variable, v is a splitting value of the splitting variable, $N_\textrm{s}$ is the number of all training samples in the current node. Xleft and Xright are the training sample sets of the left and right child nodes, respectively. $\bar{y}_{\textrm{left}}$ and $\bar{y}_{\textrm{right}}$ are the average values of the target variables of the left and right node samples, respectively. yi , yj are the current sample target variables of the left and right nodes, respectively. wi , wj denote the Spearman's rank correlation coefficient between yi , yj and the current RUL true value, respectively.

The improved random forest algorithm can evaluate the importance of each feature by calculating its importance in multiple CART trees and further select feature combinations. The specific process of the automatic feature combination selection model studied in this paper is shown in figure 3.

Figure 3.

Figure 3. Automatic feature combination extraction model.

Standard image High-resolution image

The automatic feature combination algorithm studied in this paper follows the process outlined below:

  • (1)  
    For an original dataset X with dimensions of (m, n), where m is the number of samples and n is the number of features, random sampling with replacement is performed to construct the sub-datasets $x_1,x_2,\ldots ,x_s$ , where the dimension of xi is (m, k), $i = 1,2,\ldots ,s$ and k < n.
  • (2)  
    CART models are trained using the sub-training set. The optimal split variable and split point are found by optimizing to minimize the gini impurity function $I_\textrm{G}$, as represented in equation (17). By repeating this process, multiple CART models can be generated.
    Equation (17)
    Where $x^*$ is the solution of the optimization problem $\operatorname{argmin}_{x,v}I_\textrm{G}$, $v^*$ is the solution of the optimization problem $\operatorname{argmin}_{x,v}I_\textrm{G}$ associated with $x^*$, xi is the candidate bearing degradation eigenvalue, and vij is the specific partition condition associated with xi .
  • (3)  
    All the trained CART models are used to predict each feature.
  • (4)  
    To evaluate the importance of features, the mean of all CART models predicted values is used as the score of each feature.
  • (5)  
    Based on the scores calculated in step 4, all features are sorted in descending order.
  • (6)  
    By establishing a feature selection threshold, feature combinations meeting the threshold criteria are automatically chosen in sequential order.

3.2. RMAGRU

In order to address the issue of overfitting and the limited modeling of long-term dependencies by GRU when dealing with high-dimensional data, this paper proposes a new derived model of GRU called RMAGRU, which incorporates both MHA and residual connections to enhance the model's focus on input information and its ability to model long-term dependencies. The network architecture of RMAGRU is illustrated in figure 4. RMAGRU can adaptively fuse input features and aggregate relevant information, better describing the relationships between data. Essentially, the input of each layer's hidden state undergoes a residual multi-head attention calculation, capturing important degradation characteristics from the previous layer's hidden state.

Figure 4.

Figure 4. RMAGRU structure.

Standard image High-resolution image

The internal structure of the residual multi-head attention mechanism (RMHA) is demonstrated in figure 5. In the operation of RMHA, the MHA is used to calculate the attention weighted representation of different parts of the hidden state input. These attention-weighted representations are added to the initial hidden state input, thus establishing a residual connection to preserve the original input information. The output of the residual connection goes through the BN layer for batch normalization and is used as the output of RMHA for the subsequent calculation.

Figure 5.

Figure 5. Residual multi-head attention mechanism.

Standard image High-resolution image

The calculation of the residual MHA, based on equations (10)–(13), can be defined as follows:

Equation (18)

Where, RMHA is the score calculated by the residual MHA, BN is batch normalization.

Incorporating equations (9)–(13), (18), the calculation process of an RMAGRU neuron at time t can be described as follows:

Equation (19)

Where, $\mathcal H^R$ is the non-linear transformation function of the GRU neuron. ${Q_i}_{t-1}$, ${K_i}_{t-1}$, and ${V_i}_{t-1}$ are the query subspace, key subspace, and value subspace of the ith attention head the previous time step, respectively. ${\alpha_i}_{t-1}$ is the score of the ith attention head at the previous time step. $h_{t - 1}^*$ is the output of the hidden state from the previous time step after passing the RMHA in the current layer.

As equation (19) indicates, the operation of RMAGRU can be divided into two parts. Firstly, the RMAGRU model is able to effectively guide the hidden state information from the previous time step and enhance the network's memory capacity by utilizing the residual MHA. Secondly, similar to GRU, RMAGRU is capable of effectively extracting degradation information from the hidden state and adjusting it based on update criteria. Compared to traditional RNN models, RMAGRU's hidden state contains more degradation information and the problem of gradient vanishing is alleviated through residual connections.

The input of feature data goes through multiple layers of RMAGRU to learn and extract relevant information. The output of the previous RMAGRU at each time step is used as the input to the current time step, which is divided into two branch outputs via RMHA. One branch continues the forward propagation and performs the RMAGRU internal computation, while the other branch multiplies the output of RMHA with the weight coefficient βt for the current time step to obtain the attention weight γt for the current time step. The value of βt is determined by equation (20), which makes $\beta_t\gt 1$, in line with the property that the importance of the degraded features increases during the degradation process.

Equation (20)

Where,

Equation (21)

Equation (22)

Where, ot is the RUL true value at time t.

All RMAGRU units are used to compute the attention weights γ for all time steps and then concatenate them together and normalized. Finally, the concatenated attention weights are multiplied with the output hout of the last hidden layer and subjected to batch normalization. This process can be defined as equations (23)–(25).

Equation (23)

Equation (24)

Equation (25)

The final process of predicting RUL from the fully connected layer output can be defined as equation (26).

Equation (26)

Where, $\hat o$ is the predicted value, ϕ is the activation function and LeakyReLU is chosen as the activation function in this paper.

RMAGRU network uses mean squared error (MSE) as the loss function, defined as:

Equation (27)

where, N is the length of the output vector, and o is the true RUL label.

The RUL label in this study is defined as the running time of the bearing from the current time until it fails, which can be calculated from equation (28),

Equation (28)

where, T is the total running time of the bearing from integrity to failure, and tcr is the current running time of the bearing.

3.3. General steps of the proposed method

The proposed RUL prediction method combines automatic feature combination extraction and RMAGRU network. The detailed steps are as follows:

  • (1)  
    Collect vibration data of the monitored bearing using sensors. The sampling interval is ts and the total sampling time is t. Assume there are n samples in total.
  • (2)  
    Extract features from the vibration data of all datasets to obtain a set of features $F = \left[ {{F_1},{F_2},\ldots ,{F_{50}}} \right]$ that reflects the degradation information of the bearing, where ${F_i} = {[{f_{i1}},{f_{i2}},\ldots ,{f_{in}}]^\intercal},i = 1,2,\ldots ,50$.
  • (3)  
    Use the feature combination automatic selection method to select the feature combinations $F_{\textrm{final}} = [{F_1},{F_2}, \cdots ,{F_m}]$ with the strongest importance from the feature set F, where m is the number of finally selected features.
  • (4)  
    Normalize the Ffinal selected in the previous step.
    Equation (29)
  • (5)  
    Define the input matrix Z of the RMAGRU network and the corresponding RUL label $Y = {[\begin{array}{*{20}{c}} {{y_1}}&{{y_2}}& \cdots &{{y_m}} \end{array}]^\intercal}$.
    Equation (30)
    Where, ${z_i} = \left[ {\begin{array}{*{20}{c}} {{F_{1i}}}&{{F_{2i}}}& \cdots &{{F_{mi}}} \end{array}} \right]$.
  • (6)  
    Build the RMAGRU network to learn and predict the RUL of the bearing.
  • (7)  
    Set the input to be the first k vectors of matrix Z with the output label $y_{k+1}$, set the sliding step length to 1, use MSE as the loss function during training, use backpropagation to train the model and optimize its parameters, and calculate the predicted RUL for the validation dataset.
  • (8)  
    Repeat Step 7 in a continuous iterative optimization process until the desired performance is achieved to obtain the best training model.
  • (9)  
    The test set feature combination Ffinal is fed into the trained RMAGRU model, which predicts the test set RUL and compares the predicted RUL with the actual RUL.

4. Experiment

4.1. Description of experimental dataset

The dataset used in the experiments is from the PRONOSTIA bearing dataset provided by the FEMTO-ST Research Institute, which includes several accelerated bearing degradation experiments performed on the PRONOSTIA platform [43]. Detailed experimental apparatus for the bearing test stand can be found in figure 6.

Figure 6.

Figure 6. PRONOSTIA platform.

Standard image High-resolution image

Three different operating conditions are set in the PRONOSTIA rolling bearing accelerated degradation experiment. The operating conditions are defined based on the torque applied on the bearing, the speed of rotation of the shaft, and the instantaneous measurements of the radial force applied on the bearing. The characterization of bearing degradation is reflected through data collected by a vibration sensor. The vibration sensor consists of two micro-accelerometers that are perpendicular to each other, with one aligned along the vertical axis and the other along the horizontal axis. Both accelerometers are radially located on the bearing's outer ring. The vibration signal is sampled at a frequency of 25.6 kHz, with a duration of 0.1 s for each sample and a sampling interval of 10 s. The experiment is terminated once the amplitude of the vibration data surpasses 20 g. This dataset provides six datasets that ran until failure and eleven datasets that did not fully degrade to failure for testing the model. Additionally, the actual RUL of the test signal is provided. Table 2 presents detailed description of the vibration data collected from 17 rolling bearings.

Table 2. Datasets of IEEE 2012 PHM prognostic challenge.

 Conditions 1Conditions 2Conditions 3
Load (N)400042005000
Speed (RPM)180016501500
Training setBearing1_1Bearing2_1Bearing3_1
 Bearing1_2Bearing2_2Bearing3_2
Test setBearing1_3Bearing2_3Bearing3_3
 Bearing1_4Bearing2_4 
 Bearing1_5Bearing2_5 
 Bearing1_6Bearing2_6 
 Bearing1_7Bearing2_7 

The raw vibration signal of bearing 1_1 is shown in figure 7. It is evident that the raw data volume is exceptionally large, and directly inputting it into a neural network for training would inevitably affect training speed and might even make training infeasible. Therefore, feature extraction in advance is necessary.

Figure 7.

Figure 7. The raw vibration signal of bearing 1_1.

Standard image High-resolution image

Figure 8 shows 25 kinds of features extracted from the raw horizontal vibration signal of bearing 1_1 according to table 1. It is noticeable that in the extracted time-domain and frequency-domain features, some features have similar degradation characteristics, while others may have less pronounced degradation traits or even potentially interfere with the learning of degradation features. Hence, feature selection is imperative. In this article, bearings 1_1, 1_2, 2_1, 2_2, 3_1, 3_2 are used to train the model, and the remaining 11 datasets are used as the test set.

Figure 8.

Figure 8. 25 features extracted from the horizontal vibration signal of bearing 1_1.

Standard image High-resolution image

4.2. Evaluation index

In this paper, the proposed model is evaluated using three metrics: root mean squared error (RMSE), mean absolute error (MAE) and a scoring function named SCORE. MAE and RMSE both indicate the accuracy of the predicted value compared to the actual value, where a lower error signifies a more precise prediction. SCORE is a comprehensive metric, and higher values indicate greater accuracy of the prediction. They are defined as follows:

Equation (31)

Equation (32)

Equation (33)

Equation (34)

Equation (35)

Where, $\textrm{PBR}_i$ is the predicted value of the model, ${\textrm{TBR}_i}$ is the actual value, n is the number of samples in the testing dataset.

4.3. Comparing framework settings

The experimental study is conducted on a computer with Intel(R) Core(TM) i5-13600K CPU @ 3.5GHz, 32GB RAM, and NVIDIA GeForce RTX 3070 Ti Graphics Card. The operating system is Windows 10, and the RUL prediction is implemented under the framework of torch1.7 based on Python3.9. To prove the effectiveness of our method, several recent two-stage RUL prediction frameworks are employed for comparison. In the first stage, the proposed AFCE method is compared with five other feature extraction methods, which are time domain feature extraction (TDF), frequency domain feature extraction (FDF), joint time–frequency domain feature extraction (TFDF), hybrid feature extraction (MF), and Convolutional Autoencoder feature extraction (CAE). In the second stage, these frameworks can be grouped based on the differences in neural network structures: LSTM-based framework, GRU-based framework, TCN-based framework and RMAGRU-based framework. In this comparative study, experiments are divided into two groups.

For the first group of experiments, in terms of TDF, since the sampling frequency of the bearing vibration signal is 25.6 kHz, the duration of each sampling is 0.1 s, and the sampling interval is 10 s, the 2560 data sampled each time reflect the degradation characteristics of the bearing at this time. Therefore, we calculate and normalize the time-domain characteristics of this section of signal data according to $f_1-f_{11}$ in table 1. Finally, 11 kinds of characteristic values can be obtained after calculation for each section of vibration signal with the size of $2560\times1$, and they are merged into the eigenvector of $1\times11$. Similarly, in terms of frequency domain feature extraction, we extract 12 normalized frequency domain features to represent the degradation features of bearings according to the features of each $2560\times1$ vibration signal from $f_{12}-f_{23}$ in table 1, and combine these features to obtain a feature vector of size $1\times12$. Time–frequency joint domain feature extraction involves f24 and f25 of table 1 to extract two types of joint domain features, resulting in a feature vector of size $1\times2$ for each $2560\times1$ vibration signal. Hybrid feature extraction involves extracting 25 features such as time domain, frequency domain, and joint time–frequency domain, and generating a feature vector of size $1\times25$ for each $2560\times1$ vibration signal according to $f_1-f_{25}$ of table 1. The CAE is to extract features from the collected data by using a one-dimensional convolutional network with convolution kernel of 16 and step size of 1, and compress the $2560\times1$ vibration signal into a feature vector of size $1\times1$. The automatic feature combination extraction method refers to what is presented in section 3.

For the second group of experiments, we evaluate the performance of our RMAGRU model in comparison to three other commonly used deep learning methods: LSTM, GRU, and TCN. To ensure fairness, we ensure that all hyperparameters consistent across all networks during training. During model training, a feature selection threshold of 36 is used. Backpropagation is employed for learning, and the optimizer used is Adam with a learning rate of $1\times 10^{-5}$. The batch size is fixed to 32, and we ran the training for a total number of 300 epochs. Details of all the hyperparameters used are presented in table 3.

Table 3. The hyperparameters of model training.

HyperparameterValue
Learning rate $1\times 10^{-5}$
Batch Size32
Training epochs300
Sequence length64
Early stop30
Number of layers4
Number of hidden units28
OptimizerAdam

4.4. Comparison of methods

In this section, we demonstrate the figures and experimental outcomes of the aforementioned frameworks, and analyze the impact of various key hyperparameters on RUL prediction performance.

In the first group of experiments, six different feature extraction methods are used to extract bearing degradation features and input them into three prediction models, respectively. Considering Bearing1_3 and Bearing1_4 as examples, figure 9 exhibits the prediction outcomes of the five feature extraction methods combined with the GRU model and our proposed feature extraction method combined with the GRU model. Additionally, table 5 presents the prediction results of the GRU model on the testing set for all 11 bearings, and it can be seen that for some test bearings, some feature extraction methods can obtain the best results. However, from the overall results, the AFCE method achieves good results, with the overall average values of RMSE and MAE as low as 17.21% and 14.89%, respectively, while the RMSE of other feature extraction methods is higher than 21% and MAE is higher than 18%. In order to better prove the effectiveness of the AFCE method, the five feature extraction methods are input into the LSTM model and the TCN model for testing. Table 4 reports the prediction results obtained using the LSTM model. It can be seen that out of the 11 test bearings, the AFCE method achieved the best results in 8 bearings, with the overall average RMSE of 18.19% and the overall average MAE of 15.22%. The average RMSE and average MAE of other methods for bearing prediction are between 21% and 23% and 18.25% and 19.55%, respectively. In addition, the proposed AFCE method is compared with different feature extraction methods, and the prediction results under the TCN model are shown in table 6, in most test bearings, the error of this method is the smallest, and the average error of all test bearings is lower than other feature extraction methods.

Figure 9.

Figure 9. Prediction results of different feature extraction methods in GRU model.

Standard image High-resolution image

Table 4. The results of different feature extraction methods in LSTM model.

Model FeatureLong-short term memory(LSTM)      
TDFFDFTFDFMFCAEAFCE (Proposed)
MetricRMSEMAERMSEMAERMSEMAERMSEMAERMSEMAERMSEMAE
Bearing1_323.8719.9921.3918.9922.4719.0219.9417.1523.2519.59 19.36 16.99
Bearing1_426.4022.8317.9614.1026.5022.9715.9713.1226.7223.13 15.04 12.50
Bearing1_526.3622.8034.4426.1827.4023.6627.5523.3526.2122.69 26.09 21.75
Bearing1_626.4422.8821.6115.5926.5022.9328.2323.8526.3922.86 12.18 9.46
Bearing1_724.2119.8217.6411.9923.6218.7115.5212.1123.1318.99 12.53 9.09
Bearing2_324.6520.16 9.31 7.55 23.4219.0121.6117.3823.5519.2112.119.99
Bearing2_421.7218.6120.2317.4921.8018.8023.2320.0521.3618.40 10.80 9.57
Bearing2_524.9021.3835.3228.0424.9921.4731.1925.75 24.54 21.15 28.3122.94
Bearing2_621.5118.4418.8615.8422.4919.2121.8018.4121.0518.12 16.28 12.54
Bearing2_7 13.46 11.6225.7824.7213.53 11.59 19.2015.7214.0212.0130.8030.06
Bearing3_319.1616.5024.1821.2119.1016.81 15.07 13.8119.2016.6316.59 12.57
Average22.9719.5522.4318.3422.8919.4721.7618.2522.6719.34 18.19 15.22

Table 5. The results of different feature extraction methods in GRU model.

Model FeatureGate recurrent unit(GRU)      
TDFFDFTFDFMFCAEAFCE (Proposed)
MetricRMSEMAERMSEMAERMSEMAERMSEMAERMSEMAERMSEMAE
Bearing1_322.0218.7814.8711.1922.6119.2221.4817.0525.9121.43 10.22 8.05
Bearing1_426.3622.8520.4917.9025.3422.0627.4623.4333.1327.35 12.62 10.80
Bearing1_526.7323.07 18.88 15.99 22.1119.1730.1625.1522.3819.0520.1517.43
Bearing1_626.5822.9920.3515.2524.4621.2528.8724.2327.2023.43 16.87 14.41
Bearing1_723.3118.5819.2515.0620.9018.14 13.56 10.72 16.3112.9224.5219.10
Bearing2_324.3119.7427.1925.6123.2519.5323.2717.9421.2017.21 7.44 6.39
Bearing2_421.4018.4828.2026.94 19.37 16.36 24.1420.0423.2819.1723.6521.49
Bearing2_525.2321.53 15.27 13.14 23.9420.7635.5528.8723.4919.9320.5916.55
Bearing2_621.7918.5931.3126.8616.9814.5024.3120.2327.2222.47 11.86 9.14
Bearing2_713.3611.4237.1136.87 13.31 11.20 21.9218.1517.5914.4827.0326.84
Bearing3_318.9216.49 12.43 7.16 22.3517.2720.0917.1727.3422.5314.3413.59
Average22.7319.3222.3019.2721.3318.1324.6220.2724.1020.00 17.21 14.89

Table 6. The results of different feature extraction methods in TCN model.

Model FeatureTemporal convolutional network(TCN)      
TDFFDFTFDFMFCAEAFCE (Proposed)
MetricRMSEMAERMSEMAERMSEMAERMSEMAERMSEMAERMSEMAE
Bearing1_320.8217.63 7.79 5.84 23.2819.4873.4943.3322.8719.4023.1719.53
Bearing1_426.2722.7117.6515.1926.6222.9923.3415.6227.2523.41 17.53 11.75
Bearing1_527.9823.8628.1122.6117.3513.48 15.13 11.26 26.1222.4926.9123.31
Bearing1_629.6624.9123.4918.7626.8523.4028.1520.7426.4922.91 17.48 15.26
Bearing1_730.5925.8620.3015.8824.5821.9513.0311.8621.7218.29 8.35 6.87
Bearing2_338.1834.5921.9019.2322.8918.45 15.01 11.73 22.2018.1726.5925.63
Bearing2_421.1518.6411.549.8719.1416.6516.7214.4720.9518.27 8.72 7.85
Bearing2_527.9723.3634.7030.1626.4823.3523.2119.9224.0720.82 21.90 18.86
Bearing2_631.3026.4927.7225.42 11.96 9.56 13.6812.1420.7817.8820.3918.56
Bearing2_715.0212.4330.5630.0717.8716.3554.7353.3415.1212.53 13.65 11.70
Bearing3_327.7422.8214.9313.3420.0916.6526.1424.8119.7716.91 13.42 11.68
Average26.9723.0321.7018.7621.5618.3927.5121.7522.4919.19 18.01 15.55

Based on the findings presented in tables 46, our proposed AFCE method outperformed other feature extraction methods in terms of RUL prediction results among different deep network structures. This result reflects that AFCE can retain features that contain more detailed degradation information, enabling deep network models to make more use of features for autonomous learning and achieving more accurate RUL prediction. It should be noted that the performance of our automatic feature combination extraction method in the GRU model is slightly inferior. This may be because the GRU model has poor ability to handle long sequences compared to other deep network models, and there is also a problem of gradient disappearing. Overall, our AFCE approach can substantially enhance the performance of the entire predictive model. This result confirms the effectiveness of our automatic feature combination extraction method.

In the second group of experiments, we use the AFCE method to extract feature combinations and input them into four different prediction models. To provide a comprehensive assessment of model performance, we add a new evaluation metric named SCORE. Taking Bearing2_6 and Bearing3_3 as examples, figure 10 and table 7 show the prediction results of the deep models based on AFCE.

Figure 10.

Figure 10. Prediction results of different network models method in AFCE.

Standard image High-resolution image

Table 7. Quantitative results of various models under AFCE.

ModelLSTMGRUTCNRMAGRU (Proposed)
MetricRMSEMAESCORERMSEMAESCORERMSEMAESCORERMSEMAESCORE
Bearing1_3 8.57 6.35 59.85 28.8724.1530.5421.6719.6834.5916.0914.3142.05
Bearing1_415.2112.8140.0018.6816.4731.8612.218.1649.40 11.92 7.83 51.20
Bearing1_532.5927.4020.3430.4625.7521.85 16.81 14.16 38.36 19.3417.0231.89
Bearing1_612.8910.6937.5922.7819.1129.5217.6913.8728.00 12.59 9.22 44.02
Bearing1_718.2715.3342.2019.5418.7137.6714.7311.1244.32 13.65 10.69 48.15
Bearing2_324.4020.2141.7016.0313.3048.0120.3919.3338.24 9.86 7.76 50.37
Bearing2_410.338.4947.2813.5211.2645.58 9.76 7.14 48.26 16.2615.2040.41
Bearing2_534.6929.0521.1926.6522.9825.3323.0320.4227.97 20.99 17.63 33.30
Bearing2_624.8221.3733.8216.1814.0841.2411.0510.1041.45 6.63 5.61 51.27
Bearing2_726.4724.154.5619.14 16.94 10.44 34.4034.300.05 18.35 17.633.42
Bearing3_321.0615.6949.3415.3413.8134.3610.188.4345.45 8.76 7.66 50.38
Average20.8517.4136.1720.6517.8732.4017.4515.1636.01 14.04 11.87 40.59

Figure 11 illustrates the predicted absolute error of figure 10. Several trends can be observed from the figure. Firstly, the prediction error performance of GRU is not consistently stable, showing obvious fluctuations, sometimes the error is small, sometimes the error is large, displaying a volatile pattern. Secondly, LSTM exhibits relatively large absolute errors in the early prediction stages, which can go as high as 0.4. However, in the later prediction stages, the error gradually decreases, demonstrating a more stable performance. Nevertheless, it has a relatively poorer performance in the early prediction stages. In contrast, both TCN and RMAGRU methods demonstrate relatively stable prediction absolute errors throughout the entire degradation phase of the bearing. It's worth noting that the average prediction absolute error for RMAGRU is lower than that for TCN, indicating that RMAGRU outperforms TCN in the RUL prediction task. As indicated in table 7, the AFCE-RMAGRU model outperformed other frameworks in terms of predictive accuracy, which further demonstrates the excellent performance of this framework in RUL prediction for rolling bearings.

Figure 11.

Figure 11. Absolute error of prediction for different network models method in AFCE.

Standard image High-resolution image

After comparing the prediction results between the GRU and RMAGRU models, it is discovered that in the majority of cases, the RMAGRU model is able to predict more accurately than the GRU model. This suggests that the RMAGRU model has a stronger capacity in handling lengthy sequences and mitigating gradient vanishing, which highlights its exceptional performance advantages.

In our experiments, four key parameters are optimized: the threshold of the number of features, the number of attention heads, the sequence length, and the learning rate. We obtain the average RUL prediction results on the test set by adjusting different parameter values and display them in figure 12. The horizontal axis in the figure represents different parameter values, the left vertical axis represents the predicted RMSE, while the right vertical axis represents the predicted SCORE. A smaller RMSE and a higher score indicate a higher prediction accuracy, from which the optimal parameters can be inferred.

Figure 12.

Figure 12. AFCE-RMAGRU prediction results of different hyperparameters.

Standard image High-resolution image

According to figure 12(a), when the feature threshold value is small, the prediction error is large, and as the feature threshold increases, the prediction error decreases. When the feature threshold value is 36, the RUL prediction effect reaches the best. Figures 12(b) and (c) suggest that appropriate sequence length and attention head number can enhance the model's predictive capabilities, but the relationship between these two hyperparameters and performance is not monotonically increasing. Excessive sequence length and attention head number can also lead to a decrease in computational efficiency and overfitting risk. After comprehensive consideration, the optimal attention head number is 8, and the optimal sequence length is 64. At the same time, according to figure 12(d), within the parameters of this experiment, the influence of varying learning rates on RUL prediction is small. Therefore, while ensuring accuracy, we chose the optimal learning rate of $1\times 10^{-5}$ to minimize training time costs.

4.5. Results and discussion

In this section, the proposed RMAGRU network-based RUL prediction framework is compared with other cutting-edge approaches. In the comparison, we use the RNN-based gated dual attention unit method (GDAU) [31], the LSTM method based on macro-micro attention (MMALSTM) [27], and the time-domain convolutional network based on soft-thresholding and attention (TCN-SA) [21] as controls. We employ MAE and RMSE as evaluation metrics and tested them on the same dataset. As shown in table 8, based on the experimental results, it can be concluded that the proposed framework surpasses other methods in both prediction accuracy and robustness. The comparison results further indicate that the MRAGRU network effectively leverages the information of input data and hidden layer states, and has a broad range of potential applications in RUL prediction.

Table 8. Performance comparison of different prediction networks.

MetricGDAUMMALSTMTCN-SARMAGRU
MAE14.819.615.711.87
RMSE18.723.719.714.04

According to the experimental results in the previous section, the feature combination module and the residual MHA are very important for improving prediction accuracy and robustness. The feature combination module can extract different features from the input data and automatically select better degraded feature combinations, which helps the GRU network to better capture the complex and dynamic features of rolling bearings. By utilizing residual multi-head attention, the RMAGRU network can enhance its ability to model long-term dependencies in time series data and ultimately bolster the model's robustness.

5. Conclusion

In this paper, a rolling bearing RUL prediction framework based on residual multi-head attention GRU network with automatic feature combination extraction is proposed. The framework extracts a series of bearing degradation features, designs a new feature combination extractor to obtain the optimal bearing degradation feature combination, and develops an RMAGRU network that combines residual MHA with GRU network to better capture the complex and dynamic features of bearings. The experimental outcomes demonstrate that the suggested framework surpasses other contemporary techniques in terms of prediction accuracy and robustness. Furthermore, it can achieve high-precision predictions under various operating conditions and fault degrees. Therefore, the framework has important industrial application value, which can help enterprises to better maintain and manage equipment, and improve equipment reliability and safety.

Future work will focus on optimizing the network structure to reduce model training time and exploring the bearing life prediction problem under different working condition data distributions. To improve the practical applicability of the model when faced with changing working conditions, we plan to combine transfer learning with RMAGRU to better adapt to different data features under variable operating conditions. The implementation of these works will further improve the precision and resilience, making the proposed framework more practical and reliable.

Acknowledgments

This work is supported by National Natural Science Foundation of China (Grant No. 62101314), and the Ningbo Major Science and Technology Project (20212ZDYF020021).

Data availability statement

The data supporting this study is publicly available in IEEE phm 2012 data challenge at http://www.femto-st.fr/en/Research-departments/AS2M/Research-groups/PHM/IEEE-PHM-2012-Data-challenge.php.

Ethics approval

Since the data used for this research does not involve human or animal participants, it does not violate any moral ethics.

Conflicts of interest

No potential conflict of interest was reported by the authors.

Please wait… references are loading.
10.1088/1361-6501/ad1652