This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A novel time–frequency Transformer based on self–attention mechanism and its application in fault diagnosis of rolling bearings

Yifei Ding Minping Jia [email protected] Qiuhua Miao Yudong Cao School of Mechanical Engineering, Southeast University, Nanjing 211189, PR China
Abstract

The scope of data-driven fault diagnosis models is greatly extended through deep learning (DL). However, the classical convolution and recurrent structure have their defects in computational efficiency and feature representation, while the latest Transformer architecture based on attention mechanism has not yet been applied in this field. To solve these problems, we propose a novel time–frequency Transformer (TFT) model inspired by the massive success of vanilla Transformer in sequence processing. Specially, we design a fresh tokenizer and encoder module to extract effective abstractions from the time–frequency representation (TFR) of vibration signals. On this basis, a new end-to-end fault diagnosis framework based on time–frequency Transformer is presented in this paper. Through the case studies on bearing experimental datasets, we construct the optimal Transformer structure and verify its fault diagnosis performance. The superiority of the proposed method is demonstrated in comparison with the benchmark models and other state-of-the-art methods.

keywords:
Fault diagnosis , Deep learning , Transformer , Self–attention mechanism , Rolling bearings

1 Introduction

Currently, rotating machinery is widely used in aviation, aerospace, shipbuilding, automobile and other industrial fields, acting as the power source and support of many industrial systems [1]. The rolling bearing is a key vulnerable part of rotating machinery, which is prone to failure under a harsh working environment and alternating load [2]. Therefore, the research of rolling bearing fault diagnosis is of great importance to ensure the safety and reliability of facilities [3].

The ultimate goal of fault diagnosis is to recognize the status of the target component in the machine and thus determine whether the machine needs maintenance. Generally, existing fault diagnosis methods consist of two categories: model-based [4] approaches and data-driven [5] methods. Model-based methods often require a lot of prior knowledge, making difficult to accurately establish the diagnosis model of composite components under complex conditions. Data-driven methods [5] aim to convert the data provided by sensors into a parametric or non-parametric correlation model. Data-driven methods can effectively and rapidly process machinery signals, providing accurate diagnosis results with low requirement for prior expertise. Therefore, they are becoming more and more attractive with the development of various intelligent approach. The most common intelligent fault diagnosis method is currently developed on machine learning (ML) approaches such as kk-nearest neighbor (kk-NN) [6], support vector machine (SVM) [2], self-organized map (SOM) network [7], etc. Most intelligent fault diagnosis methods for rolling bearings are built on the processing and analysis of vibration signals [1], which employ a discrimination model with the input of man-made features extracted from acquired raw signals. Note that these man-made features include: 1) time domain statistic moment [8] such as root-mean-square (RMS), kurtosis, skewness, etc., 2) frequency spectrum [9] processed by fast Fourier transform (FFT), power spectrum estimation, etc., 3) time–frequency domain energy [10] obtained with empirical mode decomposition (EMD), variational mode decomposition (VMD), etc., and even 4) fusion features [11] extracted with principal component analysis (PCA), etc.

Recently, the development of deep learning (DL) allowed us to automatically learn representations from a large amount of data, thus avoiding the need for manual design features [12]. Shao et al. [3] proposed a rolling bearing fault diagnosis method based on a deep belief network (DBN). Mao et al. [13] combined auto-encoder (AE) and extreme learning machine (ELM) to diagnose fault mode of rolling bearings, which utilizes FFT spectrum of vibration signals as input. Based on DL architecture, the convolutional neural network (CNN) which is specifically designed for variable and complex signals, has shown great merits in feature extraction. Xu et al. [14] introduced VMD and deep convolutional neural networks (DCNN) to perform fault classification of the rolling bearing in wind turbines. Jia et al. [15] employed a deep normalized convolutional neural network (NCNN) for imbalanced fault classification of machinery and analyzed its mechanism via visualization. Besides, recurrent neural network (RNN), another significant DL model, shows an advantage in learning internal features from the input of sequences, so it is also widely used in the field of diagnosis in dependence of time-series vibration signals. Liu et al. [16] offered a fault diagnosis framework of rolling bearings with recurrent neural network (RNN) and auto-encoder (AE). To solve the problem of convergence difficulty and gradient extinction of common RNNs, Chen et al. [17] introduced the long short-term memory network (LSTM), a variant of RNN, to the prediction of mechanical state. Zhao et al. [18] came up with an end-to-end fault diagnosis method based on LSTM neural network, which can directly classify the raw process data without specific feature extraction and classifier design. In addition to the above listed, a large number of fault diagnosis methods based on deep learning are constantly proposed [19, 20]. CNN and RNN are the two most common architectures in the field of DL and DL-based fault diagnosis. However, these two types of neural networks have their own defects. For example, RNNs and their variants are not suitable for parallel computation, which is very inefficient for training on large-scale datasets. In addition, RNN variants still cannot completely avoid long-range dependence problem, that is, the difficulty in establishing an effective connection between distant sequences [21]. As for CNN, another pillar of DL, it also suffers from some shortcomings such as the lack of capturing the relationships between targets, the equal treatment of all pixel points and lack of pertinence [22]. Besides, the local receptive field of convolutional kernel results in the need for numerous convolutional layers to be stacked in order to obtain global information [23].

Nowadays, attention mechanism, which can associate different positions of a sequence to compute its unique representation, has now been successfully applied in a variety of tasks including natural language processing (NLP), computer vision (CV), and even fault diagnose [24]. Long et al. [25] introduced a motor fault diagnosis method using attention mechanism and improved AdaBoost. Li et al. [24] leveraged the attention mechanism to improve the data-driven diagnosis approach and visualized its effect on learned knowledge. The attention mechanism assists the deep network to focus more on informative data segments and ignore those that contribute less to the final output. However, all these attempts only embed attention mechanism as an auxiliary module into backbone models such as CNN or RNN, which cannot completely avoid the defects of these classical models.

To take full advantage of the attention mechanism, Vaswani et al. [26] proposed a new architecture based only on attention mechanism – Transformer, which abandons all the recurrent and convolutional structures. We call this version vanilla Transformer. The proposal of Transformer has set off a revolution in the field of NLP. So far, there have been numerous varieties of Transformer, and a large number of successful practices have been put into machine translation, sentence generation, etc [27]. Computer vision and other fields are also attempting to introduce and improve Transformer to meet the new challenges [28]. Considering that fault diagnosis often needs to process the signal sequence and extract its internal correlation, Transformer-like models should have yielded unusually brilliant results in this field. Besides, even when dealing with two-dimensional input such as time–frequency representation, Transformer is good at grasping its inherent temporal correlation. Unfortunately, having said that, Transformer has not been used in fault diagnosis and related fields.

To better model the temporal information in bearing signals, construct long-distance dependence and extract more effective hidden representation from its time–frequency representation (TFR), a new model named time–frequency Transformer (TFT) is proposed in this paper. Constrained by the difficulty in extracting useful features directly from the raw vibration signal [24], we apply synchrosqueezed wavelet transform (SWT) [29] to obtain time–frequency representations (TFRs). SWT has been applied to many prognostic and health management (PHM) studies of bearings due to their good performance in non-stationary vibration signal processing [30, 31, 32, 33]. Then, TFT is proposed to provide a discriminate model between TFRs and bearing fault modes. We designed a novel tokenizer focused on TFRs, and an encoder composed of Transformer blocks to establish hidden representations. Thus, we proposed an end-to-end approach for fault diagnosis. Through the case studies on bearing experimental datasets, we constructed the optimal Transformer structure and verified the performance of the proposed diagnosis method. Through comparison with the benchmark models and other state-of-the-art methods, superiority of the proposed method is proved. The main contributions of this paper can be listed as follows.

  1. 1.

    We proposed a novel time–frequency Transformer, which can avoid some drawbacks of classical models such as RNN and CNN, to extract effective information from time–frequency representation with only attention mechanism.

  2. 2.

    We proposed an end-to-end fault diagnosis framework based on time–frequency Transformer and synchrosqueezed wavelet transform, which proves to be effective and superior on bearing experimental datasets.

2 Preliminaries

This section will briefly introduce the vanilla Transformer proposed by Vaswani et al. [26] in 2017. Variants of Transformer applicated in the fields of natural language processing (NLP) and computer vision (CV) will also be reviewed.

2.1 Transformer

Recurrent models have shown a good capability to process sequence input in the form of [x1,x2,,xt]\left[x_{1},x_{2},\dots,x_{t}\right]. Along the direction of the input sequence, they generate a sequence of hidden states hth_{t}, as a function of the previous hidden state ht1h_{t-1} and the input token xtx_{t} for position tt. At each step the model is auto-regressive, consuming the previously generated symbols as additional input when generating the next. This inherently sequential nature precludes parallelization within training examples, which becomes critical at longer sequence lengths. Vaswani et al. [26] proposed Transformer, a new architecture relying entirely on attention mechanism to draw global dependencies between input and output. Transformer completely abandons the traditional recurrent structure to realize the parallel calculation of sequence input. In addition, convolution operation which is difficult to globally model the relationship between local features is eliminated from in Transformer [21].

Refer to caption

Fig. 1: Architecture of the vanilla Transformer [26].

Transformer is a multilayer structure by stacking Transformer blocks, whose vanilla form is shown in Fig. 1. The Transformer block is characterized by a multi-head self–attention mechanism, a position-wise feed-forward network, layer normalization [34] module and residual connector [35]. The input to the Transformer is often a tensor of shape b×n\mathbb{R}^{b}\times\mathbb{R}^{n}, where bb is the batch size, nn is the sequence length (note that the difference from the dimension of sequence). The input first passes through an embedding layer, which converts each one-hot token into an embedding of dd dimensions to obtain a new tensor, i.e., b×n×d\mathbb{R}^{b}\times\mathbb{R}^{n}\times\mathbb{R}^{d}. Then, the new tensor is added to a sinusoidal position encoding, and passes through a multi-head self–attention module. The input and output of the multi-head self–attention are connected by a layer normalization layer and a residual connector. The combined output is then sent to a two-layer position-wise feed-forward network, which similarly connects the input and output through a residual connector and a layer normalization layer. Such sublayer residual connectors with layer normalization have the following form

Xout=LayerNorm(FA/FF(Xin)+Xin)X_{out}={\rm LayerNorm}(F_{A/FF}(X_{in})+X_{in}) (1)

where FA/FFF_{A/FF} denotes multi-head self–attention or position-wise feed-forward layers.

2.1.1 Multi-head self–attention

The multi-head self–attention mechanism is a key defining characteristic of Transformer models, whose mechanism behind can be viewed as learning an alignment, that is, each token in the sequence attempts to gather information from others [36]. Generally, through linear transformations on a group of input embeddings XX with dimension dmodeld_{\rm model}, we get queries Qs=XWsqQ_{s}=XW_{s}^{q} and keys Ks=XWskK_{s}=XW_{s}^{k} with dimension dkd_{k}, and values Vs=XWsvV_{s}=XW_{s}^{v} with dimension dvd_{v}. A single-head scaled dot-product attention calculates the dot products of all queries and keys, divides each by scaling factor dk\sqrt{d_{k}}, and apply a softmax function to obtain the weights on the values.

As(Qs,Ks,Vs)=softmax(QsKsdk)VsA_{s}\left(Q_{s},K_{s},V_{s}\right)={\rm softmax}\left(\frac{Q_{s}K_{s}^{\top}}{\sqrt{d_{k}}}\right)V_{s} (2)

However, instead of mapping with only one version of linear transformations, it is more beneficial to project the input tokens to different queries, keys and values hh times through different learned linear transformations. Thus, the so-called multi-head self–attention is introduced. Through the parallel self–attention calculation on queries, keys, and values of each projected version, hh different outputs headi{\rm head}_{i} are obtained. These headi{\rm head}_{i} are concatenated and once again projected, resulting in the multi-head self–attention

Ah(X)=concat(head1,,headh)WOwhereheadi=As(XWiq,XWik,XWiv)\begin{split}&A_{h}\left(X\right)={\rm concat}\left({\rm head}_{1},\dots,{\rm head}_{h}\right)W^{O}\\ &{\rm where\ \rm head}_{i}=A_{s}\left(XW_{i}^{q},XW_{i}^{k},XW_{i}^{v}\right)\end{split} (3)

where Wiqdmodel×dkW_{i}^{q}\in\mathbb{R}^{d_{\rm model}\times d_{k}}, Wikdmodel×dkW_{i}^{k}\in\mathbb{R}^{d_{\rm model}\times d_{k}} and Wivdmodel×dvW_{i}^{v}\in\mathbb{R}^{d_{\rm model}\times d_{v}} denote ii-th version of linear projection on embeddings XX to obtain different queries, keys and values, respectively. WOhdv×dmodelW^{O}\in\mathbb{R}^{h\cdot d_{v}\times d_{\rm model}} denotes the linear projection on concatenated multi-head. Note that dk=dv=dmodel/hd_{k}=d_{v}=d_{\rm model}/h in Transformer.

2.1.2 Position-wise feed-forward layers

The output of multi-head self–attention module is then passed through a two-layer feed-forward network, whose hidden layer is activated by ReLU. This feed-forward layer operates on each position independently hence the term position-wise.

FF(XA)=ReLU(0,XAW1+b1)W2+b2FF\left(X_{A}\right)={\rm ReLU}\left(0,X_{A}W_{1}+b_{1}\right)W_{2}+b_{2} (4)

where W1dmodel×dffW_{1}\in\mathbb{R}^{d_{\rm model}\times d_{ff}}, W2dff×dmodelW_{2}\in\mathbb{R}^{d_{ff}\times d_{\rm model}}, b1dffb_{1}\in\mathbb{R}^{d_{ff}}, b2dmodelb_{2}\in\mathbb{R}^{d_{\rm model}} denotes the weights and bias of two layers, respectively.

2.1.3 Transformer block

Transformer composes of several stacked Transformer blocks. A Transformer block usually contains a multi-head self–attention module and a position-wise feed-forward module, both of which use a residual connector and layer normalization to get the combined output of the module.

XA=LayerNorm(Ah(X)+X)XFF=LayerNorm(FF(XA)+XA)\begin{split}&X_{A}={\rm LayerNorm}\left(A_{h}\left(X\right)+X\right)\\ &X_{FF}={\rm LayerNorm}\left(FF\left(X_{A}\right)+X_{A}\right)\end{split} (5)

where XAX_{A} and XFFX_{FF} are the output of the multi-head self–attention module and the position-wise feed-forward module, respectively. Note that the stacked multiple Transformer blocks take the same structure, while do not share the same parameters.

2.1.4 Transformer Mode

It is important to note the differences in the mode of usage of the Transformer block. Transformers generally can be divided into three categories, named: 1) encoder-only (e.g., for classification), 2) decoder-only (e.g., for language modeling), and 3) encoder-decoder (e.g., for machine translation). The vanilla Transformer proposed by Vaswani uses an encoder-decoder structure (as shown in Fig. 1) for machine translation, which is a Seq2Seq problem. The decoder of vanilla Transformer uses Transformer blocks that are different from the aforementioned those for the encoder. The proposed method in this paper, however, adopts an encoder-only structure, so details of decoders are not introduced.

2.2 Transformer variants

Recurrent Neural Networks, especially LSTM [37] and GRU [38], have been widely used for sequence modeling and inference problems such as machine translation and language modeling before Transformer was proposed [27]. The proposal of Transformer brings brand new solutions to the NLP community. Devlin et al. [39] proposed the bidirectional Transformers (BERT), which is one of the most powerful models in the NLP field. Besides, models such as XLNet [40] and GPT [41] further expand the application of Transformer. Tay et al. [42] surveyed a dizzying number of efficient Transformer variants, providing an organized and comprehensive overview of existing work and models across multiple application fields.

Transformer is used for sequence modeling to solve several problems of traditional recurrent structures: 1) Difficulty in training parallelization. 2) Difficulty in modeling long-range dependencies. In many sequence transduction tasks, learning long-range dependence is a key challenge. The recurrent models cannot establish a direct connection between non-adjacent tokens, which greatly limits the overall understanding of the input sequence. 3) The problem of gradient explosion and gradient vanishes. The proposal of LSTM and GRU alleviated this problem but did not fundamentally eliminate the weight accumulation caused by multiple recurrences.

The great success of Transformer in NLP has also attracted the attention of researchers in the field of CV. For many years, convolutional neural network (CNN) is the fundamental pillar in CV. However, this convolution operating on the pixel matrix also has some defects, e.g., it is difficult to capture the relationship between targets, and treats all pixels equally without pertinence [22]. Many attempts [43, 44, 45] introduce attention mechanisms into visual tasks. The trend of using Transformer as neural network module is more and more obvious. Dosovitskiy et al. [28] proposed a vision Transformer (ViT) for supervised image classification. It divides the image into several patches, and patches are treated the same way as tokens (words) in an NLP application. Wu et al. [22] input convolution image features into Transformer as tokens, and used them for image classification and object detection. Guo et al. [46] put forward a point cloud Transformer (PCT) for point cloud learning to solve 3D computer vision problems.

RNN and CNN, as the two most important network structures in DL, have been widely used in the field of bearing PHM. The vibration signals of bearings contain abundant temporal correlation, which can be well modeled by RNN to give effective fault diagnosis or residual useful life (RUL) prediction. CNN can extract useful fault or degradation information from the feature matrix composed of time, frequency and time–frequency domain features. A large number of model implementations based on CNN are also derived as reviewed in Section 1. Sometimes, scholars also build a hybrid model combining these two structures. A large number of attempts broaden the application of DL in the field of bearing PHM. However, as mentioned above, these two models are now dwarfed for their inherent defects by the emergence of Transformer. To the best of our knowledge, no Transformer variant has been proposed to model the vibration signal and its TFRs, which can be useful for bearing PHM.

3 Time–frequency Transformer

This section will introduce the proposed time–frequency Transformer (TFT) in detail, whose overall diagram is shown in Fig. 2. The network architecture is mainly composed of a tokenizer, an encoder and a classifier.

Refer to caption

Fig. 2: The proposed time–frequency Transformer.

3.1 Tokenizer

The vanilla Transformer accepts a 1D token sequence and obtains token embeddings with a tokenizer by dictionary query. However, the input token here to be processed is time–frequency representation (TFR), rather than the words that can be queried in a dictionary. Thus, a specific tokenizer needs to be designed. To process 2D TFR data, we will design a new tokenizer module, which mainly includes flattening, segmentation, linear mapping and adding position encodings.

3.1.1 Token embedding

To obtain the tokens sequence, TFR is segmented into several patches along the time direction. Specifically, given a TFR xNt×Nf×C{\rm x}\in\mathbb{R}^{N_{t}\times N_{f}\times C}, where NtN_{t} and NfN_{f} are the length in time and frequency direction, respectively. CC is the number of channels, which generally refers to the stacked layers of multi-sensor signals. We first reshape x\rm x into sequence xNt×(NfC){\rm x}^{\prime}\in\mathbb{R}^{N_{t}\times(N_{f}\cdot C)} to flatten multiple channels, and cut it along the time direction to get a patch sequence [xp1,xp2,,xpNt]\left[{\rm x}_{p}^{1},{\rm x}_{p}^{2},\dots,{\rm x}_{p}^{N_{t}}\right] with length NtN_{t}, where xpiNfC{\rm x}_{p}^{i}\in\mathbb{R}^{N_{f}\cdot C}, i=1,,Nti=1,\dots,N_{t} . Then, to obtain the sequence of token embeddings with dimension dmodeld_{\rm model}, a learnable linear transformation is used to obtain the projected patches sequence xt{\rm x}_{t}. This process is expressed as:

xreshapexpxt=xpWt\begin{split}&{\rm x}\stackrel{{\scriptstyle\rm reshape}}{{\longrightarrow}}{\rm x}_{p}\\ &{\rm x}_{t}={\rm x}_{p}W_{t}\end{split} (6)

where WtNf×dmodelW_{t}\in\mathbb{R}^{N_{f}\times d_{\rm model}} denotes the learnable linear mapping. Such processing is based on the view that TFR is formed by splicing the instantaneous spectrum of the signal over a period of time. This is different from gridding segmentation on images by ViT [28], because we believe that grid cutting TFR will not retain instantaneous spectrum estimation at a certain time in each patch. The segmented TFR patches can be regarded as a sequence of instantaneous spectrum in a period of time, and processing such temporal sequence is the strength of Transformer-based structure.

3.1.2 Class token

Similar to the class token in BERT [39], a randomly initialized trainable embedding xt0=xclassdmodel{\rm x}_{t}^{0}={\rm x}_{\rm class}\in\mathbb{R}^{d_{\rm model}} is added to the beginning of the embedded token sequence. Thus, an embedding sequence xt=[xclass;xp1Wt,xp2Wt,,xpNtWt]{\rm x}_{t}=\left[{\rm x}_{\rm class};{\rm x}_{p}^{1}W_{t},{\rm x}_{p}^{2}W_{t},\dots,{\rm x}_{p}^{N_{t}}W_{t}\right] with length Nt+1N_{t}+1 is obtained. Note that output zN0{\rm z}_{N}^{0} of the class token after processed by the subsequent Transformer encoder will serve as the hidden representation of TFR.

3.1.3 Position encoding

Since the Transformer contains no recurrence or convolutional operations, to make full use of the sequence order, we should inject some information about the relative or absolute position of the tokens into the embedding sequence. Vanilla Transformer employs a kind of sinusoid position encoding based on corpus dictionary and token location [26], which is not suitable for the problem we are trying to solve. We propose to use a learnable position encoding Epos(Nt+1)×dmodelE_{\rm pos}\in\mathbb{R}^{(N_{t}+1)\times d_{\rm model}} to extract position information more flexibly through the learning process. The position encodings and token embeddings are added to get the input embeddings

z0=[xclass;xp1Wt,xp2Wt,,xpNtWt]+Epos{\rm z}_{0}=\left[{\rm x}_{\rm class};{\rm x}_{p}^{1}W_{t},{\rm x}_{p}^{2}W_{t},\dots,{\rm x}_{p}^{N_{t}}W_{t}\right]+E_{\rm pos} (7)

Two types of learnable position encodings, 1D and 2D, are considered. 1) The 1D position encoding E1dpos(Nt+1)×dmodelE_{\rm 1d\ pos}\in\mathbb{R}^{(N_{t}+1)\times d_{\rm model}} is broadcasted from a vector e1dNt+1e_{\rm 1d}\in\mathbb{R}^{N_{t}+1}, i.e., E1dpos=broadcast(e1d)E_{\rm 1d\ pos}={\rm broadcast}(e_{\rm 1d}). 1D position encoding can only encode the relative and absolute position information among Nt+1N_{t}+1 tokens since the vector elements in the same token share the same encoding. 2) 2D position encoding E2dpos(Nt+1)×dmodelE_{\rm 2d\ pos}\in\mathbb{R}^{(N_{t}+1)\times d_{\rm model}} is a learning matrix with dimensions (Nt+1)×dmodel(N_{t}+1)\times d_{\rm model}, so it can simultaneously encode the location information between and within the tokens. Subsequent case studies will analyze the performance of these two encodings.

3.2 Encoder

In TFT, the encoder can be regarded as a feature extraction structure, which undertakes the task of mining category related information from the input embeddings sequence. The Transformer encoder, which is composed of NN Transformer blocks, takes the embedded sequence z0{\rm z}_{0} as the input. Our TFT model basically employs the vanilla Transformer block structure, that is, multi-head self–attention module and position-wise feed-forward layers with residual connector and layer normalization, which has been described in Section 2. Transformer block and multi-head self–attention mechanism inside are the key defining characteristics of Transformer-like models. Besides, we also made the following improvements.

3.2.1 GeLU activation

To improve the convergence of the network, we adopt Gaussian error linear units (GeLU) activation [47] in feed-forward layers instead of ReLU activation used in vanilla Transformer. GeLU is defined as the product of input xx and mask mBernouli(Φ(x))m\sim{\rm Bernouli(\Phi(x))}, where Φ(x)=P(Xx)\Phi(x)=P(X\leq x), X𝒩(0,1)X\sim\mathcal{N}(0,1) is the cumulative distribution function of the standard normal distribution. This distribution is chosen since neuron inputs tend to follow a normal distribution, especially with layer normalization. In this setting, inputs have a higher probability of being “dropped” as xx decreases, so the transformation applied to xx is stochastic yet depends upon the input.

GeLU(x)=xP(Xx)=xΦ(x)=x12[1+erf(x/2)]{\rm GeLU}(x)=xP\left(X\leq x\right)=x\Phi\left(x\right)=x\cdot\frac{1}{2}\left[1+erf\left(x/\sqrt{2}\right)\right] (8)

We can estimate GeLU as

0.5x(1+tanh[2/π(x+0.044715x3)])0.5x\left(1+{\rm tanh}\left[\sqrt{2/\pi}\left(x+0.044715x^{3}\right)\right]\right) (9)

if greater feedforward speed is worth the cost of exactness. GeLU and ReLU are plotted in Fig. 3. It can be seen that GeLU activation is continuously differentiable and has more obvious nonlinearity than the non-differentiable ReLU activation at x=0x=0.

Refer to caption

Fig. 3: The ReLU and GeLU (μ=0,σ=1\mu=0,\sigma=1).

3.2.2 Transformer blocks in TFT

The embedding sequence z0=[z00;z01;;z0Nt]{\rm z}_{0}=\left[{\rm z}_{0}^{0};{\rm z}_{0}^{1};\dots;{\rm z}_{0}^{N_{t}}\right] obtained by tokenizer goes through multiple Transformer blocks to extract the connection among the tokens using self–attention mechanism. An encoder contains NN Transformer blocks can be represented as

zl=LayerNorm(Ah(zl1)+zl1),l=1,,Nzl=LayerNorm(FF(zl1)+zl1),l=1,,N\begin{split}&{\rm z}_{l}^{\prime}={\rm LayerNorm}\left(A_{h}\left({\rm z}_{l-1}\right)+{\rm z}_{l-1}\right),l=1,\dots,N\\ &{\rm z}_{l}={\rm LayerNorm}\left(FF\left({\rm z}_{l-1}^{\prime}\right)+{\rm z}_{l-1}^{\prime}\right),l=1,\dots,N\end{split} (10)

where Ah()A_{h}(\cdot) denotes the multi-head self–attention function with hh heads in Eq. (2). FF()FF(\cdot) denotes the position-wise feed-forward module with dmodeld_{\rm model} dimensions input and output, and dffd_{ff} dimensions hidden layer. That is, the multi-head self–attention modules and the feed-forward modules are alternately connected. The extensive use of residual connectors and layer normalization reduces the training difficulty of deep network, which is conducive to faster and more stable convergence.

In TFT, the class token used to output classification features is regarded as the next token to be predicted, that is, the output of token characterizing category is regarded as an auto-regressive prediction problem. This kind of setting is born out of the idea that obtaining the information needed by the words to be predicted from the front words when Transformer is first used in NLP tasks. When applied to classification problems, this solution also enables Transformer to better establish the relationship between the input sequence and the class token to be predicted. Thus, the hidden features zN0{\rm z}_{N}^{0} of the class token output through each layer of the encoder can contain sufficient information in the input sequence. In the last layer of the encoder, the hidden features zN0{\rm z}_{N}^{0} will be served as the output of the encoder for subsequent classification.

3.3 Classifier

The function of classifier is to map the hidden features with category information to one-hot encoding of class labels. The classifier consists of two layers of feed-forward multi-layer perceptron (MLP) and a softmax activation,

CLA(zN0)=softmax(GeLU(zN0Wc1+bc1)Wc2+bc2){\rm CLA}\left({\rm z}_{N}^{0}\right)={\rm softmax}\left({\rm GeLU}\left({\rm z}_{N}^{0}W_{c1}+b_{c1}\right)W_{c2}+b_{c2}\right) (11)

where Wc1dmodel×dffW_{c1}\in\mathbb{R}^{d_{\rm model}\times d_{ff}}, Wc2dmodel×NclaW_{c2}\in\mathbb{R}^{d_{\rm model}\times N_{cla}}, bc1dffb_{c1}\in\mathbb{R}^{d_{ff}}, bc2Nclab_{c2}\in\mathbb{R}^{N_{cla}} are the weights and biases of the two layers, respectively. NclaN_{cla} is the number of categories. To reduce the number of hyperparameters, the hidden layer uses the same hidden dimension dffd_{ff} as in encoder.

Generally speaking, we can easily know the probability of each category according to the softmax output of the network. Thus, the category of input samples can be predicted.

3.4 Training of TFT

Training of TFT follows the general deep learning scheme that using stochastic gradient descent (SGD) and error back-propagation (BP) algorithm to minimize the empirical risk. Given a training set 𝒟={xi,yi}i=1n\mathcal{D}=\left\{{\rm x}_{i},{\rm y}_{i}\right\}_{i=1}^{n} contains nn samples, the network adopts the cross-entropy (CE) loss function which is suitable for the classification problem

𝒥(θ)=1ni=1nCE(yi,y^i)\mathcal{J}\left(\theta\right)=\frac{1}{n}\sum_{i=1}^{n}\mathcal{L}_{CE}\left({\rm y}_{i},\hat{\rm y}_{i}\right) (12)

where yi{\rm y}_{i} and y^i\hat{\rm y}_{i} are the expected and estimated output of sample xix_{i}. θ\theta denotes the trainable parameters of TFT and CE()\mathcal{L}_{CE}\left(\cdot\right) denotes the cross-entropy loss function.

  1. 1.

    Optimizer: We use Adam optimizer [48] for TFT training. Adam wields an adaptive gradient strategy to accelerate training error convergence and allow the training trajectory to cross non-smooth regions of the loss landscape [49].

  2. 2.

    Regularization: We introduce dropout [50] into each sub-module of the network, which only plays a role in the training stage. By cutting off the connections of some neurons, dropout forces the network to learn more robust parameters to reduce overfitting. In addition, we use label smoothing [51], mainly through soft one-hot to add noise, which reduces the weight of the real sample label category in the calculation of the loss function, to suppress the over-fitting effect.

The detailed TFT training steps are shown in Algorithm 1.

Input: Training set 𝒳={xi,yi}i=1nS\mathcal{X}=\left\{{\rm x}_{i},{\rm y}_{i}\right\}_{i=1}^{n_{S}}, where xiNt×Nf×C{\rm x}_{i}\in\mathbb{R}^{N_{t}\times N_{f}\times C}, batch size nbn_{b}
1
2Initialize {W(l),b(l)}\left\{W^{(l)},b^{(l)}\right\} of TFT;
3 for epoch=1,2,…,max_epoch do
4       for step=1,2,…,max_step do
5             //Tokenizer
6            for each xi{\rm x}_{i} in {xi}i=1nb\left\{{\rm x}_{i}\right\}_{i=1}^{n_{b}} do
7                  
8                  Flatten and slice xi{\rm x}_{i} into patches sequence, obtain xi,p=[xi,p1Wt;;xi,pNtWt]{\rm x}_{i,p}=\left[{\rm x}_{i,p}^{1}W_{t};\ldots;{\rm x}_{i,p}^{N_{t}}W_{t}\right];
9                   Linear projection, obtain [xi,t1;;xi,tNt]\left[{\rm x}_{i,t}^{1};\ldots;{\rm x}_{i,t}^{N_{t}}\right];
10                   Add class token, obtain xi,t=[xclass;xi,p1Wt;;xi,pNtWt]{\rm x}_{i,t}=\left[{\rm x}_{\rm class};{\rm x}_{i,p}^{1}W_{t};\ldots;{\rm x}_{i,p}^{N_{t}}W_{t}\right];
11                   Add position encoding, obtain zi,0=xi,t+Epos{\rm z}_{i,0}={\rm x}_{i,t}+E_{pos};
12                   Stack batches, obtain input sequences Zo{\rm Z}_{o};
13                  
14             end for
15            // Encoder
16            for block l=1,…,N do
17                  
18                  Zl=LayerNorm(Ah(Zl1)+Zl1){\rm Z}_{l}^{\prime}={\rm LayerNorm}\left(A_{h}\left({\rm Z}_{l-1}\right)+{\rm Z}_{l-1}\right);
19                   Zl=LayerNorm(FF(Zl)+Zl){\rm Z}_{l}={\rm LayerNorm}\left(FF\left({\rm Z}_{l}^{\prime}\right)+{\rm Z}_{l}^{\prime}\right);
20                  
21             end for
22            // Classifier
23            Obtain hidden representation of class token ZN0{\rm Z}_{N}^{0};
24             Y^=softmax(GeLU(ZN0Wc1+bc1)Wc2+bc2)\hat{{\rm Y}}={\rm softmax}\left({\rm GeLU}\left({\rm Z}_{N}^{0}W_{c1}+b_{c1}\right)W_{c2}+b_{c2}\right);
25            
26            // Calculate loss and gradient descent
27            Cross entropy CE(Y,Y^)\mathcal{L}_{CE}\left({\rm Y},\hat{\rm Y}\right);
28             Batch loss 𝒥(Y,Y^;W(l),b(l))\mathcal{J}\left({\rm Y},\hat{\rm Y};W^{(l)},b^{(l)}\right);
29             Calculate gradients 𝒥W(l)\frac{\partial\mathcal{J}}{\partial W^{(l)}},𝒥b(l)\frac{\partial\mathcal{J}}{\partial b^{(l)}};
30             Apply gradients W(l)W(l)η𝒥W(l)W^{(l)}\leftarrow W^{(l)}-\eta\frac{\partial\mathcal{J}}{\partial W^{(l)}},b(l)b(l)η𝒥b(l)b^{(l)}\leftarrow b^{(l)}-\eta\frac{\partial\mathcal{J}}{\partial b^{(l)}};
31            
32       end for
33      
34 end for
Output: Weights and biases {W(l),b(l)}\left\{W^{(l)},b^{(l)}\right\}
Algorithm 1 Training of time–frequency Transformer.

4 Fault diagnosis framework based on TFT

To improve the generalization performance of intelligent fault diagnosis and speed up learning and inference, a new fault diagnosis method of rolling bearings based on TFT is proposed. Specifically, the process framework of the fault diagnosis method based on TFT is shown in Fig. 4, its specific diagnostic steps can be described as follows:

  1. 1.

    Collecting vibration signals from the rolling bearings.

  2. 2.

    Converting the collected vibration signals into TFRs through synchrosqueezed wavelet transforms (SWT), and label the samples for training.

  3. 3.

    The relevant hyperparameters and model structure of the TFT are determined.

  4. 4.

    The established model is fully trained and then applied to identify test samples.

  5. 5.

    Outputting the fault diagnosis results and evaluating the performance of the proposed method.

Refer to caption


Fig. 4: The framework of the proposed method.

5 Case studies and analysis

To verify and analyze the effectiveness of the proposed TFT and its application in bearing fault diagnosis, we implement two case studies in this section. The two bearing datasets used are collected from accelerated bearing life tester (ABLT-1A). The testing system is mainly composed of a test head, sensors, test bearings, electronic control system, computer monitoring system and data acquisition system. The specific real scene is shown in Fig. 5. In our implementation, MATLAB is used for signal processing, while TensorFlow 2.1 framework is used for deep learning. All programs run on a computer with the following configuration: AMD Ryzen 2600, NVIDIA RTX 1060, 16GB RAM.

In the following case studies, we use three most popular deep learning models as benchmark models, as follows: 1) Multilayer perceptron (MLP) [52], as known as deep neural network (DNN). 2) Convolutional neural networks (CNN). To improve the convergence performance of deep convolutional networks, we employ ResNet18 [35] with residual connection. 3) Recurrent neural network (RNN). To alleviate the long-range dependence problem of classical RNN, we use the improved GRU [38].

Refer to caption

Fig. 5: ABLT-1A.

5.1 Case 1: ABLT-1A Bearing Dataset 6308

5.1.1 Dataset description

The rolling bearing HRB6308 is selected as the experimental bearing in this experiment. The faulty bearing is installed in the first channel of the sensor, and the other three normal bearings are installed in the remaining channels of the sensor. The vibration signal of the fault or normal rolling bearing is collected by the monoaxial vibration acceleration sensor with a sampling frequency of 12,800Hz. The experimental dataset is detailed as follows. Under zero-load conditions, seven types of fault conditions such as normal (6308N), inner ring fault (6308IRF), inner ring weak fault (6308IRWF), outer ring fault (6308ORF), outer ring weak fault (6308ORWF), inner and outer ring compound fault (6308IORF), and inner and outer ring compound weak fault (6308IORWF) are simulated. Accordingly, vibration data were collected for each failure type of the bearings, which were running at 1,050 rpm. For each failure mode interception, 2000 samples of the length 1024 were obtained, totaling 2000×7=140002000\times 7=14000 samples. In addition, 60% of all datasets are used as training dataset, 20% as validation dataset for model selection and cross validation, and 20% as test dataset for final test. In each training and test, the datasets are randomly divided to ensure the comprehensive evaluation of the model performance.

5.1.2 Data preprocessing

While the raw vibration signal contains sufficient health condition information of bearings, it is not clear to be used for fault diagnosis directly [24]. Considering the non-stationarity of vibration signals, synchrosqueezed wavelet transforms (SWT) is used to process the raw data to obtain TFRs. The resolution of TFRs obtained by SWT is 1280×25601280\times 2560, which is too large to input a network. This will significantly increase scale of the network and computational expense. Similar to Ref. [53], the bicubic interpolation algorithm [54] is used to resize the resolution to 224×224224\times 224, which is the common input size. Finally, input images with a shape of 224×224×1224\times 224\times 1 are obtained. Vibration signals and corresponding TFRs of bearings with different fault modes are drawn in Fig. 6. These TFRs will be the input of the proposed fault diagnosis model.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Fig. 6: Signals and SWT results of bearings (a) 6308N (b) 6308IRF (c) 6308IRWF (d) 6308ORF (e) 6308ORWF (f) 6308IORF (g) 6308IORWF.

5.1.3 Model selection

In this section, we will determine the structure and hyperparameter selection of the model based on the cross-validation on ABLT-1A Bearing Dataset 6308, and evaluate the influence of these hyperparameter selections on the model size and generalization performance. To increase the feasibility of the research work, all parameters and diagnostic results are cross-validated for multiple average verification and analysis. The hyperparameters we need to determine include: (A) embedding dimension dmodeld_{\rm model} and hidden dimension dffd_{ff}, (B) number of attention heads hh, (C) number of Transformer blocks NN, (D) dropout rate rdpr_{dp}, and (E) selection of position encoding. Each model was trained 10 times for cross-validation. Specifically, the average of the results of cross-validation is employed. The test results of each group are shown in Table 1, where base denotes the final selected model for subsequent research, average acc denotes mean testing accuracy, and params denotes the total number of trainable parameters in the model. The empty item in the params column indicates that this hyperparameter does not affect the total number of trainable parameters. It can be seen from the table that the selection of these parameters has a certain impact on the model scale and performance. In particular, different dimensions and encoder layers will greatly affect the scale of the model.

Table 1: Model selection and influence of some hyperparameters.
dmodeld_{\rm model} dffd_{ff} hh NN rdpr_{dp} Position coding Average acc % Params num
base 64 256 8 6 0.1 1D 99.94 335,016
A 16 32 99.14 18,456
32 64 99.21 62,280
64 128 99.52 226,728
128 512 99.73 1,292,392
256 1024 99.93 5,074,920
B 1 94.46
2 97.05
4 99.22
16 99.59
C 2 98.57 135,080
4 99.72 235,048
8 99.91 434,984
D 0.01 98.24
0.5 74.95
E None 97.77 334,791
2D 99.78 349,191

In addition, for the embedding dimension dmodeld_{\rm model}, hidden layer dimension dffd_{ff} and the number of attention heads hh that have a great impact on performance, the statistical box chart of test results is shown in Fig. 7. Too small embedding dimension and hidden layer dimension cannot provide the network with enough parameterization ability, so the network performance is poor. In contrast, too large dimension will lead to over-parameterization of the network. At this time, our samples will not be enough to train the over-parameterized network, which will lead to the degradation of the generalization performance. This is also called over-fitting. Besides, we notice that unreasonable selection of hyperparameters will not only reduce the average accuracy, but also make the network performance more unstable, that is, the error distribution is more dispersed. The number of attention heads also shows a similar pattern. Finally, the optimal network structure and hyperparameter selection of TFT are shown in Table 2.

Refer to caption
Refer to caption
Fig. 7: The influence (a) embedding dimension and hidden dimension (b) the number of attention heads on the accuracy of time–frequency Transformer.
Table 2: Optimal structure and hyperparameters of the proposed time–frequency Transformer.
Value
Input size [224, 224, 3]
Batch size 32
Max epochs 100
Learning rate lrlr 5e-5
Optimizer Adam
Label smoothing rate ϵls\epsilon_{ls} 0.1
Num of encoder layers NN 6
Embedding dimension dmodeld_{\rm model} 64
Hidden dimension dffd_{ff} 256
Num of attention heads hh 8
Dropout raterdpr_{dp} 0.1
Position encoding 1D

For a fair comparison, the parameter settings of the three benchmark models are also standardized. The detailed model structure and hyperparameter settings are shown in Table 3.

Table 3: Structure and hyperparameter setting of three benchmark models.
Model Structure (units and activation) Hyperparameter
DNN Dense (512, activation=’ReLU’) Dropout ( ) Dense (128, activation=’ReLU’) Dropout ( ) Dense (num_class, activation=’ReLU’)
Dropout rate rdpr_{dp}
Max epochs = 100
Batch size = 32
Optimizer = Adam(lr=lr=1e-5)
CNN ResNet18 ( ) GlobalAveragePooling2D ( ) Dense(num_class)
Dropout rate rdpr_{dp}
Max epochs = 100
Batch size = 16
Optimizer = Adam (lr=lr=5e-5)
GRU GRU (224, dropout) × 6 Dense (128, activation=’ReLU’) Dense (num_class, activation=’ReLU’)
Dropout rate rdpr_{dp}
Max epochs = 100
Batch size = 32
Optimizer = Adam (lr=lr=5e-5)

5.1.4 Diagnosis results

Based on the optimal network structure and hyperparameter setting, the TFT model is trained on ABLT-1A Bearing Dataset 6308. As can be seen from Fig. 8, after about 40 epochs, accuracy and loss values of the training and validation datasets have been very stable, and the model has started to converge, indicating the strong convergence ability of the TFT. Note that in the early stage of training, the training loss is larger than the validation loss because the use of dropout limits the model capacity in training. With the process of training, dropout will drive the network to learn more robust features. Finally, the training loss and validation loss of the network are basically stable at the same value, indicating the good generalization ability of the network. The regularization technique we used fully guarantees the robust generalization of the network.

Refer to caption
Refer to caption
Fig. 8: The (a) loss and (b) accuracy curves of the proposed TFT.

Then, the trained model is used to classify the testing dataset to evaluate performance. Training and test have been repeated 10 times on TFT and three benchmark models. We draw the results in Table 4. Besides, the total number of trainable parameters and the average training time are also calculated to comprehensively compare the performance of the models. Comparing the test performance of the 4 models, the proposed TFT achieves the best prediction accuracy. Its maximum prediction accuracy can reach 100%, and the average accuracy is also the highest. The accuracy variance of TFT is smaller, indicating that the prediction result is more stable. The performance of RNN is second only to TFT, whose maximum accuracy is 100% and average accuracy is lower than that of TFT. Meanwhile, the variance of RNN is larger, so the result is not as stable and reliable as TFT. The prediction accuracy of DNN and CNN is significantly lower than that of TFT and RNN, which is obviously related to the characterization ability of the model itself. DNN and CNN are just simple multiplication or convolution operation on inputs, lacking the grasp of temporal information.

Table 4: Test performance, size and training time usage of the models on ABLT-1A Bearing Dataset6308.
Model Mean accuracy Best accuracy Std Params num Training time (s)
TFT 99.94% 100.00% 0.05 335,016 690
DNN 80.15% 85.71% 3.81 25,757,191 740
CNN 92.56% 97.83% 0.55 11,176,839 1030
RNN 97.03% 100.00% 1.56 1,844,103 1800

Furthermore, we compare the scale and training time of these models. DNN contains the most trainable parameters because of its internal fully connected structure. This will lead to the model over parameterized, thus make it easier to overfit when the number of samples is limited. Interestingly, the parameters of CNN are slightly less than that of DNN, but the training time is much longer than that of DNN. This is obviously due to the fact that although convolutional structure reduces the number of trainable parameters through weight sharing, calculation amount corresponding to each parameter increases significantly. The number of trainable parameters in RNN is much less than that in CNN and DNN, but the training time of RNN is the longest for its non-parallel computation.

Generally speaking, the accuracy of TFT and RNN which can grasp the temporal information is higher. But for RNN, this comes at the cost of a very long training time. The results show that the proposed TFT method based on the Transformer structure can process the time series information well with higher prediction accuracy. Moreover, TFT can realize parallel computing, so its training is much faster than models with recurrent structure.

Fig. 9 shows the confusion matrix of the best and the worst test results of TFT, with the accuracy of 100% and 99.88%, respectively. The columns represent the predict labels while the rows represent the true labels for different health states. In the best case, the accuracy is 100% for each health condition. In the worst one, the accuracy is 100% for most conditions except for IORWF and IRWF. A small number of samples with IORWF and IRWF were misjudged as IRWF and ORWF, respectively, indicating greater challenge of weak fault classification. It is also worth noting that, even in the worst case, TFT does not fail to identify the three non-weak fault states and the normal state.

Refer to caption
Refer to caption
Fig. 9: Confusion matrix of (a) the best and (b) the worst results of time–frequency Transformer on ABLT-1A Bearing Dataset 6308.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Fig. 10: The 2D visualization result of learned features via (a) TFT (b) DNN (c) CNN (d) RNN.

To visualize the extracted features of these models, t-distributed stochastic neighbor embedding (t-SNE) [55] is used to simplify the extracted high-dimensional features of the last hidden layer extracted by the above four methods into two-dimensional vector distribution, the visualization results of the test samples are shown in Fig. 10. As indicated that only the hidden features extracted by the TFT algorithm can accurately separate all faults.

5.1.5 Visualization of attention weights

Since the proposed TFT extracts features from TFRs completely based on attention mechanism, it is necessary to explore its mechanism by attention visualization. The attention mechanism adopted by Transformer is mainly used to exert different degrees of attention on the tokens and form the relationship between different tokens. Here, we try to show the attention degree of attention mechanism on different tokens, namely attention weight. Firstly, the attention weight tensors of the first and last self–attention layers are derived. Note that the attention weight is not the output of the attention layer, but the weight softmax(QsKs/dk){\rm softmax}\left(Q_{s}K_{s}^{\top}/\sqrt{d_{k}}\right) of the input in Eq. 2. Since the calculation results of multi-head attention are concatenated in the network, we sum the weight matrix of hh attention heads. Furthermore, the target of attention heads is tokens, and our TFT uses time–frequency cutting patches as tokens. Therefore, the attention mechanism actually works along the time direction. The attention mechanism and the TFRs of bearing signals are drawn in Fig. 11. These figures show the normalized attention weight of the first and last attention layers on different tokens. The larger the value is, the greater the attention weight is.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Fig. 11: Attention weights visualization of bearings (a) 6308N (b) 6308IRF (c) 6308IRWF (d) 6308ORF (e) 6308ORWF (f) 6308IORF (g) 6308IORWF.

It can be seen from Fig. 11 that the attention weight distribution of different fault samples in the first transformer block is almost the same. This obviously adheres to our intuition since the network cannot distinguish different fault types in the input layer, but “observe” different samples with the same strategy. With the layer-by-layer attention processing, the network will be able to attach different attention weights to different fault types. Combined with the TFRs, it can be found that the last Transformer block focuses on the tokens with larger values, that is, it pays more attention to the time when the amplitude is more obvious. Through such concentration of attention, the TFT model can effectively grasp the characteristic information from the TFRs, and observe each token with different weights. Thus, TFT can accurately extract the key features of different fault types and avoid the interference of fault independent factors.

5.1.6 Anti-noise robustness test

To evaluate the anti-noise performance and robustness of the designed algorithm and fault diagnosis method, we operate an anti-noise robustness test. Specially, different degrees of signal-to-noise (SNR) is added into the original bearing signals. The same experimental setup was used to evaluate the variation in performance of each fault diagnosis methods above with different SNR Settings. By adding different degrees of noise to the vibration signal, this experiment evaluates the ability of these fault diagnosis methods to adapt to a noise-changing environment. As shown in Fig. 12, it is worth noting that the proposed method achieves the best performance. When SNR is bigger than 5, that is, under the condition of relatively small noise, the proposed TFT can provide an acceptable prediction accuracy. With the SNR decreasing, the accuracy of all these methods decreases. When the SNR is smaller than -5, the accuracy of DNN and CNN almost decreased to a meaningless value. Note that the meaningless value means the accuracy almost equal to 14.29% of guess blindly. To sum up, the proposed TFT method has relatively high diagnosis accuracy under noise interference environment, and its accuracy is higher than those of other benchmark methods. Also, the proposed TFT method is less disturbed by noise, and the accuracy decreases more slowly with the increase of noise.

Refer to caption

Fig. 12: Average diagnostic accuracy for different fault diagnosis methods under different SNRs.

5.2 ABLT-1A Bearing Dataset 6205

5.2.1 Dataset and benchmark test description

The rolling bearing HRB6205 is selected as the experimental bearing in this experiment. Different the previous case, a triaxial acceleration sensor with the sampling frequency of 12,000Hz is used to collect vibration signals of fault or normal rolling bearing. Thus, the original vibration signal is converted into three-channel digital signal by a data acquisition card. In addition, we have taken into account the different rotational speeds of 1200 rpm, 1500 rpm, 1750 rpm, 2000 rpm. The specific experimental dataset is described as follows. Under zero-load conditions, six types of fault conditions such as normal (6205N), inner ring fault (6205IRF), outer ring fault (6205ORF), ball fault (6205BF), inner and outer ring compound fault (6205IORF), and outer ring and ball compound fault (6205ORBF) are simulated respectively. We obtained 1000 samples with a length of 1024 for each health state and each speed. That is, a total of 1000×6×4=240001000\times 6\times 4=24000 samples. To distinguish the samples under different speeds, the dataset is divided into four subsets according to the speed, and each subset contains 6000 samples. The benchmark will be conducted on each sub-dataset. 60% of each subset set is used as training dataset, 20% as validation dataset, and 20% as test dataset for final test. In each training and testing, the subsets are randomly divided to ensure the comprehensive evaluation of the model performance.

The proposed TFT model and three benchmark methods in Case 1 are used in the benchmark test. The model structure and hyperparameter setting are basically the same as the optimal setting in Case 1. Similarly, SWT is used to process the vibration signals of each sample to obtain the TFRs. Differently, it should be noted that the signals from the three channels are processed separately and then stacked into multi-channel TFRs to make full use of the multi-channel information in the dataset. After downsampling by bicubic interpolation algorithm, the input data with a shape of 224×224×3224\times 224\times 3 is obtained. The proposed TFR and CNN can directly process multi-channel input, while DNN and RNN need flattening operation in the input layer.

5.2.2 Diagnosis with multi-channel TFRs

Based on the optimal network structure and hyperparameter setting, the TFT model is trained respectively on four subsets of Bearing Dataset 6205. In addition, the three benchmark models are fully trained for comparison. The fully trained models are applied to the diagnosis of test set samples in each subset, and the average results are shown in Table 5. Note that the test for each subset is repeated five times, and the table shows the average accuracy of all subsets.

It can be seen that the accuracy of TFT is higher when multi-channel data is used. Both TFT and CNN can directly input multi-channel data and construct feature graphs fusing multi-channel information. When DNN and RNN are dealing with larger dimension input data, the large network size obviously limits the generalization ability. Besides, the large amount of multi-channel input data does not significantly increase the training time of TFT and CNN due to the parallel computing capability. Particularly, hardly has the training time of TFT increased. Since RNN has no parallel processing ability and no learnable embedding module, its training time is greatly increased. It can be seen that compared with other benchmark models, the proposed TFT can better deal with multi-channel TFRs and obtain better diagnosis performance with lower computational cost and network size.

Table 5: Test performance, size and training time usage of the models on ABLT-1A Bearing Dataset 6205.
Average Params num Training time (s)
TFT 99.97% 363,431 750
DNN 67.17% 77,137,286 1,510
CNN 92.65% 11,182,598 1,320
RNN 96.38% 16,368,134 3,700

5.2.3 Diagnosis across different conditions

Bearing Dataset 6205 contains the vibration signals of bearings collected at different speeds, which enables us to test the performance of TFT diagnosis under multiple speed conditions. Different from the previous benchmark test, which divided the dataset into four subsets, we mixed and scrambled all samples with different speeds in this test. Therefore, the mixed dataset contains 24000 samples in total. Based on the optimal TFT structure and hyperparameter setting, the model training is repeated 10 times and used for inference of test samples. The final result shows that the average classification accuracy is 99.87%, and the highest accuracy is 99.92%. The confusion matrix of the best results is shown in Fig. 13. Only a few samples with the real label of IORF were misclassified into BF.

To further investigate the fault feature extraction of TFT under multiple speed conditions, t-SNE is used to visualize the hidden features as shown in Fig. 14. Different colors in the figure represent different health states, while different markers represent different speeds. It can be seen that the hidden features after highly abstracted with TFT have good distinguishability. The samples of different health states can be well distinguished. Furthermore, the samples of the same state, even if with different speeds, can concentrate well. Only in the samples with IORF and ORBF faults, the 1200r samples are slightly separated from the samples with other speeds, but this does not prevent the classifier from classifying them into one category. In addition, several IORF 1500r samples that were misclassified as BF failures can also be clearly observed.

Refer to caption

Fig. 13: Confusion matrix of the best results of time–frequency Transformer on multiple speed ABLT-1A Bearing Dataset 6308.

Refer to caption

Fig. 14: The 2D visualization result of learned features from different rotate speed.

5.3 Comparison with state-of-the-art

The above comparisons with three representative benchmark models significantly demonstrates the advantages of TFT in terms of feature extraction capability, network size, training speed, classification accuracy, and robustness. These comparisons are mainly aimed at the performance play of the network itself, while may be not sufficient for specific target tasks. Therefore, comparison with other fault diagnosis methods in practical application is necessary to verify the effectiveness of TFT-based diagnosis.

In this section, several state-of-the-art techniques for fault diagnosis are employed to comparison on Dataset6308 and Dataset 6205. These comparison methods include Li’s extreme learning machine (ELM) [56], Zhang’s support vector machine (SVM) [57], Xu’s deep convolutional neural networks (DCNN) [14], Chen’s convolutional neural networks (CNN) [58], Shao’s deep belief network (DBN) [3], Chen’s long short-time memory (LSTM) network [17], Liu’s GRU-based non-linear predictive denoising autoencoder (GRU-NP-DAE) [16], Shao’s auto-encoder (AE) [19] and Mao’s auto-encoder extreme learning machine (AE-ELM) [13]. Most of these methods are based on the latest CNN and RNN type models and utilize different inputs, e.g. vibration signals, time domain statistics, frequency domain, and time–frequency domain features, respectively.

Table 6: Performance of the comparison methods on the two datasets.
Diagnose methods Signal processing Dataset6308 Dataset6205 Average
ELM [56] MRSVD (multi-resolution SVD) 79.58% 71.95% 75.77%
SVM [57] EEMD (ensemble EMD) 88.90% 87.94% 88.42%
DCNN [14] VMD 93.67% 95.82% 94.75%
CNN [58] DWT (discrete wavelet transform) 97.54% 98.16% 97.85%
DBN [3] Time-domain statistic features 98.20% 99.75% 98.98%
LSTM [17] EMD 99.12% 97.54% 98.33%
AE-ELM [13] FFT 91.03% 87.58% 89.31%
AE [19] Vibration signals 93.59% 92.81% 93.20%
GRU-NP-DAE [16] Vibration signals 96.17% 95.38% 95.78%
TFT SWT 99.94% 99.97% 99.96%

These state-of-the-art methods are used for Dataset 6308 and Dataset 6205, respectively, whose results are shown in Table 6. The performance of these methods varies using different data processing methods and diagnostic models. Among them, models with the ability to grasp temporal information, such as RNNs and their variants, clearly bring more significant performance gains compared to traditional ELMs and SVMs. Besides, TFT outperforms other DL-based approaches even all receive time–frequency inputs, which is consistent with the benchmark in case studies before. Among all these state-of-the-art solutions, the proposed TFT achieves the highest classification accuracy on both datasets, which further proves the superiority of the proposed method on rolling bearing fault diagnosis.

6 Conclusion

In this paper, we proposed a new time–frequency Transformer (TFT) inspired by vanilla Transformer architecture to process time–frequency representations (TFRs). Based on TFT, this paper presented an end-to-end fault diagnosis framework for rolling bearings. In this method, the vibration signals of rolling bearing are processed by SWT to obtain multi-channel TFRs, and then the TFRs are input into TFT to extract discriminative hidden features and accurately classify fault modes.

The proposed TFT has the following characteristics: 1) TFT completely abandons the tedious recurrence structure and convolution operation commonly used in traditional DL-based frameworks, completely relying on multi-head self–attention mechanism and feed-forward neural network layers. Compared with classical CNN and RNN, this greatly improves the parallel computing ability of the network and reduces the network scale. 2) The feature extraction structure based on self–attention mechanism with residual connector can accurately focus on the effective feature areas in the input graphs. Compared with RNNs that with long-range dependent defects, this model can better establish the relationship in input sequences.

The effectiveness of the proposed fault diagnosis method of rolling bearing based on TFT is verified by case studies on several experimental data in this paper. Compared with other benchmark models and state-of-the-art methods, the superiority of this method is verified. The diagnosis framework has the following advantages: 1) Compared with other methods based on classical deep learning model, this method has higher diagnosis accuracy and faster training speed. 2) This method can make better use of the collected multi-channel signals, and contribute this advantage to higher diagnostic accuracy and efficiency. 3) This method can adapt to a certain degree of noise environment and provide effective fault diagnosis under strong noise conditions. 4) This method can be used for fault diagnosis under multiple working conditions (speeds).

Considering that there is still much room for improvement, our future work will focus on the following. 1) We will try to apply Transformer architecture to the prognostic field. 2) We will try to further improve the model, such as using convolution operation in the tokenizer module to improve the local receptive field.

Acknowledge

The authors gratefully acknowledge the financial support of the National Natural Science Foundation of China (No. 52075095).

References

  • [1] X. Zhao, M. Jia, J. Bin, T. Wang, Z. Liu, Multiple-Order Graphical Deep Extreme Learning Machine for Unsupervised Fault Diagnosis of Rolling Bearing, IEEE Trans. Instrum. Meas. 70 (2021). doi:10.1109/TIM.2020.3041087.
  • [2] X. Yan, M. Jia, A novel optimized SVM classification algorithm with multi-domain feature and its application to fault diagnosis of rolling bearing, Neurocomputing 313 (2018) 47–64. doi:10.1016/j.neucom.2018.05.002.
  • [3] H. Shao, H. Jiang, X. Zhang, M. Niu, Rolling bearing fault diagnosis using an optimization deep belief network, Meas. Sci. Technol. 26 (11) (2015). doi:10.1088/0957-0233/26/11/115002.
  • [4] P. M. Frank, S. X. Ding, T. Marcu, Model-based fault diagnosis in technical processes, Transactions of the Institute of Measurement and Control 22 (1) (2000) 57–101. doi:10.1177/014233120002200104.
  • [5] Z. Gao, C. Cecati, S. X. Ding, A Survey of Fault Diagnosis and Fault-Tolerant Techniques—Part I: Fault Diagnosis With Model-Based and Signal-Based Approaches, IEEE Trans. Ind. Electron. 62 (6) (2015) 3757–3767. doi:10.1109/TIE.2015.2417501.
  • [6] S. Dong, X. Xu, R. Chen, Application of fuzzy C-means method and classification model of optimized K-nearest neighbor for fault diagnosis of bearing, J Braz. Soc. Mech. Sci. Eng. 38 (8) (2016) 2255–2263. doi:10.1007/s40430-015-0455-9.
  • [7] D. Xiao, J. Ding, X. Li, L. Huang, Gear Fault Diagnosis Based on Kurtosis Criterion VMD and SOM Neural Network, Appl. Sci. 9 (24) (2019) 5424. doi:10.3390/app9245424.
  • [8] P. Durkhure, A. Lodwal, Fault Diagnosis of Ball Bearing Using Time Domain Analysis and Fast Fourier Transformation.
  • [9] O. R. Seryasat, M. A. shoorehdeli, F. Honarvar, A. Rahmani, Multi-fault diagnosis of ball bearing using FFT, wavelet energy entropy mean and root mean square (RMS), in: 2010 IEEE International Conference on Systems, Man and Cybernetics, 2010, pp. 4295–4299. doi:10.1109/ICSMC.2010.5642389.
  • [10] Y. Yu, YuDejie, C. Junsheng, A roller bearing fault diagnosis method based on EMD energy entropy and ANN, Journal of Sound and Vibration 294 (1) (2006) 269–277. doi:10.1016/j.jsv.2005.11.002.
  • [11] Q. Wang, Y. B. Liu, X. He, S. Y. Liu, J. H. Liu, Fault Diagnosis of Bearing Based on KPCA and KNN Method (2014). doi:10.4028/www.scientific.net/AMR.986-987.1491.
  • [12] R. Salakhutdinov, Deep learning, in: S. A. Macskassy, C. Perlich, J. Leskovec, W. Wang, R. Ghani (Eds.), The 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’14, New York, NY, USA - August 24 - 27, 2014, ACM, 2014, p. 1973. doi:10.1145/2623330.2630809.
    URL https://doi.org/10.1145/2623330.2630809
  • [13] W. Mao, J. He, Y. Li, Y. Yan, Bearing fault diagnosis with auto-encoder extreme learning machine: A comparative study, Proceedings of the Institution of Mechanical Engineers, Part C: Journal of Mechanical Engineering Science 231 (8) (2017) 1560–1578. doi:10.1177/0954406216675896.
  • [14] Z. Xu, C. Li, Y. Yang, Fault diagnosis of rolling bearing of wind turbines based on the Variational Mode Decomposition and Deep Convolutional Neural Networks, Appl. Soft Comput. J. 95 (2020). doi:10.1016/j.asoc.2020.106515.
  • [15] F. Jia, Y. Lei, N. Lu, S. Xing, Deep normalized convolutional neural network for imbalanced fault classification of machinery and its understanding via visualization, Mech. Syst. Signal Process. 110 (2018) 349–367. doi:10.1016/j.ymssp.2018.03.025.
  • [16] H. Liu, J. Zhou, Y. Zheng, W. Jiang, Y. Zhang, Fault diagnosis of rolling bearings with recurrent neural network-based autoencoders, ISA Transactions 77 (2018) 167–178. doi:10.1016/j.isatra.2018.04.005.
  • [17] Z. Chen, Y. Liu, S. Liu, Mechanical state prediction based on LSTM neural netwok, in: 2017 36th Chinese Control Conference (CCC), 2017, pp. 3876–3881. doi:10.23919/ChiCC.2017.8027963.
  • [18] H. Zhao, S. Sun, B. Jin, Sequential Fault Diagnosis Based on LSTM Neural Network, IEEE Access 6 (2018) 12929–12939. doi:10.1109/ACCESS.2018.2794765.
  • [19] H. Shao, H. Jiang, H. Zhao, F. Wang, A novel deep autoencoder feature learning method for rotating machinery fault diagnosis, Mechanical Systems and Signal Processing 95 (2017) 187–204. doi:10.1016/j.ymssp.2017.03.034.
  • [20] H. Zhang, R. Wang, R. Pan, H. Pan, Imbalanced Fault Diagnosis of Rolling Bearing Using Enhanced Generative Adversarial Networks, IEEE Access 8 (2020) 185950–185963. doi:10.1109/access.2020.3030058.
  • [21] J. F. Kolen, S. C. Kremer, Gradient Flow in Recurrent Nets: The Difficulty of Learning LongTerm Dependencies, in: A Field Guide to Dynamical Recurrent Networks, 2010. doi:10.1109/9780470544037.ch14.
  • [22] B. Wu, C. Xu, X. Dai, A. Wan, P. Zhang, Z. Yan, M. Tomizuka, J. Gonzalez, K. Keutzer, P. Vajda, Visual Transformers: Token-based Image Representation and Processing for Computer Vision, ArXiv200603677 Cs Eess (2020). arXiv:2006.03677.
  • [23] A. Gulati, J. Qin, C.-C. Chiu, N. Parmar, Y. Zhang, J. Yu, W. Han, S. Wang, Z. Zhang, Y. Wu, R. Pang, Conformer: Convolution-augmented Transformer for Speech Recognition, ArXiv200508100 Cs Eess (2020). arXiv:2005.08100.
  • [24] X. Li, W. Zhang, Q. Ding, Understanding and improving deep learning-based rolling bearing fault diagnosis with attention mechanism, Signal Process. 161 (2019) 136–154. doi:10.1016/j.sigpro.2019.03.019.
  • [25] Z. Long, X. Zhang, L. Zhang, G. Qin, S. Huang, D. Song, H. Shao, G. Wu, Motor fault diagnosis using attention mechanism and improved adaboost driven by multi-sensor information, Measurement 170 (2021) 108718. doi:10.1016/j.measurement.2020.108718.
  • [26] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, I. Polosukhin, Attention is all you need, in: I. Guyon, U. von Luxburg, S. Bengio, H. M. Wallach, R. Fergus, S. V. N. Vishwanathan, R. Garnett (Eds.), Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, 2017, pp. 5998–6008.
    URL https://proceedings.neurips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html
  • [27] N. Shazeer, A. Mirhoseini, K. Maziarz, A. Davis, Q. V. Le, G. E. Hinton, J. Dean, Outrageously large neural networks: The sparsely-gated mixture-of-experts layer, in: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings, OpenReview.net, 2017.
    URL https://openreview.net/forum?id=B1ckMDqlg
  • [28] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, N. Houlsby, An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale, ArXiv201011929 Cs (2020). arXiv:2010.11929.
  • [29] I. Daubechies, J. Lu, H. T. Wu, Synchrosqueezed wavelet transforms: An empirical mode decomposition-like tool, Appl. Comput. Harmon. Anal. 30 (2) (2011) 243–261. doi:10.1016/j.acha.2010.08.002.
  • [30] J. Wen, H. Gao, S. Li, L. Zhang, X. He, W. Liu, Fault diagnosis of ball bearings using Synchrosqueezed wavelet transforms and SVM, in: Proceedings of 2015 Prognostics and System Health Management Conference, PHM 2015, 2016. doi:10.1109/PHM.2015.7380084.
  • [31] C. Li, V. Sanchez, G. Zurita, M. Cerrada Lozada, D. Cabrera, Rolling element bearing defect detection using the generalized synchrosqueezing transform guided by time-frequency ridge enhancement, ISA Trans. 60 (2016) 274–284. doi:10.1016/j.isatra.2015.10.014.
  • [32] C. Yi, Y. Lv, H. Xiao, T. Huang, G. You, Multisensor signal denoising based on matching synchrosqueezing wavelet transform for mechanical fault condition assessment, Meas. Sci. Technol. 29 (4) (2018). doi:10.1088/1361-6501/aaa50a.
  • [33] W. Liu, W. Chen, Z. Zhang, A Novel Fault Diagnosis Approach for Rolling Bearing Based on High-Order Synchrosqueezing Transform and Detrended Fluctuation Analysis, IEEE Access 8 (2020) 12533–12541. doi:10.1109/ACCESS.2020.2965744.
  • [34] J. L. Ba, J. R. Kiros, G. E. Hinton, Layer Normalization, ArXiv160706450 Cs Stat (2016). arXiv:1607.06450.
  • [35] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in: 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, IEEE Computer Society, 2016, pp. 770–778. doi:10.1109/CVPR.2016.90.
  • [36] D. Bahdanau, K. Cho, Y. Bengio, Neural machine translation by jointly learning to align and translate, in: Y. Bengio, Y. LeCun (Eds.), 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015.
    URL http://arxiv.org/abs/1409.0473
  • [37] S. Hochreiter, J. Schmidhuber, Long Short-Term Memory, Neural Comput. 9 (8) (1997) 1735–1780. doi:10.1162/neco.1997.9.8.1735.
  • [38] J. Chung, Ç. Gülçehre, K. Cho, Y. Bengio, Gated feedback recurrent neural networks, in: F. R. Bach, D. M. Blei (Eds.), Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6-11 July 2015, Vol. 37 of JMLR Workshop and Conference Proceedings, JMLR.org, 2015, pp. 2067–2075.
    URL http://proceedings.mlr.press/v37/chung15.html
  • [39] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, BERT: Pre-training of deep bidirectional transformers for language understanding, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Association for Computational Linguistics, Minneapolis, Minnesota, 2019, pp. 4171–4186. doi:10.18653/v1/N19-1423.
    URL https://www.aclweb.org/anthology/N19-1423
  • [40] Z. Yang, Z. Dai, Y. Yang, J. G. Carbonell, R. Salakhutdinov, Q. V. Le, Xlnet: Generalized autoregressive pretraining for language understanding, in: H. M. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché-Buc, E. B. Fox, R. Garnett (Eds.), Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 2019, pp. 5754–5764.
    URL https://proceedings.neurips.cc/paper/2019/hash/dc6a7e655d7e5840e66733e9ee67cc69-Abstract.html
  • [41] T. B. Brown, B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell, S. Agarwal, A. Herbert-Voss, G. Krueger, T. Henighan, R. Child, A. Ramesh, D. M. Ziegler, J. Wu, C. Winter, C. Hesse, M. Chen, E. Sigler, M. Litwin, S. Gray, B. Chess, J. Clark, C. Berner, S. McCandlish, A. Radford, I. Sutskever, D. Amodei, Language models are few-shot learners, in: H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan, H. Lin (Eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020.
    URL https://proceedings.neurips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html
  • [42] Y. Tay, M. Dehghani, D. Bahri, D. Metzler, Efficient Transformers: A Survey, ArXiv200906732 Cs (2020). arXiv:2009.06732.
  • [43] F. Wang, M. Jiang, C. Qian, S. Yang, C. Li, H. Zhang, X. Wang, X. Tang, Residual attention network for image classification, in: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, IEEE Computer Society, 2017, pp. 6450–6458. doi:10.1109/CVPR.2017.683.
    URL https://doi.org/10.1109/CVPR.2017.683
  • [44] J. Hu, L. Shen, G. Sun, Squeeze-and-excitation networks, in: 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, IEEE Computer Society, 2018, pp. 7132–7141. doi:10.1109/CVPR.2018.00745.
    URL http://openaccess.thecvf.com/content_cvpr_2018/html/Hu_Squeeze-and-Excitation_Networks_CVPR_2018_paper.html
  • [45] H. Zhang, I. J. Goodfellow, D. N. Metaxas, A. Odena, Self-attention generative adversarial networks, in: K. Chaudhuri, R. Salakhutdinov (Eds.), Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, Vol. 97 of Proceedings of Machine Learning Research, PMLR, 2019, pp. 7354–7363.
    URL http://proceedings.mlr.press/v97/zhang19d.html
  • [46] M.-H. Guo, J.-X. Cai, Z.-N. Liu, T.-J. Mu, R. R. Martin, S.-M. Hu, PCT: Point Cloud Transformer, ArXiv201209688 Cs (2020). arXiv:2012.09688.
  • [47] D. Hendrycks, K. Gimpel, Gaussian Error Linear Units (GELUs) (2016). arXiv:1606.08415.
  • [48] D. P. Kingma, J. Ba, Adam: A method for stochastic optimization, in: Y. Bengio, Y. LeCun (Eds.), 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015.
    URL http://arxiv.org/abs/1412.6980
  • [49] J. Zhang, T. He, S. Sra, A. Jadbabaie, Why gradient clipping accelerates training: A theoretical justification for adaptivity, in: 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020, OpenReview.net, 2020.
    URL https://openreview.net/forum?id=BJgnXpVYwS
  • [50] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, R. Salakhutdinov, Dropout: A simple way to prevent neural networks from overfitting, Tech. rep. (2014). doi:10.5555/2627435.2670313.
  • [51] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, Z. Wojna, Rethinking the inception architecture for computer vision, in: 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, IEEE Computer Society, 2016, pp. 2818–2826. doi:10.1109/CVPR.2016.308.
    URL https://doi.org/10.1109/CVPR.2016.308
  • [52] D. Rumelhart, G. Hinton, R. Williams, Learning internal representations by error propagation, in: Readings in Cognitive Science, Elsevier, 1988, pp. 399–421. doi:10.1016/B978-1-4832-1446-7.50035-2.
  • [53] Y. Ding, M. Jia, Y. Cao, Remaining Useful Life Estimation Under Multiple Operating Conditions via Deep Subdomain Adaptation, IEEE Trans. Instrum. Meas. 70 (2021) 1–11. doi:10.1109/TIM.2021.3076567.
  • [54] R. G. Keys, Cubic Convolution Interpolation for Digital Image Processing, IEEE Trans. Acoust. Speech Signal Process. 29 (6) (1981) 1153–1160. doi:10.1109/TASSP.1981.1163711.
  • [55] L. Van der Maaten, G. Hinton, Visualizing data using t-SNE., J. Mach. Learn. Res. 9 (11) (2008).
  • [56] Y. Li, X. Wang, J. Wu, Fault diagnosis of rolling bearing based on permutation entropy and Extreme Learning Machine, in: 2016 Chinese Control and Decision Conference (CCDC), 2016, pp. 2966–2971. doi:10.1109/CCDC.2016.7531490.
  • [57] X. Zhang, Y. Liang, J. Zhou, Y. zang, A novel bearing fault diagnosis model integrated permutation entropy, ensemble empirical mode decomposition and optimized SVM, Measurement 69 (2015) 164–179. doi:10.1016/j.measurement.2015.03.017.
  • [58] R. Chen, X. Huang, L. Yang, X. Xu, X. Zhang, Y. Zhang, Intelligent fault diagnosis method of planetary gearboxes based on convolution neural network and discrete wavelet transform, Computers in Industry 106 (2019) 48–59. doi:10.1016/j.compind.2018.11.003.