In the modern day industry, machine systems are becoming more complex and fulfill critical tasks. To enhance their reliability, the condition of the system should be monitored. Any rotating machine, e.g. a pump, compressor, or steam turbine, will eventually reach a point of poor health. One effective strategy for enhancing their reliability and cost-effective maintenance is to utilize Condition Monitoring (CM) and Prognostics and Health Management (PHM). The aim is to identify unexpected anomalies, faults, and failures .
Prognostics systems built on a data-driven approach acquire data in-situ using a network of sensors that monitor the system 
. The dataset evolving from the measurements of the sensors usually has a high dimensionality and may also contain unwanted interference and noise. This dimensionality of the data has a direct impact on the training time as well as the accuracy of neural network-based models. A workaround for dealing with such a problem is to reduce the dimensionality of the input signal by extracting features that carry the health information of the system. Several methods have been proposed for achieving this, for instance, extracting temporal features by computing the root mean square (RMS), Skewness, Kurtosis, or Peak to Peak distance.
The main attributes of an ideal anomaly detection and condition monitoring system include the ability to collect useful features and the utilization of these features to identify the deteriorating condition of the machine by observing the deviation from the normal (healthy) behavior. Manual feature engineering methods require domain knowledge combined with “trial and error” strategies. However, the advent of deep learning and recent progress in autoencoder based models has provided an alternative way for feature extraction and dimensionality reduction. By stacking up layers to form deep autoencoders and by reducing the number of units in the hidden layers, it is expected that hidden units will extract features that will represent the data. The best features for the task at hand are learned directly from the data, thus avoiding ad-hoc trial and error strategies. Since the autoencoder creates a reduced representation of the data, it seems intuitive to use this representation for anomaly detection. The assumption is that the autoencoder only learns to map normal data points or inliers and does not include anomalies in the trained representation. Hence, trying to reconstruct anomalous data points will fail and incur huge losses. Finally, this reconstruction error can be used as a basis for calculating anomaly scores and labeling of unseen data points.
For rotating machine (RM) anomaly detection in general, it is much easier to collect large amounts of data than to accurately obtain their corresponding labels. Particularly in cases where faults or degradation evolve naturally over time. Correctly assigning labels is susceptible to the data ambiguity issue, especially at a pivotal stage when the machines exhibit early signs of failure but are far from obvious when equated with the fully developed faults. Moreover, anomalies are rare events, having prior knowledge about all possible anomalies is almost impossible, so the attention is shifted from anomalous to normal states for which data is available in large quantities. The motivation is that by modeling the normal behavior of the machine the system will also be able to detect anomalies that have not been observed previously.
We propose a framework for RM condition monitoring and anomaly detection based on Long short-term memory (LSTM) autoencoder networks. We demonstrate its applicability in both anomaly detection and condition monitoring by evaluating it on two real-world datasets. The advantages of the autoencoder based approach include the ability to work without any preprocessing, without any predetermined transformations such as FFT, without any manual feature engineering, without any feature selection, and the fact that it does not limit itself to the preidentified anomalies; it has the potential to detect new anomalies which have never been seen before. We also show that the time-domain features learned in the process can be used to enhance the performance of a simpler detection model such as Isolation Forest and thus can obviate the need for manual feature extraction. The flowchart of the proposed method is displayed in Fig.1. The measured vibration signals will be preprocessed and afterward, the training of the model is completed that is able to extract features and reconstruct the signal for the final anomaly detection.
The remainder of this paper is organized as follows. We will discuss related works in Section II. Section III discusses our proposed LSTM-based autoencoder method for robust feature extraction and anomaly detection. Sections IV and V discuss the experimental results, while the conclusions and future works are presented in Section VI.
Ii Related Work
A plethora of studies has been conducted in the field of RM fault/anomaly detection and time-series anomaly detection in general 
. Features and their generation are important concepts in data analysis and anomaly detection. The use and selection of features are crucial for measuring differences in data thus detecting anomalies. Traditionally, features extracted from time or frequency domain have been used for monitoring the condition or computing the remaining useful life (RUL) of the machine[5, 6]
. Many extracted features are influenced by operating conditions and are insensitive to anomalies. Researchers have tried to increase the performance of anomaly detection methods for RMs by performing multivariate analysis while using multiple features together, computed 21 bearing features using signal processing techniques. Others have tried to achieve this by constructing features for specific failure modes and performing multivariate analysis based on them as did the authors in  by constructing 5 features each of which was sensitive to a different failure mode. However, such manually extracted features are not generalizable and fail to provide useful machine health information in especially unknown cases.
More recently, 
tried to classify various bearing classes by using a set of time-frequency domain features and artificial neural networks.
replaced the manual feature extraction step by applying 1-D convolutional neural networks (CNN) to raw motor signals. The evaluation performed on bearing fault detection demonstrated the superiority of their approach compared to conventional feature extraction methods. trained an autoencoder for feature extraction and used these features to train a supervised fault detection classification model. 
studied several one-class classifiers such as nearest neighbors and k-means for detecting faulty rotor bars in an induction motor. They concluded that the k-nearest neighbor method stood out among all the tested methods.
Deep neural network-based architectures in particular autoencoders are successfully employed for supervised classification of faults into different fault categories by using time or frequency domain features extracted using prior knowledge . A probabilistic framework for anomaly detection in natural gas consumption time series is introduced in . However, the prediction method predicts the consumption levels using other independent variables and does not incorporate the temporal information that the data had to offer.  employed an vanilla autoencoder to detect anomalies in the electric power system by embedding the temporal information using sliding windows. Reconstruction errors obtained on sliding windows were used to compute anomaly scores. The Inclusion of temporal information using a sliding window works well in some cases but is not scalable usually.
The current RM anomaly detection methods, as discussed above, face one or several of the following limitations: labor-intensive manual feature extraction, the requirement of accurately labeled datasets, or failure to incorporate temporal information. Time series data such as sensor data is best modeled as a sequence where the data point at each timestep is dependent on the previous data points. LSTM-based autoencoders are capable of dealing with the time-series sequences and can take variable length input. To overcome issues described above, in this study, we train an LSTM-based autoencoder over the vibration signals to autonomously monitor the condition of RMs and extract time-domain features to provide an alternative for manual feature extraction. To the best of our knowledge, this is the first time that an unsupervised method based on LSTM-autoencoders is used for identifying fault/anomalies in accelerometer vibration signals.
Iii Rotating Machine Anomaly Detection
We address the problem of anomaly detection in abnormal vibratory phenomena captured through the accelerometer sensors mounted on RMs which should indicate a deterioration of the system. As opposed to traditional health prognostic systems that usually encapsulate feature extraction and anomaly detection as distinct blocks, the proposed system takes directly raw time-series vibration signals as input and it can efficiently learn optimal features and based on these features determine the system’s health.
In the case of multi-dimensional input, the set of M sensors
(also called generators) are used to capture the behavior of a RM. This measured data is fed as input to an LSTM-based autoencoder (LSTM-AE) that is trained over the vibration signals via batch gradient descent to minimize a reconstruction error term between an original signal and its reconstruction. Specifically, the encoder maps an input vectorx
to a lower-dimensional hidden representationh
by an affine mapping following a nonlinearity and the decoder correspondingly generates an estimationx’ of the input vector x.
AEs belong to the unsupervised representation learning class which try to model the data distribution through the discovery of a set of latent representations, also called embeddings, whose variations capture most of the structure of the original data distribution . These hidden-layer units or low-dimensional embeddings force the AE model to learn the key representations from the original vibration signal. The encoder generates a rich non-linear set of features from the sensory data and the decoder learns to reconstruct the original signal using these features. The motivation to use autoencoders is their ability to detect anomalies based on the fact that anomalies are quite rare and deviate greatly from the general pattern in normal healthy data. The model is trained with the aim to learn the normal behavior of a RM, thus not recognizing anomalies.
The encoder and decoder are two main components of the network and both are based on Long short-term memory network (LSTM) units. LSTMs are a type of recurrent neural network (RNN) that can integrate the temporal information into the network and maintain a hidden state vector which acts as a memory for the past information.
We observe the input data sequence denoted by , where , meaning that for each index N there is a d-dimensional time-series sequence with T timesteps each. We use the RNN to process the variable input sequence and to extract the sequential information from the time-series data. For each sequence this is given by;
where , and are memory state, input and output vectors from last step and , and
are input weights, recurrent weights and the bias. Tanh is used as a non-linear activation function, whose output range lies in the interval.
The input sequences are passed through the encoder part of the LSTM network which encodes an input sequence or batch of sequences using LSTM units and updates its hidden state. The output of the encoder is given by;
where is the output of the LSTM-encoder, represents the parameter set of the encoder and to avoid the vanishing and exploding gradients issue,
for both encoder and decoder is chosen as ReLU activation function.
The representations obtained from mapping the vibration signal to lower-dimensional embeddings through the encoder are used by the decoder to reproduce the original signal. The Output from the last encoder of the network becomes the input to the LSTM-decoder network:
where the set of parameters of the decoder is represented by and is the reconstructed input which is used to compute the reconstruction error (RE) = , also known as the Mean Squared Error. The error is required to update the network’s encoder and decoder parameters and later, compute the anomaly scores.
The AE model should be sensitive enough to reproduce the original signal but insensitive enough to the training data and noise, such that the model learns a generalizable representation of the data. During training, the focus of the model lies in learning the normal behavior of the machine, hence anomalies are not included in the training data. Consequently, during prediction, a slight deviation from the normal behavior would increase the reconstruction error (RE). So, monitoring the increase in reconstruction error gives the possibility to detect anomalies but also anticipate the fault in advance by identifying the degradation point, the timestamp after which the RE starts to increase.
Iii-a Model Training
The raw vibration data is normalized first to have zero mean and unit variance. We divide the RM’s vibration data into four sets: a training set (), two validation sets ( & ), and a test set (). The distribution of data among these four sets is as follows: 70%, & 5% (each) and 20%. consists only of normal sequences and is used for training the LSTM-AE model. NASA bearing dataset (dataset-1, details in section IV
) does not provide explicit labels for each sequence, so we assume initial 70% of the data to be normal as machines have a low probability of being in a faulty state, meaning that we assume anomalies to occur rarely compared to the normal data. According to the ground truth in the provided datasets, failures occurred only at the end of each run-to-failure experiment and thus the first 70% is used for the training of the model. The architecture of the LSTM-AE is designed as such to allow the model to have enough capability for feature learning and secondly, the number of units of the next hidden layer is set smaller than that of the previous layer so that feature learning can be viewed as a signal compression process, as shown in Fig.2 (a). Here, the training process with the train set
is depicted, together with the compression process of the raw signal within the autoencoder. The weights of the LSTM-AE are updated via stochastic gradient descent using mini-batches and the Adam
optimizer is utilized to speed up the training process. Batch normalization allows faster and stable training of deep neural networks. To restrict the model from overfitting, dropout is used and to avoid exploding gradients, gradient clipping is applied. Model hyperparameters are learned using the Bayesian optimization method, including the size of mini-batches, learning rate, and weight decay.
Iii-B Anomaly Monitoring Process
The detection process of anomalies consists of converting reconstruction errors into anomaly scores(AS) for each input sequence and using these scores to obtain a threshold, characterizing the normal behavior of the machine.
After the model has been trained, sequences in are passed through the model to get the reconstruction errors which in turn are used to estimate the parameters of a Normal distribution ( ) using maximum likelihood estimation (MLE), similar to . The probability of obtaining the reconstruction error
is given by the value of the Normal distribution at. Using the & , the anomaly score for a datapoint is computed as follows:
where is the desired anomaly score, is RE obtained for a sequence and
are the mean and variance of a multivariate Gaussian distribution.
During the initialization phase, an anomaly score threshold is also learned using a validation set that may contain examples from anomalous sequences alongside with the normal sequences. Unseen sequences within are classified as follows: if a sequence has an anomaly score it will be labeled as an anomaly, otherwise as normal. This is also depicted in Fig. 2 (b) where the raw measured signal is input to the trained autoencoder model that generates reconstruction errors and anomaly scores. These are then used within the described decision process to determine whether the signal was normal or anomalous.
The assumption here is, that as the monitored equipment degrades or faces a failure, this disrupts the normal working of the machine and affects the interaction between different variables which can be measured by accelerometers, especially in form of vibrations. As the sensor values start deviating from the normal working condition of the machine, it is expected to see an increased error in the reconstruction of the input. By monitoring the reconstruction error and the anomaly score, an indication of the health of the monitored machine can be derived.
Iii-C Feature Extraction
Feature learning is a critical step in improving the performance of anomaly detection models due to the multidimensionality of data that is input into the model. In general, machine vibration signals comprise a stationary vibration part, a random vibration part, and noise . We study two methods for the inspection of vibration signals: automated feature extraction based on LSTM-AE and manual feature extraction based on classic signal processing methods. Isolation Forest is used to evaluate the effectiveness of these two feature extraction methods.
Isolation Forest is a random forest-based anomaly detection algorithm that utilizes isolation to determine anomalies in data. The Isolation Forest (IF) algorithm is based on Decision Forests, an ensemble method that uses the averages of outputs from many different trees.
Iii-C1 LSTM-AE based Feature Extraction
One of the main characteristics of an autoencoder that is more powerful for finding intrinsic data structures by reducing data dimensionality through non-linear transformations than Principal Component Analysis (PCA).
The encoder from the trained LSTM-AE is used to perform feature extraction. The process is done by reducing the number of units in the hidden layer, it is expected that the hidden units in the encoder network will extract features that will represent the data. To learn more abstract features, multiple AEs are stacked together to form a stacked AE, in which the output of each hidden layer is connected to the input of successive hidden layers. A stacked AE applies dimensionality reduction in a hierarchical manner, obtaining more abstract features in higher hidden layers which lead to a better reconstruction of the data .
Iii-C2 Manual Feature Extraction
Following time-domain statistical features are generally used to detect incipient machine faults/anomalies: mean (
), root mean square (RMS), percentiles (25th, 50th & 75th ), max absolute value, standard deviation (),peek-to-peek, skewness, kurtosis, entropy, and AR-coefficients. The last five features are described through the following equations:
There are two datasets considered for the evaluation of the proposed method. We evaluate and test the applicability of the proposed method in Condition Monitoring using the IMS bearing dataset (Dataset-1). In addition, the effectiveness to find anomalies in unsupervised settings is evaluated using a private industry dataset from vibration data of RMs. Evaluations performed on these two datasets demonstrate the ability of the method to detect anomalies within the industry dataset and perform condition monitoring on the bearing dataset, as the IMS bearing dataset doesn’t provide explicit labels but instead contains a degrading health state scenario of the bearing under experiment. Below, we introduce the datasets and illustrate the performance of the proposed method on these datasets.
This dataset was gathered from a run-to-failure experimental setting, involving four bearings and is subdivided into three datasets, each of which consists of the vibration signals from these four bearings . These sets were collected from three test-to-failure experiments which were performed independently, and failures occurred at the end of the test. Table I represents the properties of the collected data from these experiments.
|Set #||Batches||Batch Size||Anomaly|
|Set1||2156||4 x 20480||B3 and B4|
|Set2||984||4 x 20480||B1|
|Set3||6324||4 x 20480||B3|
This dataset consists of data from five different RMs (RM-1 to RM-5) which are of the same kind but build-wise unique. Data from each RM contains 3-dimensional vibration signals measured over time, captured using the accelerometer sensors attached to the housing of the RM. The signal recordings were taken in batches and the batch sizes vary, depending upon the RM. Considering the condition of the machine at the time of recording, labels are assigned to each batch, 0 representing normal and 1 anomalous, respectively. Table II presents the summarized statistics of the industry dataset. The difference between normal and anomalous signals is shown in Fig. 3, vibrations in all three directions (x, y, z) differ significantly for anomalous signals compared to normal vibration signals. These differences can have a variety of characteristics as the amplitude of different frequencies differs between the two signals.
V Experiments and Results
We demonstrate our method’s applicability in condition monitoring & anomaly detection with three general settings where in the first one, the autoencoder is trained with 70% of the data, 10% (split between two validation sets) is used to calculate the threshold, and remaining 20% of it is kept for evaluation to determine the prediction ability of the proposed approach for an anomaly detection task. In the second experimental setting, each run-to-failure experiment is simulated by training the LSTM-AE model with the available data from the non-failing RMs to monitor the condition of the machine in which failure occurred during the run-to-failure experiment. The trained model is used to monitor the status of the faulty machines to demonstrate that once the model is trained, it is capable of being applied to different RMs of a similar kind, enabling the transfer of rare anomaly knowledge from RMs to other RMs of a similar kind.
The third setting consists of a method for performing automated feature extraction from vibration signals. Features extracted using this setting are compared against a set of handcrafted features by training Isolation Forest on these two features sets separately.
V-a Setting 1: Online Prognostic
In this setting, as discussed, the autoencoder learns to reconstruct the normal behavior of an RM using only 70% of the available healthy/normal data in order to evaluate the prediction performance of the model in an online monitoring phase. The threshold is obtained using a validation set of size 5% to classify the samples as normal/healthy or anomalous. Anomaly scores for each data sample are calculated using the reconstruction errors and data points with anomaly scores larger than are labeled as anomalies. The point at which the anomaly score crosses the threshold and starts to increase gradually is considered as the degradation point.
Here we study 4 cases of failing bearings (B1 to B4) of the different sub-datasets (S1 to S3), S1-B3, S1-B4, S2-B1, and S3-B3, for early fault detection. The model trained on individual bearings is able to predict the degradation point and capture the propagation of fault in the simulated run-to-failure experiment as is depicted in Fig. 4. As the fault in every failing bearing occurs at the end, Fig. 4 displays the anomaly scores of the last 620, 450, and 780 batch samples for S1-B3 & S1-B4, S2-B1, and S3-B3, respectively. For every failing bearing in each dataset, the proposed method generates a trend corresponding to the health status of the bearing based on the vibration sensory data. As the bearing health starts to degrade, the anomaly score tends to increase and once it passes the degradation point (marked with the filled circles) it increases gradually. By following this trend and abrupt changes in the scores one can feasibly detect the failure. The degradation point for each bearing was calculated by identifying the abrupt change and continuous increasing trend in anomaly scores. Sample no. 2039, 1704, 667, and 5241 were identified as degradation points for S1-B3, S1-B4, S2-B1, and S3-B3, respectively. The circle on the score line indicates the degradation starting point.
The second dataset consists of 5 different RMs with anomalies, for each machine there is a different number of anomalies (see Table II
). To detect anomalies from each individual machine, we conduct five different experiments (one for each machine) using only the non-anomalous (normal) data to train the autoencoder network. The anomaly detection results are evaluated using, Precision, Recall, True positive rate (TPR), False positive rate (FPR), and F1-score. These evaluation metrics are weighted per class.
As described, model performance is evaluated over the test set consisting of 20% of the data. For instance, the test set for RM-1 consisted of 145 (92 normal, 53 anomalous) sequences in total. Each sample from the test set was reconstructed using the trained autoencoder network and was labeled as anomalous if its anomaly score was larger than the threshold, normal otherwise. The trained model was able to detect 52/53 anomalies without raising many false alarms (just one false positive). Fig. 5 displays the anomaly scores for each data point; sequences below the threshold (marked by the red line) are classified as normal and data points with anomaly scores larger than the threshold are labeled as anomalies. Evaluation scores (precision, recall, TPR, FPR, F1-score) for individual RMs from the second dataset, averaged over 10 repetitions of the experiments are presented in Table III in which the model performs very well with F1-scores close to 1 and is able to detect anomalies without generating many false positives which is indicated by the high TPR and low FPR. Based on the results, the LSTM-AE can effectively extract discriminative features directly from the raw vibration data and achieve a competitive anomaly detection rate. For the test data, an overall detection F1-score of 99% and TPR over 93% for each machine is obtained.
V-B Setting 2: Condition Monitoring
In dataset-1, during each run-to-failure experiment, only one out of 4 bearings faced failure while the other 3 remained in healthy condition except for experiment-1 in which a fault occurred in two bearings at the end. The question of interest is whether the knowledge extracted from the healthy bearings can be applied to detect the deteriorating condition of a faulty bearing and raise alarm well before the total failure actually happens. To validate this, in this setting only the data from the healthy bearings is used to train the model and this trained model is then used to evaluate the condition of the previously unseen faulty bearings.
Three simulations, one per sub-dataset are performed as follows: The Set1 model is trained using only training data from the non-failing bearings S1-B1 & S1-B2 with the aim to learn the functioning of healthy bearing and apply this learned knowledge to monitor and detect the deteriorating condition of the failing bearings S1-B3 & S1-B4. Similarly, for Set2, the training data comprised of S2-B2, S2-B3, and S2-B4 and while the test set consisted of S2-B1 and for Set3, training data consists of S3-B1, S3-B2, S3-B4 and test set of S3-B3. Fig. 6 displays the output anomaly scores of the four failing bearings from these experiments. The trend it generates for each bearing is in accordance with the ground truth, from the beginning until near the end the bearings remain in healthy condition, thus low anomaly scores are correctly calculated. The fault starts to appear only at the very end, which is captured by the model by a continuous increase in the anomaly scores.
A low anomaly score corresponds to healthy behavior while an upward trend (high anomaly score) highlights the abnormal behavior of the machine. Arrival and propagation of fault of for every failing bearing is shown in Fig. 6. It is clearly visible that the proposed approach is able to identify the initiation of the faulty trend as well as the increasing effect of the deterioration for all four bearings. Fig. 7 visualizes the anomaly scores and visually indicates the health status of the four faulty bearings S1-B3, S1-B4, and S2-B1 and S3-B3 in parts B3, B4, B1, and B3, respectively. The color bar represents the anomaly score, from 0 (blue) to 1 (red).
The same setting was applied to our second dataset where the results were obtained by combining the normal data points from four RMs (RM-2 to RM-5). We divided each RM’s data into three sets: training (70%), validation (10%) and test (20%). Train set T was created by combining the four and was used to train the model. The validation sets were used to calculate the threshold using the anomaly scores for each RM . The final anomaly detection model trained on T was then evaluated on each set and the obtained results are presented in Table IV, in which the model performs very well in detecting anomalies across four different machines without raising any false alarms. It is worthwhile to notice that combining the training data from multiple RMs increases the performance of the model compared to the performance of the model on individual RMs. The trained LSTM-AE is able to model the healthy working behavior of the RMs and is sensitive to anomalies, as it can be seen in Fig. 8 which shows that the identified anomalies match well with the ground truth. The model can reconstruct the normal signals very well, indicated by a low anomaly score, while it fails to do so whenever it encounters anomalies.
V-C Setting 3: Feature Extraction
We employ the same LSTM-based autoencoder model to automatically extract time-domain features from vibration signals. The automatically extracted features are compared with a set of manually crafted features, by training an Isolation Forest on both of these two sets and monitoring the performance of the model.
Results obtained from the IMS bearing dataset using the proposed method are compared with the IF-based anomaly detection trained only on manually extracted features. The comparison of results between two Isolation Forest based models is shown in Table V, which shows the total number of anomalies detected by the models in each bearing. It’s evident that the S1-B1, S1-B4, S2-B1, and S3-B3 are having comparatively more anomalous samples than the rest of the bearings as these four bearings result in failure according to the ground truth provided. The comparison illustrates that the latent features extracted by the LSTM-AE perform better than the manually engineered features demonstrating the effectiveness of the proposed approach in providing a robust set of features that boost the performance of the anomaly detection model. This can be confirmed by looking at the number of anomalies detected by each model; for each failing bearing, the IF model based on automatically extracted features detects a greater number of anomalies than the model using the handcrafted set of features. One interpretation of this could be that LSTM-AE provides a set of features that are more sensitive and thus can detect the subtle changes in the vibrations when the fault appears initially and starts developing gradually until the failure finally happens and bearing stops functioning. While on the other hand manually handcrafted features only start labeling signals as anomalous when the fault has fully developed already, and vibrations deviate significantly from the normal signals.
|Bearing||No. of Anomalies (A_FE)||No. of Anomalies (M_FE)|
Multi-step anomaly detection results, where the first step consists of automated features extraction using LSTM-AE and in the second step these features are utilized to train an anomaly detection model (Isolation Forest), for the second dataset are summarized in Table VI. It is important to mention that each RM except for RM-1 in dataset-2 contains very few anomalies (see Table II). IF based on features provided by LSTM-AE (AFEIF) outperforms IF trained using the manually engineered features (MFEIF) for all RMs in terms of precision. Although, for RM-3 and RM-5 MFEIF performed better in terms of TPR. However, it is worthwhile to notice that the model based on automatically engineered features results in less false positives in comparison to the model trained on handcrafted features which had significantly more false positives. As previously stated, the studies that have used handcrafted features obviously are not able to carry the complete system health representation and may not represent the characteristics of the underlying vibration signal under all circumstances. As seen from the results, fault/anomaly detection performance of the conventional methods such as Isolation Forest depends highly on the carefully crafted features. Consequently, this limits the general applicability of these methods.
The experimental results obtained in all three settings are able to support the earlier claims that LSTM-AE trained only on the normal/healthy signals is not just able to monitor the health condition of an individual RM but knowledge learned can also be effectively used to detect anomalies in similar yet different RMs. Additionally, it is also shown that latent representations obtained from LSTM-AE can be used as an alternative to manual feature extraction, which requires prior domain and signal processing knowledge.
Vi Conclusion and Future Work
In conclusion, we have demonstrated an unsupervised method for automated feature extraction from raw vibration signals that can be used for detecting faults and anomalies in rotating machines. A typical condition monitoring system requires feature extraction and decision about the health of the system. The feature extraction component of such systems involves the implementation of signal processing methods for preprocessing the data before using it for anomaly detection. The proposed method fuses the features extraction and anomaly detection modules within one condition monitoring system. Encoding the time-dependent as well as inherent features of the measured vibration time series into a hidden state of an LSTM-based autoencoder enabled the usage of the resulting reconstruction error to be used in an anomaly detection setup. The experimental results on two real-world datasets illustrate the effectiveness of the proposed model and show that LSTM-based autoencoders can extract salient features from vibration signals and achieve high accuracy in fault diagnosis. The model was able to detect the deteriorating condition of all four failing bearings for IMS dataset and achieved an overall F1 score of 99.6% for the second dataset. As demonstrated in the experiments, the latent representations obtained from the LSMT-AE carry system health information and achieve a higher F1 score than the manually extracted features (97% and 93% respectively), indicating a better representation of the characteristics of the vibration signal for the proposed method.
For future work, other alternatives for choosing a threshold value shall be investigated. Also more advanced types of autoencoders such as variational autoencoder can be tried out, which alternatively try to model the data distribution instead of learning to re-create the normal data points, which means datasets with mixed examples (anomalous non-anomalous) could also be used for training the model. This paper will help design further advanced feature extraction methods from rotating machines and help build models to monitor the condition of such machines with no or very minimal human supervision.
-  Marcos E Orchard and George J Vachtsevanos. A particle-filtering approach for on-line fault diagnosis and failure prognosis. Transactions of the Institute of Measurement and Control, 31(3-4):221–246, 2009.
-  Ruey S Tsay, Daniel Pena, and Alan E Pankratz. Outliers in multivariate time series. Biometrika, 87(4):789–804, 2000.
-  Patrick Nectoux, Rafael Gouriveau, Kamal Medjaher, Emmanuel Ramasso, Brigitte Chebel-Morello, Noureddine Zerhouni, and Christophe Varnier. Pronostia: An experimental platform for bearings accelerated degradation tests. 2012.
-  Charu C Aggarwal and Saket Sathe. Outlier ensembles: An introduction. Springer, 2017.
-  Hai Qiu, Jay Lee, Jing Lin, and Gang Yu. Wavelet filter-based weak signature detection method and its application on rolling element bearing prognostics. Journal of sound and vibration, 289(4-5):1066–1090, 2006.
Diego Alejandro Tobon-Mejia, Kamal Medjaher, Noureddine Zerhouni, and Gerard
A data-driven failure prognostics method based on mixture of gaussians hidden markov models.IEEE Transactions on reliability, 61(2):491–503, 2012.
-  Zhanguo Xia, Shixiong Xia, Ling Wan, and Shiyu Cai. Spectral regression based fault feature extraction for bearing accelerometer sensor signals. Sensors, 12(10):13694–13719, 2012.
-  Jing Tian, Carlos Morillo, Michael H Azarian, and Michael Pecht. Motor bearing fault detection using spectral kurtosis-based feature extraction coupled with k-nearest neighbor distance analysis. IEEE Transactions on Industrial Electronics, 63(3):1793–1803, 2015.
-  Jaouher Ben Ali, Nader Fnaiech, Lotfi Saidi, Brigitte Chebel-Morello, and Farhat Fnaiech. Application of empirical mode decomposition and artificial neural network for automatic bearing fault diagnosis based on vibration signals. Applied Acoustics, 89:16–27, 2015.
-  Turker Ince, Serkan Kiranyaz, Levent Eren, Murat Askar, and Moncef Gabbouj. Real-time motor fault detection by 1-d convolutional neural networks. IEEE Transactions on Industrial Electronics, 63(11):7067–7075, 2016.
-  Jiedi Sun, Changhong Yan, and Jiangtao Wen. Intelligent bearing fault diagnosis method combining compressed data acquisition and deep learning. IEEE Transactions on Instrumentation and Measurement, 67(1):185–195, 2017.
-  Roozbeh Razavi-Far, Maryam Farajzadeh-Zanjani, Shokoofeh Zare, Mehrdad Saif, and Jafar Zarei. One-class classifiers for detecting faults in induction motors. In 2017 IEEE 30th Canadian Conference on Electrical and Computer Engineering (CCECE), pages 1–5. IEEE, 2017.
-  Feng Jia, Yaguo Lei, Jing Lin, Xin Zhou, and Na Lu. Deep neural networks: A promising tool for fault characteristic mining and intelligent diagnosis of rotating machinery with massive data. Mechanical Systems and Signal Processing, 72:303–315, 2016.
-  Hermine N Akouemo and Richard J Povinelli. Probabilistic anomaly detection in natural gas time series data. International Journal of Forecasting, 32(3):948–956, 2016.
-  Marco Martinelli, Enrico Tronci, Giovanni Dipoppa, and Claudio Balducelli. Electric power system anomaly detection using neural networks. In International Conference on Knowledge-Based and Intelligent Information and Engineering Systems, pages 1242–1248. Springer, 2004.
Guillaume Alain and Yoshua Bengio.
What regularized auto-encoders learn from the data-generating
The Journal of Machine Learning Research, 15(1):3563–3593, 2014.
-  Klaus Greff, Rupesh K Srivastava, Jan Koutník, Bas R Steunebrink, and Jürgen Schmidhuber. Lstm: A search space odyssey. IEEE transactions on neural networks and learning systems, 28(10):2222–2232, 2016.
-  Pankaj Malhotra, Anusha Ramakrishnan, Gaurangi Anand, Lovekesh Vig, Puneet Agarwal, and Gautam Shroff. Lstm-based encoder-decoder for multi-sensor anomaly detection. arXiv preprint arXiv:1607.00148, 2016.
-  GK Singh and Saleh Al Kazzaz Sa’ad Ahmed. Vibration signal analysis using wavelet transform for isolation and identification of electrical faults in induction machine. Electric Power Systems Research, 68(2):119–136, 2004.
-  Fei Tony Liu, Kai Ming Ting, and Zhi-Hua Zhou. Isolation forest. In 2008 Eighth IEEE International Conference on Data Mining, pages 413–422. IEEE, 2008.
-  Tyler Manning-Dahan. Pca and autoencoders, 2018.
Pascal Vincent, Hugo Larochelle, Isabelle Lajoie, Yoshua Bengio, Pierre-Antoine
Manzagol, and Léon Bottou.
Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion.Journal of machine learning research, 11(12), 2010.
-  J Lee, H Qiu, G Yu, and J Lin. Rexnord technical services (2007).’bearing data set’, ims, university of cincinnati. nasa ames prognostics data repository, 2009.
-  John Duchi, Elad Hazan, and Yoram Singer. Adaptive subgradient methods for online learning and stochastic optimization. Journal of machine learning research, 12(7), 2011.