Layerwise Perturbation-Based Adversarial Training for Hard Drive Health Degree Prediction
With the development of cloud computing and big data, the reliability of data storage systems becomes increasingly important. Previous researchers have shown that machine learning algorithms based on SMART attributes are effective methods to predict hard drive failures. In this paper, we use SMART attributes to predict hard drive health degrees which are helpful for taking different fault tolerant actions in advance. Given the highly imbalanced SMART datasets, it is a nontrivial work to predict the health degree precisely. The proposed model would encounter overfitting and biased fitting problems if it is trained by the traditional methods. In order to resolve this problem, we propose two strategies to better utilize imbalanced data and improve performance. Firstly, we design a layerwise perturbation-based adversarial training method which can add perturbations to any layers of a neural network to improve the generalization of the network. Secondly, we extend the training method to the semi-supervised settings. Then, it is possible to utilize unlabeled data that have a potential of failure to further improve the performance of the model. Our extensive experiments on two real-world hard drive datasets demonstrate the superiority of the proposed schemes for both supervised and semi-supervised classification. The model trained by the proposed method can correctly predict the hard drive health status 5 and 15 days in advance. Finally, we verify the generality of the proposed training method in other similar anomaly detection tasks where the dataset is imbalanced. The results argue that the proposed methods are applicable to other domains.
Nowadays, increasing numbers of industrial and academic institutes rely on data centers to store and process their data. The crash of data centers may incur tremendous loss or even catastrophic consequences. The reliability and the availability of data centers are of the utmost importance to data center administrators. However, the complex architecture and functionality of data centers lead to a serious problem of IT equipment failures, among which hard drives are the most frequently failing components [1, 2]. Hence, it is in high demand to take measures to handle the hard drive failure issue.
The self-monitoring, analysis and reporting technology (SMART)  has been implemented in almost all hard drives to monitor and analyze the internal attributes of hard drives. The previous researches demonstrate that the impending hard drives failure manifests itself through SMART statistics . It is feasible to predict the impending failures by using SMART statistics. To improve failure prediction performance, many efforts have been made based on SMART attributes, including analyzing the failure behaviors of hard drives [5, 2], and designing machine learning algorithms for predicting hard drive failures [6, 7, 8]. Most of these works focused on the proactive failure prediction, which forecasts hard drive failures in advance and gives a binary result identifying the hard drive as healthy or faulted. However, hard drives usually fail gradually rather than abruptly. To harness the potential of gradual change, it is necessary to develop a health prediction model which can predict the heath status of hard drives rather than merely providing a simple binary result.
Predicting the health status of hard drives is not a trivial task. Most hard drives undergo a deterioration process before they finally fail. The SMART statistics begin to accumulate deviation from the normal state days before the final failure. Hence, we need to extract the long-term temporal dependency in the SMART statistics to make an accurate prediction of the hard drive status. The most stubborn problem in the prediction of health status is that the data to be used in the model is highly imbalanced. Although hard drive failures occur frequently in data centers, the failure records are much fewer than the healthy records, merely accounting for less than 3% of the total records . Due to the overfitting and biased fitting problems of most statistical and machine learning algorithms, the prediction model trained by the imbalanced SAMRT data is easy to be biased fitted to the healthy records and over fitted to the failure records, which results in a poor predictive performance. Some novel methods are desirable to tackle imbalance issue and establish a high quality prediction model by using the SMART statistics.
In order to extract the long-term temporal dependency embedded in the SMART statistics, we build a deep neural network based on the long short-term memory unit (LSTM)  which specializes in processing sequential data . Nonetheless, the deep neural networks (DNNs) are notorious for the overfitting and biased fitting problems on the highly imbalanced datasets like the hard drive records used in this work. To solve this problem, we propose two strategies in this paper.
Firstly, we introduce the adversarial training strategy. The adversarial training [11, 12] is a strong regularization method that injects the perturbations affecting the neural network’s inference in the most sensitive way during the training phase. The traditional adversarial training injects adversarial perturbations into the inputs to force the networks to learn a better distribution of the training data and avoid the overfitting problem [13, 14]. However, only injecting perturbations into the inputs may restrict the effectiveness of the adversarial training. In this paper, we enhance the adversarial training by enabling layerwise perturbation where the adversarial perturbation can be injected into any layer of the neural network rather than just the input layer.
Another more effective approach to mitigate the overfitting and biased fitting is to increase the number of positive samples, i.e., the failure records. It is obvious that the anomaly of SMART statistics is evident when hard drives are close to failure. Therefore, we can label the records close to failure as the failure records directly. But for the records dozens of days ahead of the final failure, they may be less informative for the failure or even demonstrate the same features as the healthy records do. It is inappropriate to label these records as failure records arbitrarily. Labeling these records is a time consuming task and relies on the expert knowledge, which makes it unfeasible in reality. Hence, in traditional supervised training, these records are discarded in spite of the high probability that they show fault features. To fully harness these potential failure records, we extend the proposed adversarial training into the semi-supervised setting where the records far before the final failure are regarded as unlabeled data during the training phase.
Based on the above two strategies, we propose a Layerwise Perturbation-based Adversarial Training (LPAT) method to train our hard drive status prediction model. Note that although the LPAT is designed for predicting hard drive health status in this work, it can be used in other similar anomaly detection problems to mitigate the overfitting and biased fitting. The main contributions are summarized as follows:
We design a novel LPAT method for hard drive status prediction tasks. Instead of only adding perturbation to inputs of neural networks, LPAT can flexibly add perturbation to a specific layer or all layers to better address the overfitting and biased fitting problem. We further introduce an approach based on Kullback-Leibler divergence to utilize unlabeled data to improve the performance of the prediction model.
To the best of our knowledge, it is the first work toward designing a layerwise perturbation-based adversarial training with the semi-supervised setting for deep learning models. We use this model to predict hard drive status in both supervised and semi-supervised settings.
Thorough experiments on two hard drive datasets demonstrate that LPAT can improve the prediction performance in both supervised and semi-supervised settings. Specifically, the model can predict hard drive health status 5 days and 15 days before failure by using sequential SMART statistics, which is more useful in practice than only predicting healthy or failed. In addition, we apply the proposed methods to a image recognition task and a sequential analysis task to verify their generality.
The rest of the paper is organized as follows. We define the prediction problem in Section II. Then, Section III presents the proposed methodology. In Section IV, we evaluate our methods thoroughly by conducting groups of experiments. Section V summarizes related work. Finally, Section VI concludes this paper and describes the future work.
Ii Problem Definition
In this work, we aim to build a model trained by LPAT to predict the status of hard drives based on their SMART statistics. The training dataset is denoted as , where is the training sample consisting of continuous SMART records from the day to the day . Each day has a feature vector of SMART attributes, and denotes health degrees of the hard drive. We predict hard drives by three different health degrees
Iii-a LSTM Based Neural Network
The proposed prediction model includes two dense (fully connected) layers, followed by a LSTM layer and a dense layer. LSTM is the basic building block used to extract the long-term temporal dependency in sequential SMART statistics. Fig. 1 is an illustration of the component in the LSTM layer. It is formulated as:
where . are the , and , respectively. represents the memory cell which is designed to counteract the problem of vanishing/exploding gradient, and thus enable extracting long-term temporal dependency. The is for resetting the memory cells. The and the are for controlling the input and the output of the memory cells. when , where is the sigmoid activation function. is a proposed update to the cell state. is the input from the lower dense layer at the time step . is the output of current block at the time step . In the LSTM units, the weight matrices , and the bias vectors are the parameters to learn for .
Iii-B Layerwise Perturbation-Based Adversarial Training
Previous works on adversarial training concluded that training a DNN with adversarial examples acts as a regularizer and improves the robustness of the neural network on the test dataset. In order to mitigate the overfitting and biased fitting problem caused by the highly imbalanced data, we design LPAT to train the prediction model. Instead of only injecting perturbation into inputs, LPAT generates adversarial samples at the time series inputs and the intermediate layers. Fig. 2 illustrates how LPAT works when training the model.
In Fig. 2, for a model consisting of hidden layers, is the input layer, and is the output of the layer, where is the dimension of the layer. Each layer has a gradient accumulation layer serving two functions: (1) it temporarily stores the backpropagation gradients on the output of the layer, which is denoted by the yellow line; and (2) it computes the adversarial perturbations for the layer and adds the perturbation to as denoted by the blue line. Then, the neural network performs feedforward process again to compute its new output. The training process can be formed as a min-max problem. The adversarial samples apply the worst perturbation to maximize the error of the model, while the model tries to be robust to such perturbations through minimizing the error caused by the adversary. The min-max problem for the layer can be formulated as an additional cost function:
where is the magnitude of the perturbation, is the non-negative cross entropy between two distributions representing the prediction error caused by the perturbation, and is a one-hot distribution of the corresponding label . The additional cost function can be simplified as:
where is a constant set to the current parameters of the model. In general, the exact calculation for is intractable for most DNNs. We follow the method proposed in  to linearly approximate with the norm constraints as:
Iii-C Extension of Semi-Supervised Setting
We extend the proposed adversarial training into the semi-supervised setting to utilize the records relatively far before the final failure. Due to the difficulty of labeling these records, we regard them as unlabeled samples during the training. Then, the adversarial perturbation is incalculable by using Eq. (8) without the label . To depict the prediction error, we introduce the Kullback-Leibler (KL) divergence  which measures the divergence between the current output distribution and the perturbed output distribution . The layerwiase adversarial perturbation for the layer can be calculated as:
is differential with and . However, the maximum value of cannot be computed directly as the first derivative has the minimum value when . Hence, we approximate it with the second-order Taylor series:
where is the Hessian matrix . Then, is the first dominant eigenvector of with magnitude , and can be represented as:
where represents the unit vector in the direction of . Given a randomly sampled unit vector , according to the power iteration method , is the convergence of the iteration , and can be approximated as follow:
where . The preliminary test shows that one-time iteration provides sufficient performance as multi-time iterations in our problem. So we approximate the convergence by the one-time power iteration . Then, the approximation of is:
Hitherto, we get the adversarial perturbation to the DNN model in both the supervised setting and the semi-supervised setting. During the training phase, the DNN model is trained not only to minimize the regular classification loss on labeled data but also to resist the adversarial perturbation on labeled and unlabeled data. Thus, for all the training dataset including labeled and unlabeled data of size , the full loss is given by:
where is the negative log-likelihood for the labeled data defined by Eq. . is computed on both labeled and unlabeled data:
The gradients computed from mini-batch inputs aggregated through intermediate layers are used to calculate perturbations for the current mini-batch. Since Eq. requires no label information, it is not necessary to shuffle mini-batch data to avoid overlap of labels. Note that the number of accumulation layers is not necessarily equal to the number of model layers as we can add perturbations to any layers. When , we only inject perturbations into the input. indicates that all layers of the model are perturbed. The training procedure is summarized in Algorithm LABEL:alg:1. Parameter can be set to a fixed value for all layers or set differently for different layers. During training, we add the gradient accumulation layer after each neural network layer. During testing, all the gradient accumulation layers are removed from the model.
In order to empirically evaluate the effectiveness of the proposed methods in addressing hard drive healthy degree prediction, we conduct a series of experiments on two datasets and compare it with several existing methods. Furthermore, a real mobile application is used as an experimental example to verify that LPAT can be applied to other similar anomaly detection problems where the dataset is highly imbalanced. In the following, we first introduce the preparation of the experiments. Then we present the experimental results along with the analysis.
Iv-a Data Preparation
|# of H||# of F||# of H||# of F|
Our evaluation and analysis are based on the Backblaze dataset
In this paper, we evaluate the proposed methods based on the SMART statistics of two different models of hard drives in year 2016. The first one is Seagate ST4000DM000 (ST-1). There are 33,800 healthy hard drives and 938 failed hard drives of ST-1 in the dataset. After data cleaning and aggregation , there are 30,685 healthy hard drives and 758 failed hard drives. To further verify our method, we collect another small dataset from Seagate ST8000DM002 (ST-2). There are 8660 healthy hard drives and 48 failed hard drives. After pre-processing there are 7932 healthy drives and 47 failed drives. It should be noted that ST-1 and ST-2 are different models of hard drives although they are from the same hard drive manufacture , which means that they show different degradation progressions. Table I lists the details.
Apparently, the dataset is extremely imbalanced. For example, there are only 2.47% samples that are failed in ST-1. The prediction model encounters serious overfitting and biased fitting problems if the dataset is used directly, even with the help of LPAT. Hence, we select the representative subset of healthy hard drives to construct the dataset for training and testing our model, by which the amount of healthy hard drives can be reduced without a significant loss of information about healthy hard drives. The -means clustering algorithm  is used to cluster the healthy hard drives into 10 clusters based on every day’s records of SMART attributes. Then, for each cluster, the top 30% samples closest to the centroid are selected as the representative subset of healthy hard drives.
We follow the previous works [6, 2, 18] to select distinctive SMART features. As the values of different SMART attributes vary widely, we rescale the values of each selected SMART attributes by the following formula to avoid bias to SMART attributes with large values:
where is the original value of a SMART attribute, and and are the minimum value and the maximum value of a SMART attribute.
Similar to the definitions used in [19, 18], we predict the hard drives based on their residual life. Label ‘0’ is “red alert”, which means the residual life is less than 5 days. Label ‘1’ means “going to fail” in 5 to 15 days. Label ‘2’ represents “healthy”. For the failed hard drives with more than 15-day residual life, we regard them as unlabeled data.
We split each dataset into three subsets, i.e., the training set, the validation set, and testing set. In the supervised setting, we use 80% of all labeled data as the training set and the rest as the testing set. For ST-1, we use 20% of the training set as validation set for tuning hyper-parameters. For ST-2, we do not further split the training set due to its small population. In the semi-supervised setting, we add unlabeled data into the training set and keep the validation set and the testing set unchanged. The information about the datasets is summarized in Table II.
Iv-B Baselines and Metrics
To demonstrate the performance improvement of the proposed approach, we compare it with the following method:
DT: It is a decision tree method, which creates a tree-like model to predict the value of a target variable by learning decision rules inferred from the data features.
RGF : It is a regularized greedy forests based method to predict the impending replacements of hard drives.
RNN : It is a recurrent neural network based model for predicting hard drive failure and giving health degrees, which treats the observed SMART attributes as time-sequence data.
Besides, we test the performance of the proposed LSTM based neural network when it is trained by different existing training methods.
basic: It is the proposed LSTM based neural network trained without adversarial training methods. Specifically, it includes two dense (fully connected) layers, followed by a LSTM layer and a dense layer.
basic+AT : It is the basic neural network trained by an adversarial training method which adds perturbations to inputs in the supervised setting.
basic+RDAT : The basic model is trained by an adversarial training method which adds perturbations to intermediate layer of networks in the supervised setting.
basic+VAT : We use the training method which adds perturbations to inputs from the model distribution alone without necessarily using the label information. It can be applied to the semi-supervised setting.
For these baselines, we tune parameters on the validation set and report their best results on the testing set.
For the proposed LPAT, we design three variations: (1) LPAT-Bottom where the adversarial perturbations are added to the bottom two layers, (2) LPAT-Top where the adversarial perturbations are added to the LSTM layer and the top dense layer, and (3) LPAT-All where the adversarial perturbations are added to the all layers.
We measure the overall accuracy, and each class’s precision, recall, and Macro-F1 score. For the basic prediction model, we set the two bottom dense layers with 128 units and the activation as None. The LSTM layer has 200 units. The final dense layer includes 3 units as we have 3 classes. The time sequence window for each sample is set to 20. We use RMSProp optimizer  with learning rate 0.001. The mini-batch size is 128. The epsilon is empirically set in . The hyperparameter is set between . The training epochs are set to 210.
Our methods are implemented in Tensorflow . All the experiments are trained and tested on four NVIDIA Tesla K80 GPUs. Codes are available online for reproducibility.
Iv-C Supervised Setting
We first conduct experiments in the supervised setting where all the training data are labeled. As can be seen from Table III, the proposed LPAT method achieves the best results on both datasets. It improves accuracy by and Macro-F1 by on ST-1 compared to the best baseline method. As LPAT can flexibly choose which layer and how many layers to add perturbations, it makes the training more robust and brings better performance on the testing set. On ST-2, it improves accuracy by and Macro-F1 by over the other baseline methods. These results indicate that the performance improvement is more significant on small datasets like ST-2. Considering that neural networks are more likely to encounter overfitting when trained by small datasets, we infer that our method shows good regularization and generalization ability to prevent overfitting.
RNN performs better than RGF on ST-1 while slightly worse than RGF on ST-2. ST-1 and ST-2 datasets are collected on different hard drive models whose SMART attributes are also different. RGF which automatically selects SMART attributes may perform better in feature obvious dataset . Models with the adversarial learning achieve better results than the basic one without the adversarial training. VAT slightly improves the performance than AT on the both datasets. basic+RDAT shows the best performance among the baseline methods only inferior to LPAT. It argues that adding perturbation to intermediate layers is an effective approach to improving robustness of neural networks.
We further investigate the impact of different perturbation patterns. For ST-1, LPAT-Bottom works better than only adding perturbation to the input layer while worse than LPAT-Top. Among the three perturbation patterns, LPAT-All achieves the best results, which could be partly attributed to the LSTM layer. Adding perturbations to the bottom layers before the LSTM layer will make perturbations wrapped in the LSTM layer and show less effect on the loss function. It undermines the effectiveness of adversary perturbations. Thus, LPAT-Bottom is less effective than adding perturbations after the LSTM layer. For ST-2, LPAT-Top achieves the best performance on accuracy and Macro-F1, though; its performance is only marginally better than LPAT-All. Based on these results, we speculate that adding perturbations to top or all layers can achieve better regularization ability.
Table IV details the results on each class. Note that the size of the training samples of class ‘1’ is two times larger than that of class ‘0’ on both datasets. Precision represents the ability for a classifier to capture hard drives fails in and days precisely. Recall represents the failure detection rate, i.e. the fraction of failed drives that are correctly classified as failed. The proposed methods outperform basic+AT, basic+RDAT, and basic+VAT in terms of precision and recall on both datasets. In addition, LPAT+All improves Macro-F1 by on class ‘0’ and on class ‘1’ than basic+VAT for ST-1, while LPAT+Top improves Macro-F1 by on class ‘0’ and on class ‘1’ for ST-2. These results demonstrate that adding perturbations to different layers improves the performance of the classifier. Besides, the better performance of LPAT compared to basic+RDAT on both datasets concludes that using the KL divergence to measure perturbation could further improve robustness of neural networks. Based on the observation, our methods can give operators and users more flexibility to take different actions and days before the final failure.
Iv-D Semi-Supervised Setting
In the following groups of experiments, we test the performance of the proposed method when using unlabeled data during training. Fig. 3 plots the testing loss on ST-1 when using all unlabeled and labeled data for AT, VAT, and LPAT+All. It shows that VAT and LPAT+All which utilize the unlabeled data produce lower testing losses. The testing loss of AT becomes unstable, a sign of overfitting, with the increase of the epochs. LPAT+All generates the lowest testing loss and keeps the tendency over the training. Thanks to the resistance to the adversarial perturbations at every layer, LPAT+All performs better than VAT which merely resists to the perturbations at the input.
Then we analyze the influence of different amounts of unlabeled data. The labeled data are always totally used during training in these experiments. But the amount of unlabeled data varies. We change the amount of unlabeled data used during training from 0% to 100% of all the unlabeled data. A larger value means that more unlabeled samples are used. Fig. 4 gives the effect on the overall performance.
From Fig. 4, we can see that when using unlabeled data, both the accuracy and the Macro-F1 score are improved compared with the purely supervised setting where the percent of unlabeled data used is . By extending the adversarial perturbation to the semi-supervised setting, the Macro-F1 of LPAT+All and basic+VAT are improved by and , respectively. When the amount of unlabeled data increases, the accuracy and the Macro-F1 of both LPAT+All and basic+VAT firstly increase. They achieve the highest scores around and unlabeled data, respectively. After that, the accuracy and the Macro-F1 score slightly decrease. We speculate that this is because when there are too many unlabeled data, the perturbation-based adversarial training method will focus more on resisting the perturbations to minimize the adversarial loss on unlabeled samples rather than the supervised loss. Hence, the model prioritizes being robust to perturbations rather than correctly predicting hard drives health degrees. In addition, we find that the performance of LPAT+All is almost always better than that of basic+VAT even when LPAT+All is trained without unlabeled data. It once again indicates that layerwise perturbation is a powerful approach to improving regularization and generalization of models.
Table V reports the best overall results of different methods. In order to clearly identify the performance improvement brought by the extension of semi-supervised setting, we list the performance of basic+AT and basic+RDAT which only use the label information to generate perturbation at the bottom of the table. For ST-1, LPAT+All improves Macro-F1 by than basic+VAT and by than basic+RDAT. For ST-2, LPAT+All improves Macro-F1 by than basic+VAT and by than basic+RDAT. This results show that the extension of semi-supervised setting could considerably improve the performance. Besides, the improvement brought by the layerwise perturbation is usually more notable on small dataset.
Iv-E Parameter Analysis
LPAT mainly involves two hyperparameters: and , where controls the intensity of perturbation added to different layers and controls the trade-off between the supervised loss and the adversarial loss. Fig. 5(a) shows the effects of with different values of . LPAT achieves the best result when and . We can find that the performance (i.e., Macro-F1) is not very sensitive to the parameter . When , it typically can improve the neural networks performance. But when the value of is extraordinarily large, the performance would drop substantially. The perturbation is too large for neural networks to resist. Fig. 5(b) shows the effects of with different values of . The best result appears around and . When , Macro-F1 shows a descending trend. It indicates that should not be too large otherwise the model would be trained to focus on the resistance against perturbation. In most cases , the proposed method generates better results than ‘base’. We find from the above results that the adversarial perturbation should be moderate to achieve the balance between the supervised loss and the adversarial loss.
Iv-F Generality in Other Domains
In the following experiments, we examine the generality of LPAT in other anomaly detection problems where the dataset is imbalanced. We apply the proposed methods to a image recognition task and a sequential analysis task. In the image recognition task, we construct an imbalanced dataset from MNIST. In the time series analysis task, we use a mobile application
MNIST  is a large dataset of handwritten digits which includes 60,000 training images and 10,000 testing images with 10 classes from 0-9. We choose class ‘3’ and ‘5’ to construct the imbalanced 2-class dataset. Images of ‘3’ and ‘5’ are more similar to each other than other pairs, which increases the classification difficulty. In the original training dataset, there are 6,131 images of class ‘3’ and 5,421 images of class ‘5’. We randomly select images of class ‘3’ as labeled data and images as unlabeled data. All images of class ‘5’ are used as labeled data. Then, there are 306 labeled images of class ‘3’ and 5,421 labeled images of class ‘5’ in the imbalanced training dataset we construct. For the testing dataset, we randomly select original testing images of class ‘3’ and all original testing images of class ‘5’ to construct the new testing dataset. For the basic model, we use a ReLU based neural network consisting of two hidden layers with the number of hidden units (1200, 1200) .
In order to further verify the effectiveness of LPAT in real scenario, we use a mobile application DeepMood to test LPAT’s performance. DeepMood harnesses the sequential information collected from the basic keystroke patterns and the accelerometer on the phone to predict the user’s mood disorder. The alphanumeric character typing pattern (ALPH.) and accelerometer values pattern (ACCEL.) of 40 volunteers were collected over 8 weeks. There are 722 positive samples (disorder) and 7,500 negative samples (health) in the original dataset. We mask the labels of 30% positive samples to construct the unlabeled data. We use 80% of all labeled data as the training set and the rest as the testing set. The single-view DNN proposed in  is used as the basic model.
Table VI shows the results of the image recognition task (MNIST) and the sequential analysis task (ALPH. and ACCEL.). As the two tasks are binary classification problems, we use F1-score instead of Macro-F1. The performance of four representative methods is listed. Note that basic+AT and basic+RDAT are applied to the supervised setting, and so they cannot utilize the unlabeled data. It can be found that LPAT+All demonstrates the best performance on the three datasets, which verifies the generality of our proposed methods.
On MNIST, the accuracy and the F1-score of LPAT+All are and higher than those of basic+VAT, respectively. The slight improvement is partly due to the fact that MNIST is a relatively easy task where the accuracy and the F1-score are already quite high even without adversarial training. It is difficult to further improve the performance on MNIST.
On the two datasets of the mobile application DeepMood, LPAT+All shows the ability to substantially improve the performance. The F1-score of LPAT+All is over higher than the best baseline methods on both the two datasets. Basic+RDAT which adds adversarial perturbations to intermediate layers performs better than basic+AT only adding perturbations to inputs. But the performance of basic+RDAT is slightly inferior to that of basic+VAT which can be applied in the semi-supervised setting. This observation indicates that using KL divergence and unlabeled data is a little more powerful than adding perturbations to intermediate layers on these two datasets.
V Related Work
A series of models have been proposed to predict the impending failures of hard drives. A nonparametric model was proposed by  to detect the anomalous SMART values of hard drives.  designed a Bayesian classifier to predict the hard drive failures. In order to capture the temporal features embedded in the SMART statistics, some researchers resorted to the time series analysis methods.  employed the hidden Markov model to predict the imminent failures.  used the Bayes-based model to detect the changepoint of SMART statistics, and then compacted the time sequence representation by using an exponential smoothing technique. Besides, the down sampling technology was implemented in  to solve the imbalanced dataset problem. All the above works merely attempted to predict the failure, and cannot provide further information about the status of hard drives.  tried to assess the health status of hard drives based on their residual lives. They designed a simple recurrent neural network to classify hard drives as different groups. According to the health status, different protective measures can be taken to improve the storage system reliability. However, the accuracy of their prediction is relatively poor due to the simpleness of their model.
Adversarial training is a strong regularization method which was originally introduced in [12, 11]. Their works showed that several deep neural networks are vulnerable to a very small perturbation in the direction that the model’s assignment of labels to an unseen class in the most adversarial(sensitive) way. This small perturbation is called adversarial perturbation, which has shown a better performance [26, 30] than dropout  and models trained by random perturbations [12, 26] like adding Gaussian noise . They also found that training the models to be robust against adversarial perturbations was effective to improve performance on testing dataset.  proposed the virtual adversarial method which expanded adversarial training method into semi-supervised learning areas, which improved robustness of models by utilizing the model’s posterior distribution against local perturbations around each input data point. They also demonstrated the effects in image classification tasks  and text domain .  designed noise training methods to improve performance of shallow neural networks in mobile cloud.  designed methods to improve robustness and performance of very deep neural networks such as VGGnet , InceptionV3  by perturbing intermediate layers. However, above methods either can only be used in supervised learning domains or can only add adversarial perturbations to inputs in semi-supervised learning way.
In this paper, we propose a layerwise perturbation-based adversarial training method with an application on hard drive health degree prediction. Differing from traditional methods which usually predict hard drive health in a binary fashion, the proposed method focuses on different health degrees. Additionally, the proposed LPAT can not only add adversarial perturbations to the input but also to the intermediate layers or all layers, which improves the generalization and performance of model. Besides, it utilizes unlabeled data to further improve performance on prediction. Extensive experiments demonstrate the superiority of our proposed methods.
In the future, we will explore improving performance on more healthy degrees which can give better instructions for technicians and users to take different actions. Also, our method calculates adversarial perturbations each time based on current mini-batch, which is time consuming. We will further optimize the method to reduce computation time and improve performance. In addition, we will extend the adversarial perturbation methods into more other domains which have large-scale unlabeled data such as text classification.
- Note that we choose the health degree intuitively; other definitions can be used with only a slight modification.
- S. Sankar, M. Shaw, K. Vaid, and S. Gurumurthi, “Datacenter scale evaluation of the impact of temperature on hard disk drive failures,” ACM Transactions on Storage, vol. 9, no. 2, 2013.
- M. M. Botezatu, I. Giurgiu, J. Bogojeska, and D. Wiesmann, “Predicting disk replacement towards reliable data centers,” in KDD, 2016.
- B. Allen, “Monitoring hard disks with smart,” Linux Journal, no. 117, 2004.
- Backblaze, “Hard drive reliability review for 2015,” https://www.backblaze.com/blog/hard-drive-reliability-q4-2015/, 2016.
- G. F. Hughes, J. F. Murray, K. Kreutz-Delgado, and C. Elkan, “Improved disk-drive failure warnings,” IEEE Transactions on Reliability, vol. 51, no. 3, 2002.
- J. F. Murray, G. F. Hughes, and K. Kreutz-Delgado, “Machine learning methods for predicting failures in hard drives: A multiple-instance application,” J. Mach. Learn. Res., vol. 6, no. 5, 2005.
- B. Zhu, G. Wang, X. Liu, D. Hu, S. Lin, and J. Ma, “Proactive drive failure prediction for large scale storage systems,” in MSST, 2013.
- F. Mahdisoltani, I. Stefanovici, and B. Schroeder, “Proactive error prediction to improve storage system reliability,” in USENIX ATC, 2017.
- S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Computation, vol. 9, no. 8, 1997.
- A. Graves, A. r. Mohamed, and G. Hinton, “Speech recognition with deep recurrent neural networks,” in IEEE ICASSP, 2013.
- C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. Goodfellow, and R. Fergus, “Intriguing properties of neural networks,” in ARXIV, 2013.
- I. J. Goodfellow, J. Shlens, and C. Szegedy, “Explaining and harnessing adversarial examples,” in ICLR, 2015.
- T. Miyato, A. M. Dai, and I. Goodfellow, “Adversarial training methods for semi-supervised text classification,” in ICLR, 2016.
- A. Kurakin, I. Goodfellow, and S. Bengio, “Adversarial machine learning at scale,” in ICLR, 2017.
- S. Kullback and R. A. Leibler, “On information and sufficiency,” The Annals of Mathematical Statistics, vol. 22, no. 1, 1951.
- Golub, G. H, and H. A. Van der Vorst, “Eigenvalue computation in the 20th century,” Journal of Computational and Applied Mathematics, vol. 123, no. 1, 2000.
- T. Kanungo, D. M. Mount, N. S. Netanyahu, C. D. Piatko, R. Silverman, and A. Y. Wu, “An efficient k-means clustering algorithm: Analysis and implementation,” IEEE TPAMI, vol. 24, no. 7, 2002.
- C. Xu, G. Wang, X. Liu, D. Guo, and T. Y. Liu, “Health status assessment and failure prediction for hard drives with recurrent neural networks,” IEEE Transactions on Computers, vol. 65, no. 11, 2016.
- J. Li, R. J. Stones, G. Wang, Z. Li, X. Liu, and K. Xiao, “Being accurate is not enough: New metrics for disk failure prediction,” in IEEE SRDS, 2016.
- S. Sankaranarayanan, A. Jain, R. Chellappa, and S. N. Lim, “Regularizing deep networks using efficient layerwise adversarial training,” in AAAI, 2018.
- T. Miyato, S.-i. Maeda, M. Koyama, and S. Ishii, “Virtual adversarial training: a regularization method for supervised and semi-supervised learning,” in ARXIV, 2017.
- T. Tieleman and G. E. Hinto, “Lectur e6.5-rmsprop: Divide the gradient by a running average of tts recent magnitude,” COURSERA: Neural Networks for Machine Learnings, vol. 4, no. 2, 2012.
- M. Abadi, A. Agarwal, P. Barham, E. Brevdo, Z. Chen, C. Citro, G. S. Corrado, A. Davis, J. Dean, M. Devin, et al., “Tensorflow: Large-scale machine learning on heterogeneous distributed systems,” in ARXIV, 2016.
- B. Cao, L. Zheng, C. Zhang, P. S. Yu, A. Piscitello, J. Zulueta, O. Ajilore, K. Ryan, and A. D. Leow, “Deepmood: Modeling mobile phone typing dynamics for mood detection,” in KDD, 2017.
- Y. LeCun, L. Bottou, Y. Bengio, and G. Hinton, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, 1998.
- T. Miyato, S.-i. Maeda, M. Koyama, K. Nakae, and S. Ishii, “Distributional smoothing with virtual adversarial training,” in ICLR, 2016.
- G. F. Hughes, J. F. Murray, K. Kreutz-Delgado, and C. Elkan, “Improved disk-drive failure warnings,” IEEE Transactions on Reliability, vol. 51, no. 3, 2002.
- G. Hamerly, C. Elkan, et al., “Bayesian approaches to failure prediction for disk drives,” in ICML, 2001.
- Y. Zhao, X. Liu, S. Gan, and W. Zheng, “Predicting disk failures with hmm-and hsmm-based approaches.,” in ICDM, 2010.
- Q. Wang, W. Guo, K. Zhang, A. G. Ororbia II, X. Xing, X. Liu, and C. L. Giles, “Adversary resistant deep neural networks with an application to malware detection,” in KDD, 2017.
- N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: A simple way to prevent neural networks from overfitting,” J. Mach. Learn. Res., vol. 15, no. Jan, 2014.
- P. Bachman, O. Alsharif, and D. Precup, “Learning with pseudo-ensembles,” in NIPS, 2014.
- J. Wang, J. Zhang, W. Bao, X. Zhu, B. Cao, and P. S. Yu, “Not just privacy: Improving performance of private deep learning in mobile cloud,” in KDD, 2018.
- K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in ICLR, 2015.
- C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the inception architecture for computer vision,” in CVPR, 2016.