Robust Multi-Output Learning with Highly Incomplete Data via Restricted Boltzmann Machines

Robust Multi-Output Learning with Highly Incomplete Data via Restricted Boltzmann Machines

Abstract

In a standard multi-output classification scenario, both features and labels of training data are partially observed. This challenging issue is widely witnessed due to sensor or database failures, crowd-sourcing and noisy communication channels in industrial data analytic services. Classic methods for handling multi-output classification with incomplete supervision information usually decompose the problem into an imputation stage that reconstructs the missing training information, and a learning stage that builds a classifier based on the imputed training set. These methods fail to fully leverage the dependencies between features and labels. In order to take full advantage of these dependencies we consider a purely probabilistic setting in which the features imputation and multi-label classification problems are jointly solved. Indeed, we show that a simple Restricted Boltzmann Machine can be trained with an adapted algorithm based on mean-field equations to efficiently solve problems of inductive and transductive learning in which both features and labels are missing at random. The effectiveness of the approach is demonstrated empirically on various datasets, with particular focus on a real-world Internet-of-Things security dataset.

1 Introduction

Modern machine learning techniques usually require large sets of fully observed and well labelled data for training, which is usually unrealistic in real-world applications. Co-occurrence of missing features and partially observed label information is a widely witnessed issue in industrial data analytic business. On one hand, training features are not complete or heavily corrupted due to unpredictable sensor / end-device failure, collapse of database and noisy communication channels. On the other hand, human and automatic annotators can not recognize the incomplete feature profiles. It is safe for them to leave the incomplete training instances unlabelled and provide labels to the fully observed data instances with noise-free feature profiles. Consequently, it is highly demanding for the machine learning system to tolerate co-occurrence of missing features and partially observed labels for robust learning in practical use.

We study both transductive and inductive multi-output classification. Multi-output classification, including multi-class and multi-label learning tasks, outputs class labels with higher dimensional representation. They thus define a more sophisticated learning scenario compared to binary classification. The former associates an input instance to one class of a finitely defined class set, while the latter allows one instance to be associated to multiple labels simultaneously. In the transductive case, the learning objective is to infer the missing features and labels based on the observed ones (no separate test set is required). In the inductive case, the model is trained over a set of incomplete training instances and the classification is performed by inferring the labels associated to incomplete test instances. Though the problem of multi-output learning with semi-supervised labels has been discussed in previous works [\citeauthoryearSun, Y.Zhang, and Zhou2010, \citeauthoryearBucak, Jin, and Jain2011], they all assume features of both training and test instances to be fully observed and free from noise. In our work, we consider that both features and labels are missing completely at random, which means that the mask of missing information is assumed to be statistically independent on the data distribution.

State-of-the-art accuracies for classification with incomplete or noise-corrupted features and partially observed labels are given by CLE [\citeauthoryearHan et al.2018], NoisyIMC [\citeauthoryearChiang, Hsieh, and S.Dhillon2015] and MC-1 [\citeauthoryearCabral et al.2015]. In particular, CLE and NoisyIMC are able to conduct both transductive and inductive learning for multi-label classification, while MC-1 is a transductive-only method. It is worth noting that CLE has been originally proposed to handle the issue of weak labeling in multi-label learning, where only positive labels are observed. NoisyIMC and MC-1 can work in the semi-supervised learning scenario. However, none of the three methods can handle all the challenges co-currently raised in our work. First, all of these assume the testing instances to have fully observed feature profiles. They don’t consider coping with incomplete testing instances by design. Second, all of them are designed specifically for multi-label learning and adapting them to multi-class classification is not straightforward.

More recently, methods based on Deep Latent Variable Models (DLVM) have been proposed to deal with missing data. In [\citeauthoryearMattei and Frellsen2019], the Variational Autoencoder [\citeauthoryearKingma and Welling2014] has been adapted to be trained with missing data and a sampling algorithm for data imputation is proposed. Other approaches based on Generative Adversarial Networks (GAN) by [\citeauthoryearGoodfellow et al.2014] are proposed in [\citeauthoryearYoon, Jordon, and van der Schaar2018] and [\citeauthoryearLi, Jiang, and Marlin2019]. Impressive results on image datasets are displayed for these models, at the price of a rather high model complexity and the need for a large training set. In addition these works are focused on features reconstruction, and additional specifications and fine-tuning are required to be able to take partially observed labels into account. The models specifications are quite involved and any new specificity of the dataset may increase both the cost and the difficulty in training (especially for the approaches based on GANs).

In this paper we choose to address this problem in a more economical and robust manner. We consider the old and simple architecture of the Restricted Boltzmann Machine and adapt it to the multi-output learning context (RBM-MO) with missing data. The RBM-MO method serves as a generative model which collaboratively learns the marginal distribution of features and label assignments of input data instances, despite the incomplete observations. Building on the ideas expressed in [\citeauthoryearNijman and Kappen1994, \citeauthoryearGhahramani and Jordan1994] we adapt the approach to the more effective contrastive divergence training procedure [\citeauthoryearHinton2002] and provide results on various real-world datasets. The advantage of the RBM-MO model is that of providing a robust and flexible method to deal with missing data, with little additional complexity with respect to the classic RBM. Indeed, the trained model can be naturally applied to both transductive and inductive scenarios, achieving superior multi-output classification performance then state-of-the-art baselines. Moreover, it works seamlessly with multi-class and multi-label tasks, providing a unified framework for multi-output learning.

In Section 2 we will present a brief overview of RBM. In Section 3 and 4 we will show how the RBM can be adapted to work with missing data and we will present an effective imputation procedure. In Section 5 we present the results of our experimental study. Here, in addition to presenting results on some commonly used public datasets, we will also introduce a dataset for security in the Internet-of-Things which serves as a typical example of multi-output classification with highly incomplete data [\citeauthoryearanonymous2018] and evaluate to what extent the proposed RBM-MO method can improve the quality of security services.

2 Overview of Restricted Boltzmann Machines

An RBM is a Markov random field with pairwise interactions defined on a bipartite graph formed by two layers of non-interacting variables: the visible nodes represent instances of the input data while the hidden nodes provide a latent representation of the data instances. and will denote respectively the sets of visible and hidden variables. In our setting, the visible variables will further split into two subsets and corresponding respectively to features and labels, such that . The visible variables form an explicit representation of the data and are noted . The hidden nodes serve to approximate the underlying dependencies among the visible units.

In this paper, we will work with binary hidden nodes . The variables corresponding to the visible features will be either real with a Gaussian prior or binary, depending on the data to model, and labels variables will always be binary . The energy function of the RBM is defined as

(1)

where and are biases acting respectively on the visible and hidden units and is the weight matrix that couples visible and hidden nodes. The joint probability distribution over the nodes is then given by the Boltzmann measure

(2)

where is in product form and encodes the nature of each visible variable, either with a Gaussian prior or a binary prior . is the partition function. The classical training method consists in maximizing the marginal likelihood over the visible nodes by tuning the RBM parameters via gradient ascent of the log likelihood . The update rules for the weights (and similarly for the fields) are

(3)

with the parameter corresponding to the learning rate.

The tractability of the method relies heavily on the fact that the conditional probabilities and are given in closed forms. In our case these read:

(4)
(5)

where is the logistic function. The gradient of the likelihood w.r.t. the weights (and similarly w.r.t. the fields and ) is given by

(6)

where the brackets and respectively indicate the average over the data and over the distribution (2). The positive term is directly linked to the data and can be estimated exactly with (5), while the negative term is intractable. Many strategies are used to compute this last term: the contrastive divergence (CD) approach [\citeauthoryearHinton2002] consists in estimating the term over a finite number of Gibbs sampling steps, starting from a data point and making alternate use of (4) and (5); in its persistent version (PCD) [\citeauthoryearTieleman2008] the chain is maintained over subsequent mini-batches; using mean-field approximation [\citeauthoryearMarylou, Tramel, and Krzakala2015] the term is computed by means of a low-couplings expansion.

3 Learning RBM with incomplete data

The RBM is a generative model able to learn the joint distribution of some empirical data given as input. As such, it is intrinsically able to encode the relevant statistical properties found in the training data instances that relate features and labels, and this makes the RBM particularly suitable to be used in the multi-output setting in the presence of incomplete observations. In this sense, the most natural way to deal with incomplete observations is to marginalize over the missing variables; in this section we show how the contrastive divergence algorithm can be adapted to compute such marginals.

Given a partially-observed instance , we have a new partition of the visible space , where is a subset of observed values of that can correspond both to features and labels. and denote respectively the observed and missing values of . Let’s define the following quantity

corresponding to the marginalization of the numerator of the probability distribution over the visible variables, representing the parameters of the model. The probability over the observed variables is given by

Taking the log-likelihood and then computing the gradient with respect to the weight matrix element (also similarly for the fields and ), we obtain two different expressions for and .

(7)
(8)

The gradient of the LL over the weights (6) now reads

(9)

where is the indicator function of the samples dependent set . The observed variables are pinned to the values given by the training samples. In terms of our model, the pinned variables play the role of an additional bias over the hidden variables of a RBM where the ensemble of visible variables is reduced to the missing ones. This results in the following shift to the bias of the hidden nodes:

(10)

With respect to the non-lossy case where is given in closed form, here we need to sum over the missing variables in order to estimate . This means that also the positive term of the gradient (3) is now intractable and we need to approximate it. For CD training, we can simply perform Gibbs sampling over the missing variables (keeping fixed the observed variables). The training algorithm then becomes:

1:Data: a training set of N data vectors
2:Randomly initialize the weight matrix W
3:for t = 0 to T (# of epochs) do
4:     Divide the training set in minibatches
5:     for all minibatches  do Positive term:
6:         pin variables to their correct value
7:         initialize randomly
8:         sample using and for k steps
9:         compute the positive terms in (7) and (8) Negative term:
10:         initialize randomly
11:         iterate eq. (4), (5) (k steps) to compute Full update:
12:         update with equations (3) and (3)
13:     end for
14:end for
Algorithm 1 Lossy-CDk (RBM training with Incomplete data)

We note that the extra computational burden of Lossy-CD with respect to standard CD is due only to the extra Gibbs sampling steps in the positive term. Given that the observed variables strongly bias the sampling procedure speeding up convergence, only few sampling steps are needed to compute this term. Indeed, in our experiments we observed that a single sampling step (Lossy-CD1) is enough, making the additional complexity minimal. Finally, we note that the same method can be applied to PCD and mean-field training procedures. In the first case, it is sufficient to keep track of an additional persistent chain, which requires little extra memory and no extra computational complexity. In the second case, we only need to substitute Gibbs sampling with iterative mean-field equations.

4 Mean-field based imputation with RBM

As a generative model, the trained RBM can be used to sample new data. For imputation of missing features and labels we just need to use the observed portions of our data to bias the sampling procedure in the same way as for the computation of the positive term in Alg. 1. Namely, we estimate by pinning the observed variables and iterating CD/PCD or mean-field to approximate the equilibrium values of the missing variables. In case of a high percentage of missing observations, however, we might expect the observed variables to be correlated to many different equilibrium configurations, such that the sampling could be biased towards the wrong sample. This effect is present also during training, but in that case it is mitigated by the average over the minibatch while for imputation over a single data instance it becomes relevant. To overcome this problem, we simply average over multiple imputations for each incomplete data instance. Here we highlight the fact that the equilibrium configurations of the RBM are weighted according to the empirical data distribution; as a consequence, the bias toward an incorrect sample is easily discarded and generally a small number of different imputations need to be averaged to obtain the correct result. To further reduce the number of imputations, we employ mean-field. Indeed, a good equilibrium configuration could require more CD/PCD samples (simply due to the sampling noise) while we expect a non-extensive number of fixed-points related to meaningful (equilibrium) configurations to be present [\citeauthoryearDecelle, Fissore, and Furtlehner2018] and these are directly obtained by iteration.

More in details, let and be the marginal probabilities respectively of visible labels and hidden variables to be activated and the marginal expectation of the visible features variables. Mean-field equations at lowest order (, being the size of the system) express self-consistent relations among these quantities

(11)
(12)
(13)

Higher order terms corresponding to TAP equations are discarded [\citeauthoryearMézard2017]. These equations can be efficiently solved by iteration starting from random configurations until a fixed point is reached. Observed variables are simply introduced by pinning their corresponding probabilities ( or for label variables) or their marginal expectation (for feature variables) to the observed values. In practice we run these fixed-point equations times and the imputations are obtained by simple average

In the multi-label setting, the predictor is the indicator function ( is learned, it is chosen to maximize the accuracy for known labels), while for class labels we have

Model RMSE Averaged AUC Accuracy
30% 50% 80% 30% 50% 80% 30% 50% 80%
RBM-MO(50%) 0.183 0.182 0.185 0.969 0.971 0.929 0.950 0.912 0.822
CLE(50%) 0.195 0.195 0.195 0.686 0.718 0.742 0.256 0.232 0.282
NoisyIMC(50%) 0.209 0.210 0.210 0.621 0.578 0.552 0.225 0.232 0.192
MC-1(50%) 0.334 0.335 0.337 0.495 0.493 0.500 0.110 0.111 0.112
RBM-MO(80%) 0.209 0.213 0.211 0.938 0.932 0.906 0.920 0.852 0.733
CLE(80%) 0.206 0.208 0.206 0.673 0.678 0.625 0.230 0.215 0.220
NoisyIMC(80%) 0.212 0.211 0.213 0.652 0.577 0.537 0.230 0.217 0.210
MC-1(80%) 0.334 0.334 0.335 0.500 0.501 0.500 0.112 0.110 0.110
Table 1: Transductive test on MNIST multi-class data set
Model RMSE Micro-AUC Hamming-Accuracy
30% 50% 80% 30% 50% 80% 30% 50% 80%
RBM-MO(50%) 0.131 0.137 0.123 0.943 0.934 0.888 0.919 0.907 0.873
CLE(50%) 0.130 0.130 0.131 0.905 0.893 0.898 0.885 0.871 0.878
NoisyIMC(50%) 0.132 0.133 0.133 0.865 0.863 0.858 0.845 0.841 0.848
MC-1(50%) 0.258 0.255 0.267 0.522 0.528 0.527 0.826 0.817 0.824
RBM-MO(80%) 0.160 0.158 0.158 0.875 0.867 0.826 0.856 0.858 0.832
CLE(80%) 0.129 0.129 0.128 0.913 0.897 0.899 0.889 0.875 0.876
NoisyIMC(80%) 0.133 0.134 0.134 0.853 0.857 0.849 0.839 0.835 0.826
Table 2: Transductive test on Scene multi-label data set
Model RMSE Micro-AUC Hamming-Accuracy
30% 50% 80% 30% 50% 80% 30% 50% 80%
RBM-MO(50%) 2.316 2.427 2.463 0.842 0.852 0.794 0.972 0.956 0.912
CLE(50%) 9.543 8.434 10.902 0.872 0.837 0.739 0.900 0.870 0.800
NoisyIMC(50%) 8.443 8.792 9.850 0.835 0.817 0.799 0.820 0.801 0.738
MC-1(50%) 2.220 2.180 2.128 0.590 0.607 0.573 0.697 0.700 0.693
RBM-MO(80%) 3.235 3.128 3.238 0.843 0.782 0.746 0.972 0.945 0.910
CLE(80%) 15.265 15.989 14.559 0.832 0.784 0.785 0.885 0.757 0.729
NoisyIMC(80%) 14.155 14.289 14.395 0.852 0.822 0.803 0.895 0.807 0.735
MC-1(80%) 2.776 2.423 2.476 0.536 0.540 0.536 0.691 0.688 0.687
Table 3: Transductive test on EventCat multi-label data set
Model Averaged AUC Accuracy
30% 50% 80% 30% 50% 80%
RBM-MO(50%) 0.887 0.914 0.910 0.533 0.673 0.660
CLE(50%) 0.785 0.791 0.791 0.297 0.256 0.268
NoisyIMC(50%) 0.780 0.771 0.781 0.302 0.272 0.265
RBM-MO(80%) 0.891 0.909 0.889 0.562 0.682 0.647
CLE(80%) 0.768 0.664 0.622 0.271 0.200 0.176
NoisyIMC(80%) 0.748 0.687 0.615 0.264 0.220 0.178
Table 4: Inductive test on Pendigits multi-class dataset
Model Micro-AUC Hamming-Accuracy
30% 50% 80% 30% 50% 80%
RBM-MO(50%) 0.839 0.826 0.793 0.970 0.970 0.965
CLE(50%) 0.705 0.707 0.706 0.700 0.724 0.719
NoisyIMC(50%) 0.704 0.702 0.700 0.710 0.717 0.718
RBM-MO(80%) 0.759 0.791 0.766 0.964 0.964 0.967
CLE(80%) 0.693 0.688 0.694 0.718 0.706 0.718
NoisyIMC(80%) 0.689 0.688 0.685 0.705 0.704 0.704
Table 5: Inductive test on EventCat multi-label data set

5 Experimental Study

5.1 Experimental configuration

To evaluate the efficiency of RBM-MO we compare its performance against CLE, NoisyIMC and MC-1, which provide state-of-the-art baselines.

For the transductive experiments we randomly hide features and labels of the whole dataset to generate incomplete data for training, and we compute appropriate scores for the reconstruction of missing features and labels. In the inductive test, instead, we split the whole dataset into non-overlapping training and testing sets. Concerning the training set the same protocol is used as in the transductive test. For the test set the difference is that now all labels are hidden. Once the classifier is trained, it is applied on the test set to predict the labels. We still randomly hide the entries of test features vectors, so as to form an incomplete testing set. Finally, in the splitting we use 70% of the data instances for training and the remaining 30% for testing.

We denote by , and the percentage of masked features, labels and classes labels respectively. Note that a masked class label means that all binary variables attached to the classes of a given label are masked together. These rates of masking are kept identical in the learning and test sets.

In the transductive test, we compute the Root Mean Squared Error (RMSE) to measure the reconstruction accuracy with respect to the missing feature values, defined w.r.t. the dataset as:

where and are the incomplete and reconstructed feature matrices. denotes the set of unobserved features. Furthermore, for the reconstructed labels we calculate Micro-AUC scores and Hamming-accuracy [\citeauthoryearG.Madjarov, Kocev, and Gjorgjevikj2012] in the multi-label scenario, and Averaged AUC plus Accuracy [\citeauthoryearLi and Ye2015] in the multi-class case. In the tables, we define Hamming-accuracy as 1-Hamming loss to keep a consistent variation tendency with the AUC scores. In the inductive test we only compute the scores on the reconstructed labels, since reconstructing missing features is not the goal of inductive classification.

We run the test as described times with different realizations of the missing features and labels. Average and standard deviation of the computed scores are recorded to compare the overall performances. In the tables, we use red fonts to denote the best reconstruction and classification performances among all the algorithms involved in the empirical study. The bold black font is used to highlight the performance of the proposed RBM-MO method.

For the baselines, we used grid search to choose the optimal parameter combination following the suggested ranges of parameters as in [\citeauthoryearHan et al.2018].

The RBM-MO is trained following the guidelines in [\citeauthoryearHinton2010]. We always use binary variables for the hidden layer, while in the visible layer we use binary variables for MNIST and Gaussian variables for the other datasets. In all the simulations, we fix the number of hidden nodes to 100. The learning rate is fixed to 0.001 and the size of the mini-batches to 10. During training the number of Gibbs steps is set to while for imputation we iterate the mean-field equations 10 times. As a stopping condition, we considered the degradation of the transductive AUC scores with a look-ahead of 500 epochs1.

5.2 Summary of datasets

We consider 4 publicly available datasets emerging from diverse disciplines, such as image processing and biology. These datasets cover both multi-label and multi-class learning tasks, and they are popularly used as benchmark datasets in multi-output learning research.

Dataset No. of Instances No. of Features No.of Labels No. of Classes
Yeast 2,417 103 14 -
Scene 2,407 294 6 -
Pendigits 10992 16 - 10
MNIST 70,000 784 - 10
EventCat 5,93 72 6 -
Table 6: Summary of 4 public multi-label and multi-class data sets.

In addition, we consider the challenging scenario of abnormality detection on IoT devices. The relevant dataset consists in security telemetry data collected from various network appliances (e.g. smart watches, smartphones, driving assistance systems…), each reporting a features vector whose entries indicate the occurring frequency of a specific type of alert (e.g. downloading suspicious files, login failures, unfixed vulnerabilities…). Multiple labels are assigned to each device in the collected dataset, corresponding to a variety of categories of security threats. The learning problem is cast as a multi-label classification as the same telemetry features can be related to multiple threats (e.g. scanning activity and data breaching can occur simultaneously) and the collected records for training and testing are usually highly incomplete due to various causes. For example, customers can set up their privacy control policies to limit the coverage of the telemetry data shared with security vendors, or deployed sensors can unexpectedly fail. Simultaneously, it is safe for human analysts to only label the most typical and clearly defined cyber menaces, as mislabeling can reduce usability of the provided protection service by causing false alarm and/or false negative. In the following, we will test our method on a complete realization of such a dataset that we refer to as EventCat.

Some details about the datasets are reported in Table.6.

5.3 Qualitative results on MNIST

A qualitative evaluation of the performance of the RBM-MO model is given by looking at features reconstruction for the MNIST dataset, as reported in Fig. 1. The model at hand has been trained over a dataset in which 50% of the features were missing. To assess the robustness of the method, we computed the reconstructions in the highly challenging case in which 80% of the features were missing. Apart from some smoothing due to the employment of mean-field imputations, the reconstructed samples look reasonably realistic. In general, from the qualitative point of you the results are comparable to those obtained with more complex and expensive DLVMs like MIWAE and MisGAN [\citeauthoryearMattei and Frellsen2019, \citeauthoryearLi, Jiang, and Marlin2019].

Figure 1: Features reconstruction by RBM-MO trained over an incomplete dataset with 50% missing-at-random features, whose classification accuracy has been measured to be around 91%. The first block shows some complete testing instances. The second and third block show the same testing instances after hiding respectively 50% and 80% of the pixels. The last two columns show the results of the mean-field imputations over the incomplete testing instances.

5.4 Empirical study of transductive learning on public datasets

The transductive results for MNIST (multi-class) and Scene (multi-label) datasets are reported in tables 1 and 2, while for the other datasets the full results are reported in the supplementary material. Globally the results show better performances of RBM-MO compared to the baseline methods CLE, NoisyIMC and MC-1. Going into the details, we first observe that RBM-MO is by a large margin more efficient than all of the baselines for the inference of class labels (see table 1 and the Pendigits table in the supplementary material), probably because it is able to encode more complex statistical properties. Among the baseline methods, CLE performs the best. Explicitly enforcing a predictive constraint on the subspace representation of features and labels in the model, provides more robustness against missing information.

On the multi-label problems, the situation is still in favour of RBM-MO but with less margin (table 2 and Yeast table in the supplementary material), in particular at a larger percentage of missing features.

Now if we look at the reconstruction error on these datasets we observe that RBM-MO generally achieves a higher reconstruction accuracy than the other opponents, especially on the MNIST dataset. The results verify empirically the basic motivation of using a generative model, such as the RBM, in the raised challenging learning scenarios: incomplete features and labels can provide complementary information to each other, so as to better recover the missing elements. The variance of the results is omitted in the tables by lack of space. For RBM-MO the standard deviation of the derived RMSE, AUC and accuracy scores is not larger than over the different datasets. Although the RMSE scores reported by the baseline methods look comparable to the RBM-MO ones, and in certain cases they are better, they also come with a slightly higher variance, such that the RBM-MO seems to be more efficient and robust for features reconstruction.

5.5 Empirical study of inductive learning on public datasets

Except MC-1, all the other baseline methods are used for inductive learning. As in the transductive test, we show only the mean of the derived metrics in the tables. Nevertheless, we have similar variance ranges for the computed scores as reported in the transductive test. Clearly RBM-MO is much better adapted to this setting than the baseline methods both for multi-class (table 4) and multi-label learning (Yeast and Scene tables in supplementary material). With a few exceptions at large missing rates () the results of RBM-MO are distinctively better. The baseline inductive methods CLE and NoisyIMC are specifically designed for multi-label learning. In the multi-class scenario, unlabeled instances can’t be treated properly. If a data instance is unlabeled, the whole corresponding row in the label matrix will be considered as missing. These structured missing patterns violate the assumption of random entry missing in semi-supervised learning, which thus leads to performance deterioration of the baseline methods. By comparison, RBM-MO can be adapted seamlessly to multi-class and multi-label learning, producing consistently good performances.

5.6 Transductive and inductive evaluation on EventCat data

Transductive results are reported in table 3. Compared to the three baseline approaches, RBM-MO shows an overall improvement of the combined features and labels reconstruction accuracy for all combinations of . When not the best one (at low missing percentage) RBM-MO reaches comparable features reconstruction errors to the best baseline. Meanwhile the classification performance is consistently better, with generally better AUC scores and up to higher hamming accuracy scores compared to baseline methods. Note that MC-1 achieves the best features imputation results in the transductive test when . However RBM-MO can achieve a similar features imputation accuracy while producing significantly higher labels recovery precision.

As seen in table 5, the results of inductive test of RBM-MO are equally encouraging. Even with highly incomplete training data, e.g. and , RBM-MO produces the best predictions over partially observed testing data instances. In the extreme situation with and set to 0.8, the performances of all the methods are relatively low. Although the algorithms can not work on a stand-alone basis in this situation, their output can guide human analysts to effectively narrow down the hypotheses of correlation between event profiles and threat types. Despite the difficult learning scenario, RBM-MO still achieves the highest precision, which denotes its applicability for aiding human analysts to identify useful heuristic rules for detecting malicious events.

6 Conclusion

Machine learning is witnessing a race to high complexity models eager for large data and computational power. In the context of multi-output classification in a challenging scenario - (i) learning with highly incomplete features and partially observed labels; ii) applying the learnt classifier with incomplete testing instances) - we advocate instead for simple probabilistic and interpretable models. After refining the learning of the RBM model, we give empirical evidences that it can be efficiently adapted to this context on a great variety of datasets. Experiments are conducted on both public databases and a real-world IoT security dataset, showing various sizes of training sets as well as features and labels vectors. Our approach consistently outperforms the state-of-the-art robust multi-class and multi-label learning approaches with imperfect training data, indicating good usability for practical applications.

Footnotes

  1. Full code with instructions for reproducibility and extensions is released: link hidden for blind review

References

  1. anonymous. 2018. Hidden for blind review.
  2. Bucak, S. S.; Jin, R.; and Jain, A. K. 2011. Multi-label learning with incomplete class assignments. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition 2011, 2801–2808.
  3. Cabral, R.; l. Torre, F. D.; Costeira, J. P.; and Bernardino, A. 2015. Matrix completion for weakly-supervised multi-label image classification. IEEE Transactions on Pattern Analysis and Machine Intelligence 37(1):121–135.
  4. Chiang, K.-Y.; Hsieh, C.-J.; and S.Dhillon, I. 2015. Matrix completion with noisy side information. In Proceedings of the 28th International Conference on Neural Information Processing Systems, NIPS’15, 3447–3455. Cambridge, MA, USA: MIT Press.
  5. Decelle, A.; Fissore, G.; and Furtlehner, C. 2018. Thermodynamics of restricted Boltzmann machines and related learning dynamics. Journal of Statistical Physics 172(6):1576–1608.
  6. Ghahramani, Z., and Jordan, M. I. 1994. Supervised learning from incomplete data via an em approach. In Advances in Neural Information Processing Systems 6, 120–127. Morgan Kaufmann.
  7. G.Madjarov; Kocev, D.; and Gjorgjevikj, D. 2012. An extensive experimental comparison of methods for multi-label learning. Pattern Recognition 3083–3104.
  8. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; and Bengio, Y. 2014. Generative adversarial nets. In Ghahramani, Z.; Welling, M.; Cortes, C.; Lawrence, N. D.; and Weinberger, K. Q., eds., Advances in Neural Information Processing Systems 27. Curran Associates, Inc. 2672–2680.
  9. Han, Y.; Sun, G.; Shen, Y.; and Zhang, X. 2018. Multi-label learning with highly incomplete data via collaborative embedding. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’18, 1494–1503.
  10. Hinton, G. E. 2002. Training products of experts by minimizing contrastive divergence. Neural computation 14:1771–1800.
  11. Hinton, G. 2010. A practical guide to training restricted Boltzmann machines. Momentum.
  12. Kingma, D. P., and Welling, M. 2014. Auto-encoding variational bayes. In International Conference on Learning Representations (ICLR), 4413–4423.
  13. Li, Lin, Y. W., and Ye, M. 2015. Experimental comparisons of multi-class classifiers. Informatica.
  14. Li, S. C.-X.; Jiang, B.; and Marlin, B. 2019. Learning from incomplete data with generative adversarial networks. In International Conference on Learning Representations.
  15. Marylou, G.; Tramel, E.; and Krzakala, F. 2015. Training restricted Boltzmann machines via the Thouless-Anderson-Palmer free energy. In Proceedings of the 28th International Conference on Neural Information Processing Systems, NIPS’15, 640–648.
  16. Mattei, P.-A., and Frellsen, J. 2019. MIWAE: Deep generative modelling and imputation of incomplete data sets. In Chaudhuri, K., and Salakhutdinov, R., eds., Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, 4413–4423. Long Beach, California, USA: PMLR.
  17. Mézard, M. 2017. Mean-field message-passing equations in the Hopfield model and its generalizations. Phys. Rev. E 95:022117.
  18. Nijman, M. J., and Kappen. 1994. Using boltzmann machines to fill in missing values.
  19. Sun, Y.; Y.Zhang; and Zhou, Z. 2010. Multi-label learning with weak label. In Proceedings of the 24th AAAI Conferences on Artificial Intelligence, AAAI-10, 593–598.
  20. Tieleman, T. 2008. Training restricted Boltzmann machines using approximations to the likelihood gradient. In Proceedings of the 25th International Conference on Machine Learning, ICML ’08, 1064–1071.
  21. Yoon, J.; Jordon, J.; and van der Schaar, M. 2018. GAIN: Missing data imputation using generative adversarial nets. In Dy, J., and Krause, A., eds., Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, 5689–5698. Stockholmsmässan, Stockholm Sweden: PMLR.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
""
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
   
Add comment
Cancel
Loading ...
402583
This is a comment super asjknd jkasnjk adsnkj
Upvote
Downvote
""
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters
Submit
Cancel

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test
Test description