# Noise Tolerance Under Risk Minimization

###### Abstract

In this paper we explore noise tolerant learning of classifiers. We formulate the problem as follows. We assume that there is an training set which is noise-free. The actual training set given to the learning algorithm is obtained from this ideal data set by corrupting the class label of each example. The probability that the class label of an example is corrupted is a function of the feature vector of the example. This would account for most kinds of noisy data one encounters in practice. We say that a learning method is noise tolerant if the classifiers learnt with noise-free data and with noisy data, both have the same classification accuracy on the noise-free data. In this paper we analyze the noise tolerance properties of risk minimization (under different loss functions). We show that risk minimization under 0-1 loss function has impressive noise tolerance properties and that under squared error loss is tolerant only to uniform noise; risk minimization under other loss functions is not noise tolerant. We conclude the paper with some discussion on implications of these theoretical results.

## I Introduction

In most situations of learning a classifier, one has to contend with noisy examples. Essentially, when training examples are noisy, the class labels of examples as provided in the training set may not be ‘correct’. Such noise can come through many sources. If the class conditional densities overlap, then same feature vector can come from different classes with different probabilities and this can be one source of noise. In addition, in many applications (e.g, document classification etc.), training examples are obtained through manual classification and there will be inevitable human errors and biases. Noise in training data can come about by errors in feature measurements also. Errors in feature values would imply that the observed feature vector is at a different point in the feature space though the label remains the same and hence it can also be looked at as a noise corruption of the class label. It is always desirable to have classifier design strategies that are robust to noise in training data.

A popular methodology in classifier learning is (empirical) risk minimization [1, 2]. Here, one chooses a convenient loss function and the goal of learning is to find a classifier that minimizes risk which is expectation of the loss. The expectation is with respect to the underlying probability distribution over the feature space. In case of noisy samples, this expectation would include averaging with respect to noise also.

In this paper, we study noise tolerance properties of risk minimization under different loss functions such as 0-1 loss, squared error loss, exponential loss, hinge loss etc. We consider what we call non-uniform noise where the probability of the class label for an example being incorrect, is a function of the feature vector of the example. This is a very general noise model and can account for most cases of noisy datasets. We say that risk minimization (with a loss function) is noise tolerant if the minimizers under the noise-free and noisy cases have the same probability of mis-classification on noise-free datasets. We present some analysis to characterize noise tolerance of risk minimization with different loss functions.

As we show here, the 0-1 loss function has very good noise tolerance properties, In general, risk minimization under 0-1 loss is desirable because it achieves least probability of mis-classification. However, the optimization problem here is computationally hard. To overcome this, many of the classifier learning strategies use some convex surrogates of the 0-1 loss function (e.g., hinge loss, squared error loss etc.). The convexity of the resulting optimization problems makes these approaches computationally efficient. There have been statistical analyses of such methods so that one can bound risk under 0-1 loss, of the classifier obtained as a minimizer of risk under some other convex loss [3]. The analysis we present here is completely different because the objective is to understand noise tolerance properties of risk minimization. Here we are interested in comparing minimizers of risk under the same loss function but under different noise conditions.

The rest of the paper is organized as follows. In section 2 we discuss the concept of noise tolerant learning of classifiers. In section 3, we present our results regarding noise tolerance of different loss functions. We present a few simulation results to support our analysis in section 4 and conclude in section 5.

## Ii Noise Tolerant Learning

When we have to learn with noisy data where class labels may be corrupted, we want approaches that are robust to label noise. Most of the standard classifiers (e.g. support vector machine, adaboost etc.) perform well only under noise-free training data; when there is label noise, they tend to over-fit.

There are many approaches to tackle label noise in training data. Outliers detection [4], restoration of clean labels for noisy points [5] and restricting the effects of noisy points on the classifier [6, 7] are some of the well known tricks to tackle the label noise. However all these are mostly heuristic and also need extra computation. Many of them also assume uniform noise and sometimes assume knowledge of noise variance.

A different approach would be to look for methods that are inherently noise tolerant. That is, the algorithm will handle the noisy data the same way that it would handle noise-free data. However due to some property of the algorithm, its output would be same whether the input is noise free or noisy data.

Noise tolerant learning using statistical queries [8] is one such approach. The algorithm learns by using some statistical quantities computed from the examples. That is the reason for its noise tolerance properties. However, the approach is mostly limited to binary features. Also, the appropriate statistical quantities to be computed depends on the type of noise and the type of classifier being learned.

In this paper, we investigate the noise tolerance properties of the general risk minimization strategy. We formulate our concept of noise tolerance as explained below. For simplicity, we consider only the two class classification problem.

We assume that there exists an ideal noise-free sample which is unobservable but where the class label given to each example is correct. We represent this ideal sample by , where

The actual training data given to the learning algorithms is obtained by corrupting these (ideal) noise-free examples by changing the class label on each example. The actual training data set would be , where with probability and is with probability . If , then we say that the noise is uniform. Otherwise, we say noise is non-uniform.

We note here that under non-uniform classification noise, the probability of the class label being wrong can be different for different examples. We assume, throughout this paper, that , , which is reasonable.

As a notation, we assume that the risk is defined over class of functions, , that map feature space to real numbers. This allows us to treat all loss functions through a single notation. We call any such a classifier and the class label assigned by it to a feature vector would be .

Let be a specific loss function. For any classifier , the risk under no-noise case is

where is the loss function. The expectation here is with respect to the underlying distribution of the feature vector . Let be the minimizer of .

Under the noisy case, the risk of any classifier is,

Note that has additional randomness due to noise corruption of labels and the expectation includes averaging with respect to that also. To emphasize this, we use the notation to denote risk under noisy case. Let be the minimizer of .

Definition 1: Risk minimization under loss function , is said to be noise-tolerant if , where the probability is w.r.t. the underlying distribution of .

That is, the general learning strategy of risk minimization under a given loss function, is said to be noise-tolerant if the classifier it would learn with the noisy training data has the same probability of misclassification as that of the classifier the algorithm would learn if it is given ideal or noise-free class labels for all training data. Noise tolerance can be achieved even when because we are only comparing the probability of mis-classification of and . However, is a sufficient condition for noise tolerance.

Thinking of an ideal noise-free sample allows us to properly formulate the noise-tolerance property as above. We note once again that this noise-free sample is assumed to be unobservable. Making the probability of label corruption, , to be a function of would take care of most cases of noisy data. For example, consider a 2-class problem with overlapping class conditional densities where the training data are generated by sampling from the respective class conditional densities. Then we can think of the unobservable noise-free dataset to be the one obtained by classifying the examples using Bayes optimal classifier. The labels given in the actual training dataset would not agree with the ideal labels (because of overlapping class conditional densities); however, the observed labels are easily seen to be noisy versions where the noise probability is a function of the feature vector. If there are any further sources of noise in generating the dataset given to the algorithm, these can also be easily accounted for by because the probability of wrong label for different examples can be different.

## Iii Noise Tolerance of Risk Minimization

In this section, we analyze noise tolerance property of risk minimization with respect to different loss functions.

### Iii-a 0-1 Loss Function

The 0-1 loss function is,

where denotes indicator of event .

###### Theorem 1

Assume . Then, (i). Risk minimization with 0-1 loss function is noise tolerant under uniform noise. (ii). In case of non-uniform noise, risk minimization with 0-1 loss function is noise tolerant if .

Proof: For any , let and . The risk for a function under no-noise case is

where denotes that the above integral is an expectation integral with respect to the distribution of feature vectors. Recall is the minimizer of . Let . Then risk for any in presence of noise would be

(1) | |||||

Given any , using (1) we have

(2) |

For the first part of the theorem, we consider uniform noise and hence . From Eq.(1), we now get, for any , . Hence we have, . Since is minimizer of , we have , which implies if . Thus under uniform label noise, also minimizes . This completes proof of first part of the theorem. (The fact that risk minimization under 0–1 loss function is tolerant to uniform noise is known earlier (see, e.g. [9, chap 4]))

For the second part of the theorem, is no longer constant but we assume . This implies and hence from Eq.(2), we get if . Thus, , which is minimizer of , also minimizes . This shows that risk minimization with 0-1 loss function is noise-tolerant under non-uniform noise if and completes proof of the theorem.

If , then risk minimization is, in general, not tolerant to non-uniform noise as we show by a counter example.

Example 1: Fig. 1 shows a binary classification problem where examples are generated using the
true classifier .
Let the probability distribution on the feature space be uniformly concentrated on the training dataset.
We note here that we get perfect classification if we consider quadratic classifiers.
Since we want to consider the case where , we restrict the family of classifiers over which risk is minimized to linear classifiers.

(a) Without Noise: The linear classifier which minimizes is
and .

(b) With Noise: We now introduce non-uniform label noise in the data with the noise rates as follows: , , ,
and any noise rate (less than 0.5) to rest of the points. Consider another linear classifier
. From Fig. 1,
we see that . Now using Eq.(2), we get

That is, , although .

This example proves that risk minimization with 0-1 loss is, in general, not noise tolerant if .

Remark 1: We note here that the assumption may not be very restrictive; mainly because the noise free ideal data set is only a mathematical entity and need not be observable. For example, we can take to be the Bayes optimal classifier and assume that the ideal data set is obtained by classifying samples using the Bayes optimal classifier. Then . This means that if we minimize risk under 0-1 loss function with the actual training set, then the minimizer would be .

Finally, we note that all the above analysis is applicable to empirical risk minimization also by simply taking (in Eq.(III-A) and (2)) to be the empirical distribution.

While, as shown here, 0-1 loss function has impressive noise tolerant properties, risk minimization with this loss is difficult because it is a non-convex optimization problem. In machine learning, many other loss functions are used to make risk minimization computationally efficient. We will now examine the noise tolerance properties of other loss functions.

### Iii-B Squared Error Loss Function

Squared error loss function is given by,

We first consider the case when the function is an affine function of . Let , where and .

###### Theorem 2

Risk minimization with squared error loss function for finding linear classifiers is noise tolerant under uniform noise if .

Proof: For noise-free case, the risk is, , whose minimizer is . Risk under uniform label noise () is given as

which is minimized by

Since we assume , we have . Hence we get, . Which means . Thus under uniform noise, least square approach to learn linear classifiers is noise tolerant and the proof of theorem is complete.

###### Corollary 1

Fisher Linear Discriminant (FLD) is noise tolerant under uniform label noise.

Proof: For binary classification, FLD [2] finds direction as, , which is proportional to . Here and . , represent the two classes and , denote corresponding means. FLD can be obtained as the risk minimizer under squared error loss function [2, chap 4] by choosing the target values as: and , where , and .

When the training set is corrupted with uniform label noise, let and be the two sets now and and corresponding means. Let and . New target values are: and . The empirical risk in this case is, . Equating the derivative of with respect to to zero, we get, , where is the mean of training set. Setting the gradient of with respect to to zero and using the values of and , we get,

where we have used the fact that, and . Note that for any . Thus we see that, . Thus FLD is noise tolerant under uniform label noise.

Remark 2: What we have shown is that risk minimization under squared error loss function is tolerant to uniform noise if we are learning linear classifiers. We can, in general, nonlinearly map feature vectors to a higher dimensional space so that the training set becomes linearly separable. Since uniform label noise in the original feature space should become uniform label noise in the transformed feature space, we feel that Theorem 2 should be true for risk minimization under squared error loss for any family of classifiers.

Now consider the non-uniform noise case where is not same for all . Then, the risk is minimized by, . Here, term can no longer be taken out of expectation. Hence, we may not get noise tolerance. We show that it is so by a counter example as below.

Example 2: Consider the unit circle centered at origin in and data points placed on its circumference as, . and .
Assume that the probability distribution on the feature space is uniformly concentrated on the training dataset. Let the set of classifiers contain only linear classifiers passing through origin.

(a) Without Noise: In this case, risk is minimized by . Classifier, , linearly separates the two classes. Thus .

(b) With Noise: Now let us introduce non-uniform label noise as follows.
, where and ;
for rest of the points.
In this case, risk is minimized by .
mis-classifies and
as shown in Fig. 2.
Hence .

Thus squared error loss is not noise tolerant under non-uniform noise even if the minimum risk under
noise-free case is zero and the optimal classifier is linear in parameters.

Remark 3: An interesting special case of non-uniform noise is class conditional classification noise (CCCN) [10],
where for .
Least squares method may not be tolerant to such a non-uniform noise. However, using the proof of Corollary 1, it is easy to verify
that FLD is noise tolerant under CCCN.

### Iii-C Exponential Loss Function

Exponential loss function is given by,

This is the effective loss function for adaboost. We show, through the following counter example, that exponential loss function is not tolerant to even uniform noise.

Example 3:
Let be the training dataset such that
, and , with , and .
Let the probability distribution on the feature space be uniformly concentrated on the training dataset.
Here, we find a linear classifier which minimizes the risk under exponential loss function. We consider linear classifiers
expressed as .

(a) Without Noise: The risk of a linear
classifier without label noise is written as:

By equating the derivative of to zero, we get,

correctly classifies all the points.
Thus .

(b) With Noise: Now let us introduce uniform label noise with noise rate .
The risk will be

Again equating the derivative of to zero, we get,

mis-classifies . Thus . Thus risk minimization under exponential loss is not noise tolerant even with uniform noise.

### Iii-D Log Loss Function

Log loss function is given by,

This is the effective loss function for logistic regression. Risk minimization with log loss function also is not noise tolerant. We demonstrate it using following counter example.

Example 4:
Consider the same training dataset as in Example 3.
We need to find a linear classifier, , which minimizes the risk under log loss function.

(a) Without Noise:
The risk of a linear classifier without label noise is

Equating the derivative of to zero, we get,

where . Roots of this polynomial are , and .
The only positive root is . Using this value of , we get .
classifies all the points correctly.
Thus .

(b) With Noise: Now let us introduce uniform label noise with noise rate .
The risk will be,

Equating the derivative of to zero, we get a sixth degree polynomial in which has only one positive root. This root gives us the value of . The classifier, mis-classifies . Which means .

Thus, and log loss is not noise tolerant even with uniform noise.

### Iii-E Hinge Loss Function

This is a convex loss function and has the following form.

Support vector machine is based on minimizing risk under the hinge loss. Here we show that hinge loss function is not noise tolerant using a counter example.

Example: 5 Consider the same training dataset as in Example 3.
Here we consider learning linear classifiers expressed as .

(a) Without Noise: The risk of a linear classifier with noise-free training data is

To find the minimizer of , we need to solve

The optimal solution of the above linear program is
which is also the minimizer of .
classifies all the points correctly.
Thus .

(b) With Noise:
Now we introduce uniform label noise with noise rate in the training data.
The risk of a linear classifier in presence of uniform label noise is

Minimizing of by solving the equivalent linear program as earlier, we get . The classifier mis-classifies . Thus . Thus hinge loss is not noise tolerant even under uniform noise even when the optimal classifier is linear.

## Iv Some Empirical Results

In this section, we present some empirical evidence for our theoretical results. The main difficulty in doing such simulations is that there is no general purpose algorithm for risk minimization under 0–1 loss. Here we use the CALA-team algorithm proposed in [11] which (under sufficiently small learning step-size) converges to minimizer of risk under 0–1 loss in case of linear classifiers. Hence, here we restrict the simulations only to learning of linear classifiers and hence give experimental results on Iris dataset.

Iris recognition is a three class classification problem in 4-dimensions. The first class, Iris-setosa, is linearly separable from the other two classes, namely, Iris-versicolor and Iris-virginica. We consider a linearly separable 2-class problem by combining the latter two classes as one class.

The original Iris data set has no label noise. We introduce different rates of uniform noise varying from 10% to 30%. We incorporated non-uniform label noise as follows. For every example, the probability of flipping the label is based on which quadrant (with respect to the first two features) the example falls in. The noise rate in this case is represented by a quadruple with -th element representing probability of wrong class label if the feature vector is in quadrant ().

For training, we use entire dataset with label noise inserted to it. We use the original noise-free examples for testing. We use test error rate as an indicator of the noise-tolerance. We compare CALA algorithm for risk minimization under 0–1 loss with SVM (hinge loss), linear least square (squared error loss), and logistic regression (log loss) which are risk minimization algorithms under different convex loss functions.

The results are shown in Table I. For each noise rate, we generated ten random noisy training data sets. We show the mean and standard deviation of accuracy on test set with each of the algorithms. (The CALA algorithm [11] is a stochastic one and hence has a non-zero standard deviation even in the case of no-noise data). As can be seen from the table, risk minimization under 0–1 loss has impressive noise tolerance under both uniform and non-uniform label noise. Both SVM and logistic regression have the highest accuracy under no-noise; but their accuracy drops from 98% to 89% and 91% respectively under uniform noise rate of 20%. Linear least squares algorithm achieves accuracy of 92% when there is no noise and it drops to only 91% when 20% uniform noise is added, showing that it is tolerant to uniform noise. (The performance of Fisher linear discriminant is similar to that of linear least squares: it achieves accuracy of 94%, 92.20%2.49, 91.07%2.88, 90.27%2.16 respectively on 0%, 10%, 20% and 30% uniform noise and 91.53%1.72 and 87.67%2.71 on the two cases of non-uniform noise). Also, the large standard deviations of SVM and other algorithms in the non-uniform noise case show their sensitivity to noise.

0–1 loss | hinge loss | sq. err. loss | log loss | |

Noise Rate | (CALA) | (SVM,C=) | (Least Sq.) | (LogReg) |

No Noise | 97.530.38 | 98.67 | 92.67 | 98.67 |

Uniform 10% | 97.470.98 | 93.402.92 | 92.531.33 | 92.871.47 |

Uniform 20% | 97.071.09 | 89.474.02 | 91.471.17 | 91.671.87 |

Uniform 30% | 97.071.05 | 83.736.79 | 90.131.77 | 90.071.99 |

Non-Uniform 15,20,25,30% | 96.471.49 | 89.673.18 | 91.271.49 | 91.672.07 |

Non-Uniform 30,25,20,15% | 97.001.01 | 82.477.04 | 85.805.07 | 85.935.09 |

## V Conclusion

While learning a classifier, one has to often contend with noisy training data. In this paper, we presented some analysis to bring out the inherent noise tolerant properties of the risk minimization strategy under different loss functions.

Of all the loss functions, the 0-1 loss function has best noise tolerant properties. We showed that it is noise tolerant under uniform noise and also under non-uniform noise if the risk minimizer achieves zero risk on uncorrupted or noise-free data.

If we consider the case where we think of our ideal noise-free sample as the one obtained by classifying iid feature vectors using Bayes optimal classifier, the minimum risk achieved would be zero if the family of classifiers over which the risk is minimized includes the structure of Bayes classifier. In such a case, the noise-tolerance (under non-uniform label noise) of risk minimization implies that if we find the classifier to minimize risk under 0-1 loss function (treating the labels given in our training data as correct), we would (in a probabilistic sense) automatically learn the Bayes optimal classifier. This is an interesting result that makes risk minimization under 0-1 loss a very attractive classifier learning strategy.

A problem with minimizing risk under 0-1 loss function is that it is difficult to use any standard optimization technique to minimize risk due to discontinuity of loss function. Hence, given the noise-tolerance properties presented here, an interesting problem to address is that of some gradient-free optimization techniques to minimize risk under 0-1 loss function. For the linear classifier case, the stochastic optimization algorithm proposed in [11] is one such algorithm. To really exploit the noise-tolerant property of the 0-1 loss function we need more efficient techniques of that kind and also techniques which work for nonlinear classifiers.

On the other hand, risk under convex loss functions is easy to optimize. Many generic classifiers are based on minimizing risk under these convex loss function. But it is observed in practice that in presence of noise, these approaches over-fit.

In this paper, we showed that these convex loss functions are not noise tolerant. Risk minimization under hinge loss, exponential loss and log loss is not noise tolerant even under uniform label noise. This explains the problem one faces with algorithms such as SVM if the class labels given are sometimes incorrect. We also showed that the linear least squares approach is noise tolerant under uniform noise but not under non-uniform noise. Same is shown to be true for Fisher linear discriminant.

Most algorithms for learning classifiers focus on minimizing risk under a convex loss function to make the optimization more tractable. The analysis presented in this paper suggests that looking for techniques to minimize risk under 0-1 loss function may be a promising approach for classifier design especially when we have to learn from noisy training data.

## References

- [1] R. O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification (2nd Edition). Wiley-Interscience, 2000.
- [2] C. M. Bishop, Pattern Recognition and Machine Learning (Information Science and Statistics). Springer-Verlag New York, Inc., 2006.
- [3] P. L. Bartlett and M. I. Jordan, “Convexity, classification, and risk bounds,” Journal of The American Statistical Association, vol. 101, no. 473, pp. 138–156, 2006.
- [4] C. E. Brodley and M. A. Friedl, “Identifying mislabeled training data,” Journal of Artificial Intelligence Research, vol. 11, pp. 131–167, August 1999.
- [5] S. Fine, R. Gilad-bachrach, S. Mendelson, and N. Tishby, “Noise tolerant learnability via the dual learning problem,” in NSCT, June 1999.
- [6] L. Xu, K. Crammer, and D. Schuurmans, “Robust support vector machine training via convex outlier ablation,” in Proceedings of the 21st National Conference on Artificial Intelligence (AAAI-06), (Boston, Massachusetts), July 2006.
- [7] Q. Tao, G. Wu, F. Wang, and J. Wang, “Posterior probability support vector machines for unbalanced data,” IEEE Transactions in Neural Networks, vol. 16, pp. 1561–1573, November 2005.
- [8] M. Kearns, “Efficient noise-tolerant learning from statistical queries,” Journal of the ACM, vol. 45, pp. 983–1006, November 1998.
- [9] M. Thathachar and P. Sastry, Networks of Learning Automata: Techniques for Online Stochastic Optimization. Boston, MA: Kluwar, 2004.
- [10] G. Stempfel, L. Ralaivola, and F. Denis, “Learning from noisy data using hyperplane sampling and sample averages,” Technical Report 3564, HAL - CNRS, France, 2007.
- [11] P. S. Sastry, G. D. Nagendra, and N. Manwani, “A continuous-action learning automata team for noise tolerant learning of half spaces,” IEEE Transaction on System, Man and Cybernetics Part B: Cybernetics, vol. 40, pp. 19–28, February 2010.