Strong Consistency of Reduced means Clustering
Abstract
Reduced means clustering is a method for clustering objects in a lowdimensional subspace. The advantage of this method is that both clustering of objects and lowdimensional subspace reflecting the cluster structure are simultaneously obtained. In this paper, the relationship between conventional means clustering and reduced means clustering is discussed. Conditions ensuring almost sure convergence of the estimator of reduced means clustering as unboundedly increasing sample size have been presented. The results for a more general model considering conventional means clustering and reduced means clustering are provided in this paper. Moreover, a new criterion and its consistent estimator are proposed to determine the optimal dimension number of a subspace, given the number of clusters.
Consistency of RKM Clustering
clustering \kwddimension reduction \kwdmeans
1 Introduction
The aim of cluster analysis is the discovery of a finite number of homogeneous classes from data. In some cases, a cluster structure is considered to lie in a lowdimensional subspace of data, and the following procedure is applied:
 Step .

Principal component analysis (PCA) is performed, and the first few components are obtained.
 Step .

Conventional means clustering is performed for the principal scores on the first few principal components.
This twostep procedure is called “tandem clustering” by Arabie & Hubert (1994) and has been discouraged by several authors (e.g., Arabie & Hubert, 1994; Chang, 1983; De Soete & Carroll, 1994). Because the first few principal components of PCA do not necessarily reflect the cluster structure in data, the appropriate clustering result may not be obtained by using the tandem clustering approach. Figure 1 shows that the first two principal components do not reflect the cluster structure, and the clustering result of the tandem clustering is incorrect.
De Soete & Carroll (1994) proposed reduced means (RKM) clustering. RKM clustering simultaneously determines the clusters of objects on the basis of the means criterion and the subspace that is informative about the cluster structure in data on the basis of component analysis. In other words, for given data points in , the fixed cluster number and the dimension number of subspace , RKM clustering is defined by the minimization problem of the following loss function:
(1) 
where and is a columnwise orthonormal matrix. For some clustering methods related to means clustering, several authors have discussed their statistical properties (e.g., Abraham et al., 2003; GarcíaEscudero et al., 1999; Pollard, 1981; Pollard, 1982; von Luxburg et al., 2008). However, because RKM clustering is proposed in the framework of descriptive statistics, the statistical properties are not discussed. When data points are independently drawn from a population distribution , the objective function is rewritten as
where is a set containing or fewer points in , and is the empirical measure obtained from the data. For each fixed and , the strong law of large numbers (SLLN) shows that
Thus, we wish to ensure that the global minimizer of converges almost surely to the global minimizers of , say the population global minimizers.
In this paper, the strong consistency of RKM under i.i.d. sampling is proven. For this purpose, the framework of the proof of the strong consistency of the means clustering approach proposed by Pollard (1981) is used; in this framework, the existence and uniqueness of the population global minimizers are assumed for consistency. Conditions for the existence of the global minimizers are not discussed. For RKM clustering, the uniqueness of the population global minimizers cannot be assumed because RKM clustering has rotational indeterminacy. Therefore, the sufficient condition for the existence of the population global minimizers must be derived; it is also necessary to establish that the distance between the sample estimator and the set of global minimizers converges almost surely to zero, as the sample size approaches infinity.
This paper is organized as follows. In Section 2, the original algorithm of RKM clustering and visualization of the result are described. Then, the relationship between the conventional means clustering method and RKM clustering is presented. The notation and some properties of RKM, including the rotational indeterminacy, is introduced in Section 3. The uniform SLLN and continuity of the objective function of RKM clustering are presented in Section 4. In Section 5, conditions for the existence of the population global minimizers are determined, and a theorem regarding the strong consistency of RKM clustering is stated. In Section 6, the main proof of the consistency theorem is explained. In Section 7, a new criterion and its consistent estimator are proposed to determine the optimal dimension number of a subspace, given the number of clusters. Moreover, the effectiveness of the criterion through numerical experiments are illustrated.
2 Reduced means clustering
2.1 Algorithm and visualization of reduced means clustering
Let be a data matrix and be row vectors of , where is the number of objects and is the number of variables. The number of clusters and components to which the variables are reduced are denoted by and , respectively. RKM clustering is defined as the minimizing problem of the following criterion:
(2) 
where and denote the usual Euclidean norm and Frobenius norm, respectively, is a binary membership matrix that specifies cluster membership for each objects, is a columnwise orthonormal loading matrix, is a centroid matrix, and is a centroid of the th cluster for each . For example, this problem can be solved by the following alternating least square algorithm:
 Step .

First, initial values are chosen for and .
 Step .

is expressed as the singular value decomposition of , where is a orthonormal matrix, is a diagonal matrix, and is a columnwise orthonormal matrix. is updated by .
 Step .

For each and each , we update by
 Step .

is updated using .
 Step .

Finally, the value of the function for the present values of , and is computed. When the present values have decreased the function value, , and are update in accordance with Steps –. Otherwise, the algorithm has converged.
Other formulations and algorithms for RKM clustering have been presented by De Soete & Carrol (1994) and Timmerman et al. (2010).
The algorithms for RKM clustering monotonically decrease the function . As shown below, because is bounded, the solution for each iteration converges to a local minimum point. Because of the binary constraint on , the solutions of these algorithms may often be local minimums. To prevent this, many random starts are required to be used.
The objective function can be decomposed into two terms:
(3) 
The first term of equation is the objective function of the PCA, and the second term is the means criterion in a low dimensional subspace. Thus, for optimal solutions , and , we have . Using the optimal solutions , , and , the lowdimensional representation of the objects and cluster centers can be obtained:
(4) 
Using and , a biplot reflecting the cluster structure can be presented. Figure 2 shows the biplot of the RKM clustering for the same data as that used in Figure 1.
2.2 The relationship between the conventional means and the RKM clusterings
The objective function of the conventional means clustering method is given by
(5) 
where is an cluster center matrix. is expressed as the singular value decomposition of , where is an orthonormal matrix, is an diagonal matrix, and is a columnwise orthonormal matrix. Function can be expressed as
Considering and as a lowdimensional centroid matrix and a loading matrix , respectively, function is equivalent to the objective function of RKM, . Thus, RKM clustering includes the conventional means clustering analysis as a special case.
3 Preliminaries
Let be a probability space and be independent random variables with a common population distribution on ; let be the empirical measure based on . For typographical convenience, the set of all columnwise orthonormal matrices are denoted by , and , where is the cardinality of . Thus, the parameter space is denoted by . denotes the dimensional closed ball of radius centered at the origin. For each , define and . Let be a nonnegative decreasing function and be a probability measure on . For each finite subset and each , the loss function of RKM with is defined by
Write
For , both descriptions and are used. In addition, and . For each , and . The parameters and are used to emphasize that and are dependent on the index . One of the measurable estimators in will be denoted by or . Similarly, we will also denote one of the measurable estimators in by or . To illustrate the existence of measurable estimators, see Section 6.7 of Pfanzagl (1996).
Let be the distance between two matrices based on Frobenius norm and the Hausdorff distance, which is defined for finite subsets as
Moreover, let be the product distance with and . In this paper, the distance between and is defined as
To clarify the minimization procedures, the function must satisfy some regularity conditions. As proposed by Pollard (1981), it is assumed that is continuous, and . Moreover, to control the growth of , it is assumed that
For each and each ,
Therefore, as long as is finite, is also finite for each and each .
Let be a orthonormal matrix, i.e., . For each and each ,
It follows that is not a singleton when , thus suggesting that RKM clustering has rotational indeterminacy.
4 The uniform SLLN and the continuity of
Proposition 1.
Let be an arbitrary number. Let denote the class of all integrable functions on of the form
where takes all values over . Suppose that . Then,
(6) 
Proof.
DeHardt (1971) provided the sufficient condition for the uniform SLLN ; for all , there exists a finite class of functions such that for each , and exist in with and .
An arbitrary is selected, and denotes the surface of the sphere on of radius centered at the origin. To find such a finite class , is defined as the finite set of satisfying
and as the finite sets of satisfying
Define . Take as the finite class of functions of the form
where takes all values over and is defined as zero for all negative .
For given and , there exists with for each and each with . Corresponding to each , choose
and
Because is a monotone function and
for each and each , these functions ensure that .
If we choose to be greater than ,
The second term would be less than if is sufficiently large. Moreover, because is uniform continuous on a bounded set, the first term can be less than if is sufficiently small. Thus, the uniform SLLN is proven. ∎
Similarly, the continuity of on can be proven.
Proposition 2.
Let be an arbitrary number. Suppose that . Then, is continuous on .
Proof.
If are select such that and , then for each , there exists with , and furthermore,
(7) 
for . When a sufficiently large and a sufficiently small are selected, the last bound is less than . For each , there also exists with . Therefore, the other inequality necessary for the continuity is obtained by interchanging and in the inequality . ∎
5 The consistency theorem
5.1 The existence of the population global optimizers
The aim of this paper is to prove that, for a fixed measure satisfying some natural assumptions, the infimum distance between the (measurable) estimator with and parameters achieving converges almost surely to , as the sample size goes to infinity. However, there may be no such parameters. Thus, before providing the consistency theorem, the sufficient condition for the existence of parameters achieving in is provided. The following proposition ensures the existence of such parameters. The proof and some details about the proposition are given in Appendix A.
Proposition 3.
Suppose that and that for . Then, .
5.2 Strong consistency of reduced means clusterings
If the parameter space is , the strong consistency of RKM clustering can be proven. Note that since is compact, we have and the identification condition:
where .
Proposition 4.
Suppose that . Then, for each ,
Proof.
Since the uniform SLLN and the continuity of , the proof of this proposition is given by the similar argument of the proof of the following consistency theorem. ∎
In a study by Pollard (1981), the uniqueness of the parameter is also assumed for the strong consistency theorem. As discussed in Section 3, we cannot assume the uniqueness condition. Thus, the condition that for is assumed instead of the uniqueness condition.
This condition is equivalent to the distinctness condition that has distinct points for all . Indeed, suppose that there exists such that have or fewer distinct points; that is, . There exists such that and . Then, , which contradicts to . Thus, the condition that for implies the distinctness condition. Moreover, this condition is equivalent to since for each satisfying .
The following main theorem gives the sufficient condition for the strong consistency of the estimator of RKM clustering.
Theorem 1.
Suppose that and that for . Then, ,
6 Proof of Theorem 1
Because almost sure convergence is dealt with, null sets of elements exists for which the convergence does not hold. Hereafter, denotes the set obtained by avoiding a proper null set from . In the first step of the proof, when is sufficiently large, the estimators of the cluster centers are contained within a compact ball that does not depend on . For convenience, it is assumed that as . When is bounded, the proof is a little complicated.
First, we prove the following lemma.
Lemma 1.
Suppose that . Then, there exists such that
Proof.
Select an appropriate value to satisfy the condition that the ball has positive measure, i.e., . Let be sufficiently large for satisfying and
(8) 
From the definition of , for any set containing at most points and any . The parameter is chosen such that it only consists of the origin. Then, by SLLN,
for each .
Let . By the axiom of choice, for an arbitrary there exists a subsequence such that and . Thus,
On the other hand, because . Therefore, we have and , which is a contradiction. Therefore, , that is,
∎
Without loss of generality, all can be assumed contain at least one point of when is sufficiently large. The next lemma shows that for sufficiently large , there exists such that the closed ball contains all estimators of centers. When , the next lemma is obviously satisfied.
From the results in Section 4 and using the same arguments in the final part of this section, the conclusions of the theorem are proven when .
Lemma 2.
Under the assumption of the theorem, there exists such that
Proof.
Choose sufficiently large to satisfy the inequality and
(9) 
where is selected to ensure . Note that for .
Suppose that contains at least one center outside and consider the effect on by deleting such outside centers from for all . From Lemma 1, all contain at least one center on when is sufficiently large, say . In the worst case, the cluster of should contain all sample points belonging to clusters outside . Because these points must be outside , the increment of due to the deletion of centers outside from would be at most
Denote the set obtained by deleting centers outside from by . For each , is contained in , and thus,
For each satisfying and each , we have
and
Thus,
for all . Note that
by Proposition 4.
Let . By the axiom of choice, for an arbitrary there exists a subsequence such that and . For any with or fewer points and any ,