Values of quasi-clustering criteria.
Two clustering problems are considered. We consider a lot of different clusters of the same data for a given number of clusters. Data clustering is understood as their stable partition into a given number of sets. Clustering is considered stable if the corresponding partitioning remains unchanged with its minimum change. How to create a new cluster based on ensemble clusterings? The second problem is the following. A definition of the committee synthesis as ensemble clustering is introduced. The sets of best and worst matrices of estimates are considered. Optimum clustering is built on the basis of the clusterings obtained as being closest to the set of the best estimation matrices or as the most distant from the set of worst-case matrices of estimates. As a result, the problem of finding the best committee clustering is formulated as a discrete optimization problem on permutations.
There are many different approaches to solving the problems of clustering multidimensional data: based on the optimization of internal criteria (indices) [1, 2], hierarchical clustering , centroid-based clustering , density-based clustering , distribution-based clustering , and many others. There are well-known books and papers on clustering [7, 8, 9, 10].
This section is devoted to one approach to the creation of stable clusterings and the processing of their sets. A natural criterion is considered, which is applicable to any clustering method. In work , various criteria (indices) are proposed, optimizing which clustering is built with a definite look “what is clustering?” In this chapter, we use a criterion based on stability. If we really got clustering, that is, a solution for the whole sample, the partitioning should not change with a small change in the data. Criteria are introduced for the quality of the partition obtained. If the criterion value is less than one, then the partition is unstable. Let us obtain for the same data clusterings. How to create a new ensemble clustering based on the partitions? Previously, a committee method for building ensemble clusterings was proposed [12, 13, 14, 15]. Let there be results of cluster analysis of the same data for clusters. The committee method of building ensemble clustering makes it possible to build such clusters, each of which is the intersection of “many” initial clusters. In other words, we find such clusters whose objects are “equivalent” to each other according to several principles. As initial clusterings, one can take stable ones. Finally, we consider a video-logical approach to building the initial coarse clusterings.
2. Criteria for stability of clustering
Let the sample of objects be given and is the clustering of the sample into clusters obtained by some method, Speaking of clustering, we mean applying a method to a sample without focusing on the method itself. Is partition of a sample by this method clustering or here some kind of stopping criterion is satisfied? For example, an extremum of some functional is obtained or the maximum number of operations in the iterative process is fulfilled. We will use the following thesis as the main one. If the resulting partition is indeed clustering, then it must be the same clustering for any minimal change in the sample . Let be arbitrary, ensemble then the sample partition must be clustering. The fact of “coincidence” of clusterings and will be called identity, the clusterings themselves are identical and denoted it as . In this case, it is natural to call a partition as stable clustering if the partitions and coincide for all . In the case of non-identity of some individual with , we will call as quasi-clustering.
If , then in this case, we will talk about stable clustering or simply clustering. Suppose that for some the condition is not satisfied, and is the clustering of the sample obtained from the partition using as the initial approximation. Then can significantly differ from . We will use as a function of the proximity between clustering and partitioning the value . Note that to calculate proximity it is required to find the maximum matching in a bipartite graph, for which there is a polynomial algorithm . If does not exist, we will assume that .
2.1. Method of minimizing the dispersion criterion
It is known that in order to minimize the dispersion criterion, it suffices to satisfy inequalities
for any clusters and , arbitrary , where , .
We establish the conditions for the identity of the partitions and . In the case [considering (Eq. (1))] to satisfy the condition inequalities.
must be satisfied. In the case inequalities must be satisfied.
2.2. -means method
Let the clustering be obtained by -means method, that is, . In the case of equality, the object is considered to belong to a cluster with a lower number. Then, is satisfied if under and under .
2.3. Method of hierarchical agglomeration grouping
We confine ourselves to the case of an agglomeration hierarchical grouping. To find the value of the criterion , you can calculate the partitioning , partitions , and compare with each . Here it is possible to save in the calculation of without carrying through the clustering for some of . Indeed, let there be clustering of the sample into clusters, . is a partition obtained by the clustering algorithm . The main property of the hierarchical grouping is that for any there is for which . In this case, if at some step for some the condition does not hold for all , then the condition will not be fulfilled.
We give some examples illustrating the stability criteria introduced.
Below are the results obtained for model samples. The method of clustering based on the minimization of the dispersion criterion  has been used. As the initial data, we used samples of a mixture of two two-dimensional normal distributions with independent features, different , and . Examples are shown in Figures 1–3 (images of the samples in question) and in Tables 1 and 2. Figure 1 represents a sample of 200 objects for which all the criteria , (), () are equal to 1, and the resulting clustering into two clusters is stable clustering. Here we used distributions with parameters , and .
Further, with the same parameters , experiments were carried out for .
Then, we used distributions with parameters , , . In this case, we have the case of strongly intersecting distributions. Formally, the clustering method gives a quasi-clustering, approximately corresponding to the partitioning of the original sample (Figure 3) into two sets by a diagonal from the upper left corner of the picture to the lower right. The values of the criteria in Table 2 were obtained.
Data clustering of  and criteria values , (), (). The following data from classification problem of electromagnetic signals were considered: , , , . We give the values of the stability criteria obtained. Figure 4 shows the visualization  of the sample. The accuracy of the supervised classification methods was about 87% of the correct answers. However, the clustering of data turned out to be only quasi-clustering (Table 3).
3. Committee synthesis of ensemble clustering
The problem is as follows. There are clusterings for the same number of clusters. How to choose from them the only one or build a new clustering from the available ones? In the supervised classification problem (with the help of a collective solution of a set of algorithms) there is a criterion according to which one can choose an algorithm from existing ones or build a new algorithm. This is a supervised classification error. This direction in the theory of classification appeared in the early 1970s of the last century [20, 21], then was created an algebraic approach , various correctors were appeared. The key in the algebraic approach is the creation in the form of special algebraic polynomials of a correct (error-free) algorithm based on a set of supervised classification algorithms. Some algebraic operations on matrices of “degrees of belonging” of recognized objects are used. Various types of correctors were also created [22, 23, 24, 25], when the problem of constructing (and applying) the best algorithm is also solved in two stages. First, the supervised classification algorithms are determined, and then the corrector. This can be, for example, the problem of approximating a given partial Boolean function by some monotonic function. In recent decades, there are conferences on multiple classifier systems, these issues are reflected in the books [21, 10]. How to choose or create the best clustering using a finite set of given solutions? Here, all problems are connected primarily with the absence of a single generally accepted criterion. Each clustering algorithm finds such “source” clusters of objects that are “equivalent” to each other. In this chapter, it is proposed to build such a clustering of the initial data, the cluster solutions of which have a large intersection with the initial clusters.
Let the sample of objects for supervised classification and classes are given. In the theory of supervised classification, the following definition of the supervised classification algorithm exists . Let be equal to 1 when the object is classified by the algorithm as and 0 otherwise: . Here the intersection of classes is allowed. Unlike the supervised classification problem, when clustering a sample, we have freedom in the designation of clusters.
It is clear that this definition defines a class of equivalent matrices for some matrix.
The number of clusters and the length of the control sample are considered to be given. This definition emphasizes the fact that in an arbitrary partition of a sample into clusters, we have complete freedom in the numbering of clusters. In what follows we shall always consider matrices of dimension .
Let there be given algorithms for clustering and their solutions for sample . We denote an arbitrary element of the clustering .
Therefore, we have or set , .
There are two problems.
Construction of the mapping on, , (that is, the construction of some kind of clustering).
Finding the optimal element in (i.e. finding the best clustering in ).
It is clear that .
The general scheme of collective synthesis is shown in Figure 5.
We note that the total number of possible values is bounded from above by a quantity . Let be the operator that performs permutation of columns of matrices with the help of a substitution , is the set of all operators . We believe that . We continue to the -dimensional case . We denote is the extension of . From the definition of the adder it follows that . Further, we have and finally . Therefore, the product defines the desired mapping and specifies some ensemble clustering. It is necessary to determine the optimal element from , find it and .
We introduce definitions of potentially best and worst-case solutions. As the “ideal” of the collective solution, we will consider the case when all algorithms give us essentially the same partitions or coverings.
As the distance between two numerical matrices, we consider the function .
Denote by the set of all contrast matrices, and by the set of all blurred matrices. We introduce definitions for estimating the quality of matrices.
The set (where ) is called the mean blurred matrix.
Figure 6 illustrates the sets of contrasting and blurred matrices. Arrows indicate some elements of sets.
Let us show that + = for any . We write . If then , and +. If then , and +.
Summing over all the set of values of pairs of indices , we get that + = .
We consider the problem of finding optimal ensemble clusterings for the criterion (2). It is clear that .
We introduce the notations , , , . Let , be some permutation of the set . A set of permutations uniquely determines the matrix of estimates.
We will further assume that the “initial” matrix of the algorithm corresponds to the permutation . is the matrix of the algorithm corresponding to some permutation . Then .
Consider , .
Then . We convert this expression.
The identity is valid. Get .
Thus, minimizing a function is equivalent to maximizing the second sum of the expression.
After applying the permutations , the sets change. We introduce the notations , , , .
Figure 7 schematically shows the changes in sets .
Since the second sum is always not positive, we have an upper bound. We consider the problem of minimizing a function . We write out all possible variants of the function in the form of a table in Figure 8. Then the minimum of this function is reduced to finding the maximum matching of the bipartite graph, for finding which we can use the polynomial Hungarian algorithm .
It is clear that . Now we can propose the following heuristic algorithm for steepest descent.
1. We calculate .
2. We find for each .
If , then apply the found permutations and go to step 1).
If then the END of algorithm.
NOTE. We note that our algorithm does not even find a local minimum of the criterion . Nevertheless, this algorithm is very fast, its complexity at each iteration is estimated as .
4. The algorithm of collective k-means
Results of clustering by algorithms of sampling of objects to clusters solutions are obtained, which we can write in the form of a binary matrix . We assume that the cluster numbers in each algorithm are fixed. Then any horizontal layer number of this three-dimensional matrix will denote the results of object clustering. As an ensemble clustering of the sample , we can take the result of clustering the “new” descriptions—the layers of the original matrix . As a method of clustering, we take the method of minimizing the dispersion criterion. Let there be a lot of clusterings with heuristic clustering algorithms, then we calculate their sample mean as the solution of the problem . Where do we obtain . Note that this method makes it possible to calculate such ensemble clusterings that the sets of heuristic clustering of the objects of some cluster of the collective solution will be close to each other in the Euclidean metric. The committee synthesis of collective decisions provides more interpretable solutions. Indeed, if are separate solutions of heuristic clustering algorithms, then the cluster of collective solution will be the “intersection” of many some original clusters .
5. Man-machine (video-logical) clustering method
In the problems of ensemble clustering synthesis considered earlier, we did not consider the number of initial clustering algorithms, their quality and their proximity. Ensemble clustering was built and reflected only the opinion of the collective decisions that we used. “Internal” indices  reflect the person’s ideas about clustering. You can think up examples of data when known internal criteria lead to degenerate solutions.
At the same time, a person has the ability to cluster visual sets on a plane without using any proximity functions, criteria and indices. The following idea was realized. A person can personally cluster projections of sets of points from into . Having made such clusterings under different projections, we can construct generally speaking various clusterings, which we submit to the input of the construction of the collective solution. The person himself “does not see” the objects in , but can exactly solve the clustering tasks on the plane. Thus, here we use precise solutions, but of various partial information about the data. Consider this video-logical method on one model example.
A sample of two normal distributions with independent characteristics was considered. The first feature of the first distribution (200 objects) had zero expectation and the standard deviation, the first attribute of the second distribution (200 objects) had these values equal to 5. All the other 49 attributes for all objects had , . That is, the two sets had equal distributions for 49 features and one informative feature. Clustering of the entire sample by minimizing dispersion is shown in Figure 9. Black and gray points on sample visualization represent the objects of the first and second clusters. Here the fact of informative character of the first feature is lost.
The program of the video-logical approach worked as follows. With the help of a single heuristic approach, all projections are automatically ordered according to the descending criteria of the presence of two clusters. Next we as experts consider some projections and with the help of the mouse we select in each of them two clusters. Figure 10 shows two such examples. Note that the first feature was present in all projections. It was used “manually” as the defining area for the dense location of objects. Then 10 “manual” clustering went to the program entrance for the committee synthesis of the collective solution. Note that only two objects were erroneously clustered.
This chapter consists of two parts. First, clustering criteria based on sustainability are introduced. Next, we propose an approach to processing the sets of obtained partitions of the same sample. As the initial clustering, it is better to use stable clustering. It is shown how a person can be used in the construction of the committee synthesis of ensemble clustering.
The reported study was funded by RFBR according to the research project No 17-01-00634 and No 18-01-00557.
Halkidi M, Batistakis Y, Vazirgiannis M. Cluster validity methods: Part 1. SIGMOD Record. 2002; 31(2):40-45. DOI: 10.1145/601858.601862
Aggarwal C, Reddy C. Data Clustering: Algorithms and Applications. CRC Press; 2014
Duda R, Hart P, Stork D. Pattern Classification. 2nd ed. New York: Wiley; 2000
Lloyd S. Least squares quantization in PCM (PDF). IEEE Transactions on Information Theory. 1982; 28(2):129-137. DOI: 10.1109/TIT.1982.1056489
Kriegel H, Kröger P, Sander J, Zimek A. Density-based clustering. WIREs Data Mining and Knowledge Discovery. 2011; 1(3):231-240. DOI: 10.1002/widm.30
Dempster A, Laird N, Rubin D. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, Series B. 1977; 39(1):1-38 JSTOR 2984875. MR 0501537
Jain A, Dubes R. Algorithms for Clustering Data. Englewood Cliffs: Prentice-Hall, Inc.; 1998
Kaufman L, Rousseeuw P. Finding Groups in data: An Introduction to Cluster Analysis. New York: Wiley; 2009
Aggarwal C. Data Mining: The Textbook. Yorktown Heights/New York: IBM T.J. Watson Research Center; 2015. 771 p. DOI: 10.1007/978-3-319-14142-8
Kuncheva L. Combining Pattern Classifiers: Methods and Algorithms. Hoboken: Wiley; 2004. DOI: 10.1002/9781118914564
Desgraupes B. Clustering indices. University Paris Ouest. Lab Modal'X; 2013
Ryazanov V. Commitee synthesis of algorithms for recognition and classification. Journal of Computational Mathematics and Mathematical Physics. 1981; 21(6):1533-1543. DOI: 10.1016/0041-5553(81)90161-0
Ryazanov V. On the synthesis of classification algorithms on finite sets of classification algorithms (taxonomy). Journal of Computational Mathematics and Mathematical Physics. 1982; 22(2):429-440. DOI: 10.1016/0041-5553(82)90049-0
Ryazanov V. One approach for classification (taxonomy) problem solution by sets of heuristic algorithms. In: Proceedings of the 9-th Scandinavian Conference on Image Analysis; 6–9 June 1995; Uppsala; 1995(2). pp. 997-1002
Biryukov A, Shmakov A, Ryazanov V. Solving the problems of cluster analysis by collectives of algorithms. Journal of Computational Mathematics and Mathematical Physics. 2008; 48(1):176-192. DOI: 10.1134/S0965542508010132
Kuhn H. The Hungarian method for the assignment problem. Naval Research Logistics Quarterly. 1955; 2:83-97. DOI: 10.1002/nav.3800020109
Ryazanov V. Estimations of clustering quality via evaluation of its stability. In: Bayro-Corrochano E, Hancock E, editors. CIARP 2014. LNCS. Vol. 8827; 2014. pp. 432-439. DOI: 10.1007/978-3-319-12568-8_53
Ryazanov V. About estimation of quality of clustering results via its stability. Intelligent Data Analysis. 2016; 20:S5-S15. DOI: 10.3233/IDA-160842
Sigillito V, Wing S, Hutton L, Baker K. Classification of radar returns from the ionosphere using neural networks. Johns Hopkins APL Technical Digest. 1989; 10:262-266
Rastrigin L, Erenstein R. Collective decision-making in pattern recognition. Avtomatica i telemehanika. 1975; 9:133-144
Method of committees in pattern recognition. Sverdlovsk, IMM AN USSR; 1974
Yu Z. On the algebraic approach to solving problems of recognition or classification. Problems of Cybernetics. 1978; 33:5-68
Zuev Y. The method of increasing the reliability of classification in the presence of several classifiers, based on the principle of monotony. Journal of Computational Mathematics and Mathematical Physics. 1981; 21(1):157-167
Krasnoproshin V. About the optimal corrector of the set of recognition algorithms. Journal of Computational Mathematics and Mathematical Physics. 1979; 19(1):204-215
Zhuravlev Y. Selected Scientific Works. Moscow: Publishing House Magister; 1998. p. 420