Next Article in Journal
Learning to Classify DWDM Optical Channels from Tiny and Imbalanced Data
Next Article in Special Issue
Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
Previous Article in Journal
Belief and Possibility Belief Interval-Valued N-Soft Set and Their Applications in Multi-Attribute Decision-Making Problems
Previous Article in Special Issue
Information-Corrected Estimation: A Generalization Error Reducing Parameter Estimation Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Summarizing Finite Mixture Model with Overlapping Quantification

Graduate School of Information Science and Technology, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656, Japan
*
Author to whom correspondence should be addressed.
Submission received: 28 September 2021 / Revised: 7 November 2021 / Accepted: 8 November 2021 / Published: 13 November 2021
(This article belongs to the Special Issue Information Theory and Machine Learning)

Abstract

:
Finite mixture models are widely used for modeling and clustering data. When they are used for clustering, they are often interpreted by regarding each component as one cluster. However, this assumption may be invalid when the components overlap. It leads to the issue of analyzing such overlaps to correctly understand the models. The primary purpose of this paper is to establish a theoretical framework for interpreting the overlapping mixture models by estimating how they overlap, using measures of information such as entropy and mutual information. This is achieved by merging components to regard multiple components as one cluster and summarizing the merging results. First, we propose three conditions that any merging criterion should satisfy. Then, we investigate whether several existing merging criteria satisfy the conditions and modify them to fulfill more conditions. Second, we propose a novel concept named clustering summarization to evaluate the merging results. In it, we can quantify how overlapped and biased the clusters are, using mutual information-based criteria. Using artificial and real datasets, we empirically demonstrate that our methods of modifying criteria and summarizing results are effective for understanding the cluster structures. We therefore give a new view of interpretability/explainability for model-based clustering.

1. Introduction

1.1. Motivation

Finite mixture models are widely used for modeling data and finding latent clusters (see McLachlan and Peel [1] and Fraley and Raftery [2] for overviews and references). When they are used for clustering, they are typically interpreted by regarding each component as a single cluster. However, the one-to-one correspondence between the clusters and mixture components does not hold when the components overlap. This is because the clustering structure then becomes more ambiguous and complex. Let us illustrate this using a Gaussian mixture model estimated for the Wisconsin breast cancer dataset in Figure 1 (details of the dataset and estimation are discussed in Section 8.2). A number of the components overlap with one another, which makes it difficult to estimate the shape of distribution or number of clusters. Therefore, we need an analysis of the overlaps to correctly interpret the models.
We address this issue from two aspects. In the first aspect, we consider merging mixture components to regard several components as one cluster. We repeatedly select the most overlapping pairs of components to merge them. In this procedure, it is important how the degree of overlap is measured. A number of criteria for measuring cluster overlaps have been proposed [4,5,6], but they have not yet been compared theoretically. We give a theoretical framework for comparing merging criteria by defining three essential conditions that any method for merging clusters should satisfy. The more conditions any method satisfies, the better it is. From this viewpoint, we evaluate the existing criteria (entropy (Ent) [4], directly estimated misclassification (DEMP) [5] probability, mixture complexity (MC) [7]). We also modify these existing criteria so that they can satisfy more essential conditions.
In the second aspect, we consider how to summarize the merging results quantitatively. After merging mixture components, we obtain two types of clustering structures; those among the upper-components and those among sub-components within each upper-component, as illustrated in Figure 2. These structures might be still ambiguous because the upper-components are determined to be the different clusters, but they may overlap; the sub-components are determined to belong to the same cluster, but they may be scattered in the cluster. Therefore, we need to evaluate the degree to which the upper- and sub-components are discriminated as different clusters. We realize this using the notions of mixture complexity (MC) [7] and normalized mixture complexity (NMC). They give real-valued quantification of the number of effective clusters and the degree of their separation, respectively. We therefore develop a novel method for cluster summarization.
Our hypotheses in this paper are summarized as follows:
  • Modifying merging criteria based on essential conditions can improve the ability to find cluster structures in the mixture model.
  • Cluster summarization based on MC and NMC effectively describes the clustering structures.
We empirically verify them by experiments, using artificial and real datasets.

1.2. Significance and Novelty of This Paper

The significance and novelty of this paper is summarized below.

1.2.1. Proposal of Theoretical Framework for Evaluating Merging Criteria

We give a theoretical framework for evaluating merging methods by defining the essential conditions. They are necessary conditions that any merging criterion should satisfy: (1) the criterion should take the best value when the components are entirely overlapped, (2) it should take the worst value when the components are entirely separated, and (3) it should be invariant with respect to the scale of the weights. We empirically confirm that the more essential conditions any merging method satisfies, the better the clustering structure obtained in terms of larger interdistances and smaller intradistances.

1.2.2. Proposal of Quantitative Clustering Summarization

We propose a method for quantitatively summarizing clustering results based on MC and NMC. MC is an extended concept of the number of clusters into a real number from the viewpoint of information theory [7]. It quantifies the diversity among the components, considering their overlap and weight bias. NMC is defined by normalizing MC to remove the effects of weight bias. It quantifies the degree of the scatter of the components based only on their overlap. Furthermore, MC and NMC have desirable properties for clustering summarization: they are scale invariant and can quantify overlaps among more than two components. We empirically demonstrate that our MC-based method effectively summarizes the clustering structures. We therefore give a novel quantification of clustering structures.

2. Related Work on Finite Mixture Models and Model-Based Clustering

In this section, we present related work on finite mixture models and model-based clustering in four parts: roles of overlap, model, optimization, and visualization. The overlap has a particular impact on the construction of models.

2.1. Roles of Overlap

There has been widespread discussion about the roles of overlap in finite mixture models. One argues that the overlap is emerged to represent various distributions. While this flexibility is beneficial for modeling the data, various issues arise in applying them to clustering. For example, McLachlan and Peel [1] pointed out that some skew clusters required more than one Gaussian component to be represented. Moreover, Biernacki et al. [8] pointed out that the number of mixture components selected for estimating densities was typically more than that of clusters because of overlapping. Model selection methods based on clustering (complete) likelihood, such as the integrated complete likelihood (ICL) [8], the normalized maximum likelihood (NML) [9,10], and the decomposed normalized maximum likelihood (DNML) [11,12], have been proposed to obtain less-overlapping mixtures so that one component corresponds to one cluster. However, they have problems in that they need to define the shape of the clusters in advance. This leads to a trade-off between shape flexibility and component overlap in model-based clustering.
Others argue that the overlap represents that the data belong to more than one cluster. For example, in clustering documents by their topics, the data may have several topics. Such issues have been widely discussed in the field of overlapping clustering. For example, Banerjee et al. [13] extended the mixture model to allow the data to belong to multiple clusters based on membership matrices. Fu and Banerjee [14] considered the product of cluster distributions to represent multiple memberships of the data. Xu et al. [15] proposed methods for describing more complex memberships by calculating correlation weights between the data and the cluster. While these methods allow complex relationships between the data and the clusters, cluster shapes become simple.
The overlap is also used for measuring the complexity of clustering structures in the concept of MC [7]. It is a non-integer valued quantity, which implies the uncertainty of determining the number of clusters. MC was introduced in the scenario of change detection in [7]. This paper gives a new application scenario of MC in the context of quantifying clustering structures. Moreover, this paper also newly introduces NMC as a variant of MC, which turns out to be most effective in this context.

2.2. Model

We discuss the issue of constructing models achieving both flexible cluster shapes and interpretability. Allowing each cluster to have complex shapes is a solution to tackle this. For example, mixtures of non-normal distributions have been proposed for this purpose, as reviewed by Lee and McLachlan [16]. Modeling each cluster as a finite mixture model, called the mixture of mixture model or multi-layer mixture model, has been considered in this regard. Various methods have been proposed to estimate such mixture models based on maximum likelihood estimation [17,18] and Bayesian estimation [19,20]. However, additional parameters are required for assigning sub-components to upper-clusters in many cases because changes of assignment do not change the overall distribution. Merging mixture components [4,5,6] is an alternative way of the composition of mixture models using single-layer estimations. In this approach, the criteria to measure the degree of component overlap have to be identified. Although various concepts have been developed to measure the degree of overlap, such as entropy [5], misclassification rate [4,6], and unimodality [4], they have not been satisfactorily compared yet.

2.3. Optimization

Merging components has also been discussed in the scenario of optimizing parameters in the mixture models. Ueda et al. [21] proposed splitting and merging mixture components to obtain better estimations, and Minagawa et al. [22] revised their methods to search the models with higher likelihoods. Zhao et al. [23] considered randomly swapping the mixture components during optimization, which allows a more flexible search than splitting and merging components. Because these methods aim only to optimize the models, there remains the problem of interpreting them.
We also refer to the agglomerative hierarchical clustering as a similar approach to merging components. Our methods are similar to the Bayesian hierarchical clustering methods [24,25] in that the number of merging is automatically decided. However, our approaches can not only create clusters, but also evaluate their shape and closeness under the assumption that the mixture models are given.

2.4. Visualization

Methods of interpreting clustering structures have been studied along with visualization methods. Visualizing the values of criteria with a dendrogram is useful for understanding cluster structures among sub-components [6]. Class-preserved projections [26] and parametric embedding [27] were proposed for visualizing structures among upper-clusters by reducing data dimension. We present a method to interpret both structures uniformly based on the MC and NMC.

3. Merging Mixture Components

We assume that data x N = x 1 , , x N and a finite mixture model are given. The probability distribution of the model f is written as follows:
f ( x ) = k = 1 K ρ k g k ( x ) ,
where K denotes the number of components, ρ 1 , , ρ K denote the mixture proportions of each component summing up to one, and g ( x | θ 1 ) , , g ( x | θ K ) denote the probability distributions. We assume that the data x N are independently sampled from f. The random variable X following f is called an observed variable, because it can be observed as a data point. We also define the latent variable  Z Z { 1 , , K } as the index of the component from which X originated. The pair ( X , Z ) is called a complete variable. The distribution of the latent variable P ( Z ) and the conditional distribution of the observed variable P ( X | Z ) can be given by the following:
P ( Z = k ) = ρ k , P ( X | Z = k ) = g k ( X ) .
In the case that f is not known, we will replace f by its estimation f ^ under the assumption that f ^ is so close to f that x N can be approximately regarded as samples from f ^ .
We discuss identifying cluster structures in x N and f by merging mixture components as described below. First, we define a criterion function denoted as Crit : Z × Z R , which measures the degree of overlap or closeness between two components. For simplicity, we change the sign of the original definitions as needed so that Crit takes smaller values as the components are closer. Then, we choose the closest two components that minimize the criterion and merge them. By repeating the merging process several times, we finally obtain clusters. We show the pseudo-code and computational complexity of this procedure in Appendix A.

4. Essential Conditions

In this section, we propose three essential conditions that the criteria should satisfy, so that the criteria can be compared in terms of the conditions. To establish the conditions, we restrict the criteria to those that can be calculated from the posterior probability of the latent variables { γ k ( x n ) } k , n defined as follows:
γ k x n P Z = k | X = x n = ρ k g k x n f x n ,
where k is the index of the component. After merging the components i and j, the posterior probability can be easily updated as follows:
γ i j x n P Z { i , j } | X = x n = γ i x n + γ j x n .
Note that some other merging methods reestimate the distribution of the merged components as a single component [4]. We do not consider these in this study because they lack the benefit that the merged components can have complex shapes.
For later use, we define Best ( Crit ) and Worst ( Crit ) as the best and worst values that the criteria can take:
Best ( Crit ) min Crit ( i , j ) w . r . t . γ k , n k , n , Worst ( Crit ) max Crit ( i , j ) w . r . t . γ k , n k , n ,
where { γ k , n } k , n is a set of K × N real values in [ 0 , 1 ] that satisfies k γ k , n = 1 for all n.
We formulate the three conditions. They provide natural and minimum conditions on the behaviors in the extreme cases that the components are entirely overlapped or separated and on the scale invariance of the criteria. The conditions for the moderate cases that the components partially overlap should be investigated in further studies.
First, we define the condition that a criterion should take the best value when the two components entirely overlap. It is formally defined as follows.
Definition 1.
If a criterion satisfies that
n , g i x n = g j x n Crit ( i , j ) = Best ( Crit ) ,
then, we say that it satisfies the condition BO (best in entirely overlap).
Next, we define the condition that the criterion should take the worst value when the two components are entirely separated.
Definition 2.
We consider that the sequence of the models { f t = k ρ k , t g k , t } t = 1 satisfies the following:
n , g i , t x n g j , t x n 0
as t . We define Crit t ( i , j ) as the criterion value based on f t . Then, if (1) implies that
lim t Crit t ( i , j ) Worst ( Crit ) ,
we say that it satisfies the condition WS (worst in entirely separate).
Note that this definition is written using limits in case that the distribution of the components has support in the entire space, such as the Gaussian distributions.
Finally, we define the condition that the value of the criterion should be invariant with the scale of mixture proportions.
Definition 3.
We consider that the components i and j are isolated from the other components, i.e., the sequence of the models { f t = k ρ k , t g k , t } t = 1 satisfies the following:
g i , t x n + g j , t x n g k , t x n 0
for all k i , j and n as t . In addition, we consider another sequence of the mixture model { f ¯ t = k ρ ¯ k , t g k , t } t = 1 with different scales on the mixture proportions of the components i and j, i.e., ρ ¯ k , t = a ρ k , t ( k = i , j ) holds for some a > 0 . We define Crit ¯ t ( i , j ) as the criterion value based on f ¯ ( t ) . Then, we say that the criterion satisfies the condition SI (Scale invariance) if for any a, the following holds:
lim t Crit t ( i , j ) = lim t Crit ¯ t ( i , j ) .

5. Modifying Merging Methods

In this section, we introduce the existing merging criteria and propose new criteria by modifying them so that they can satisfy more essential conditions.

5.1. Entropy-Based Criterion

First, we introduce the entropy-based criterion (Ent) proposed by Baudry et al. [5]. It selects the components that reduce the entropy of the latent variable the most. This criterion, denoted as Crit Ent , is formulated as follows:
Crit Ent ( i , j ) n = 1 N Ψ γ i x n + Ψ γ j x n Ψ γ i j x n ,
where Ψ ( x ) x log x .
However, it violates the conditions BO and SI. Therefore, we propose to modify it in two regards. First, we correct the scale of the weights to make Crit Ent satisfy SI. We propose a new criterion Crit NEnt 1 defined as follows:
Crit NEnt 1 ( i , j ) Crit Ent ( i , j ) N ρ ˜ i + ρ ˜ j ,
where ρ ˜ k n γ k ( x n ) / N . This satisfies the condition SI.
Next, we propose removing the effects of the weight biases to make Crit NEnt 1 satisfy BO. We further introduce a new criterion Crit NEnt 2 defined as follows:
Crit NEnt 2 ( i , j ) Crit NEnt 1 ( i , j ) H ˜ i , j ( Z ) , H ˜ i , j ( Z ) k { i , j } Ψ ρ ˜ k ρ ˜ i + ρ ˜ j .
This satisfies all conditions: BO, WS, and SI.

5.2. Directly Estimated Misclassification Probabilities

Second, we introduce the criterion named directly estimated misclassification probabilities (DEMP) [4]. It selects the components with the highest misclassification probabilities. The criterion is formulated as follows:
Crit DEMP ( i , j ) max M ˜ j , i , M ˜ i , j ,
where
M ˜ j , i P ˜ z ^ ( X ) = j | Z = i n γ i x n 1 z ^ x n = j N ρ ˜ i , z ^ ( x ) = arg max k = 1 , , K γ k x .
However, this violates the condition BO when z ^ ( x n ) is not i or j for some n. Therefore, we modify it by restricting the choice of the latent variable to component i or j. We define z ^ i , j ( x ) as follows:
z ^ i , j ( x ) arg max k = i , j γ k x n
and define Crit DEMP 2 by replacing z ^ ( x ) with z ^ i , j ( x ) in the definition of Crit DEMP . Then, this satisfies all essential conditions.

5.3. Mixture Complexity

Finally, we propose a new criterion based on mixture complexity (MC) [7]. MC is an extended concept of (the logarithm of) the number of clusters into a real value considering the overlap and bias among the components. It is defined based on information theory, and formulated as follows:
MC γ k x n k , n ; w n n k = 1 K Ψ ρ ˜ k n = 1 N w n W k = 1 K Ψ γ k x n ,
where { w n } n denotes the weights of the data x N , W n w n denotes their sum, and ρ ˜ k is redefined as ρ ˜ k n w n γ k ( x n ) / W . Examples of MC for mixtures of two components are shown in Figure 3. In them, the exponential of the MCs take values between 1 and 2, according to the uncertainty in the number of clusters induced by the overlap or weight bias between the components.
We first propose a new merging criterion Crit MC to select the components whose MCs are the smallest. It is defined as follows:
Crit MC ( i , j ) MC γ k x n γ i j x n k { i , j } , n ; γ i j x n n .
However, this does not satisfy the condition WS because of the effects of the weight biases. Therefore, we modify it by removing the biases to propose a new criterion, which we call the normalized mixture complexity (NMC) Crit NMC . The criterion is defined as follows:
Crit NMC ( i , j ) Crit MC ( i , j ) H ˜ i , j ( Z ) .
It satisfies all conditions BO, WS, and SI. Note that it is equivalent to Crit NEnt 2 because Crit NMC = 1 + Crit NEnt 2 .
We summarize the relationships between the criteria and the essential conditions in Table 1. The modification led to the fulfillment of many conditions.

6. Stopping Condition

We also propose a new stopping condition based on NMC. First, we calculate the NMC for the (unmerged) mixture model f defined as follows:
NMC 0 MC γ k x n k , n ; { 1 } n H ˜ ( Z ) .
Since it represents the average degree of separation in the components of f, it can be used for the stopping condition for merging. Then, before merging components i and j, we compare Crit NMC ( i , j ) to NML 0 . If Crit NMC ( i , j ) NML 0 , then the merging algorithm halts without merging components i and j. Otherwise, the algorithm merges components i and j and continues further.
Note that this stopping criterion can be applied when a criterion other than Crit NMC is used. In this case, we use the criterion to search the two closest components and use NMC to decide whether to merge them.

7. Clustering Summarization

In this section, we propose methods to quantitatively explain the merging results, using the MC and NMC.
We consider that a mixture model with K-component is merged into L upper- components. We define the sets I 1 , , I L that partition { 1 , , K } as the sets of the indices that are contained in each upper-component. Then, the MC and NMC among the upper-components, denoted as MC ( up ) and NMC ( up ) , respectively, can be calculated as follows:
MC ( up ) MC k I l γ k x n l , n , { 1 } n , NMC ( up ) MC ( up ) l Ψ τ ˜ l ,
where τ ˜ l denotes the weight of the upper-component l calculated as follows:
τ ˜ l 1 N k I l γ k x n = k I l ρ ˜ k .
For each l, the MC and NMC in the sub-components within the upper-component l, written as MC ( l ) and NMC ( l ) , respectively, can be calculated as follows:
MC ( l ) MC γ k x n k I l γ k x n k I l , n ; k I l γ k x n n , NMC ( l ) MC ( l ) k I l Ψ ρ ˜ l ( k ) ,
where ρ ˜ l ( k ) denotes the relative weight of the sub-component k I l calculated as ρ ˜ l ( k ) ρ ˜ k / k I l ρ ˜ k . NMC is undefined if the denominator is 0.
MC and NMC quantify the degree to which the components are regarded as clusters in different ways: larger values indicate that the components definitely look like different clusters. MC quantifies this by measuring (the logarithm of) the number of clusters continuously, considering the ambiguity induced by the overlap and weight bias among the components. It takes a value between 0 and the logarithm of the number of the components. In contrast, NMC measures the scattering of the components based only on their overlap. It takes a value between 0 and 1. They have also the desirable properties that they are scale invariant and can quantify overlaps among more than two components.
Therefore, we propose the summarization of clustering structures by listing MC ( up ) , NMC ( up ) , component weights, MC ( l ) , and NMC ( l ) in a table, which we call the clustering summarization. The clustering summarization is useful for evaluating the confidence level of the clustering results.
We show an example of the clustering summarization using the mixture model illustrated in Figure 4. In this example, there are four Gaussian components as illustrated in Figure 4a, and two merged clusters on the left and right sides as illustrated in Figure 4b–d. The clustering summarization is presented in Table 2. For the upper-components, the exponential of MC is almost two, and the NMC is almost one. This indicates that two upper-components can be definitely regarded as different clusters. For both sub-components, the exponential of MC is larger than one. This indicates that they have more complex shapes than a single component. Moreover, the structures within Component 1 are more complex than those in 2, because the MC and NMC are larger.

8. Experiments

In this section, we present the experimental results to demonstrate the effectiveness of merging the mixture components and modifying the criteria.

8.1. Analysis of Artificial Dataset

To reveal the differences among the criteria, we conducted experiments with artificially generated Gaussian mixture models. First, we randomly created a two-dimensional Gaussian mixture model f = k = 1 K ρ k N ( x ; μ k , Σ k ) as follows:
K 50 , ρ 1 , , ρ K Dir ( 1 , , 1 ) , μ 1 , , μ K i . i . d . N μ ; [ 0 , 0 ] , 3 2 × I 2 , a 1 , b 1 , , a K , b K i . i . d . U [ 0.5 , 1.5 ] , Σ k a k , 0 , 0 , b k ( k = 1 , , K ) ,
where Dir ( α , , α ) denotes the Dirichlet distribution, and U [ m , M ] denotes the uniform distribution from m to M. Then, we sampled 5000 points x 5000 from f, and ran the merging algorithms without stopping conditions. The algorithms were evaluated using the (maximum) intra-cluster distance D intra and (minimum) inter-cluster distance D inter defined as follows:
D intra max k = 1 , , K n γ k x n x n μ ˜ k 2 n γ k x n , D inter min 1 i < j K μ ˜ i μ ˜ j 2 ,
where μ ˜ 1 , , μ ˜ K denote the centers of the components defined as
μ ˜ k n γ k x n x n n γ k x n .
The clustering structure is said to be better, as D intra is smaller and D inter is larger. Both distances are measured with several K and compared among the algorithms with different criteria. Although we may obtain better results for these metrics by using them as merging criteria in a similar way as used in hierarchical clustering [28,29], we used them only for comparison rather than optimizing them.
The experiments were performed 100 times by randomly generating f and the data. Accordingly, the ranking of the criteria was calculated for each distance. Table 3 presents the average rank of each criterion. As seen from the table, the modifications of the criteria improved the rank. In addition, DEMP2 and NMC, satisfying all essential conditions, were always in the top three. These results indicate the effectiveness of the essential conditions.
To further investigate the relationships between the essential conditions and resulting cluster structures, we illustrated the cluster obtained in a trial where the intra-cluster distance was the largest in Figure 5. For the criterion Ent, one cluster continued to grow. This is because Ent lacks the condition SI, and is advantageous for larger clusters. For the criterion NEnt1, the growth of the larger clusters was mitigated by adding the condition SI to Ent. Nevertheless, the intra-cluster distances were still large because NEnt lacked the condition BO. It tended to create unnecessarily large clusters because it tended to merge larger and more distant components rather than smaller and closer components. The criterion NMC improved such a disadvantage by adding the condition BO to NEnt1. For the criterion MC, distant components were merged, as the condition WS was not satisfied. NMC overcame this by adding the condition WS to MC. The differences between DEMP and DEMP2 were unclear in Figure 5c,d, and both criteria elucidated the cluster structure well because they satisfied relatively many conditions. We conclude that the essential conditions are effective for obtaining better cluster structures.

8.2. Analysis of Real Dataset

We discuss the results of applying the merging algorithms and clustering summarization to eight types of real datasets with true cluster labels. The details of the datasets and processing are described in Appendix B.

8.2.1. Evaluation of Clustering Using True Labels

First, we compared the clustering performance of the merging algorithms by measuring similarity between estimated and true cluster labels. Formally, given the dataset { x n } n and the true labels { z n } n , we first estimated the clustering structures using { x n } n without seeing { z n } n , and obtained the estimated labels { z ^ n } n . We define K and K ^ as the number of the true and estimated clusters. Then, we evaluated the similarity between { z n } n and { z ^ n } n using the adjusted Rand index (ARI) [30] and F-measure. ARI takes values between -1 and 1, and F-measure takes values between 0 and 1. Their larger value corresponds to better clustering. Both indices can be applied when the number of true and estimated clusters is different.
To run the merging algorithms, the mixture models should be estimated first. In our experiments, we estimated them by the variational Bayes Gaussian mixture model with K = 20 [31] implemented in the Scikit-learn package [32]; we adopted this, as it exhibited good performance in our experiments. We used the prior distributions of the mixture proportions as the Dirichlet distributions with α = 0.1 , and we set the other parameters for prior distributions as the default values in the package. For each dataset, we fitted the algorithm ten times with different initializations and used the best one.
We compared the merging algorithms with three types of model-based clustering algorithms based on the Gaussian mixture model, which are summarized in Table 4. First, we estimated the number of components, using BIC [33]. It selects a suitable model for describing the densities, and the mixture components tend to overlap. Nevertheless, it has been widely used for clustering by regarding each component as a cluster. Second, we estimated the number of clusters using DNML [11,12]. It selects a model whose components can be regarded as clusters by considering the description length of the latent and observed variables. Finally, we estimated the clusters as the mixture of Gaussian mixture models implemented by Malsiher-Walli et al, [20]. By fixing two integers K and L, K Gaussian mixture models were estimated with L components. The number of clusters was automatically adjusted by shrinking the redundant clusters. As in the original paper, we set K = 30 , L = 15 (and some specific parameters in the paper) for the DLB dataset and K = 10 , L = 5 for the other datasets.
We estimated the models ten times and compared the average score among the methods. The average number of clusters are listed in Table 5, and F-measure and ARI are listed in Table 6 and Table 7.
Two clusters that achieved the best score and that were obtained by the heuristics proposed in Section 6 are described. The best scores of the merging algorithms exceeded those of all other methods for six out of eight datasets. In particular, the merging methods satisfying many essential conditions, such as DEMP, DEMP2, and NMC, obtained high scores with a smaller number of clusters. Therefore, it can be said that the merging algorithms with more essential conditions are effective for elucidating the clustering structures. Moreover, the scores with NMC-based stopping conditions exceeded those of all other methods for four out of eight datasets.
To further investigate the relationships between the performances of the algorithms and the shapes of the datasets, we estimated the proportion of outliers based on the k-nearest neighbor distances D nn ( 5 ) . We calculated the ratio of the 5-nearest neighbor distance D nn ( 5 ) ( x n ) and its average ( 1 / N ) n D nn ( k ) ( x n ) for each data point, and we plotted the proportions for which the ratio exceeded 2.0, 3.0, 4.0, and 5.0 in Figure 6. As seen from the figure, the datasets where the merging methods did not work well, such as AIS, DLB, WSC, and YST, contained relatively many outliers. This is reasonable because the merging algorithms do not aim to merge distant clusters. We can conclude that the merging methods are particularly effective when the datasets have fewer outliers or when we want to find the aggregated clusters.

8.2.2. Results of Clustering Summarization

Next, we analyzed the results of the merging methods using the clustering summarization proposed in Section 7. As examples, we show one result obtained using the NMC and NMC-based stopping conditions for the Flea beetles and Wisconsin breast cancer datasets. The clustering results are summarized in Table 8 and Table 9, respectively. For the upper-components in Flea beetles dataset, the exponential of MC ( up ) was close to 3.0, and NMC ( up ) was close to 1.0; we see that the effective number of clusters was around three, and the clusters were well-separated. Components 2 and 3 were unmerged, and the exponentials of MC and NMC of Component 1 were close to 1.0 and 0.0, respectively. This indicates that each cluster can be represented by almost a single Gaussian distribution. Furthermore, the (exponentials of) MC and the NMC of the upper-components in the Wisconsin cancer dataset were 1.66 and 0.763, respectively. It can be expected that the situation was a partial overlap of the two clusters. For Components 1 and 2, NMCs were relatively large. This shows that partially separated components are needed to describe each component. MC of Component 2 was smaller than that of Component 1. Then, it is expected that Component 2 had simpler shapes than Component 1; however, the former seemed to have small components that might be outliers because NMC was larger. Plots of the predicted clusters are illustrated for the Flea beetles and Wisconsin breast cancer datasets in Figure 7 and Figure 8, respectively. We observe that the predictions described previously match to the actual plots. Therefore, we can reveal significant information about the clustering structures by observing the clustering summarizations.

8.2.3. Relationships between Clustering Summarization and Clustering Quality

Finally, we confirmed that MC and NMC in the sub-components were also related to the quality of classification. To confirm this, we conducted additional experiments discussed below. First, we ran the merging algorithms until K = 1 without the stopping conditions. Then, for every merged clusters created at K = K start , , 1 , we counted the number of data points classified into them. We define N C ( k ) as the number of points with true labels k classified into the merged cluster C. Then, we evaluated the quality of the cluster C using the entropy calculated as follows:
H C = k = 1 K N C ( k ) k N C ( k ) log N C ( k ) k N C ( k ) ,
where the cluster C for k N C ( k ) = 0 were omitted. This takes values between 0 and log K . Smaller values are preferred, because H C becomes small when most of the points within the component share the same cluster label. We calculated the MC/NMC and H C within the clusters for all datasets and merging algorithms, and we plotted the relationships between them in Figure 9. Note that the unmerged clusters were omitted because the NMC could not be defined. From the figure, it is evident that both MC and NMC had positive correlations with H C . The correlation coefficients were 0.794 and 0.637 for MC and NMC, respectively. This observation is useful in applications. If the obtained cluster has smaller MC and NMC, then we can confirm that it contains only one group. Otherwise, we need to assume that it contains more than one group. Therefore, we conclude that MC and NMC indicate the confidence level of the cluster structures.

9. Discussion

To improve the interpretability of the mixture models with overlap, we have established novel methodologies to merge the components and summarize the results.
For merging mixture components, we proposed essential conditions that the merging criteria should satisfy. Although there have been studies creating some rules in the clustering approach [34,35], they have not been applied to clustering by merging components. The proposed essential conditions for merging criteria contributed to comparing and modifying existing criteria. The limitation of our conditions is that they only provide the necessary conditions for extreme cases, where the components are entirely overlapped or separated. The conditions for the moderate cases that the components partially overlap should be investigated in further studies.
We also proposed a novel methodology to interpret the merging results based on clustering summarization. While previous studies [6,26,27] have focused on interpreting the structures among sub-components or upper-clusters only, our methods can quantify both structures uniformly based on the MC and NMC. They represented the overview of the structures in the mixture models by evaluating how much the components were distinguished based on the degree of overlap and weight bias.
We verified the effectiveness of our methods, using artificial and real datasets. In the artificial data experiments, we confirmed that the intra- and inter-cluster distances were improved corresponding to the modification of the criteria. Further, by observing the clusters with maximum intra-cluster distance, we found that the essential conditions were helpful to prevent the clusters from merging distant components or growing too much. In the real data experiments, we confirmed that the best scores of the proposed methods were better than the comparison methods for many datasets, and the scores obtained using the stopping condition were also better for the datasets containing relatively smaller outliers. In addition, we confirmed that the clustering summary was helpful to interpret the merging results. It was related to the shape of the clusters, weight biases, and the existence of the outliers. Further, we found that the MC and NMC within the components were also related to the quality of the classification. Therefore, the clustering summary also represented the confidence level of the cluster structures.

10. Conclusions

We have established the framework of theoretically interpreting overlapping mixture models by merging the components and summarizing merging results. First, we proposed three essential conditions for evaluating cluster-merging methods. They declared necessary properties that the merging criterion should satisfy. In this framework, we considered Ent, DEMP, and MC and their modifications to investigate whether they satisfied the essential conditions. The stopping condition based on NMC was also proposed.
Moreover, we proposed the clustering summarization based on MC and NMC. They quantify how overlapped the clusters are, how biased the clustering structure is, and how scattered the components are in a respective cluster. We can conduct this analysis from higher level clusters to lower level components to give a comprehensive survey of the global clustering structure. We then quantitatively explained the shape of the clusters, weight biases, and existence of the outliers.
In the experiments, we empirically demonstrated that the modification of the merging criteria improved the ability to find better clustering structures. We also investigated the merging order for each criterion and found that the essential conditions were helpful to prevent the clusters from merging distant components or growing too much. Further, we confirmed, using the real dataset, that the clustering summary revealed varied information in the clustering structure, such as the shape of the clusters, weight biases, the existence of the outliers, and even the confidence level of the cluster structures. We believe that this methodology gives a new view of the interpretability/explainability for model-based clustering.
We have studied how to interpret the overlapping mixture models after they were estimated. It remains for future study to apply merging criteria even in the phase of estimating mixture models.

Author Contributions

Conceptualization, S.K. and K.Y.; methodology, S.K. and K.Y.; software, S.K.; validation, S.K. and K.Y.; formal analysis, S.K.; investigation, S.K. and K.Y.; resources, S.K.; data curation, S.K.; writing—original draft preparation, S.K. and K.Y.; writing—review and editing, S.K. and K.Y.; visualization, S.K.; supervision, K.Y.; project administration, K.Y.; funding acquisition, K.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by JST KAKENHI JP19H01114 and JST-AIP JPMJCR19U4.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All of the datasets used in this paper can be obtained in the manner described in the README file at https://github.com/ShunkiKyoya/summarize_cluster_overlap. Moreover, all of the experimental results can be reproduced by executing the .ipynb files.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Details of the Merging Algorithm

We show the pseudo-code and computational complexity of the merging algorithm. First, the pseudo-code of merging mixture components is shown in Algorithm A1.
Algorithm A1 Merging mixture components
Require: data xN, finite mixture model f, criterion function Crit.
 1: while (The number of components) > 1 do
 2:   i , j : = arg min i < j Crit ( i , j )
 3:  if a certain stopping condition is satisfied then
 4:    return The current components.
 5:  end if
 6:  Merge components i and j.
 7: end while
 8: return The current components.
Next, we discuss the computational complexity in this algorithm given x N and f below. First, the cost of calculating { γ k ( x n ) } k , n can be written as O ( T dist N K ) , where T dist is the cost to calculate f ( x ) for a point. To merge components, it is needed to repeat updating { Crit ( i , j ) } i , j and { γ k ( x n ) } k , n at most ( K 1 ) times. The cost for updating { Crit ( i , j ) } i , j and { γ k ( x n ) } k , n are O ( T crit K 2 ) and O ( N ) , respectively, where T crit is the cost to calculate Crit ( i , j ) for a pair of the components. Overall, we need O ( K ( T dist + T crit K 2 + N ) ) to complete the algorithm.
For the criteria referred to in this section, their computational complexity T crit are O ( N ) for Ent, NEnt1, DEMP2, MC, and NMC (NEnt2), and O ( N K ) for DEMP.

Appendix B. Details of the Datasets in the Real Data Experiment

The datasets used in the real data experiment are summarized in Table A1. We show the detail and preprocessing of them below. All variables in the datasets are normalized after they are selected.
Table A1. Summary of the real dataset, where N denotes the number of points, d denotes the number of features, and K denotes the number of true clusters.
Table A1. Summary of the real dataset, where N denotes the number of points, d denotes the number of features, and K denotes the number of true clusters.
DatasetAbbreviation ( N , d ) K
AISAIS(202, 3)2
Flea beatlesBTL(74, 2)3
CrabsCRB(200, 5)4
DLBCLDLB(7932, 3)4
EcoliECL(327, 6)5
SeedsSDS(210, 7)3
Wisconsin breast cancerWSC(569, 3)2
YeastYST(626, 3)2
The AIS dataset [36] consists of the physical measurements of athletes who trained at the Australian Institute of Sport. Two cluster labels are male and female. As did Lee and McLachlan [16] and Malsiner-Walli et al. [20], we use three variables: BMI, LBM, and body fat percentage (BFat).
The Flea beetles dataset [37] consists of two physical measurements (width and angle) of flea beetles. Three cluster labels are the different species, named Concinna, Heikertingeri, and Heptapotamica.
The Crabs dataset [38] describes five morphological measurements (frontal lobe size, rear width, carapace length, carapace width, and body depth) of 200 crabs. Four cluster labels are formed by combining two color forms and two sexes (male and female).
The DLBCL dataset [39] contains fluorescent intensities of multiple conjugated antibodies (markers) on the cells derived from the lymph nodes of patients diagnosed with DLBCL (diffuse large B-cell lymphoma). As did Lee and McLachlan [40] and Malsiner-Walli et al. [20], we consider four labels corresponding to the cell populations.
The Ecoli dataset [41,42] contains cellular localization sites of proteins. We consider five variables named mcg, gvh, aac, alm1, and alm2. Binary attributes are omitted here. For the labels, we consider five localization sites named cp, im, imU, om, and pp. The other localization sites are omitted because there are little data assigned to them.
The Yeast dataset [41,42] also describes cellular localization sites of proteins. As did Franczac et al. [43] and Malsiner-Walli et al. [20], we select three variables and two cluster labels from the dataset. For the variables, we consider three attributes of proteins, named mcg, alm, and vac. For the labels, we consider two localization sites named CYT and ME3.
The Seeds dataset [44] consists of the seven geometric parameters of grains: area, perimeter, compactness, length of kernel, width of the kernel, asymmetry coefficient, and length of kernel groove. Three cluster labels are kernels belonging to different varieties of wheat: Kama, Rosa, and Canadian.
The Wisconsin breast cancer dataset [3] describes characteristics of the cell nuclei in the images of breast masses. Two cluster labels are benign and malignant. As did Fraley and Raftery [2] and Malsiner-Walli et al. [20], we select three variables: extreme area, extreme smoothness, and mean texture.

References

  1. McLachlan, G.J.; Peel, D. Finite Mixture Models; Wiley Series in Probability and Statistics: New York, NY, USA, 2000. [Google Scholar]
  2. Fraley, C.; Raftery, A.E. How Many Clusters? Which Clustering Method? Answers via Model-Based Cluster Analysis. Comput. J. 1998, 41, 578–588. [Google Scholar] [CrossRef]
  3. Mangasarian, O.L.; Street, W.N.; Wolberg, W.H. Breast Cancer Diagnosis and Prognosis via Linear Programming. Operat. Res. 1995, 43, 570–577. [Google Scholar] [CrossRef] [Green Version]
  4. Hennig, C. Methods for Merging Gaussian Mixture Components. Adv. Data Anal. Class. 2010, 4, 3–34. [Google Scholar] [CrossRef]
  5. Baudry, J.P.; Raftery, A.E.; Celeux, G.; Lo, K.; Gottardo, R. Combining Mixture Components for Clustering. J. Comput. Graph. Stat. 2010, 19, 332–353. [Google Scholar] [CrossRef] [Green Version]
  6. Melnykov, V. Merging Mixtute Components for Clustering Through Pairwise Overlap. J. Comput. Graph. Stat. 2016, 25, 66–90. [Google Scholar] [CrossRef]
  7. Kyoya, S.; Yamanishi, K. Mixture Complexity and Its Application to Gradual Clustering Change Detection. arXiv 2020, arXiv:2007.07467. [Google Scholar]
  8. Biernacki, C.; Celeux, G.; Govaert, G. Assessing a Mixture Model for Clustering With the Integrated Completed Likelihood. IEEE Trans. Patt. Anal. Mach. Intell. 2000, 22, 719–725. [Google Scholar] [CrossRef]
  9. Hirai, S.; Yamanishi, K. Efficient Computation of Normalized Maximum Likelihood Codes for Gaussian Mixture Models With Its Applications to Clustering. IEEE Trans. Inform. Theory 2013, 59, 7718–7727. [Google Scholar] [CrossRef]
  10. Hirai, S.; Yamanishi, K. Correction to Efficient Cotampution of Normalized Maximum Likelihood Codes for Gaussian Mixture Models With Its Applications to Clustering. IEEE Trans. Inform. Theory 2019, 65, 6827–6828. [Google Scholar] [CrossRef]
  11. Wu, T.; Sugawara, S.; Yamanishi, K. Decomposed Normalized Maximum Likelihood Codelength Criterion for Selecting Hierarchical Latent Variable Models. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; Association for Computing Machinery: New York, NY, USA, 2017. [Google Scholar]
  12. Yamanishi, K.; Wu, T.; Sugawara, S.; Okada, M. The Decomposed Normalized Maximum Likelihood Code-Length Criterion for Selecting Hierarchical Latent Variable Models. Data Mining Know. Discov. 2019, 33, 1017–1058. [Google Scholar] [CrossRef]
  13. Banerjee, A.; Krumpelman, C.; Basu, S.; Mooney, R.J.; Ghosh, J. Model-based Overlapping Clustering. In Proceedings of the 11th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, USA, 21–24 August 2005; Association for Computing Machinery: New York, NY, USA, 2005. [Google Scholar]
  14. Fu, Q.; Banerjee, A. Multiplicative Mixture Models for Overlapping Clustering. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008. [Google Scholar]
  15. Xu, Y.; Yang, Y.; Wang, H.; Hu, J. An Overlapping Clustering Approach with Correlation Weight. In Proceedings of the International Joint Conference on Rough Sets, Olsztyn, Poland, 3–7 July 2017; Springer: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  16. Lee, S.X.; McLachlan, G.J. Model-Based Clustering and Classification With Non-Normal Mixture Distributions. Stat. Method Appl. 2013, 22, 427–454. [Google Scholar] [CrossRef]
  17. Li, J. Clustering Based on a Multi-layer Mixture Model. J. Comput. Graph. Stat. 2004, 14, 547–568. [Google Scholar] [CrossRef]
  18. Di Zio, M.; Guarnera, U.; Rocci, R. A Mixture of Mixture Models For a Classifitation Problem. Comput. Stat. Data Anal. 2007, 51, 2573–2585. [Google Scholar] [CrossRef]
  19. Yarebakan, H.Z.; Rajwa, B.; Dundar, M. The Infinite Mixture of Infinite Gaussian Mixtures. In Advances in Neural Information Processing Systems; Curran Associates Inc.: Montréal, QC, Canada, 2014; Volume 27. [Google Scholar]
  20. Malsiner-Walli, G.; Frühwirth-Schnatter, S.; Grün, B. Identifying Mixtures of Mixtures Using Bayesian Estimation. J. Comput. Graph. Stat. 2017, 26, 285–295. [Google Scholar] [CrossRef]
  21. Ueda, N.; Nakano, R.; Ghahramani, Z.; Hingon, G.E. SMEM Algorithm for Mixture Models. Neur. Comput. 2000, 12, 2109–2128. [Google Scholar] [CrossRef] [PubMed]
  22. Minagawa, A.; Tagawa, N.; Tanaka, T. SMEM Algorithm Is Not Fully Compatible with Maximum-Likelihood Framework. Neur. Comput. 2002, 14, 1261–1266. [Google Scholar] [CrossRef]
  23. Zhao, Q.; Hautamäki, V.; Kärkkäinen, I.; Fränti, P. Random Swap EM algorithm for Gaussian Mixture Models. Pattern Recognit. Lett. 2012, 33, 2120–2126. [Google Scholar] [CrossRef]
  24. Heller, K.A.; Ghahramani, Z. Bayesian Hierarchical Clustering. In Proceedings of the 22nd International Conference on Machine Learning, Bonn, Germany, 7–11 August 2005; Association for Computing Machinery: New York, NY, USA, 2005. [Google Scholar]
  25. Telgarsky, M.; Dasgupta, S. Agglomerative Bregman Clustering. In Proceedings of the 29th International Conference on Machine Learning, Edinburgh, UK, 26 June–1 July 2012; Association for Computing Machinery: New York, NY, USA, 2012. [Google Scholar]
  26. Dhillon, I.S.; Modha, D.S.; Spangler, W.S. Class Visualization of High-Dimensional Data With Applications. Comput. Stat. Data Anal. 2002, 41, 59–90. [Google Scholar] [CrossRef]
  27. Iwata, T.; Saito, K.; Ueda, N.; Stromsten, S.; Griffiths, T.L.; Tenenbaum, J.B. Parametric Embedding for Class Visualization. Neural Comput. 2007, 19, 2536–2556. [Google Scholar] [CrossRef]
  28. Ward, J.H., Jr. Hierarchical Grouping to Optimize an Objective Function. J. Am. Stat. Associat. 1963, 58, 236–244. [Google Scholar] [CrossRef]
  29. Sneath, P.H.A.; Sokal, R.R. Numerical Taxonomy: The Principles and Practice of Numerical Classification; W. H. Freeman: San Francisco, CA, USA, 1973. [Google Scholar]
  30. Hubert, L.; Arabie, P. Comparing Partitions. J. Class. 1985, 2, 193–218. [Google Scholar] [CrossRef]
  31. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: New York, NY, YSA, 2006. [Google Scholar]
  32. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  33. Schwarz, G. Estimating the Dimension of a Model. Annals Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  34. Puzicha, J.; Hofmann, T.; Buhman, J.M. A theory of proximity based clustering: structure detection by optimization. Pattern Recognit. 2000, 33, 617–634. [Google Scholar] [CrossRef]
  35. Kleinberg, J. An Impossibility Theorem for Clustering. In Advances in Neural Information Processing Systems 15; Curran Associates Inc.: Vancouver, BC, Canada, 2002. [Google Scholar]
  36. Cook, R.D.; Weisberg, S. An Introduction to Regression Graphics; Wiley-Interscience: New York, NY, USA, 1994. [Google Scholar]
  37. Lubischew, A.A. On the Use of Discriminant Functions in Taxonomy. Biometrics 1962, 18, 455–477. [Google Scholar] [CrossRef]
  38. Campbell, N.A.; Mahon, R.J. A Multivariate Study of Variation in Two Species of Rock Crab of The Genus Leptograpsus. Austral. J. Zool. 1974, 22, 417–425. [Google Scholar] [CrossRef]
  39. Aghaeepour, N.; Finak, G.; Consortium, T.F.; Consortium, T.D.; Hoos, H.; Mosmann, T.R.; Brinkman, R.; Gottardo, R.; Scheuermann, R.H. Critical Assessment of Automated Flow Cytometry Data Analysis Techniques. Nat. Meth. 2013, 10, 228–243. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Lee, S.X.; McLachlan, G.J. EMMIXuskew: An R Package for Fitting Mixtures of Multivariate Skew t Distributions via the EM Algorithm. J. Stat. Softw. 2013, 55, 1–22. [Google Scholar] [CrossRef] [Green Version]
  41. Nakai, K.; Kanehisa, M. Expert System for Predicting Protein Localization Sites in Gram-Negative Bacteria. Protein. Struct. Funct. Genet. 1991, 11, 95–110. [Google Scholar] [CrossRef]
  42. Nakai, K.; Kanehisa, M. A Knowledge Base for Predicting Protein Localization Sites in Eukaryotic Cells. Genomics 1992, 14, 897–911. [Google Scholar] [CrossRef]
  43. Franczak, B.C.; Browne, R.P.; and, P.D.M. Mixtures of Shifted Asymmetric Laplace Distributions. IEEE Trans. Patt. Anal. Mach. Intell. 2014, 36, 1149–1157. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Charytanowicz, M.; Niewczas, J.; Kulczycki, P.; Kowalski, P.A.; ukasik, S.; Żak, S. Complete Gradient Clustering Algorithm for Features Analysis of X-Ray Images. Informat. Technol. Biomed. 2010, 69, 15–24. [Google Scholar]
Figure 1. Estimated Gaussian components for the Wisconsin breast cancer dataset [3].
Figure 1. Estimated Gaussian components for the Wisconsin breast cancer dataset [3].
Entropy 23 01503 g001
Figure 2. Upper-components and sub-components.
Figure 2. Upper-components and sub-components.
Entropy 23 01503 g002
Figure 3. Examples of MC for mixtures of two components. Images are obtained from [7].
Figure 3. Examples of MC for mixtures of two components. Images are obtained from [7].
Entropy 23 01503 g003
Figure 4. Example of the merged mixture model. Images are obtained from [7].
Figure 4. Example of the merged mixture model. Images are obtained from [7].
Entropy 23 01503 g004
Figure 5. Scatter plots for the cluster with K = 20 whose intra-cluster distance is the largest. The thickness of the color corresponds to the posterior probabilities. The numbers in the parenthesis show D intra .
Figure 5. Scatter plots for the cluster with K = 20 whose intra-cluster distance is the largest. The thickness of the color corresponds to the posterior probabilities. The numbers in the parenthesis show D intra .
Entropy 23 01503 g005
Figure 6. The proportions of the data x n that satisfy D nn ( 5 ) ( x n ) / [ ( 1 / N ) n D nn ( 5 ) ( x n ) ] > 2.0, 3.0, 4.0, 5.0.
Figure 6. The proportions of the data x n that satisfy D nn ( 5 ) ( x n ) / [ ( 1 / N ) n D nn ( 5 ) ( x n ) ] > 2.0, 3.0, 4.0, 5.0.
Entropy 23 01503 g006
Figure 7. Predicted cluster labels for the Flea beetles dataset.
Figure 7. Predicted cluster labels for the Flea beetles dataset.
Entropy 23 01503 g007
Figure 8. Predicted cluster labels for the Wisconsin breast cancer dataset.
Figure 8. Predicted cluster labels for the Wisconsin breast cancer dataset.
Entropy 23 01503 g008
Figure 9. Scatter plots of the MC/NMC and the entropy of the true cluster label.
Figure 9. Scatter plots of the MC/NMC and the entropy of the true cluster label.
Entropy 23 01503 g009
Table 1. Summary of the relationships between the criteria and the essential conditions. Check marks are attached to the conditions that are satisfied.
Table 1. Summary of the relationships between the criteria and the essential conditions. Check marks are attached to the conditions that are satisfied.
Before ModificationAfter Modification
criterionBOWSSIcriterionBOWSSI
Ent NEnt1
DEMP(✓)DEMP2
MC NMC = NEnt2
Table 2. Example of a clustering summarization.
Table 2. Example of a clustering summarization.
Upper-Components
MC (exp):0.647 (1.91)
NMC:0.933
Component 1Component 2
Weight:0.494Weight:0.506
MC (exp):0.566 (1.76)MC (exp):0.324 (1.38)
NMC:0.817NMC:0.467
Table 3. Average ranks of the criteria. For each K, the best rank is denoted in boldface.
Table 3. Average ranks of the criteria. For each K, the best rank is denoted in boldface.
D intra
K403020105
Ent6.006.006.005.945.58
NEnt13.674.153.934.154.18
DEMP4.304.564.072.572.13
DEMP22.402.522.452.121.91
MC2.252.043.214.534.94
NMC2.371.731.351.692.27
D inter
K403020105
Ent5.025.145.385.505.19
NEnt14.294.234.264.563.99
DEMP4.994.974.822.882.63
DEMP23.563.553.172.852.48
MC2.092.062.353.904.86
NMC1.061.041.031.291.84
Table 4. Overview of the comparison methods.
Table 4. Overview of the comparison methods.
AbbreviationMethodReference
GMM + BICGMM and BIC criterion[33]
GMM + DNMLGMM and DNML criterion[11,12]
MixMixMixture of Gaussian mixture models[20]
Table 5. Estimated number of clusters. Merge (best F-measure) is the number of clusters when F-measure is highest. Merge (best ARI) is the number of clusters when ARI is highest. Merge (NMC) is the number of clusters obtained by the NMC-based stopping condition.
Table 5. Estimated number of clusters. Merge (best F-measure) is the number of clusters when F-measure is highest. Merge (best ARI) is the number of clusters when ARI is highest. Merge (NMC) is the number of clusters obtained by the NMC-based stopping condition.
DatasetAISBTLCRBDLCECLSDSWSCYST
K 23445322
GMM + BIC3.02.63.06.64.02.03.03.5
GMM + DNML1.01.01.02.84.01.02.01.0
MixMix2.71.21.07.44.53.22.12.8
Merge
(Best F-measure)
Ent18.020.019.117.619.019.419.218.6
NEnt12.03.05.710.28.96.75.38.8
DEMP2.03.04.34.96.24.73.42.3
DEMP22.03.04.34.95.74.12.72.0
MC2.03.05.63.64.97.12.72.0
NMC3.03.04.37.16.44.32.02.1
Merge
(Best ARI)
Ent19.020.019.417.619.019.419.218.6
NEnt13.03.05.73.27.06.75.38.8
DEMP2.43.04.34.96.24.93.62.3
DEMP22.03.04.35.06.14.52.72.0
MC2.03.05.63.64.97.14.02.0
NMC4.03.04.37.06.44.42.02.1
Merge
(NMC)
Ent19.020.017.817.619.018.119.218.6
NEnt14.03.03.66.94.74.95.68.9
DEMP4.83.03.39.15.75.34.48.0
DEMP25.03.03.39.25.75.23.68.5
MC5.03.06.211.39.08.66.19.7
NMC4.03.03.16.44.33.22.96.9
Table 6. F-measure for the real datasets. For each merging algorithm, scores that exceed all comparison methods are denoted in boldface.
Table 6. F-measure for the real datasets. For each merging algorithm, scores that exceed all comparison methods are denoted in boldface.
DatasetAISBTLCRBDLCECLSDSWSCYST
GMM + BIC0.9120.8050.8100.7340.7870.7940.8570.864
GMM + DNML0.6710.5900.4000.9030.7870.5000.9140.850
MixMix0.9250.5780.4000.7610.8290.8490.9470.826
Merge
(Best)
Ent0.9160.9860.8660.9310.8740.9000.8970.867
NEnt10.9010.9860.8890.9220.8600.9280.9040.868
DEMP0.9060.9860.8770.9520.8740.9080.9050.942
DEMP20.9060.9860.8770.9520.8750.9120.9050.944
MC0.9310.9860.8630.9210.8700.8860.8860.928
NMC0.9160.9860.8930.9490.8750.9380.9450.942
Merge
(NMC)
Ent0.8920.9860.8220.9310.8740.8520.8970.867
NEnt10.8920.9860.8280.9050.8230.8220.9040.868
DEMP0.8220.9860.8230.7580.8670.8860.8810.870
DEMP20.8050.9860.8220.7540.8670.8920.8800.820
MC0.8030.9860.8310.7060.8600.8780.8580.771
NMC0.8920.9860.8280.9160.8480.8100.9250.878
Table 7. ARI for the real datasets. For each merging algorithm, scores that exceed all comparison methods are denoted in boldface.
Table 7. ARI for the real datasets. For each merging algorithm, scores that exceed all comparison methods are denoted in boldface.
Dataset ( K )AISBTLCRBDLCECLSDSWSCYST
GMM + BIC0.7430.6030.5950.5060.5900.5420.6170.516
GMM + DNML0.0000.0000.0000.8700.5900.0000.6850.000
MixMix0.7510.1100.0000.5010.6730.6230.7990.589
Merge
(Best)
Ent0.7000.9580.7070.9130.7590.7670.6880.508
NEnt10.7010.9580.7390.8520.7190.8100.6880.511
DEMP0.6660.9580.7320.9340.7630.7820.7340.769
DEMP20.6570.9580.7310.9360.7630.7880.7320.773
MC0.7410.9580.7000.8490.7440.7450.6640.709
NMC0.7000.9580.7480.9280.7690.8320.7910.760
Merge
(NMC)
Ent0.7000.9580.6260.9130.7590.6570.6880.508
NEnt10.7000.9580.6420.8340.6380.5940.6880.511
DEMP0.5760.9580.6400.5230.7540.7280.6700.514
DEMP20.5450.9580.6390.5210.7540.7450.6700.402
MC0.5340.9580.6600.4520.7000.7280.6330.313
NMC0.7000.9580.6430.8780.7250.5750.7320.524
Table 8. Clustering summarization for the Flea beetles dataset.
Table 8. Clustering summarization for the Flea beetles dataset.
Upper-Components
MC (exp):0.963 (2.62)
NMC:0.897
Component 1Component 2Component 3
Weight:0.440Weight:0.268Weight:0.293
MC:0.057MC:0.000MC:0.000
(exp)(1.06)(exp)(1.00)(exp)(1.00)
NMC:0.209NMC:-NMC:-
Table 9. Clustering summarization for the Wisconsin breast cancer dataset.
Table 9. Clustering summarization for the Wisconsin breast cancer dataset.
Upper-Components
MC (exp):0.509 (1.66)
NMC:0.763
Component 1Component 2
Weight:0.387Weight:0.613
MC (exp):0.714 (2.04)MC (exp):0.270 (1.31)
NMC:0.613NMC:0.676
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kyoya, S.; Yamanishi, K. Summarizing Finite Mixture Model with Overlapping Quantification. Entropy 2021, 23, 1503. https://0-doi-org.brum.beds.ac.uk/10.3390/e23111503

AMA Style

Kyoya S, Yamanishi K. Summarizing Finite Mixture Model with Overlapping Quantification. Entropy. 2021; 23(11):1503. https://0-doi-org.brum.beds.ac.uk/10.3390/e23111503

Chicago/Turabian Style

Kyoya, Shunki, and Kenji Yamanishi. 2021. "Summarizing Finite Mixture Model with Overlapping Quantification" Entropy 23, no. 11: 1503. https://0-doi-org.brum.beds.ac.uk/10.3390/e23111503

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop