Next Article in Journal
DFTSA-Net: Deep Feature Transfer-Based Stacked Autoencoder Network for DME Diagnosis
Next Article in Special Issue
Region Adaptive Single Image Dehazing
Previous Article in Journal
A Novel Method for Colorectal Cancer Screening Based on Circulating Tumor Cells and Machine Learning
Previous Article in Special Issue
Significance Support Vector Regression for Image Denoising
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Penalized Matrix Normal Mixture Model for Clustering Matrix Data

Department of Mathematics and Statistics, Chonnam National University, 77 Yongbong-ro, Buk-gu, Gwangju 61186, Korea
*
Author to whom correspondence should be addressed.
Submission received: 19 August 2021 / Revised: 17 September 2021 / Accepted: 22 September 2021 / Published: 26 September 2021
(This article belongs to the Special Issue Information Theory in Signal Processing and Image Processing)

Abstract

:
Along with advances in technology, matrix data, such as medical/industrial images, have emerged in many practical fields. These data usually have high dimensions and are not easy to cluster due to their intrinsic correlated structure among rows and columns. Most approaches convert matrix data to multi dimensional vectors and apply conventional clustering methods to them, and thus, suffer from an extreme high-dimensionality problem as well as a lack of interpretability of the correlated structure among row/column variables. Recently, a regularized model was proposed for clustering matrix-valued data by imposing a sparsity structure for the mean signal of each cluster. We extend their approach by regularizing further on the covariance to cope better with the curse of dimensionality for large size images. A penalized matrix normal mixture model with lasso-type penalty terms in both mean and covariance matrices is proposed, and then an expectation maximization algorithm is developed to estimate the parameters. The proposed method has the competence of both parsimonious modeling and reflecting the proper conditional correlation structure. The estimators are consistent, and their limiting distributions are derived. We applied the proposed method to simulated data as well as real datasets and measured its clustering performance with the clustering accuracy (ACC) and the adjusted rand index (ARI). The experiment results show that the proposed method performed better with higher ACC and ARI than those of conventional methods.

1. Introduction

Among the clustering methods, the Gaussian mixture model, which is model-based clustering, is widely used thanks to its easy interpretability. Since most clustering approaches suffer from the curse of dimensionality in “high dimension/low sample size data”, so does model-based clustering. The problem is further exacerbated in a Gaussian mixture model that has to estimate multiple covariance matrices. For example, given a p-dimensional random sample X i = ( X i 1 , , X i p ) T , i = 1 , , n for a multivariate Gaussian mixture model with G components, we need to estimate a total of G p ( p + 1 ) / 2 + G p + G 1 parameters. So, the number of parameters to be estimated grows quickly with dimension p. To address this problem, a penalized Gaussian mixture clustering model has been proposed that imposes constraints on the mean or precision matrix. In particular, Ref. [1] suggests a penalized multivariate normal mixture clustering model for variable selection by regularizing the mean structure. In each cluster, there may be some variables that are not relevant to distinguish it from other clusters. So, finding the non-informative (noise) variables and removing them may largely improve interpretability. Additionally, Ref. [2] suggests a penalized multivariate normal mixture clustering model for achieving better clustering performance by regularizing the precision matrices. In another way, Ref. [3] adopts the idea of a penalized covariance estimation from [4] and applies a lasso-type penalty to the off-diagonal component of the precision matrices.
However, as the dimension p increases, the computational speed extremely slows down, making it difficult to estimate the parameters practically for these methods. In recent years, along with advances in technology, matrix data have emerged in many practical fields. The conventional Gaussian mixture method converts these matrix data into vector data. So, the dimension of the data can be extremely large, i.e., a p × q matrix will be converted to a vector of p q dimension. In addition, clustering a matrix dataset is difficult because correlations between variables or occasions can change across occasions. To solve the aforementioned issues, Ref. [5] suggests finite mixtures of matrix normal distributions for matrix data. Direct analysis of three-way data structures by using a matrix normal distribution allows considering and estimating correlations between variables and occasions. Recently, the matrix mixture model was extended for various non-normal matrix data: Refs. [6,7] proposed finite mixture of matrix skewed distributions, and [8] introduced two matrix-variate distributions—both elliptical heavy-tailed generalizations of the matrix-variate normal distribution that are used in a finite mixture model. Ref. [9] suggested a penalized matrix normal mixture model by imposing a penalty for the mean matrix. This method can capture the row-wise and column-wise correlation simultaneously and has the ability to find the sparsity nature that is inherent in the signals and images. However, the approach of [9] shows weakness for high dimension/low sample size data because it cannot reduce the number of precision parameters to estimate.
In a multivariate normal distribution, the conditional correlation between two variables, given the other variables, is explicitly determined by the precision matrix. That is, if the ( i , j ) th element of the precision matrix is zero, then the conditional correlation between the ith and the jth variables given the other variables is zero, and they are conditionally independent. Ref. [10] introduced the idea of covariance selection, which consists of inferring a sparse estimation of the precision matrix and interpreting its sparsity structure as a conditional dependency between variables. A Gaussian graphical model is a network whose values on the vertices follow a centered multivariate normal distribution, and has been applied to analyze high-dimensional data with small sample size through the regularized estimation of the precision matrix [11]. A notable approach for the precision matrix estimation is the graphical lasso algorithm [12].
In this paper, the work of [9] is extended by further regularizing the precision matrix as in the Gaussian graphical model, to cope better with the curse of dimensionality for large size images. Specifically, we propose a penalized matrix normal mixture model with lasso-type penalty terms in both mean and covariance matrices, and then an expectation maximization algorithm is developed to estimate the parameters. The estimators are consistent and their limiting distributions are derived. The experiment results show that the proposed approach attains much improved clustering performance in comparison with conventional methods.
The capability of the proposed model is compared with two popular distance-based methods and two conventional model-based methods in Appendix A Table A1. The major contributions of this research to overcome the limitations of the conventional methods are as follows:
(1)
Since the data type for the Gaussian mixture model is a vector and K-means and spectral clustering methods use dissimilarity and similarity between the observations as input, these methods do not have the modeling ability for the row-wise/column-wise covariance structure of matrix data (images). On the other hand, both [9] and the proposed model can capture the row-wise and column-wise correlation of image pixels simultaneously by estimating the row covariance U and column covariance V, where the covariance matrix Σ is the Kronecker product of U and V. This capability of modeling separable row-wise/column-wise covariance structure helps for better clustering of image data.
(2)
Because the proposed model is a model-based clustering method, it has the sound mathematical basis and the interpretability of results with the estimated parameters, whereas centroid-based and similarity-based methods do not.
(3)
If variables are measured on different scales and one variable has much larger values than the others, then the first variable will dominate the clustering. Therefore, it is preferred to use a clustering method that is invariant against affine transformations. K-means and similarity-based methods using the Euclidean distance are not invariant against affine transformations, whereas Gaussian mixture models are.
(4)
Model-based clustering methods have free parameters to estimate and often have difficulty in the estimation of parameters when the data are of a high dimension but the sample is small. Given p × q , random matrix data X, and the number of clusters G, the conventional Gaussian mixture model needs to estimate a total of G p q ( p q + 1 ) / 2 + G p q + G 1 parameters because it converts the data into a vector of p q dimension. The method [9] is regularized on the mean matrix, and thus it needs to estimate G p ( p + 1 ) / 2 + G q ( q + 1 ) / 2 + G p q + G 1 r parameters, where r is the number of the mean parameters values that are shrunken to zero. Since the proposed method is regularized further on the precision matrix, that is, on the row and column covariances, U and V, with shrunken elements, it needs to estimate G p ( p + 1 ) / 2 + G q ( q + 1 ) / 2 + G p q + G 1 r * parameters, where r * is the number of the mean and precision parameters values that are shrunken to zero. The range of r and r * is 0 r G p q and r r * G p ( p + 1 ) / 2 + G q ( q + 1 ) / 2 + G p q , respectively. Therefore, the proposed approach is a more parsimonious model that has the fewest parameters and is able to cope better with the curse of dimensionality for large size images.
(5)
There often exist correlations between the rows and the columns of matrix data. A separable covariance structure of the proposed model can effectively reduce parameters, speed up calculations, and provide practical interpretability. Therefore, the proposed method has the competence of both parsimonious modeling to overcome the curse of dimensionality and reflecting the proper conditional correlation structure for high dimension/low sample size data in comparison with the conventional methods.
The paper is organized as follows. In Section 2, the matrix normal mixture model is explained and then the penalized matrix normal mixture model with the mean and precision matrices is introduced. Next, a new EM algorithm for the precision matrices of the proposed model, a model selection method, k-fold cross-validation, and the asymptotic theory are developed. In Section 3, the proposed method is applied to both simulation data and real data. In the simulation data experiment, the clustering performance is compared with that of the K-means clustering method [13], the spectral clustering method [14], and [9]’s method. The consistency of the estimates for the proposed model is shown, using the averaged spectral norm of the difference (SL) and the averaged Frobenius norm of the difference (FL) [3]. In the real data experiment, the clustering performance is also compared with that of the K-means clustering method, the spectral clustering method and the method of [9]. In Section 4, we summarize the results and discuss future work. The Appendix A, Appendix B and Appendix C provides a comparative overview table and the proofs of the penalized maximum likelihood estimator of the mean matrix for the proposed model and the asymptotic theory.

2. Method

A mixture model using matrix normal distribution was previously introduced by [5]. The matrix normal mixture model [15,16,17] is reviewed is shown in Section 2.1. Then, our penalized matrix normal mixture model, estimation, asymptotic theory, and model selection are presented in Section 2.2, Section 2.3 and Section 2.4.

2.1. Matrix Normal Mixture Model (MNMM)

2.1.1. Definition of MNMM

The p × q random matrix X has a matrix normal distribution, with mean matrix M and row covariance U and column covariance V, denoted by X N p , q ( M , U , V ) , if the p.d.f. of X is given by the following:
f ( X | M , U , V ) = ( 2 π ) p q 2 | V | p 2 | U | q 2 e x p 1 2 t r ( V 1 ( X M ) T U 1 ( X M ) ) ,
where M R p × q , U R p × p , V R q × q and | A | , t r ( A ) are the determinant value and the trace of a matrix A, respectively. The p.d.f. (1) can be converted to the p.d.f. of the vectorization of X as follows:
v e c ( X ) N p q ( v e c ( M ) , V U ) ,
where v e c is the vectorization operator and ⊗ is the Kronecker product. The matrix normal distribution is a generalization of the matrix variate, i.e., it can effectively handle three-way data. In particular, the separable covariance structure has an important advantage, which is reducing the number of parameters. For example, given the p × q random matrix X, the dimension of inseparable covariance matrix Σ is p q × p q . However, the dimension of separable covariance matrix Σ = V U is p × p + q × q .

2.1.2. Maximum Likelihood Estimation

The log-likelihood function for given i.i.d observations X 1 , X 2 , , X n of N p , q ( M , U , V ) is as follows:
L ( M , U , V ; X ) = n p q 2 l o g 2 π n p 2 l o g | V | n q 2 l o g | U | 1 2 i = 1 n t r ( V 1 ( X i M ) T U 1 ( X i M ) .
The maximum likelihood estimators (m.l.e.) for M, U, V are as follows:
M ^ = X ¯ , U ^ ( t + 1 ) = 1 n q i = 1 n ( X i X ¯ ) ( V ^ ( t ) ) 1 ( X i X ¯ ) T , V ^ ( t + 1 ) = 1 n p i = 1 n ( X i X ¯ ) T ( U ^ ( t + 1 ) ) 1 ( X i X ¯ ) .
Note that m.l.e., U ^ and V ^ are of the no closed-form solution, and a V ^ and b U ^ , which is a constant multiple of U ^ and V ^ , are also solutions. However, a V ^ b U ^ must be the same as V ^ U ^ [15]. U ^ and V ^ can be obtained by the iterative algorithm of (2) until the convergence criterion is satisfied. The criterion is set as U ^ ( t ) U ^ ( t + 1 ) F / U ^ ( t ) F < 0.01 and V ^ ( t ) V ^ ( t + 1 ) F / V ^ ( t ) F < 0.01 .

2.1.3. EM Algorithm of MNMM

Given i.i.d p × q observations X 1 , X 2 , , X n from the mixture of matrix normal distributions with G groups, the p.d.f. of the matrix normal mixture is as follows:
g ( X ) = j = 1 G π j f ( X | Θ j ) ,
where Θ j is the set of parameters corresponding to the jth group, i.e., Θ j = ( M j , U j , V j ) , π 1 , , π G are the weights of belonging to the group g, and f is the matrix normal distribution with mean M j , row covariance U j , column covariance V j . Then, the log-likelihood function of MNMM is as follows:
L ( Θ ; X ) = i = 1 n l o g j = 1 G π j f ( X i | Θ j ) .
It is difficult to obtain the maximum log-likelihood value by solving the first derivative Equation (3) because the p.d.f is defined as the sum in logarithm. So, the EM algorithm is widely used to estimate parameters by maximizing the conditional expectation for the latent variable z, given the data instead of the log-likelihood function [18].
In the E-step, the posterior probability of the latent variables z i j is defined by Bayes’ theorem as follows:
τ ^ i j ( t ) = π ^ j ( t ) f ( X i | Θ ^ j ( t ) ) l = 1 G π ^ l ( t ) f ( X i | Θ ^ l ( t ) ) .
In the M-step, the parameters, Θ ^ ( t + 1 ) , that maximize the expected conditional log-likelihood given data are found as follows:
Θ ^ ( t + 1 ) = a r g m a x Θ ^ j i = 1 n j = 1 G τ ^ i j ( t ) l o g π j f ( X i | Θ j ) .
We obtain the following estimates by using the partial derivative and some algebra.
π ^ j ( t + 1 ) = i = 1 n τ ^ i j ( t ) n , M ^ j ( t + 1 ) = i = 1 n τ ^ i j ( t ) X i i = 1 n τ ^ i j ( t ) , U ^ j ( t + 1 ) = i = 1 n τ ^ i j ( t ) ( X i M ^ j ( t + 1 ) ) ( V ^ j ( t ) ) 1 ( X i M ^ j ( t + 1 ) ) T q i = 1 n τ ^ i j , V ^ j ( t + 1 ) = i = 1 n τ ^ i j ( t ) ( X i M ^ j ( t + 1 ) ) T ( U ^ j ( t + 1 ) ) 1 ( X i M ^ j ( t + 1 ) ) r i = 1 n τ ^ i j .
Similar to the m.l.e. of the matrix normal distribution, U ^ j and V ^ j are obtained by updating the previously estimated parameters in the EM algorithm until a convergence criterion is satisfied.

2.2. Penalized Matrix Normal Mixture Model (PMNMM)

Let M g ( k , l ) be the ( k , l ) th element of the mean matrix of group g, M g . If M g ( k , l ) is the same for all groups, i.e., M 1 ( k , l ) = = M G ( k , l ) , then the ( k , l ) th element of the mean matrix cannot differentiate the clusters. In this case, when the data are standardized, the ( k , l ) th element of the standardized mean matrix of each cluster will be 0, and is known to be the noise variable (non-informative) for clustering [2]. Therefore, penalizing on the elements of mean matrix of the matrix normal mixture was proposed by [9] for identifying the underlying spatial structure and improving the signal-to-noise ratio (SNR) of the local field potentials (LFP) signals. However, in order to improve the clustering performance in high-dimension/low sample size data, the regularization of the precision matrix is inevitable because the number of parameters required to estimate the multiple covariance matrices becomes very large. In this paper, the method of [9] is extended by further regularizing the precision matrix to cope better with the curse of dimensionality and reflect the proper conditional correlation structure for matrix data.

2.2.1. Penalized Log-Likelihood Function of PMNMM

The penalized log-likelihood function is considered as follows:
L p ( Θ ; Λ , X ) = i = 1 n l o g j = 1 G π j f ( X i | Θ j ) λ 1 j = 1 G | M j | 1 λ 2 j = 1 G | U j 1 | 1 λ 3 j = 1 G | V j 1 | 1 ,
where λ 1 , λ 2 , λ 3 > 0 are the tuning parameters, and | A | 1 is the L1 norm, which is the sum of all absolute values of elements of A. Setting λ 2 , λ 3 = 0 is equivalent to the penalized matrix normal mixture model with lasso regularization of the mean matrix proposed in [9]. Additionally, setting λ 1 , λ 2 , λ 3 = 0 is equivalent to the non-regularized matrix mixture model proposed in [5].

2.2.2. EM Algorithm of PMNMM

The EM algorithm is employed as in Section 2.1.3. The E-step is the same as Equation (4). In the M-step, the parameter estimates can be obtained by solving the constraint optimization problem as follows:
Θ ˜ ( t + 1 ) = a r g m a x M j , U j & V j > 0 L ( Θ ; X ) λ 1 j = 1 G | M j | 1 λ 2 j = 1 G | U j 1 | 1 λ 3 j = 1 G | V j 1 | 1 ,
where U j & V j > 0 means that U j and V j are symmetric and positive definite matrices. The estimate of M can be obtained by the subgradient approach as shown in Appendix B as follows:
( M ˜ j ( t + 1 ) ) ( k , l ) = s i g n ( M ^ j ( t ) ) ( k , l ) a b s ( M j ^ ( t ) ) λ 1 i = 1 n τ i j U j ^ ( t ) 1 p × q V j ^ ( t ) ( k , l ) + , j = 1 , , G ,
where M j ^ ( t ) = i = 1 n τ ^ i j X i i = 1 n τ ^ i j is the m.l.e. of MNMM, A ( k , l ) + = m a x ( A ( k , l ) , 0 ) , 1 p × q is p × q the matrix with all components of 1, and s i g n ( A ) ( k , l ) is the function that extracts the sign of a real number, and a b s ( A ) is the absolute value of A. Then, we focus on the constraint optimization problem with respect to each U and V. To maximize the log-likelihood with the constraint, the problem is redefined as the constraint optimization with respect to U j 1 and V j 1 as follows:
a r g m a x U j > 0 j = 1 G q i = 1 n τ ^ i j ( t ) 2 l o g | U j 1 | q i = 1 n τ ^ i j ( t ) 2 t r ( S ˜ j ( U ) U j 1 ) λ 2 j | U j 1 | 1
with
S ˜ j ( U ) = i = 1 n τ ^ i j ( t ) ( X i M ˜ j ( t ) ) ( V ^ j 1 ) ( t ) ( X i M ˜ j ( t ) ) T q i = 1 n τ ^ i j ( t )
and
a r g m a x V j > 0 j = 1 G p i = 1 n τ ^ i j ( t ) 2 l o g | V j 1 | p i = 1 n τ ^ i j ( t ) 2 t r ( S ˜ j ( V ) V j 1 ) λ 3 j | V j 1 | 1
with
S ˜ j ( V ) = i = 1 n τ ^ i j ( t ) ( X i M ˜ j ( t ) ) T ( U ^ j 1 ) ( t + 1 ) ( X i M ˜ j ( t ) ) p i = 1 n τ ^ i j ( t ) .
However, for Equations (6)–(9), the subgradient approach is not possible because the solution of the constraint optimization problem must satisfy the conditions that U j and V j are symmetric and positive definite matrices. So, the graphical lasso algorithm [12] is applied to maximize the following constraint optimization function with respect to positive definite matrix Σ .
l o g | Σ 1 | t r ( S Σ 1 ) λ | Σ 1 | 1 ,
where S is the sample covariance matrix. Therefore, we can apply the algorithm by substituting S = S ˜ j ( U ) , Σ = U , λ = 2 λ 2 / q i = 1 n τ ^ i j ( t ) and S = S ˜ j ( V ) , Σ = V , λ = 2 λ 3 / p i = 1 n τ ^ i j ( t ) to estimate U j 1 and V j 1 , respectively. The graphical lasso algorithm is implemented with R package glasso [12]. Then the m.l.e. of U 1 and V 1 are obtained by iterating E-step and M-step until the penalized log-likelihood converges. To prevent the convergence of the parameter estimation from prematurely stopping, Aitken’s acceleration [19] is used. Aitken’s acceleration computes the acceleration coefficient at iteration t as follows:
a ( t ) = l ( t ) l ( t 1 ) l ( t 1 ) l ( t 2 ) ,
where l ( t ) is the penalized log-likelihood value at iteration t. The accelerated maximum of the penalized log-likelihood at iteration t is as follows:
l ^ ( t ) = l ( t 1 ) + l ( t ) l ( t 1 ) 1 a ( t ) .
The iteration stops when 0 ( l ^ ( t ) l ( t ) ) / | l ( t ) | ϵ for very small positive number ϵ . We set ϵ = 0.001 .
The iterative parameter estimation procedure needs the initial starting values. So the K-means clustering method is used to obtain the initial parameter values. For the fixed number of clusters G, the vectorized data are divided into G clusters, using K-means. Then, based on a group label that is predicted, the m.l.e. are obtained for each cluster; we use those values as the initial value set for the EM algorithm. Additionally, in the process of estimating the parameters that maximize the objective function, the EM algorithm has to be run multiple times because local maximum values may exist.

2.3. Asymptotic Theory

The consistency of the penalized likelihood estimator for the mean of the mean-regularized matrix normal mixture model [9] under the following mild conditions was shown by the direct application of Theorem 1 in [20]. That is, the parameter space Ψ d 1 , d 2 must be defined as follows:
Ψ d 1 , d 2 = { ( π 1 , , π G , M 1 , , M G , V 1 U 1 , , V G U G ) Ψ d 1 , d 2 : σ i ( U h ) σ i ( V h ) = c h f o r i = 1 , , m i n p , q a n d h = 1 , , G }
to guarantee the identifiability of the matrix normal mixture model.
Since Theorem 1 in [20] is applicable to all lasso type penalized likelihood estimators, the penalized likelihood estimators of the proposed model are also consistent under the condition (10). We further establish the asymptotic distributions of the estimators, and prove their consistency in another way. We derive asymptotic distributions of the proposed estimators, using the techniques in [4]. The proof is given in Appendix C.
Theorem 1.
Let X 1 , , X n be a random sample from a mixture matrix normal distribution. Let M ˜ j , U ˜ j * , and V ˜ j * be the penalized maximum likelihood estimator of M j , U j 1 , and V j 1 , respectively, j = 1 , , G . Assume that the true parameter value M j , U j , V j Ψ ¯ d 1 , d 2 , and n λ i λ i 0 0 as n , i = 1 , 2 , 3 ; then,
(1) n ( M ˜ j M j ) d a r g m i n H ( V 1 ) ,
where
V 1 ( H ) = t r ( U j 1 H V j 1 W 1 T ) t r ( V j 1 H T U j 1 W 1 ) + t r ( V j 1 H T U j 1 H ) + λ 10 l m H ( l , m ) s i g n ( M ˜ j ( l , m ) ) I ( M ˜ j ( l , m ) 0 ) + | H ( l , m ) | I ( M ˜ j ( l , m ) = 0 ) ,
in which H is a p × q matrix, H ( l , m ) is the ( l , m ) th element of H, W 1 is a random p × q matrix such that v e c ( W 1 ) N ( 0 , Λ 1 ) , and Λ 1 is such that the following holds:
c o v ( W 1 ( i , j ) , W 1 ( i , j ) ) = c o v ( X ( i , j ) M j ( i , j ) , X ( i , j ) M j ( i , j ) ) ,
(2) n ( U ˜ j * U j 1 ) d a r g m i n L = L ( V 2 ) ,
where
V 2 ( L ) = t r ( L U j L U j ) + t r ( L W 2 ) + λ 20 l m L ( l , m ) s i g n ( U ˜ j * ( l , m ) ) I ( U ˜ j * ( l , m ) 0 ) + | L ( l , m ) | I ( U ˜ j * ( l , m ) = 0 ) ,
in which L is a p × p matrix, L ( l , m ) is the ( l , m ) th element of L, W 2 is a random p × p matrix such that v e c ( W 2 ) N ( 0 , Λ 2 ) , and Λ 2 is such that the following holds:
c o v ( W 2 ( i , j ) , W 2 ( i , j ) ) = c o v ( x i * m i * ) V j 1 ( x j * m j * ) T / q , ( x i * m i * ) V j 1 ( x j * m j * ) T / q ,
where x i * and m i * is the ith row of the matrix X and M j , respectively,
(3) n ( V ˜ j * V j 1 ) d a r g m i n R = R ( V 3 ) ,
where
V 3 ( R ) = t r ( R V j R V j ) + t r ( R W 3 ) + λ 30 l m R ( l , m ) s i g n ( V ˜ j * ( l , m ) ) I ( V ˜ j * ( l , m ) 0 ) + | R ( l , m ) | I ( V ˜ j * ( l , m ) = 0 ) ,
in which R is a q × q matrix, R ( l , m ) is the ( l , m ) th element of R, W 3 is a random q × q matrix such that v e c ( W 3 ) N ( 0 , Λ 3 ) , and Λ 3 is such that the following holds:
c o v ( W 3 ( i , j ) , W 3 ( i , j ) ) = c o v ( x i m i ) T U j 1 ( x j m j ) / p , ( x i m i ) T U j 1 ( x j m j ) / p ,
where x i and m i is the ith column of the matrix X and M j , respectively.
Corollary 1.
Assume that M j , U j , V j Ψ ¯ d 1 , d 2 and n λ i λ i 0 0 as n , i = 1 , 2 , 3 , then M ˜ j , U ˜ j * , and V ˜ j * are consistent.
Proof. 
Since n ( M ˜ j M j ) converges in a distribution, n ( M ˜ j M j ) = O p ( 1 ) and | M ˜ j M j | 1 = O p ( 1 / n ) = o p ( 1 ) as n . Therefore, the proposed estimator M ˜ j is n consistent. The proof for U ˜ j * and V ˜ j * is same as for M ˜ j .  □

2.4. Model Selection

In the penalized mixture model, the selection of the optimal hyperparmeters is very important. The proposed penalized matrix normal mixture model has four hyperparameters ( G , λ 1 , λ 2 , λ 3 ) . A grid search method is used to select the optimal set of ( G , λ 1 , λ 2 , λ 3 ) with the largest penalized log-likelihood. However, even if the grid search method is applied, a model fitted with a fixed dataset is often not reliable due to the overfitting problem. Generally, the k-fold cross-validation (CV) method is used to solve these problems. In k-fold CV, the data X = X 1 , X 2 , , X n are split into K mutually exclusive subsets denoted by Y 1 , Y 2 , , Y K . Then, one of the K subsets is selected as the test data, the model with the remaining data is applied, and the penalized log-likelihood value for the test data is calculated with the estimated parameters. After using each of the subsets as test data in turn, the final score is obtained by averaging the K penalized log-likelihood calculated values as follows:
L p ( K ) = 1 K k = 1 K L p ( Y k | Y k ) ,
where Y k is the dataset excluding Y k . For example, in 4-fold cross-validation, the mean of the penalized log-likelihood is obtained as shown in Figure 1.
Therefore, using the k-fold cross-validation method, we can select the optimal parameter set with the largest mean of the penalized log-likelihood among different combinations of grid parameter values.

3. Experiments

3.1. Simulation Studies

In Scenarios 1 and 2, we generate 20 × 20 data from two matrix normal distributions N 400 ( v e c ( M 1 ) , V 1 U 1 ) , N 400 ( v e c ( M 2 ) , V 2 U 2 ) , each with a different type of mean structure. One has a rectangle shape and the other has a cross shape, as in Figure 2.
We fix G = 2 in the simulation study, and the ranges of the tuning parameter λ 1 , λ 2 , λ 3 are set to be λ 1 = 20 , 10 , 5 , 1 , 0.5 , 0.1 , λ 2 = 1 , 0.1 , 0.01 , 0.001 , λ 3 = 1 , 0.1 , 0.01 , 0.001 , and the optimal ( λ 1 , λ 2 , λ 3 ) set is determined with the 4-fold CV method. The sample size n is set to be 100, 200, 300, and 1000. To compare the performance for different degrees of sparsity, we consider different precision structures for Scenario 1 and Scenario 2, banding and AR(1) model, respectively:
Scenario 1: Banding precision structure
( V 1 U 1 ) = Σ 1 1 = 1 i f i = j 0.2 i f | i j | = 1 0 o t h e r w i s e , 1 i , j 400 .
( V 2 U 2 ) = Σ 2 1 = 1 i f i = j 0.2 i f | i j | = 1 0.3 i f | i j | = 2 0 o t h e r w i s e , 1 i , j 400 .
Scenario 2: AR(1) precision structure
( V 1 U 1 ) = Σ 1 1 = 0 . 5 | i j | , 1 i , j 400 .
( V 2 U 2 ) = Σ 2 1 = 0 . 4 | i j | , 1 i , j 400 .
The proposed model is compared with the conventional methods: K-means, spectral clustering, and the model of [9]. The criteria for performance are ACC and adjusted rand index (ARI) [21]. ACC shows the closeness of a predicted value to a standard or known value, and ARI is a clustering evaluation method that indicates perfect clustering if it is 0 and random clustering if it is 0. Additionally, the two norms are considered to show the consistency of the parameter estimates of the model: the averaged spectral norm and Frobenius norm of the difference between the estimated precision matrix and the truth.
S L = 1 G j = 1 G Σ ^ j 1 Σ j 1 s , F L = 1 G j = 1 G Σ ^ j 1 Σ j 1 F = 1 G j = 1 G i j ( Σ ^ j 1 ( i , j ) Σ j 1 ( i , j ) ) 2 ,
where A s is the largest singular value of matrix A, and A F is the Frobenius norm.
Since the clustering results depend on the generated data, each scenario is run 50 times. ACC, ARI, SL, and FL are calculated each time, and their average results are reported in Table 1, Table 2, Table 3 and Table 4, respectively. Figure 3, Figure 4, Figure 5 and Figure 6 show the violin plots of the ACC and ARI for each method with a box plot and underlying distribution by sample size in Scenario 1 and 2.
In Scenario 1, Figure 3 and Figure 4 and Table 1 show that the proposed model has the highest ACC (0.995) and ARI (0.98) for sample size 100; the method of [9] follows next. In addition, as n increases, the performance of PMNMM and the method of [9] become similar. To see if there is any significant difference among the means of the ACC and ARI, the Kruskal–Wallis test [22] is applied. The p-value of the test for significant difference among the means of ACC and ARI is 0, so we can confirm that the ACC and ARI means of the models differ under the significance level α = 0.05 . Then, to compare pairwise difference, the Dunn test [23] is applied. The null hypothesis of the Dunn test is that there is no difference between the two groups. As a result, PMNMM and the method of [9] is better than both K-means and the spectral method (p-value = 0), and the spectral method is better than K-means (p-value = 0) in all sample sizes. For a small sample case (n = 100), the proposed model shows better performance (p-value = 0) because there are many zero-valued elements of the precision matrix in the outside of the band, which are estimated to be relatively better than other methods. For the other sample size, p-values of the Dunn test between the ACC of PMNMM and the method of [9] is 0.9436, 1, and 1, respectively, and 0.8309, 1, 1 for ARI, respectively. Because the null hypothesis cannot be rejected, the performance of the two models is considered to be the same. As shown in Figure 3 and Figure 4, the ACC and ARI values of the K-means method, spectral method and the method of [9] are more spread than those of PMNMM. This implies that the proposed method produces more stable clustering results.
Table 2 shows the mean measured errors between estimated parameters and true parameters for 50 runs in Scenario 1. As the sample size increases, the FL value of the estimated mean matrix decreases, and similarly, the SL and FL values of the estimated precision matrix decrease. To see if there is any significant difference among values for each sample size, the Kruskal–Wallis test and the Dunn test are applied. With the Kruskal–Wallis test, the p-values for the means of FL (mean), SL (precision), and FL (precision) are 0, so the means of measured errors for three norms differ among different sample sizes under the significance level α = 0.05 . Next, with the Dunn test, we can confirm that the means of each error decrease as the sample size increases, as p-values are smaller than the significance level α = 0.05 for two different increasing sample sizes. This implies that the proposed method has consistency.
In Scenario 2, similar to the previous process, we can check if there is any significant difference among the means of the ACC and ARI. However, under the AR(1) precision structure, the results are slightly different. For the sample size n = 100, the p-values of the Kruskal–Wallis test are 0.19 and 0.17 for ACC and ARI, respectively, so we cannot reject the null hypothesis that the means of the ACC and ARI for all models are the same. For the other sample sizes, the p-value of the Kruskal–Wallis test is 0, so the Dunn test is applied. For n = 200, the p-value of the Dunn test between PMNMM and spectral is 0.3911, so the null hypothesis cannot be rejected. On the contrary, the p-value of the Dunn test between PMNMM and the method of [9] is 0, so PMNMM is better than the method of [9]. The sparse precision matrix is not shown clearly for the data with n = 100 sample size in Scenario 2, and this hinders PMNMM from exploring sparsity in the precision matrix, showing similar performance as the conventional method. However, for n = 300, the p-values of the Dunn test between any pair of methods are 0. We can confirm that PMNMM is better than the K-means and spectral methods, as well as the method of [9]. For n = 1000, PMNMM and the method of [9] have the same performance (p-values = 1, 1 for ACC and ARI, respectively), and are better than the other two methods, but the performance of the other two methods is the same (p-values = 0.3433, 0.3294).
Table 4 shows the mean measured errors between estimated parameters and true parameters in scenario 2. Table 4 shows similar results as Table 2, and the Kruskal–Wallis test and the Dunn test are applied in the same way as before. As a result, for the Kruskal–Wallis test, the p-values for the test of significant difference among the means of each measured error are 0, and for the Dunn test, the p-values for all pairwise comparisons are 0, so all null hypotheses are rejected under the significance level α = 0.05 . This implies that the proposed method has consistency.
In the simulation studies, the two precision structures are considered: banding and AR(1). The experimental results of Scenario 1 showed that the performance of the proposed method is better than the other methods in all sample size cases; we can confirm the consistency of the proposed method through the measured errors with the increasing sample size. The penalized matrix normal mixture model in the banding precision matrix showed good performance, even with a very small sample size. However, in the AR(1) precision structure of Scenario 2, a larger number of samples was needed to achieve good performance, and the value of the measured error was higher when compared with Scenario 1. This implies that the proposed model works better for data with a more sparse precision structure.

3.2. Real Data Studies

  • Datasets
    We evaluated the proposed model on the CIFAR 10 and Fashion-MNIST datasets:
    CIFAR 10
    The CIFAR 10 dataset (https://www.cs.toronto.edu/~kriz/cifar.html, accessed on 21 September 2021) contains well-known image data. There are 60,000 images of 32 × 32 × 3 size with 10 types—airplane, automobile, bird, cat, deer, dog, frog, horse, ship, truck—and each type has 6000 images, respectively. The dataset is divided into a training set with 50,000 images and a test set with 10,000 images.
    Fashion-MNIST
    The Fashion-MNIST dataset (https://github.com/zalandoresearch/fashion-mnist, accessed on 21 September 2021) is a grayscale image dataset, designed to replace the original MNIST dataset. There are 70,000 images of 28 × 28 size with 10 types—T-shirt/top, trouser, pullover, dress, coat, sandal, shirt, sneaker, bag, ankle boot— and each type has 7000 images, respectively. The dataset is divided into a training set with 60,000 images and a test set with 10,000 images.

3.2.1. CIFAR 10 Dataset

In the experiment, two types of images are selected: frog and ship (Figure 7). For the performance comparison of the proposed method, the training images are transformed to grayscale, and 500 images are randomly selected from the training set for each type. Inaddition, the K-means and spectral methods, as well as the method of [9] are the conventional methods to be compared with PMNMM; the number of clusters G is set to 2; and the ranges of the tuning parameter λ 1 , λ 2 , λ 3 are set to λ 1 = 1 , 0.5 , 0.1 , λ 2 = 1 , 0.5 , 0.1 , 0.05 , 0.01 , λ 3 = 1 , 0.5 , 0.1 , 0.05 , 0.01 . Using the training set, the optimal set of ( λ 1 , λ 2 , λ 3 ) selected through the 4-fold CV method is (1, 0.01, 0.01), and the performance of the proposed model is evaluated on the testing set. A total of 500 images are randomly selected from the test set, and the models are fitted to obtain ACC and ARI, respectively. This process is repeated 50 times to calculate the average ACC and ARI of each model, and the results are shown in Table 5. Then, the Kruskal–Wallis test and the Dunn test are applied to see if there is any significant difference between the methods. As a result, the p-value of the Kruskal–Wallis test is 0, the ACC and ARI of the PMNMM are higher than those of K-means and spectral method with the Dunn test (p-value = 0), but PMNMM and the method of [9] have the same performance. The null hypothesis of the Dunn test between the K-means and spectral methods is not rejected (p-value = 0.075, 0.085 for ACC and ARI, respectively) under the significance level α = 0.05 .
To compare PMNMM with [9]’s method further, a more extreme low sample situation is set up by randomly selecting 10 images from the training set for each type. The results are shown in Table 6.
Table 5 and Table 6 show that the performance does not improve significantly even if the sample size increases because the two groups are difficult to be distinguished. There is also little performance difference between the two methods because the weight of the penalty imposed on the precision matrix is small.

3.2.2. Fashion Dataset

We consider applying the methods to the fashion data, which can be distinguished more clearly. Similar to the CIFAR 10 data application, two types of images are selected—shirt and sandal—and 500 images are randomly selected from the training set for each type (Figure 8). The ranges of the tuning parameter λ 1 , λ 2 , λ 3 are set to λ 1 = 1 , 0.5 , 0.1 , λ 2 = 5 , 3 , 1 , 0.5 , 0.1 , λ 3 = 5 , 3 , 1 , 0.5 , 0.1 . Using the training set, the optimal set of ( λ 1 , λ 2 , λ 3 ) selected through the 4-fold CV method is (0.1, 0.1, 0.1). The results for ACC and ARI are shown in Table 7. It is confirmed that PMNMM and the method of [9] have almost the same performance for the sample size n = 1000. Then, the Kruskal–Wallis and Dunn tests are applied; the p-value of the Kruskal–Wallis test is 0; the ACC and ARI of the PMNMM are higher than those of K-means and spectral methods with the Dunn test (p-value = 0); and the null hypothesis of the Dunn test between the K-means and spectral methods is not rejected (p-value = 1) under the significance level α = 0.05 .
To compare PMNMM with the method of [9] further, a more extreme low sample situation is set up by randomly selecting 10 images from the training set for each type. The results are shown in Table 8.
Table 8 shows that the proposed method works better than the method of [9] that imposes a penalty only on the mean matrix for extremely small sample size data.

4. Conclusions

We have proposed a new penalized mixture model with the penalties imposed both on the mean matrix and precision matrix of the matrix normal mixture model. The conventional Gaussian mixture model for matrix data was challenging because it has to estimate one big covariance structure; therefore, the calculation speed becomes very slow as the dimension increases. On the contrary, the proposed method is a parsimonious model since it reduces a large number of parameters by regularizing the row and column variance matrices of the matrix normal distribution, especially imposing a lasso penalty on the precision matrix. Therefore, the proposed method can reflect the sparse conditional correlation structure and work even with high-dimensional data with a small sample size. The asymptotic distributions of the parameter estimators were derived, and their consistency was proved. We have shown the superior performance of the proposed model by comparing it with three conventional methods (method of [9], K-means method, and spectral method) through both synthetic and real data experiments. A performance comparison was done with ACC and ARI, and PMNMM was significantly better than the K-means and spectral methods by a nonparametric test. In the simulation studies, we evaluated the clustering performance for two different degree of sparsity of precision matrix. For the banding precision structure, the ACC performance of the proposed method reached 0.995 and 0.999, with the small sample size, n = 100 and 200, respectively, which are higher than the other methods. For the AR(1) precision structure, it is similar to that of [9] with n = 100 and higher than those of the other methods with n = 200, 300, and 1000. In conclusion, when the true precision matrix is sparse, that is, the variables have strong conditionally independent structure, the proposed model is well fitted to a small sample size data. For the real datasets, the ACC performance of the proposed method reaches 0.74 and 0.95 for clustering two types of CIFAR 10 and Fashion-MNIST images, respectively, which are higher than or equal to those of the other methods.
In the proposed model, calculation speed is an important factor, as mentioned above. Thus, the graphical lasso algorithm [12] was applied. A faster algorithm to estimate a large precision matrix would be beneficial for the proposed model to improve the calculation speed. In addition, to cluster a 3D image or higher dimensional data, new penalized mixture models for the tensor structure need to be developed. A penalized matrix non-Gaussian (such as skew-normal, skew-t) mixture model is worth developing for clustering more general types of matrix data.

Author Contributions

Conceptualization, J.H. and J.B.; methodology, J.H. and J.B.; software, J.H.; validation, J.H.; formal analysis, J.H.; investigation, J.H.; resources, J.H.; writing—original draft preparation, J.H.; writing—review and editing, J.B.; visualization, J.H.; supervision, J.B.; project administration, J.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research of the first author was supported by BK21 FOUR (Fostering Outstanding Universities for Research, NO. 5120200913674) funded by the Ministry of Education (MOE, Korea) and National Research Foundation of Korea (NRF) and was supported by the Ministry of Education of the Republic of Korea and the National Research Foundation of Korea (NRF-2018S1A6A3A04042721). The research of the second author was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2018R1D1A1B07049729).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on https://www.cs.toronto.edu/~kriz/cifar.html, accessed on 21 September 2021 and https://github.com/zalandoresearch/fashion-mnist, accessed on 21 September 2021. The R code for the proposed method is available on request from the first author.

Acknowledgments

The authors thank two reviewers for their valuable comments for the improvement of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. A Comparative Overview Table

Table A1. A comparative overview table between the proposed method and the conventional methods.
Table A1. A comparative overview table between the proposed method and the conventional methods.
K-MeansSpectralGaussian MixtureMixture of Matrix Normal with Regularized Mean ([9])Mixture of Matrix Normal with Regularized Mean/Covariance (Proposed Method)
Data typeVectorVector/MatrixVectorMatrixMatrix
Clustering typeCentroid-basedSimilarity-basedModel-basedModel-basedModel-based
Affine transformation invarianceXXOOO
Number of parameters ( p × q data, number of clusters: G)NANA G p q ( p q + 1 ) / 2 +
G p q + G 1
G p ( p + 1 ) / 2 +
G q ( q + 1 ) / 2 +
G p q + G 1 r
G p ( p + 1 ) / 2 +
G q ( q + 1 ) / 2 +
G p q + G 1 r *
Modeling ability for covariance structure of matrix dataXXXOO
r: the number of mean parameters values that are shrunken to zero ( 0 r G p q ) ; r * : the number of mean and precision parameters values that are shrunken to zero; ( r r * G p ( p + 1 ) / 2 + G q ( q + 1 ) / 2 ) + G p q ) ; NA: not applicable.

Appendix B

Proof  of Equation (5).
L p ( Θ ; Λ ) M j = M j i = 1 n τ i j 1 2 t r ( V j 1 ( X i M j ) T U j 1 ( x i M j ) ) λ 1 | M j | 1 = 0 ,
using the formula X t r ( A X T B X ) = B X A T + B T X A (KB Petersen, 2012 [24]), we obtain the following:
M j i = 1 n τ i j 1 2 t r ( V j 1 ( X i M j ) T U j 1 ( x i M j ) ) λ 1 | M j | 1 = i = 1 n τ i j U j 1 ( X i M j ) V j 1 λ 1 s i g n ( M j ) = 0 .
Some algebraic manipulations are applied as follows:
i = 1 n τ i j U j 1 ( X i M j ) V j 1 i = 1 n τ i j λ 1 i = 1 n τ i j s i g n ( M j ) = 0 . U j 1 i = 1 n τ i j X i i = 1 n τ i j V j 1 U j 1 M j V j 1 λ 1 i = 1 n τ i j s i g n ( M j ) = 0 . i = 1 n τ i j X i i = 1 n τ i j M j λ 1 i = 1 n τ i j U j s i g n ( M j ) V j = 0 .
Because i = 1 n τ i j X i i = 1 n τ i j is the estimate of M j for the non-penalty model, the following equation is derived:
M ^ j = M ˜ j λ 1 i = 1 n τ i j U j s i g n ( M ˜ j ) V j = s i g n ( M ˜ j ) | M ˜ j | λ 1 i = 1 n τ i j U j 1 p × q V j .

Appendix C

Proof of Theorem 1.
Let U j * and V j * be the inverse matrix of U j and V j , respectively. Then, the penalized Q function can be rewritten in the M-step as follows.
Q p = j = 1 G i = 1 n p q 2 l o g ( 2 π ) + p 2 l o g | V j * | + q 2 l o g | U j * | 1 2 t r V j * ( X i M j ) T U j * ( X i M j ) λ 1 j = 1 G | M j | 1 λ 2 j = 1 G | U j * | 1 λ 3 j = 1 G | V j * | 1 .
We took a similar approach as that which [4] used to derive the asymptotic distribution of the precision matrix estimator in the Gaussian graphical model.  □

Appendix C.1. The Limiting Dist. of M ˜ j

Let M ˜ j be the matrix that minimizes the following:
1 2 i = 1 n τ i j t r V j * ( X i M j ) T U j * ( X i M j ) + λ 1 | M ˜ j | , o r i = 1 n τ i j t r V j * ( X i M j ) T U j * ( X i M j ) / i = 1 n τ i j + λ 1 * | M ˜ j | ,
where λ 1 * = 2 λ 1 / i = 1 n τ i j .
Define V 1 n ( H ) as the following:
V 1 n ( H ) = i = 1 n τ i j t r V j * ( X i M j H / n ) T U j * ( X i M j H / n ) / i = 1 n τ i j + λ 1 * l = 1 p m = 1 q | M ˜ j ( l , m ) + H ( l , m ) / n | i = 1 n τ i j t r V j * ( X i M j ) T U j * ( X i M j ) / i = 1 n τ i j + λ 1 * l = 1 p m = 1 q | M ˜ j ( l , m ) | .
Note the following:
i = 1 n τ i j t r V j * ( X i M j H / n ) T U j * ( X i M j H / n ) / i = 1 n τ i j i = 1 n τ i j t r V j * ( X i M j ) T U j * ( X i M j ) / i = 1 n τ i j = t r V j * ( i = 1 n τ i j ( X i M j i = 1 n τ i j ) ) T U j * ( H / n ) t r V j * ( H / n ) T U j * ( i = 1 n τ i j ( X i M j i = 1 n τ i j ) ) + t r ( V j * H T U j * H / n ) = t r U j * H V j * ( M ^ j M j ) T / n t r V j * H T U j * ( M ^ j M j ) / n + t r ( V j * H T U j * H ) / n ,
where M ^ j = i = 1 n τ i j X i / i = 1 n τ i j , which is the m.l.e. of M j for the non-penalized matrix normal mixture model.
Together with the fact that
λ 1 * l = 1 p m = 1 q ( | M ˜ j ( l , m ) + H ( l , m ) / n | | M ˜ j ( l , m ) | ) = λ 1 * n l = 1 p m = 1 q H ( l , m ) s i g n ( M ˜ j ( l , m ) ) I ( M ˜ j ( l , m ) 0 ) + | H ( l , m ) | I ( M ˜ j ( l , m ) = 0 ) ,
n V 1 n ( H ) can be rewritten as follows:
t r ( U j 1 H V j 1 W 1 n T ) t r ( V j 1 H T U j 1 W 1 n ) + t r ( V j 1 H T U j 1 H ) + n λ 1 * l = 1 p m = 1 q H ( l , m ) s i g n ( M ˜ j ( l , m ) ) I ( M ˜ j ( l , m ) 0 ) + | H ( l , m ) | I ( M ˜ j ( l , m ) = 0 ) ,
where W 1 n = n ( M ^ j M j ) and v e c ( W 1 n ) d N ( 0 , Λ 1 ) . Therefore, n V 1 n ( H ) d V ( H ) .
Since both V 1 ( H ) and n V 1 n ( H ) are convex and V 1 ( H ) has a unique minimum, we have the following:
a r g m i n n V 1 n ( H ) = n ( M ˜ j M j ) d a r g m i n V 1 ( H )

Appendix C.2. The Limiting Dist. of U ˜ j *

Let U ˜ j * be the positive definite matrix that minimizes the following:
q i = 1 n τ i j 2 l o g | U ˜ j * | + q i = 1 n τ i j 2 t r ( U ˜ j * S j ( U ) ) + λ 2 | U ˜ j * | 1 ,
where S j ( U ) = i = 1 n τ i j ( X i M j ) V j * ( X i M j ) T q i = 1 n τ i j , which is the m.l.e. of U j for the non-penalized matrix normal mixture model, or minimizes the following:
l o g | U ˜ j * | + t r ( U ˜ j * S j ( U ) ) + λ 2 * | U ˜ j * | 1 ,
where λ 2 * = 2 λ 2 / ( q i = 1 n τ i j ) .
Define V 2 n ( L ) as follows:
V 2 n ( L ) = l o g | U ˜ j * + L n | + t r ( U ˜ j * + L n ) S j ( U ) + λ 2 * l m | U ˜ j * ( l , m ) + L ( l , m ) n | + l o g | U ˜ j * | t r ( U ˜ j * S j ( U ) ) + λ 2 * l m | U ˜ j * ( l , m ) | .
Note the following:
l o g | U ˜ j * + L n | l o g | U ˜ j * | = l o g | I + U j 1 / 2 L U j 1 / 2 n | = i = 1 p l o g 1 + σ i ( U j 1 / 2 L U j 1 / 2 / n ) ,
where σ j ( · ) denotes the ith-largest eigenvalue of a matrix. Since
l o g 1 + σ i ( U j 1 / 2 L U j 1 / 2 / n ) = σ i ( U j 1 / 2 L U j 1 / 2 ) n σ i 2 ( U j 1 / 2 L U j 1 / 2 ) n + o ( 1 n ) ,
we conclude the following:
l o g | U ˜ j * + L n | l o g | U ˜ j * | = i = 1 p σ i ( U j 1 / 2 L U j 1 / 2 ) n t r ( U j 1 / 2 L U j L U j 1 / 2 ) n + o ( 1 n ) = t r ( U j 1 / 2 L U j 1 / 2 ) n t r ( U j 1 / 2 L U j L U j 1 / 2 ) n + o ( 1 n ) = t r ( L U j ) n t r ( L U j L U j ) n + o ( 1 n ) .
On the other hand,
t r ( U ˜ j * + L n ) S j ( U ) t r ( U ˜ j * S j ( U ) ) = t r ( L S j ( U ) n ) = t r ( L U j ) n + t r L ( S j ( U ) U j ) n .
Together with the fact that
λ 2 * l m ( | U ˜ j * ( l , m ) + L ( l , m ) n ) | | U ˜ j * ( l , m ) | ) = λ 2 * n l m L ( l , m ) s i g n ( U ˜ j * ( l , m ) ) I ( U ˜ j * ( l , m ) 0 ) + | L ( l , m ) | I ( U ˜ j * ( l , m ) = 0 ) ,
n V 2 n ( L ) can be rewritten as follows:
t r ( L U j L U j ) + t r ( L W 2 n ) + n λ 2 * l m L ( l , m ) s i g n ( U ˜ j * ( l , m ) ) I ( U ˜ j * ( l , m ) 0 ) + | L ( l , m ) | I ( U ˜ j * ( l , m ) = 0 ) + o ( 1 ) ,
where W 2 n = n ( S j ( U ) U j ) and v e c ( W 2 n ) d N ( 0 , Λ 2 ) because S j ( U ) is the m.l.e. of U j for non-penalized matrix normal mixture model. Since both V 2 ( L ) and n V 2 n ( L ) are convex and V 2 ( L ) has a unique minimum, the following holds:
a r g m i n n V 2 n ( L ) = n ( U ˜ j * U j ) d a r g m i n V 2 ( L )

Appendix C.3. The Limiting Dist. of V j *

Let V ˜ j * be the positive definite matrix that minimizes the following:
p i = 1 n τ i j 2 l o g | V ˜ j * | + p i = 1 n τ i j 2 t r ( V ˜ j * S j ( V ) ) + λ 3 | V ˜ j * | 1 ,
where S j ( V ) = i = 1 n τ i j ( X i M j ) T U j * ( X i M j ) p i = 1 n τ i j , or minimizes the following:
l o g | V ˜ j * | + t r ( V ˜ j * S j ( V ) ) + λ 3 * | V ˜ j * | 1 ,
where λ 3 * = 2 λ 3 / ( p i = 1 n τ i j ) .
Define V 3 n ( R ) as follows:
V 3 n ( R ) = l o g | V ˜ j * + R n | + t r ( V ˜ j + R n ) S j ( V ) + λ 3 * l m | V ˜ j * ( l , m ) + R ( l , m ) n | . + l o g | V ˜ j * | t r ( V ˜ j * S j ( V ) ) + λ 3 * l m | V ˜ j * ( l , m ) |
Note the following:
l o g | V ˜ j * + R n | l o g | V ˜ j * | = l o g | I + V j 1 / 2 R V j 1 / 2 n | = i = 1 q l o g 1 + σ i ( V j 1 / 2 R V j 1 / 2 / n ) .
Since
l o g 1 + σ i ( V j 1 / 2 R V j 1 / 2 / n ) = σ i ( V j 1 / 2 R V j 1 / 2 ) n σ i 2 ( V j 1 / 2 R V j 1 / 2 ) n + o ( 1 n ) ,
we conclude the following:
l o g | V ˜ j * + R n | l o g | V ˜ j * | = i = 1 q σ i ( V j 1 / 2 R V j 1 / 2 ) n t r ( V j 1 / 2 R V j R V j 1 / 2 ) n + o ( 1 n ) = t r ( V j 1 / 2 R V j 1 / 2 ) n t r ( V j 1 / 2 R V j R V j 1 / 2 ) n + o ( 1 n ) = t r ( R V j ) n t r ( R V j R V j ) n + o ( 1 n ) .
On the other hand,
t r ( V ˜ j * + R n ) S j ( V ) t r ( V ˜ j * S j ( V ) ) = t r ( R S j ( V ) n ) = t r ( R V j ) n + t r R ( S j ( V ) V j ) n .
Together with the fact that
λ 3 * l m ( | V ˜ j * ( l , m ) + R ( l , m ) n ) | | V ˜ j * ( l , m ) | ) = λ 3 * n l m R ( l , m ) s i g n ( V ˜ j * ( l , m ) ) I ( V ˜ j * ( l , m ) 0 ) + | R ( l , m ) | I ( V ˜ j * ( l , m ) = 0 ) ,
n V 3 n ( R ) can be rewritten as follows:
t r ( R V j R V j ) + t r ( R W 3 n ) + n λ 3 * l m R ( l , m ) s i g n ( V ˜ j * ( l , m ) ) I ( V ˜ j * ( l , m ) 0 ) + | R ( l , m ) | I ( V ˜ j * ( l , m ) = 0 ) + o ( 1 ) ,
where W 3 n = n ( S j ( V ) V j ) and v e c ( W 3 n ) d N ( 0 , Λ 3 ) because S j ( V ) is the m.l.e. of V j for non-penalized matrix normal mixture model. Since both V 3 ( R ) and n V 3 n ( R ) are convex and V 3 ( R ) has a unique minimum, we have the following:
a r g m i n n V 3 n ( R ) = n ( V ˜ j * V j ) d a r g m i n V 3 ( R )

References

  1. Pan, W.; Shen, X. Penalized Model-Based Clustering with Application to Variable Selection. J. Mach. Learn. Res. 2007, 8, 1145–1164. [Google Scholar]
  2. Zhou, H.; Pan, W.; Xiaotong, S. Penalized model-based clustering with unconstrained covariance matrices. Electron. J. Stat. 2009, 3, 1473–1496. [Google Scholar] [CrossRef] [PubMed]
  3. Ruan, L.; Yuan, M.; Zou, H. Regularized Parameter Estimation in High-Dimensional Gaussian Mixture Models. Neural Comput. 2011, 23, 1605–1622. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Yuan, M.; Lin, Y. Model selection and estimation in the Gaussian graphical model. Biometrika 2007, 94, 19–35. [Google Scholar] [CrossRef] [Green Version]
  5. Viroli, C. Finite mixtures of matrix normal distributions for classifying three-way data. Stat. Comput. 2011, 21, 511–522. [Google Scholar] [CrossRef]
  6. Gallaugher, M.; McNicholas, P. A Mixture of Matrix Variate Skew-t Distributions. arXiv 2017, arXiv:1703.08882. [Google Scholar]
  7. Melnykov, V.; Zhu, X. On model-based clustering of skewed matrix data. J. Multivar. Anal. 2018, 167, 181–194. [Google Scholar] [CrossRef]
  8. Tomarchio, S.; Punzo, A.; Bagnato, L. Two new matrix-variate distributions with application in model-based clustering. Comput. Stat. Data Anal. 2020, 152, 107050. [Google Scholar] [CrossRef]
  9. Gao, X.; Shen, W.; Zhang, L.; Hu, J.; Fortin, N.; Frostig, R.; Ombao, H. Regularized matrix data clustering and its application to image analysis. Biometrics 2020, 77, 890–902. [Google Scholar] [CrossRef]
  10. Dempster, A.P. Covariance selection. Biometrics 1972, 28, 157–175. [Google Scholar] [CrossRef]
  11. Lartigue, T. Mixture of Gaussian Graphical Models with Constraints. Ph.D. Thesis, Institut Polytechnique de Paris, Palaiseau, France, 2020. [Google Scholar]
  12. Friedman, J.; Hastie, T.; Tibshirani, R. Sparse inverse covariance estimation with the graphical lasso. Biostatistics 2007, 9, 432–441. [Google Scholar] [CrossRef] [Green Version]
  13. Lloyd, S. Least squares quantization in PCM. IEEE Trans. Inf. Theory 2006, 28, 129–137. [Google Scholar] [CrossRef]
  14. Ng, A.; Jordan, M.; Weiss, Y. On Spectral Clustering: Analysis and an algorithm. In Proceedings of the Conference on Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 9–14 December 2002; pp. 849–856. [Google Scholar]
  15. Dutilleul, P. The MLE algorithm for the matrix normal distribution. J. Stat. Comput. Simul. 1999, 64, 105–123. [Google Scholar] [CrossRef]
  16. Srivastava, M.; Rosen, T.; Rosen, D. Models with a Kronecker product covariance structure: Estimation and testing. Mathematical Methods of Statistics. Math. Methods Stat. 2008, 17, 357–370. [Google Scholar] [CrossRef]
  17. Dawid, A. Some matrix-variate distribution theory: Notational considerations and a Bayesian application. Biometrika 1981, 68, 265–274. [Google Scholar] [CrossRef]
  18. Dempster, A.; Laird, N.; Rubin, D. Maximum Likelihood From Incomplete Data Via The EM algorithm. J. R. Stat. Soc. Ser. B (Methodological) 2008, 39, 1–38. [Google Scholar]
  19. Aitken, A. XXV.—On Bernoulli’s Numerical Solution of Algebraic Equations. Proc. R. Soc. Edinb. 1926, 46, 289–305. [Google Scholar] [CrossRef]
  20. Fan, J.; Li, R. Variable Selection via Nonconcave Penalized Likelihood and Its Oracle Properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
  21. Hubert, L.; Arabie, P. Comparing Partitions. J. Classif. 1985, 2, 193–218. [Google Scholar] [CrossRef]
  22. Kruskal, W.; Wallis, W. Errata: Use of Ranks in One-Criterion Variance Analysis. J. Am. Stat. Assoc. 1952, 47, 583–621. [Google Scholar] [CrossRef]
  23. Dunn, O. Multiple comparison using RANK sums. Technometrics 1964, 6, 241–252. [Google Scholar] [CrossRef]
  24. Petersen, K.B.; Pedersen, M.S. The Matrix Cookbook. 2012. Available online: https://www.math.uwaterloo.ca/~hwolkowi/matrixcookbook.pdf (accessed on 21 September 2021).
Figure 1. 4-fold cross-validation.
Figure 1. 4-fold cross-validation.
Entropy 23 01249 g001
Figure 2. Mean structure of each group.
Figure 2. Mean structure of each group.
Entropy 23 01249 g002
Figure 3. Violin plot of ACC in Scenario 1.
Figure 3. Violin plot of ACC in Scenario 1.
Entropy 23 01249 g003
Figure 4. Violin plot of ARI in Scenario 1.
Figure 4. Violin plot of ARI in Scenario 1.
Entropy 23 01249 g004
Figure 5. Violin plot of ACC in Scenario 2.
Figure 5. Violin plot of ACC in Scenario 2.
Entropy 23 01249 g005
Figure 6. Violin plot of ARI in Scenario 2.
Figure 6. Violin plot of ARI in Scenario 2.
Entropy 23 01249 g006
Figure 7. Two types of images in CIFAR 10 dataset: frog and ship.
Figure 7. Two types of images in CIFAR 10 dataset: frog and ship.
Entropy 23 01249 g007
Figure 8. Two types of images in CIFAR 10 dataset: frog and ship.
Figure 8. Two types of images in CIFAR 10 dataset: frog and ship.
Entropy 23 01249 g008
Table 1. Scenario 1: Performance comparison for PMNMM and conventional clustering methods (The values in parentheses are the standard deviations of the 50 runs’ results).
Table 1. Scenario 1: Performance comparison for PMNMM and conventional clustering methods (The values in parentheses are the standard deviations of the 50 runs’ results).
nPMNMMMethod of [9]K-MeansSpectral
ACCARIACCARIACCARIACCARI
1000.9950.980.9150.7060.7540.2880.8380.489
(0.01)(0.041)(0.07)(0.213)(0.097)(0.184)(0.097)(0.198)
2000.9990.9990.9970.9990.8750.5720.9360.763
(0.002)(0.002)(0.005)(0.018)(0.06)(0.16)(0.021)(0.071)
30011110.9320.7470.9610.849
(0.0005)(0.002)(0.001)(0.005)(0.026)(0.088)(0.014)(0.051)
100011110.9780.9160.9880.954
(0.0002)(0.0008)(0.0002)(0.0008)(0.005)(0.019)(0.003)(0.013)
Table 2. Scenario 1: Mean measured errors of the parameter estimates for the selected optimal parameter set (The values in parentheses are the standard deviations of the 50 runs’ results).
Table 2. Scenario 1: Mean measured errors of the parameter estimates for the selected optimal parameter set (The values in parentheses are the standard deviations of the 50 runs’ results).
nOptimal Parameter SetFLSLFL
λ 1 λ 2 λ 3 (Mean)(Precision)(Precision)
100200.0010.012.90.9265.34
(0.078)(0.071)(0.168)
2000.10.010.12.190.5973.9
(0.048)(0.045)(0.088)
3000.10.0010.0011.790.4653.39
(0.048)(0.028)(0.048)
10000.10.0010.0010.9830.2912.6
(0.026)(0.01)(0.025)
Table 3. Scenario 2: Performance comparison for PMNMM and conventional clustering methods (The values in parentheses are the standard deviations of the 50 runs’ results).
Table 3. Scenario 2: Performance comparison for PMNMM and conventional clustering methods (The values in parentheses are the standard deviations of the 50 runs’ results).
nPMNMMMethod of [9]K-meansSpectral
ACCARIACCARIACCARIACCARI
1000.6650.1350.6710.1370.6360.1050.6740.155
(0.094)(0.146)(0.086)(0.122)(0.1)(0.132)(0.104)(0.155)
2000.8470.5480.7530.2860.6780.1550.8110.4
(0.132)(0.334)(0.093)(0.182)(0.091)(0.143)(0.062)(0.147)
3000.9910.9640.9090.7050.7430.2690.8810.583
(0.006)(0.024)(0.097)(0.227)(0.095)(0.168)(0.03)(0.092)
10000.9990.9940.9980.9930.9680.8760.9710.888
(0.001)(0.005)(0.001)(0.004)(0.009)(0.032)(0.006)(0.022)
Table 4. Scenario 2: Mean measured errors of the parameter estimates for the selected optimal parameter set (The values in parentheses are the standard deviations of the 50 runs’ results).
Table 4. Scenario 2: Mean measured errors of the parameter estimates for the selected optimal parameter set (The values in parentheses are the standard deviations of the 50 runs’ results).
nOptimal Parameter SetFLSLFL
λ 1 λ 2 λ 3 (Mean)(Precision)(Precision)
1000.10.0010.14.511.236.77
(0.253)(0.155)(0.372)
2000.10.0010.0012.930.7975.34
(0.384)(0.061)(0.24)
3000.10.0010.0012.020.6214.71
(0.062)(0.047)(0.054)
10000.10.0010.0011.110.4224.11
(0.026)(0.013)(0.027)
Table 5. Performance comparison of PMNMM with the conventional methods for CIFAR 10 dataset (The values in parentheses are the standard deviations of the 50 runs’ results).
Table 5. Performance comparison of PMNMM with the conventional methods for CIFAR 10 dataset (The values in parentheses are the standard deviations of the 50 runs’ results).
MethodACCARI
PMNMM0.74180.2335
(0.014)(0.026)
[9]’s method0.74180.2335
(0.014)(0.026)
K-means0.66540.1088
(0.015)(0.019)
Spectral0.67510.1222
(0.017)(0.023)
Table 6. Performance comparison of PMNMM and [9]’s method for CIFAR 10 dataset (The values in parentheses are the standard deviations of the 50 runs’ results).
Table 6. Performance comparison of PMNMM and [9]’s method for CIFAR 10 dataset (The values in parentheses are the standard deviations of the 50 runs’ results).
MethodACCARI
PMNMM0.72060.1936
(0.042)(0.0753)
[9]’s method0.72040.1934
(0.043)(0.0743)
Table 7. Performance comparison of PMNMM with the conventional methods for the fashion dataset (The values in parentheses are the standard deviations of the 50 runs’ results).
Table 7. Performance comparison of PMNMM with the conventional methods for the fashion dataset (The values in parentheses are the standard deviations of the 50 runs’ results).
MethodACCARI
PMNMM0.95040.8113
(0.007)(0.024)
[9]’s method0.95000.8095
(0.014)(0.026)
K-means0.88280.5865
(0.015)(0.046)
Spectral0.88310.5870
(0.013)(0.040)
Table 8. Performance comparison of PMNMM and the method of [9] for the fashion dataset (The values in parentheses are the standard deviations of the 50 runs’ results).
Table 8. Performance comparison of PMNMM and the method of [9] for the fashion dataset (The values in parentheses are the standard deviations of the 50 runs’ results).
MethodACCARI
PMNMM0.93420.7541
(0.025)(0.086)
[9]’s method0.67540.1207
(0.026)(0.037)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Heo, J.; Baek, J. A Penalized Matrix Normal Mixture Model for Clustering Matrix Data. Entropy 2021, 23, 1249. https://0-doi-org.brum.beds.ac.uk/10.3390/e23101249

AMA Style

Heo J, Baek J. A Penalized Matrix Normal Mixture Model for Clustering Matrix Data. Entropy. 2021; 23(10):1249. https://0-doi-org.brum.beds.ac.uk/10.3390/e23101249

Chicago/Turabian Style

Heo, Jinwon, and Jangsun Baek. 2021. "A Penalized Matrix Normal Mixture Model for Clustering Matrix Data" Entropy 23, no. 10: 1249. https://0-doi-org.brum.beds.ac.uk/10.3390/e23101249

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop