Next Article in Journal
A Robust Electric Spring Model and Modified Backward Forward Solution Method for Microgrids with Distributed Generation
Previous Article in Journal
Study of Lotka–Volterra Biological or Chemical Oscillator Problem Using the Normalization Technique: Prediction of Time and Concentrations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Unified Scalable Equivalent Formulation for Schatten Quasi-Norms

1
Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, School of Artificial Intelligence, Xidian University, Xi’an 710071, China
2
Peng Cheng Laboratory, Shenzhen 518066, China
*
Author to whom correspondence should be addressed.
Submission received: 7 July 2020 / Revised: 5 August 2020 / Accepted: 5 August 2020 / Published: 10 August 2020
(This article belongs to the Section Mathematics and Computer Science)

Abstract

:
The Schatten quasi-norm is an approximation of the rank, which is tighter than the nuclear norm. However, most Schatten quasi-norm minimization (SQNM) algorithms suffer from high computational cost to compute the singular value decomposition (SVD) of large matrices at each iteration. In this paper, we prove that for any p, p 1 , p 2 > 0 satisfying 1 / p = 1 / p 1 + 1 / p 2 , the Schatten p-(quasi-)norm of any matrix is equivalent to minimizing the product of the Schatten p 1 -(quasi-)norm and Schatten p 2 -(quasi-)norm of its two much smaller factor matrices. Then, we present and prove the equivalence between the product and its weighted sum formulations for two cases: p 1 = p 2 and p 1 p 2 . In particular, when p > 1 / 2 , there is an equivalence between the Schatten p-quasi-norm of any matrix and the Schatten 2 p -norms of its two factor matrices. We further extend the theoretical results of two factor matrices to the cases of three and more factor matrices, from which we can see that for any 0 < p < 1 , the Schatten p-quasi-norm of any matrix is the minimization of the mean of the Schatten ( 1 / p + 1 ) p -norms of 1 / p + 1 factor matrices, where 1 / p denotes the largest integer not exceeding 1 / p .

1. Introduction

The affine rank minimization problem arises directly in various areas of science and engineering, including statistics, machine learning, information theory, data mining, medical imaging, and computer vision. Some representative applications include matrix completion [1], robust principal component analysis (RPCA) [2], low-rank representation [3], multivariate regression [4], multi-task learning [5], and system identification [6]. To efficiently solve such problems, we mainly relax the rank function to its tractable convex envelope, that is, the nuclear norm · (sum of the singular values, also known as the trace norm or Schatten 1-norm), which leads to a convex optimization problem [1,7,8,9]. In fact, the nuclear norm of a matrix is the 1 -norm of the vector of its singular values, and thus it can motivate a low-rank solution. However, it has been shown in [10,11] that the 1 -norm over-penalizes large entries of vectors and therefore results in a solution from a possibly biased solution space. Recall from the relationship between the 1 -norm and nuclear norm that the nuclear norm penalty shrinks all singular values equally, which also leads to the over-penalization of large singular values [12]. That is, the nuclear norm may make the solution deviate from the original solution as the 1 -norm does. Unlike the nuclear norm, the Schatten p-quasi-norm with 0 < p < 1 is non-convex, and it can give a closer approximation to the rank function. Thus, the Schatten quasi-norm minimization (SQNM) has received a significant amount of attention from researchers in various communities, such as image recovery [13], collaborative filtering [14,15], and MRI analysis [16].
Recently, several iteratively re-weighted least squares (IRLS) algorithms as in [17,18] have been proposed to approximately solve associated Schatten quasi-norm minimization problems. Lu et al. [13] proposed a family of iteratively re-weighted nuclear norm (IRNN) algorithms to solve various non-convex surrogate (including the Schatten quasi-norm) minimization problems. In [13,14,19,20], the Schatten quasi-norm has been shown to be empirically superior to the nuclear norm for many real-world problems. Moreover, [21] theoretically proved that the SQNM requires significantly fewer measurements than nuclear norm minimization as in [1,9]. However, all the algorithms have to be solved iteratively and involve singular value decomposition (SVD) in each iteration. Thus, they suffer from a high computational complexity of O ( min ( m , n ) m n ) and are not applicable for large-scale matrix recovery problems [22].
It is well known that the nuclear norm has a scalable equivalent formulation (also called the bilinear spectral penalty [9,23,24]), which has been successfully applied in many applications, such as collaborative filtering [15,25,26]. Zuo et al. [27] proposed a generalized shrinkage-thresholding operator to iteratively solve p quasi-norm minimization with 0 p < 1 . Since the Schatten p-quasi-norm of one matrix is equivalent to the p quasi-norm on its singular values, we may naturally ask the following question: can we design a unified scalable equivalent formulation to the Schatten p-quasi-norm with arbitrary p values (i.e., 0 < p < 1 )?
In this paper (note that the first version [28] of this paper appeared on ArXiv, and in the current version we have included experimental results and fixed a few typos.), we first present and prove the equivalence relation on the Schatten p-(quasi-)norm of any matrix and minimization of the product of the Schatten p 1 -(quasi-)norm and Schatten p 2 -(quasi-)norm of its two smaller factor matrices, for any p, p 1 , p 2 > 0 satisfying 1 / p = 1 / p 1 + 1 / p 2 . Moreover, we prove the equivalence between the product formulation of the Schatten (quasi-)norm and its weighted sum formulation for the two cases of p 1 and p 2 : p 1 = p 2 and p 1 p 2 . When p > 1 / 2 and by setting the same value for p 1 and p 2 , there is an equivalence between the Schatten p-(quasi-)norm of any matrix and the Schatten 2 p -norms of its two factor matrices, where a representative example is the equivalent formulation of the nuclear norm, i.e.,
X = min X = U V T U F 2 + V F 2 2 .
That is, the SQNM problems with p > 1 / 2 can be transformed into problems only involving the smooth convex norms of two factor matrices, which can lead to simpler and more efficient algorithms (e.g., [12,22,29,30]). Clearly, the resulting algorithms can reduce the per-iteration cost from O ( min ( m , n ) m n ) to O ( m n d ) , where d m , n in general.
We further extend the theoretical results of two factor matrices to the cases of three and more factor matrices, from which we can know that for any 0 < p < 1 , the Schatten p-quasi-norm of any matrix is equivalent to the minimization of the mean of the Schatten ( 1 / p + 1 ) p -norms of 1 / p + 1 factor matrices, where 1 / p denotes the largest integer not exceeding 1 / p . Note that the norms of all the smaller factor matrices are convex and smooth. Besides the theoretical results, we also present several representative examples for two and three factor matrices. In fact, the bi-nuclear and Frobenius/nuclear quasi-norms defined in our previous work [22] and the tri-nuclear quasi-norm defined in our previous work [29] are three important special cases of our unified scalable formulations for Schatten quasi-norms.

2. Notations and Background

In this section, we present the definition of the Schatten p-norm and discuss some related work about Schatten p-norm minimization. We first briefly recall the definitions of both the norm and the quasi-norm.
Definition 1.
A quasi-norm · on a vector space X is a real-valued map X [ 0 , ) with the following conditions:
  • x 0 for any vector x X , and x = 0 if and only if x = 0 .
  • α x = | α | x if α or , and x X .
  • There is a constant c 1 so that if x 1 , x 2 X we have x 1 + x 2 c ( x 1 + x 2 ) (triangle inequality).
If the triangle inequality is replaced by x 1 + x 2 x 1 + x 2 , then it becomes a norm.
Definition 2.
The Schatten p-norm ( 0 < p < ) of a matrix X R m × n (without loss of generality, we assume that m n ) is defined as
X S p = i = 1 n σ i p ( X ) 1 / p ,
where σ i ( X ) denotes the i-th singular value of X.
When p 1 , Definition 2 defines a natural norm, whereas it defines a quasi-norm for 0 < p < 1 . For instance, the Schatten 1-norm is the so-called nuclear norm, X , and the Schatten 2-norm is the well-known Frobenius norm, X F = i = 1 m j = 1 n x i j 2 . As the non-convex surrogate for the rank function, the Schatten p-quasi-norm is the better approximation than the nuclear norm [21], analogous to the superiority of the p quasi-norm to the 1 -norm [18,31].
To recover a low-rank matrix from a small set of linear observations, b R l , the general SQNM problem is formulated as follows:
min X R m × n X S p p , subject to A ( X ) = b ,
where A : R m × n R l is a general linear operator. Alternatively, the Lagrangian version of (2) is
min X R m × n λ X S p p + f A ( X ) b ,
where λ > 0 is a regularization parameter, and f ( · ) : R l R denotes a certain measurement for characterizing the loss ( A ( X ) b ) . For instance, A is the projection operator P Ω , and f ( · ) = · 2 2 in matrix completion [13,17,20,32], where P Ω is the orthogonal projection onto the linear subspace of matrices supported on Ω : = { ( i , j ) | D i j is observed } , that is, P Ω ( D ) i j = D i j if ( i , j ) Ω , and P Ω ( D ) i j = 0 otherwise. For RPCA problems [2,33,34,35,36], A is an identity operator and f ( · ) = · 1 . In the multivariate regression problem in [37], A ( X ) = A X with A being a given matrix, and f ( · ) = · F 2 . In addition, f ( · ) may be chosen as the Hinge loss in [23] or the p quasi-norm in [14].
Generally, SQNM problems (e.g., (2) and (3)) are non-convex, non-smooth, and even non-Lipschitz [38]. So far, only a few algorithms (e.g., IRLS [17,18] and IRNN [13]) have been developed to solve such challenging problems. However, since the SQNM algorithms involve SVD of large matrices in each iteration, they suffer from a high computational cost of O ( m n 2 ) , which severely limits their applicability to large-scale problems. While there have been many efforts towards fast SVD computation such as partial SVD [39], the performances of these methods are still unsatisfactory for many real applications [40]. As an alternative to reduce the computational complexity of SVD on a large matrix, one can factorize X into two smaller factor matrices (i.e., X = U V T ). According to the unitary invariant property of norms, (2) and (3) can be reformulated into two much smaller matrix optimization problems as in [23,41,42,43,44], which are still non-convex and non-smooth. Therefore, it is very important to determine how to transform challenging problems such as (2) and (3) into more tractable ones, which can be solved by simpler and more efficient algorithms as in [30,43].

3. A Unified Formulation for Schatten Quasi-Norms

In this section, we first present and prove an equivalence relation on the Schatten p-(quasi-)norm of any matrix and the Schatten p 1 -(quasi-)norm and Schatten p 2 -(quasi-)norm of its two factor matrices, where 1 / p = 1 / p 1 + 1 / p 2 with any p 1 > 0 and p 2 > 0 . Moreover, we prove the equivalence between the product formulation of the Schatten (quasi-)norm and its weighted sum formulation for the two cases: p 1 = p 2 and p 1 p 2 , respectively. For any 1 / 2 < p 1 , the Schatten p-(quasi-)norm of any matrix is equivalent to minimizing the mean of the Schatten 2 p -norms of both factor matrices, which can lead to simpler and more efficient algorithms. Finally, we extend the theoretical results of two factor matrices to the cases of three and more factor matrices. One can see that for any 0 < p < 1 , the Schatten p-(quasi-)norm of any matrix is the minimization of the mean of the Schatten ( 1 / p + 1 ) p -norms of 1 / p + 1 factor matrices.

3.1. Unified Schatten Quasi-Norm Formulations of Two Factor Matrices

In this subsection, we present some unified Schatten quasi-norm scalable formulations for the case of two factor matrices.
Theorem 1.
Each matrix X R m × n with r a n k ( X ) = r d can be decomposed into the product of two much smaller matrices U R m × d and V R n × d , i.e., X = U V T . For any 0 < p 1 , p 1 > 0 and p 2 > 0 satisfying 1 / p 1 + 1 / p 2 = 1 / p , we have
X S p = min U R m × d , V R n × d : X = U V T U S p 1 V S p 2 .
The proof of Theorem 1 is provided in Section 5.1. From Theorem 1, it is clear that for any 0 < p 1 and p 1 , p 2 > 0 satisfying 1 / p = 1 / p 1 + 1 / p 2 , the Schatten p-(quasi-)norm of any matrix X is equivalent to minimizing the product of the Schatten p 1 -(quasi-)norm and Schatten p 2 -(quasi-)norm of its two factor matrices.
In fact, we can find that p 1 and p 2 may have the same value (i.e., p 1 = p 2 = 2 p ) or different values (i.e., p 1 p 2 ). Next, we discuss the two cases for p 1 and p 2 , that is, p 1 = p 2 and p 1 p 2 .

3.1.1. The Case of p 1 = p 2

We first discuss the case when p 1 = p 2 . In fact, for any given 0 < p 1 , there exist infinitely many pairs of positive numbers p 1 and p 2 satisfying 1 / p 1 + 1 / p 2 = 1 / p , such that the equality (4) holds. By setting the same value for p 1 and p 2 (i.e., p 1 = p 2 = 2 p ), we give a unified scalable equivalent formulation for the Schatten p-(quasi-)norm as follows.
Theorem 2.
Given any matrix X R m × n of r a n k ( X ) = r d , then the following equalities hold:
X S p = min U R m × d , V R n × d : X = U V T U S 2 p V S 2 p = min U R m × d , V R n × d : X = U V T U S 2 p 2 p + V S 2 p 2 p 2 1 / p .
Remark 1.
The proof of Theorem 2 is provided in Section 5.2. From the equality in (5), we know that for any 0 < p 1 , the Schatten p-(quasi-)norm minimization problems in many low-rank matrix completion and recovery applications can be transformed into the problems of minimizing the mean of the Schatten 2 p -(quasi-)norms of two smaller factor matrices. It should be noted that when 1 / 2 < p 1 , the norms of two smaller factor matrices are convex and smooth (see Example 2 below) due to 2 p > 1 , which can lead to simpler and more efficient algorithms as in [43].
When p = 1 and p 1 = p 2 = 2 , the equalities in Theorem 2 become the following form [23].
Corollary 1.
Given any matrix X R m × n with r a n k ( X ) = r d , the following equalities hold:
X = min U R m × d , V R n × d : X = U V T U F V F = min U R m × d , V R n × d : X = U V T 1 2 ( U F 2 + V F 2 ) .
The bilinear spectral penalty in (6) has been widely used in many low-rank matrix completion and recovery problems, such as collaborative filtering [9,23], RPCA [45], online RPCA [46], and image recovery [47]. Note that the well-known equivalent formulation of the nuclear norm in Corollary 1 is just a special case of Theorem 2 when p = 1 and p 1 = p 2 = 2 . Moreover, we give two more representative examples for the case of p 1 = p 2 .
Example 1.
When p = 1 / 2 , and by setting p 1 = p 2 = 1 and using Theorem 1, we have the following property, as in our previous work [22,29].
Property 1.
For any matrix X R m × n with r a n k ( X ) = r d , the following equalities hold:
X S 1 / 2 = min U R m × d , V R n × d : X = U V T U V = min U R m × d , V R n × d : X = U V T U + V 2 2 .
In [22,29], the scalable formulations in the above equalities are known as the bi-nuclear quasi-norm. In other words, the bi-nuclear quasi-norm is also a special case of Theorem 2 when p = 1 / 2 and p 1 = p 2 = 1 .
Example 2.
When p = 2 / 3 , and by setting p 1 = p 2 = 4 / 3 and using Theorem 1, we have the following property.
Property 2.
For any matrix X R m × n with r a n k ( X ) = r d , the following equalities hold:
X S 2 / 3 = min U R m × d , V R n × d : X = U V T U S 4 / 3 V S 4 / 3 = min U R m × d , V R n × d : X = U V T U S 4 / 3 4 / 3 + V S 4 / 3 4 / 3 2 3 / 2 .

3.1.2. The Case of p 1 p 2

In this part, we discuss the case of p 1 p 2 . Unlike the case of p 1 = p 2 , we may set infinitely many different values for p 1 and p 2 . For any given 0 < p 1 , there must exist p 1 , p 2 > 0 , at least one of which is no less than 1, such that 1 / p = 1 / p 1 + 1 / p 2 . Indeed, for any 0 < p 1 , the values of p 1 and p 2 may be different (e.g., p 1 = 1 and p 2 = 2 for p = 2 / 3 ), and thus we give the following unified scalable equivalent formulations for the Schatten p-(quasi-)norm. Note that after our work [28] appeared on ArXiv, [12] also presented the same result independently as in Theorem 3.
Theorem 3.
Given any matrix X R m × n of r a n k ( X ) = r d , and any 0 < p 1 , p 1 > 0 and p 2 > 0 satisfying 1 / p 1 + 1 / p 2 = 1 / p , then the following equalities hold:
X S p = min U R m × d , V R n × d : X = U V T U S p 1 V S p 2 = min U R m × d , V R n × d : X = U V T p 2 U S p 1 p 1 + p 1 V S p 2 p 2 p 1 + p 2 1 / p = min U R m × d , V R n × d : X = U V T U S p 1 p 1 / p 1 + V S p 2 p 2 / p 2 1 / p 1 / p .
Remark 2.
The proof of Theorem 3 is given in Section 5.3. From Theorem 3, we know that Theorem 2 and Corollary 1 can be viewed as two special cases of Theorem 3 when p 1 = p 2 = 2 p and p 1 = p 2 = 2 , respectively. That is, Theorem 3 is the more general form of Theorem 2 and Corollary 1. From the equality in (8), we can see that for any 0 < p 1 , the Schatten p-(quasi-)norm minimization problem can be transformed into the one of minimizing the weighted sum of the Schatten p 1 -(quasi-)norm and Schatten p 2 -(quasi-)norm of two smaller factor matrices (see Examples 3 and 4 below), where the weights of the two terms are p 2 / ( p 1 + p 2 ) and p 1 / ( p 1 + p 2 ) , respectively.
Below we give two representative examples for the case of p 1 p 2 .
Example 3.
When p = 2 / 3 , and by setting p 1 = 1 and p 2 = 2 , we have the following property, as in our previous work [22].
Property 3.
Given any matrix X R m × n with r a n k ( X ) = r d , the following equalities hold:
X S 2 / 3 = min U R m × d , V R n × d : X = U V T U V F = min U R m × d , V R n × d : X = U V T 2 U + V F 2 3 3 / 2 .
The scalable formulation in (9) is known as the Frobenius/nuclear hybrid quasi-norm defined in our previous work [22]. It is clear that the Frobenius/nuclear hybrid quasi-norm is also a special case of Theorem 3 when p = 2 / 3 , p 1 = 1 , and p 2 = 2 . As shown in [22,29], we can design more efficient algorithms to solve the Schatten p-quasi-norm minimization with 1 / 2 p < 1 .
Example 4.
When p = 2 / 5 , and by setting p 1 = 1 / 2 and p 2 = 2 , we have the following property.
Property 4.
Given any matrix X R m × n with r a n k ( X ) = r d , the following equalities hold:
X S 2 / 5 = min U R m × d , V R n × d : X = U V T U S 1 / 2 V F = min U R m × d , V R n × d : X = U V T 4 U S 1 / 2 1 / 2 + V F 2 5 5 / 2 .

3.2. Extensions to Multiple Factor Matrices

In this subsection, we extend the unified Schatten quasi-norm formulations of two factor matrices to the case of multiple factor matrices.
Theorem 4.
Each matrix X R m × n of r a n k ( X ) = r d can be decomposed into the product of three much smaller matrices U R m × d , V R d × d and W R n × d , that is, X = U V W T . For any 0 < p 1 and p i > 0 for all i = 1 , 2 , 3 , satisfying 1 / p 1 + 1 / p 2 + 1 / p 3 = 1 / p , then
X S p = min U R m × d , V R d × d , W R n × d : X = U V W T U S p 1 V S p 2 W S p 3 = min U R m × d , V R d × d , W R n × d : X = U V W T U S p 1 p 1 / p 1 + V S p 2 p 2 / p 2 + W S p 3 p 3 / p 3 1 / p 1 / p .
The proof of Theorem 4 is provided in Section 5.4. From Theorem 4, one can see that for any 0 < p 1 and p 1 , p 2 , p 3 > 0 satisfying 1 / p 1 + 1 / p 2 + 1 / p 3 = 1 / p , the Schatten p-(quasi-)norm of any matrix is equivalent to minimizing the weighted sum of the Schatten p 1 -(quasi-)norm, Schatten p 2 -(quasi-)norm, and Schatten p 3 -(quasi-)norm of the three smaller factor matrices, where the weights of the three terms are p / p 1 , p / p 2 , and p / p 3 , respectively. Similarly, we extend the results in Theorem 4 to the case of more factor matrices as follows.
Theorem 5.
Each matrix X R m × n of r a n k ( X ) = r d can be decomposed into the product of multiple much smaller matrices U i , i = 1 , 2 , , M (i.e., X = i = 1 M U i ). For any 0 < p 1 and p i > 0 for all i = 1 , 2 , , M , satisfying i = 1 M 1 / p i = 1 / p , then
X S p = min U i : X = i = 1 M U i i = 1 M U i S p i = min U i : X = i = 1 M U i i = 1 M U i S p i p i / p i 1 / p 1 / p .
The proof of Theorem 5 is very similar to that of Theorem 4, and is thus omitted. From Theorem 5, we can know that for any 0 < p 1 and p i > 0 for all i = 1 , 2 , , M satisfying i = 1 M 1 / p i = 1 / p , the Schatten p-(quasi-)norm of any matrix is equivalent to minimizing the weighted sum of the Schatten p i -(quasi-)norms of the smaller factor matrices, where the weights for these terms are p / p i for all i = 1 , 2 , , M .
Similar to the case of two factor matrices, for any given 0 < p 1 , there exist infinitely many positive numbers p 1 , p 2 , and p 3 such that 1 / p 1 + 1 / p 2 + 1 / p 3 = 1 / p , and the equality (10) holds. By setting the same value for p 1 , p 2 , and p 3 (i.e., p 1 = p 2 = p 3 = 3 p ), we give the following unified scalable equivalent formulations for the Schatten p-(quasi-)norm.
Corollary 2.
Given any matrix X R m × n of r a n k ( X ) = r d , then the following equalities hold:
X S p = min U R m × d , V R d × d , W R n × d : X = U V W T U S 3 p V S 3 p W S 3 p = min U R m × d , V R d × d , W R n × d : X = U V W T U S 3 p 3 p + V S 3 p 3 p + W S 3 p 3 p 3 1 / p .
Remark 3.
The proof of Corollary 2 is provided in Section 5.5. From the equality in (12), we know that for any 0 < p < 1 , Schatten p-quasi-norm minimization can be transformed into the problem of minimizing the mean of the Schatten 3 p -norms (or quasi-norms) of three much smaller factor matrices. We also note that when 1 / 3 < p 1 , the norms of the three factor matrices are convex and smooth because 3 p > 1 , which can also lead to some simpler and more efficient algorithms.
Example 5.
Below we give a representative example. When p = 1 / 3 and p 1 = p 2 = p 3 = 1 , the equalities in Corollary 2 become the following forms, as in our previous work [29].
Property 5.
For any matrix X R m × n of r a n k ( X ) = r d , the following equalities hold:
X S 1 / 3 = min U R m × d , V R d × d , W R n × d : X = U V W T U V W = min U R m × d , V R d × d , W R n × d : X = U V W T U + V + W 3 3 .
From Property 5, we can see that the tri-nuclear quasi-norm defined in our previous work [29] is also a special case of Corollary 2.
Theorem 2 shows that for any p satisfying 1 / 2 < p 1 , the Schatten p-(quasi-)norm of any matrix is equivalent to minimizing the mean of the Schatten 2 p -norms of both much smaller factor matrices, as well as Corollary 2 for any p satisfying 1 / 3 < p 1 . In other words, if 1 / 2 < p 1 or 1 / 3 < p 1 , Schatten p-(quasi-)norm minimization can be transformed into a simpler problem only involving the convex and smooth norms of two or three smaller factor matrices. Moreover, we extend the results of Theorem 2 and Corollary 2 to the case of more factor matrices, as shown in Corollary 3 below. The proof of Corollary 3 is very similar to that of Corollary 2, and is thus omitted. That is, for any 0 < p < 1 , the Schatten p-quasi-norm of any matrix can theoretically be equivalent to the minimization of the mean of the Schatten ( M p ) -norms of all M factor matrices, where M = ( 1 / p + 1 ) . It needs to be strongly emphasized that the norms of all factor matrices are convex and smooth due to M p > 1 , which can help us to design simpler and more efficient algorithms.
Corollary 3.
Given any matrix X R m × n of r a n k ( X ) = r d , the following equalities hold:
X S p = min U i : X = i = 1 M U i i = 1 M U i S M p = min U i : X = i = 1 M U i i = 1 M U i S M p M p M 1 / p .

4. Numerical Experiments

We conducted numerical experiments to evaluate the performance of our scalable Schatten quasi-norm formulations. For a simple reason, we only briefly consider the following bi-nuclear (BN) quasi-norm and Frobenius/nuclear (FN) quasi-norm regularized least-squares problems:
min U , V λ 2 ( U + V ) + 1 2 P Ω ( U V T D ) F 2 , min U , V λ 3 ( 2 U + V F 2 ) + 1 2 P Ω ( U V T D ) F 2 .
Here we introduce the alternating direction method of multipliers (ADMM) for solving the above two scalable Schatten quasi-norm minimization models as in [43]. We compared our BN and FN methods with existing state-of-the-art methods such as weighted nuclear norm minimization (WNNM) [48] (http://www4.comp.polyu.edu.hk/~cslzhang/) and iteratively re-weighted nuclear norm (IRNN) [49], which usually have better performance than traditional nuclear norm solvers such as [50]. As suggested in [13], we chose the q -norm penalty, where q was chosen from the range of { 0.1 , 0.2 , , 1 } . That is, IRNN is one Schatten q-quasi-norm minimization method. The source code of IRNN can be downloaded at the link: https://sites.google.com/site/canyilu/. We selected the same values for the parameters d, ρ , and μ 0 for both our methods (e.g., d = 9 as in [51]), and all the experiments were performed on a PC with an Intel i7-7700 CPU and 32 GB RAM.
We report the inpainting results of all the methods on the lenna and pepper images with 50% random missing pixels, as shown in Figure 1 and Figure 2. The results show that both our methods produced significantly better PSNR results than the other methods. As analyzed in [22,29], our quasi-norm formulations required significantly fewer observations than traditional nuclear norm solvers (e.g., [50]), and also led to scalable optimization problems. In particular, the resulting Schatten quasi-norm formulations may be Lipschitz, while the original Schatten quasi-norm minimization is non-Lipschitz and generally NP-hard [38]. Table 1 shows the running time of all the methods. It is clear that both our methods were more than 50 times faster than the well-known methods, especially IRNN, which suffers from high computational costs due to performing SVD in each iteration.

5. Proofs of Main Results

In this section, we give the proofs for some theorems and corollaries. We first introduce several key lemmas, such as Jensen’s inequality, Hölder’s inequality, and Young’s inequality, which are used throughout our proofs.
Lemma 1 (Jensen’s inequality).
Assume that the function g ( · ) : R + R + is a continuous concave function on [ 0 , + ) . For all t i 0 satisfying i t i = 1 , and any x i R + for i = 1 , , n , then
g i = 1 n t i x i i = 1 n t i g ( x i ) .
Lemma 2 (Hölder’s inequality).
For any p , q > 1 satisfying 1 / p + 1 / q = 1 , then for any x i and y i , i = 1 , , n ,
i = 1 n | x i y i | i = 1 n | x i | p 1 / p i = 1 n | y i | q 1 / q
with equality iff there is a constant c 0 such that each x i p = c y i q .
Lemma 3 (Young’s inequality).
Let a , b 0 and 1 < p , q < be such that 1 / p + 1 / q = 1 . Then
a p p + b q q a b
with equality iff a p = b q .

5.1. Proof of Theorem 1

Before giving a complete proof for Theorem 1, we first present and prove the following lemma.
Lemma 4.
Suppose that Z R m × n is a matrix of rank r min ( m , n ) , and we denote its thin SVD by Z = L Z Σ Z R Z T , where L Z R m × r , R Z R n × r , and Σ Z R r × r . For any A R r × r satisfying A A T = A T A = I r × r , and given p ( 0 < p 1 ) , then ( A Σ Z A T ) k , k 0 for all k = 1 , , r , and
T r p ( A Σ Z A T ) T r p ( Σ Z ) = Z S p p ,
where T r p ( B ) = i B i i p .
Proof. 
For any k { 1 , , r } , we have ( A Σ Z A T ) k , k = i a k i 2 σ i 0 , where σ i 0 is the i-th singular value of Z. Then
T r p ( A Σ Z A T ) = k i a k i 2 σ i p .
Recall that g ( x ) = x p with 0 < p < 1 is a concave function on R + . By Jensen’s inequality [52], as stated in Lemma 1, and i a k i 2 = 1 for any k { 1 , , r } , we have
i a k i 2 σ i p i a k i 2 σ i p .
Using the above inequality and k a k i 2 = 1 for any i { 1 , , r } , (18) can be rewritten as
T r p ( A Σ Z A T ) = k i a k i 2 σ i p k i a k i 2 σ i p = i σ i p = T r p ( Σ Z ) = Z S p p .
In addition, when g ( x ) = x (i.e., p = 1 ), we obtain
i a k i 2 σ i p = i a k i 2 σ i ,
which means that the inequality (19) is still satisfied. □
Proof of Theorem 1.
Let U = L U Σ U R U T and V = L V Σ V R V T be the thin SVDs of U and V, respectively, where L U R m × d , L V R n × d , and R U , Σ U , R V , Σ V R d × d . X = L X Σ X R X T , where the columns of L X R m × d and R X R n × d are the left and right singular vectors associated with the top d singular values of X with rank at most r ( r d ) , and Σ X = diag ( [ σ 1 ( X ) , , σ r ( X ) , 0 , , 0 ] ) R d × d .
Recall that X = U V T (i.e., L X Σ X R X T = L U Σ U R U T R V Σ V L V T ), then O 1 , O ^ 1 R d × d satisfy L X = L U O 1 and L U = L X O ^ 1 , which implies that O 1 = L U T L X and O ^ 1 = L X T L U . Thus, O 1 = O ^ 1 T . Since L X = L U O 1 = L X O ^ 1 O 1 , we have O ^ 1 O 1 = O 1 T O 1 = I d . Similarly, we have O 1 O ^ 1 = O 1 O 1 T = I d . In addition, O 2 R d × d satisfies R X = L V O 2 with O 2 O 2 T = O 2 T O 2 = I d . Let O 3 = O 2 O 1 T R d × d , then we have O 3 O 3 T = O 3 T O 3 = I d , that is, i ( O 3 ) i j 2 = j ( O 3 ) i j 2 = 1 for i , j { 1 , 2 , , d } , where a i , j denotes the element in the i-th row and the j-th column of the matrix A. In addition, let O 4 = R U T R V , we have i ( O 4 ) i j 2 1 and j ( O 4 ) i j 2 1 for i , j { 1 , 2 , , d } .
According to the above analysis, we have O 2 Σ X O 2 T = O 2 O 1 T Σ U O 4 Σ V = O 3 Σ U O 4 Σ V . Let ϱ i and τ j denote the i-th and j-th diagonal elements of Σ U and Σ V , respectively. In the following, we consider the two cases of p 1 and p 2 , that is, at least one of p 1 and p 2 must be no less than 1, or both of them are smaller than 1. It is clear that for any 1 / 2 p 1 and p 1 , p 2 > 0 satisfying 1 / p 1 + 1 / p 2 = 1 / p , at least one of p 1 and p 2 must be no less than 1. On the other hand, only if 0 < p < 1 / 2 , there exist 0 < p 1 < 1 and 0 < p 2 < 1 such that 1 / p 1 + 1 / p 2 = 1 / p , that is, both of them are smaller than 1.
Case 1. For any 0 < p 1 , there exist p 1 > 0 and p 2 > 0 , at least one of which is no less than 1, such that 1 / p 1 + 1 / p 2 = 1 / p . Without loss of generality, we assume that p 2 1 . Here, we set k 1 = p 1 / p and k 2 = p 2 / p . Clearly, we can know that k 1 , k 2 > 1 and 1 / k 1 + 1 / k 2 = 1 . From Lemma 4, we have
X S p T r p ( O 2 Σ X O 2 T ) 1 / p = T r p ( O 2 O 1 T Σ U O 4 Σ V ) 1 / p = T r p ( O 3 Σ U O 4 Σ V ) 1 / p = i = 1 d j = 1 d τ j ( O 3 ) i j ( O 4 ) j i ϱ i p 1 / p = i = 1 d ϱ i p j = 1 d τ j ( O 3 ) i j ( O 4 ) j i p 1 / p a i = 1 d ( ϱ i p ) k 1 1 / k 1 i = 1 d j = 1 d τ j ( O 3 ) i j ( O 4 ) j i p × k 2 1 / k 2 1 / p = i = 1 d ϱ i p 1 1 / p 1 i = 1 d j = 1 d τ j ( O 3 ) i j ( O 4 ) j i p 2 1 / p 2 b i = 1 d ϱ i p 1 1 / p 1 i = 1 d j = 1 d τ j ( O 3 ) i j 2 + ( O 4 ) j i 2 2 p 2 1 / p 2 c i = 1 d ϱ i p 1 1 / p 1 j = 1 d τ j p 2 1 / p 2 ,
where the inequality a holds due to the Hölder’s inequality [52], as stated in Lemma 2. The inequality b follows from the basic inequality x y x 2 + y 2 2 for any real numbers x and y, and the inequality c relies on the facts that i ( O 3 ) i j 2 = 1 and i ( O 4 ) j i 2 1 , and we apply Jensen’s inequality (see Lemma 1) for the convex function h ( x ) = x p 2 with p 2 1 .
For any matrices U R m × d and V R n × d satisfying X = U V T , we have
X S p U S p 1 V S p 2 .
On the other hand, let U = L X Σ X p / p 1 and V = R X Σ X p / p 2 , where Σ X p is entry-wise power to p, then we obtain
X = U V T , U S p 1 p 1 = V S p 2 p 2 = X S p p with 1 / p = 1 / p 1 + 1 / p 2 ,
and
X S p = T r p ( Σ X ) 1 / p = U S p 1 V S p 2 .
Therefore, under the constraint X = U V T , we have
X S p = min U R m × d , V R n × d : X = U V T U S p 1 V S p 2 .
Case 2. For any 0 < p < 1 / 2 , there exist 0 < p ^ 1 < 1 and 0 < p ^ 2 < 1 such that 1 / p ^ 1 + 1 / p ^ 2 = 1 / p . Next, we prove that the result of Theorem 1 also holds. In fact, for any given p, there must exist p 1 > 0 and p 2 1 such that 1 / p 1 + 1 / p 2 = 1 / p and 1 / p 1 = 1 / p ^ 1 + 1 / q with q 1 . Clearly, we can know that 1 / p ^ 1 < 1 / p 1 . Let U = L X Σ X p / p 1 , V = R X Σ X p / p 2 , U 1 = L X Σ X p / p ^ 1 and V 1 = R X Σ X p / p ^ 2 , then we have X = U ( V ) T = U 1 ( V 1 ) T , which implies that
X S p = U S p 1 V S p 2 = U 1 S p ^ 1 V 1 S p ^ 1 .
Since 1 / p = 1 / p 1 + 1 / p 2 = 1 / p ^ 1 + 1 / p ^ 2 and 1 / p 1 = 1 / p ^ 1 + 1 / q , then 1 / p ^ 2 = 1 / q + 1 / p 2 . Consider any factor matrices U and V satisfying X = U V T , V = L V Σ V R V T is the thin SVD of V. Let U 1 = U U 2 T and V 1 = L V Σ V p ^ 2 / p 2 , where U 2 T = R V Σ V p ^ 2 / q , and thus it is not difficult to verify that
V = V 1 U 2 , X = U 1 V 1 T , V S p ^ 2 = U 2 S q V 1 S p 2 , U 1 S p 1 U S p ^ 1 U 2 S q ,
where the above inequality follows from (20) with q 1 . Combining (21) and (22) and for any U and V satisfying X = U V T , we have
X S p = U S p 1 V S p 2 U 1 S p 1 V 1 S p 2 U S p ^ 1 U 2 S q V 1 S p 2 = U S p ^ 1 V S p ^ 2 ,
where the first inequality follows from (20). Recall that
X S p = U 1 S p ^ 1 V 1 S p ^ 2 .
Therefore, for any 0 < p ^ 1 < 1 and 0 < p ^ 2 < 1 satisfying 1 / p = 1 / p ^ 1 + 1 / p ^ 2 , and by (23) and (24), we also have
X S p = min U R m × d , V R n × d : X = U V T U S p ^ 1 V S p ^ 2 .
In summary, for any 0 < p 1 , p 1 > 0 and p 2 > 0 satisfying 1 / p = 1 / p 1 + 1 / p 2 , we have
X S p = min U R m × d , V R n × d : X = U V T U S p 1 V S p 2 .
This completes the proof. □

5.2. Proof of Theorem 2

Proof. 
Since p 1 = p 2 = 2 p > 0 , 1 / p 1 + 1 / p 2 = 1 / p , and using Theorem 1,
X S p = min U R m × d , V R n × d : X = U V T U S 2 p V S 2 p .
Due to the basic inequality x y x 2 + y 2 2 for any real numbers x and y, we have
X S p = min U , V : X = U V T U S 2 p V S 2 p = min U , V : X = U V T U S 2 p p V S 2 p p 1 / p min U , V : X = U V T U S 2 p 2 p + V S 2 p 2 p 2 1 / p .
Let U = L X Σ X 1 / 2 and V = R X Σ X 1 / 2 , where Σ X 1 / 2 is entry-wise power to 1 / 2 . Then, we obtain
X = U V T , U S 2 p 2 p = V S 2 p 2 p = X S p p ,
which implies that
X S p = U S 2 p V S 2 p = U S 2 p 2 p + V S 2 p 2 p 2 1 / p .
The theorem now follows because
min U , V : X = U V T U S 2 p V S 2 p = min U , V : X = U V T U S 2 p 2 p + V S 2 p 2 p 2 1 / p .
This completes the proof. □

5.3. Proof of Theorem 3

Proof. 
For any 0 < p 1 , p 1 > 0 and p 2 > 0 satisfying 1 / p 1 + 1 / p 2 = 1 / p , and using Theorem 1, we have
X S p = min U R m × d , V R n × d : X = U V T U S p 1 V S p 2 .
Let k 1 = ( p 1 + p 2 ) / p 2 and k 2 = ( p 1 + p 2 ) / p 1 , and it is easy to know that 1 / k 1 + 1 / k 2 = 1 . Then
X S p = min U , V : X = U V T U S p 1 V S p 2 = min U , V : X = U V T U S p 1 p V S p 2 p 1 / p min U , V : X = U V T U S p 1 p k 1 k 1 + V S p 2 p k 2 k 2 1 / p = min U , V : X = U V T p 2 U S p 1 p 1 + p 1 V S p 2 p 2 p 1 + p 2 1 / p
where the above inequality follows from the well-known Young inequality and the monotone increasing property of the function g ( x ) = x 1 / p .
Let U = L X Σ X p / p 1 and V = R X Σ X p / p 2 , then X = U V T and
X S p = min U R m × d , V R n × d : X = U V T U S p 1 V S p 2 = U S p 1 V S p 2 = p 2 U S p 1 p 1 + p 1 V S p 2 p 2 p 1 + p 2 1 / p
which implies that
X S p = min U R m × d , V R n × d : X = U V T U S p 1 V S p 2 = min U , V : X = U V T p 2 U S p 1 p 1 + p 1 V S p 2 p 2 p 1 + p 2 1 / p = min U , V : X = U V T U S p 1 p 1 / p 1 + V S p 2 p 2 / p 2 1 / p 1 / p .
This completes the proof. □

5.4. Proof of Theorem 4

Proof. 
Let U R m × d and V ^ R n × d be any factor matrices such that X = U V ^ T , and p ^ 1 = p 1 > 0 and p ^ 2 = p 2 p 3 / ( p 2 + p 3 ) > 0 , which means that 1 / p ^ 1 + 1 / p ^ 2 = 1 / p . According to Theorem 1, we obtain
X S p = min U R m × d , V ^ R n × d : X = U V ^ T U S p ^ 1 V ^ S p ^ 2 .
Let V R d × d and W R n × d be factor matrices of V ^ (i.e., V W T = V ^ T ). Since p ^ 2 = p 2 p 3 / ( p 2 + p 3 ) , then 1 / p ^ 2 = 1 / p 2 + 1 / p 3 and
V ^ S p ^ 2 = min V R d × d , W R n × d : V ^ = ( V W T ) T V S p 2 W S p 3 .
Combining (26) and (27), we obtain
X S p = min U , V , W : X = U V W T U S p 1 V S p 2 W S p 3 .
Using the above result, we have
X S p = min U , V , W : X = U V W T U S p 1 V S p 2 W S p 3 = min U , V , W : X = U V W T U S p 1 p V S p 2 p W S p 3 p 1 / p min U , V , W : X = U V W T p 2 p 3 U S p 1 p 1 + p 1 p 3 V S p 2 p 2 + p 1 p 2 W S p 3 p 3 p 2 p 3 + p 1 p 3 + p 1 p 2 1 / p ,
where the above inequality follows from the well-known Young inequality.
Let U = L X Σ X p / p 1 , V = Σ X p / p 2 , and W = R X Σ X p / p 3 . It is easy to verify that X = U V W T . Then, we have
X S p = min U , V , W : X = U V W T U S p 1 V S p 2 W S p 3 = U S p 1 V S p 2 W S p 3 = p 2 p 3 U S p 1 p 1 + p 1 p 3 V S p 2 p 2 + p 1 p 2 W S p 3 p 3 p 2 p 3 + p 1 p 3 + p 1 p 2 1 / p .
Therefore, we have
X S p = min U , V , W : X = U V W T U S p 1 V S p 2 W S p 3 = min U , V , W : X = U V W T p 2 p 3 U S p 1 p 1 + p 1 p 3 V S p 2 p 2 + p 1 p 2 W S p 3 p 3 p 2 p 3 + p 1 p 3 + p 1 p 2 1 / p = min U , V , W : X = U V W T U S p 1 p 1 / p 1 + V S p 2 p 2 / p 2 + W S p 3 p 3 / p 3 1 / p 1 / p .
This completes the proof. □

5.5. Proof of Corollary 2

Proof. 
Since p 1 = p 2 = p 3 = 3 p > 0 and 1 / p 1 + 1 / p 2 + 1 / p 3 = 1 / p , and using Theorem 4, we have
X S p = min U , V , W : X = U V W T U S 3 p V S 3 p W S 3 p .
From the basic inequality x y z x 3 + y 3 + z 3 3 for arbitrary positive numbers x, y, and z, we obtain
X S p = min U , V , W : X = U V W T U S 3 p V S 3 p W S 3 p = min U , V , W : X = U V W T U S 3 p p V S 3 p p W S 3 p p 1 / p min U , V , W : X = U V W T U S 3 p 3 p + V S 3 p 3 p + W S 3 p 3 p 3 1 / p .
Let U = L X Σ X 1 / 3 , V = Σ X 1 / 3 , and W = R X Σ X 1 / 3 , where Σ X 1 / 3 is entry-wise power to 1 / 3 , then we have
X = U V W T , U S 3 p 3 p = V S 3 p 3 p = W S 3 p 3 p = X S p p ,
which implies that
X S p = U S 3 p V S 3 p W S 3 p = U S 3 p 3 p + V S 3 p 3 p + W S 3 p 3 p 3 1 / p .
The theorem now follows because
min U , V , W : X = U V W T U S 3 p V S 3 p W S 3 p = min U , V , W : X = U V W T U S 3 p 3 p + V S 3 p 3 p + W S 3 p 3 p 3 1 / p .
This completes the proof. □

6. Conclusions

Generally, the Schatten quasi-norm minimization problem is non-convex, non-smooth, and even non-Lipschitz, and thus most existing algorithms are too slow or even impractical for large-scale matrix recovery and completion problems. Therefore, it is very important to know how to transform such challenging problems into simpler ones. In this paper, we first presented and rigorously proved that for any p, p 1 , p 2 > 0 satisfying 1 / p = 1 / p 1 + 1 / p 2 , the Schatten p-quasi-norm of any matrix is equivalent to the minimization of the product (or weighted sum) of the Schatten p 1 -(quasi-)norm and Schatten p 2 -(quasi-)norm of two much smaller factor matrices. In particular, when p > 1 / 2 , there is an equivalence between the Schatten p-(quasi-)norm of any matrix and the Schatten 2 p -norms of its two factor matrices (i.e., Property 1). That is, the Schatten quasi-norm minimization problem with p > 1 / 2 can be transformed into a simpler one only involving the smooth norms of smaller factor matrices, which can naturally lead to simpler and more efficient algorithms (e.g., [12,22,29,30]).
We further extended the equivalence formulation of two factor matrices to the cases of three and more factor matrices, from which we can see that for any 0 < p < 1 , the Schatten p-quasi-norm of any matrix is the minimization of the mean of the Schatten ( 1 / p + 1 ) p -norms of 1 / p + 1 factor matrices. In other words, for any 0 < p < 1 , the Schatten quasi-norm minimization problem can be transformed into an optimization problem only involving the smooth norms of multiple factor matrices. Finally, we provided some representative examples for two and three factor matrices. It is clear that the bi-nuclear and Frobenius/nuclear quasi-norms in our previous work [22] and the tri-nuclear quasi-norm in our previous work [29] are three important special cases. Our future work is the theoretical analysis of the properties of the proposed unified Schatten p-norm formulations compared to the nuclear norm and the Schatten quasi-norm as in [53,54] (e.g., how many observations are sufficient for our models to reliably recover low-rank matrices).

Author Contributions

Methodology and Formal analysis, F.S. (Fanhua Shang); Formal analysis, Y.L.; Investigation, L.K.; Software, F.S. (Fanjie Shang); Visualization, H.L.; Review & editing, L.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Nos. 61876220, 61876221, 61976164, 61836009 and U1701267, and 61871310), the Project supported the Foundation for Innovative Research Groups of the National Natural Science Foundation of China (No. 61621005), the Program for Cheung Kong Scholars and Innovative Research Team in University (No. IRT_15R53), the Fund for Foreign Scholars in University Research and Teaching Programs (the 111 Project) (No. B07048), the Science Foundation of Xidian University (Nos. 10251180018 and 10251180019), the National Science Basic Research Plan in Shaanxi Province of China (Nos. 2019JQ-657 and 2020JM-194), and the Key Special Project of China High Resolution Earth Observation System-Young Scholar Innovation Fund.

Acknowledgments

We thank all the reviewers for their valuable comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Candès, E.; Recht, B. Exact Matrix Completion via Convex Optimization. Found. Comput. Math. 2009, 9, 717–772. [Google Scholar] [CrossRef] [Green Version]
  2. Candès, E.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? J. ACM 2011, 58, 1–37. [Google Scholar] [CrossRef]
  3. Liu, G.; Lin, Z.; Yu, Y. Robust Subspace Segmentation by Low-Rank Representation. In Proceedings of the 27th International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 663–670. [Google Scholar]
  4. Yuan, M.; Ekici, A.; Lu, Z.; Monteiro, R. Dimension reduction and coefficient estimation in multivariate linear regression. J. R. Stat. Soc. B 2007, 69, 329–346. [Google Scholar] [CrossRef]
  5. Argyriou, A.; Micchelli, C.A.; Pontil, M.; Ying, Y. A Spectral Regularization Framework for Multi-Task Structure Learning. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 3–6 December 2007; pp. 25–32. [Google Scholar]
  6. Liu, Z.; Vandenberghe, L. Interior-Point Method for Nuclear Norm Approximation with Application to System Identification. SIAM J. Matrix Anal. Appl. 2009, 31, 1235–1256. [Google Scholar] [CrossRef]
  7. Fazel, M.; Hindi, H.; Boyd, S.P. A Rank Minimization Heuristic with Application to Minimum Order System Approximation. In Proceedings of the 2001 American Control Conference, Arlington, VA, USA, 25–27 June 2001; pp. 4734–4739. [Google Scholar]
  8. Candès, E.; Tao, T. The power of convex relaxation: Near-optimal matrix completion. IEEE Trans. Inform. Theory 2010, 56, 2053–2080. [Google Scholar] [CrossRef] [Green Version]
  9. Recht, B.; Fazel, M.; Parrilo, P.A. Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization. SIAM Rev. 2010, 52, 471–501. [Google Scholar] [CrossRef] [Green Version]
  10. Fan, J.; Li, R. Variable selection via nonconcave penalized likelihood and its Oracle properties. J. Am. Stat. Assoc. 2001, 96, 1348–1361. [Google Scholar] [CrossRef]
  11. Zhang, C.H. Nearly unbiased variable selection under minimax concave penalty. Ann. Stat. 2010, 38, 894–942. [Google Scholar] [CrossRef] [Green Version]
  12. Xu, C.; Lin, Z.; Zha, H. A Unified Convex Surrogate for the Schatten-p Norm. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 926–932. [Google Scholar]
  13. Lu, C.; Tang, J.; Yan, S.; Lin, Z. Generalized Nonconvex Nonsmooth Low-Rank Minimization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 4130–4137. [Google Scholar]
  14. Nie, F.; Wang, H.; Cai, X.; Huang, H.; Ding, C. Robust Matrix Completion via Joint Schatten p-Norm and Lp-Norm Minimization. In Proceedings of the 2012 IEEE 12th International Conference on Data Mining, Brussels, Belgium, 10 December 2012; pp. 566–574. [Google Scholar]
  15. Aravkin, A.; Kumar, R.; Mansour, H.; Recht, B.; Herrmann, F.J. Fast methods for denoising matrix completion formulations, with applications to robust seismic data interpolation. SIAM J. Sci. Comput. 2014, 36, S237–S266. [Google Scholar] [CrossRef] [Green Version]
  16. Majumdar, A.; Ward, R.K. An algorithm for sparse MRI reconstruction by Schatten p-norm minimization. Magn. Reson. Imaging 2011, 29, 408–417. [Google Scholar] [CrossRef]
  17. Mohan, K.; Fazel, M. Iterative Reweighted Algorithms for Matrix Rank Minimization. J. Mach. Learn. Res. 2012, 13, 3441–3473. [Google Scholar]
  18. Lai, M.; Xu, Y.; Yin, W. Improved iteratively rewighted least squares for unconstrained smoothed p minimization. SIAM J. Numer. Anal. 2013, 51, 927–957. [Google Scholar] [CrossRef]
  19. Nie, F.; Huang, H.; Ding, C. Low-Rank Matrix Recovery via Efficient Schatten p-Norm Minimization. In Proceedings of the Twenty-sixth AAAI conference on artificial intelligence, Toronto, AB, Canada, 22–26 July 2012; pp. 655–661. [Google Scholar]
  20. Marjanovic, G.; Solo, V. On p Optimization and Matrix Completion. IEEE Trans. Signal Process. 2012, 60, 5714–5724. [Google Scholar] [CrossRef]
  21. Zhang, M.; Huang, Z.; Zhang, Y. Restricted p-Isometry Properties of Nonconvex Matrix Recovery. IEEE Trans. Inform. Theory 2013, 59, 4316–4323. [Google Scholar] [CrossRef]
  22. Shang, F.; Liu, Y.; Cheng, J. Scalable Algorithms for Tractable Schatten Quasi-Norm Minimization. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 2016–2022. [Google Scholar]
  23. Srebro, N.; Rennie, J.; Jaakkola, T. Maximum-Margin Matrix Factorization. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 13–18 December 2004; pp. 1329–1336. [Google Scholar]
  24. Mazumder, R.; Hastie, T.; Tibshirani, R. Spectral regularization algorithms for learning large incomplete matrices. J. Mach. Learn. Res. 2010, 11, 2287–2322. [Google Scholar] [PubMed]
  25. Mitra, K.; Sheorey, S.; Chellappa, R. Large-Scale Matrix Factorization with Missing Data under Additional Constraints. In Proceedings of the 24th Annual Conference on Neural Information Processing Systems 2010, Vancouver, BC, Canada, 6–9 December 2010; pp. 1651–1659. [Google Scholar]
  26. Recht, B.; Ré, C. Parallel stochastic gradient algorithms for large-scale matrix completion. Math. Prog. Comp. 2013, 5, 201–226. [Google Scholar] [CrossRef] [Green Version]
  27. Zuo, W.; Meng, D.; Zhang, L.; Feng, X.; Zhang, D. A Generalized Iterated Shrinkage Algorithm for Non-convex Sparse Coding. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 217–224. [Google Scholar]
  28. Shang, F.; Liu, Y.; Cheng, J. Unified Scalable Equivalent Formulations for Schatten Quasi-Norms. arXiv 2016, arXiv:1606.00668. [Google Scholar]
  29. Shang, F.; Liu, Y.; Cheng, J. Tractable and Scalable Schatten Quasi-Norm Approximations for Rank Minimization. In Proceedings of the Artificial Intelligence and Statistics, Cadiz, Spain, 9–11 May 2016; pp. 620–629. [Google Scholar]
  30. Yang, L.; Pong, T.K.; Chen, X. A Nonmonotone Alternating Updating Method for a Class of Matrix Factorization Problems. SIAM J. Optim. 2018, 28, 3402–3430. [Google Scholar] [CrossRef] [Green Version]
  31. Foucart, S.; Lai, M. Sparsest solutions of underdetermined linear systems via q-minimization for 0 < q ≤ 1. Appl. Comput. Harmon. Anal. 2009, 26, 397–407. [Google Scholar]
  32. Liu, Y.; Shang, F.; Cheng, H.; Cheng, J. A Grassmannian Manifold Algorithm for Nuclear Norm Regularized Least Squares Problems. In Proceedings of the Thirtieth Conference on Uncertainty in Artificial Intelligence, Quebec City, QC, Canada, 23–27 July 2014; pp. 515–524. [Google Scholar]
  33. Shang, F.; Liu, Y.; Cheng, J.; Cheng, H. Robust Principal Component Analysis with Missing Data. In Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, Shanghai, China, 3–7 November 2014; pp. 1149–1158. [Google Scholar]
  34. Xu, H.; Caramanis, C.; Sanghavi, S. Robust PCA via outlier pursuit. In Proceedings of the 24th Annual Conference on Neural Information Processing Systems 2010, Vancouver, BC, Canada, 6–9 December 2010; pp. 2496–2504. [Google Scholar]
  35. Chen, Y.; Jalali, A.; Sanghavi, S.; Caramanis, C. Low-rank matrix recovery from errors and erasures. IEEE Trans. Inform. Theory 2013, 59, 4324–4337. [Google Scholar] [CrossRef] [Green Version]
  36. Shang, F.; Liu, Y.; Tong, H.; Cheng, J.; Cheng, H. Robust bilinear factorization with missing and grossly corrupted observations. Inform. Sci. 2015, 307, 53–72. [Google Scholar] [CrossRef]
  37. Hsieh, C.; Olsen, P.A. Nuclear Norm Minimization via Active Subspace Selection. In Proceedings of the International Conference on Machine Learning, Beijing, China, 21–26 June 2014; pp. 575–583. [Google Scholar]
  38. Bian, W.; Chen, X.; Ye, Y. Complexity analysis of interior point algorithms for non-Lipschitz and nonconvex minimization. Math. Program. 2015, 149, 301–327. [Google Scholar] [CrossRef] [Green Version]
  39. Larsen, R. PROPACK-Software for Large and Sparse SVD Calculations. 2015. Available online: http://sun.stanford.edu/srmunk/PROPACK/ (accessed on 28 March 2020).
  40. Cai, J.F.; Osher, S. Fast Singular Value Thresholding without Singular Value Decomposition. Methods Anal. Appl. 2013, 20, 335–352. [Google Scholar] [CrossRef] [Green Version]
  41. Liu, G.; Yan, S. Active subspace: Toward scalable low-rank learning. Neural Comp. 2012, 24, 3371–3394. [Google Scholar] [CrossRef]
  42. Shang, F.; Liu, Y.; Cheng, J. Generalized Higher-Order Tensor Decomposition via Parallel ADMM. In Proceedings of the 28th AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014; pp. 1279–1285. [Google Scholar]
  43. Shang, F.; Cheng, J.; Liu, Y.; Luo, Z.Q.; Lin, Z. Bilinear Factor Matrix Norm Minimization for Robust PCA: Algorithms and Applications. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 2066–2080. [Google Scholar] [CrossRef] [Green Version]
  44. Krechetov, M.; Marecek, J.; Maximov, Y.; Takac, M. Entropy-Penalized Semidefinite Programming. arXiv 2019, arXiv:1802.04332. [Google Scholar]
  45. Cabral, R.; Torre, F.; Costeira, J.; Bernardino, A. Unifying Nuclear Norm and Bilinear Factorization Approaches for Low-rank Matrix Decomposition. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2488–2495. [Google Scholar]
  46. Feng, J.; Xu, H.; Yan, S. Online robust PCA via stochastic optimization. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 404–412. [Google Scholar]
  47. Jin, K.H.; Ye, J.C. Annihilating Filter-Based Low-Rank Hankel Matrix Approach for Image Inpainting. IEEE Trans. Image Process. 2015, 24, 3498–3511. [Google Scholar]
  48. Gu, S.; Xie, Q.; Meng, D.; Zuo, W.; Feng, X.; Zhang, L. Weighted Nuclear Norm Minimization and Its Applications to Low Level Vision. Int. J. Comput. Vis. 2017, 121, 183–208. [Google Scholar] [CrossRef]
  49. Lu, C.; Tang, J.; Yan, S.; Lin, Z. Nonconvex Nonsmooth Low Rank Minimization via Iteratively Reweighted Nuclear Norm. IEEE Trans. Image Process. 2016, 25, 829–839. [Google Scholar] [CrossRef] [Green Version]
  50. Toh, K.C.; Yun, S. An accelerated proximal gradient algorithm for nuclear norm regularized least squares problems. Pac. J. Optim. 2010, 6, 615–640. [Google Scholar]
  51. Hu, Y.; Zhang, D.; Ye, J.; Li, X.; He, X. Fast and Accurate Matrix Completion via Truncated Nuclear Norm Regularization. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2117–2130. [Google Scholar] [CrossRef] [PubMed]
  52. Mitrinović, D.S. Analytic Inequalities; Springer: Berline/Heidelberg, Germany, 1970. [Google Scholar]
  53. Xu, Z. The minimal measurement number for low-rank matrix recovery. Appl. Comput. Harmon. Anal. 2018, 44, 497–508. [Google Scholar] [CrossRef]
  54. Zhang, R.; Li, S. Optimal RIP bounds for sparse signals recovery via p minimization. Appl. Comput. Harmon. Anal. 2019, 47, 566–584. [Google Scholar] [CrossRef]
Figure 1. Image inpainting results of IRNN [49], WNNM [48], and our two methods on the lenna image with 50% random missing pixels (best viewed in color).
Figure 1. Image inpainting results of IRNN [49], WNNM [48], and our two methods on the lenna image with 50% random missing pixels (best viewed in color).
Mathematics 08 01325 g001
Figure 2. Image inpainting results of IRNN [49], WNNM [48], and our two methods on the pepper image with 50% random missing pixels (best viewed in color).
Figure 2. Image inpainting results of IRNN [49], WNNM [48], and our two methods on the pepper image with 50% random missing pixels (best viewed in color).
Mathematics 08 01325 g002
Table 1. Comparison of running time of the methods (in seconds).
Table 1. Comparison of running time of the methods (in seconds).
IRNN [49]WNNM [48]BNFN
lenna2381.04389.537.065.27
pepper2257.39382.787.355.31

Share and Cite

MDPI and ACS Style

Shang, F.; Liu, Y.; Shang, F.; Liu, H.; Kong, L.; Jiao, L. A Unified Scalable Equivalent Formulation for Schatten Quasi-Norms. Mathematics 2020, 8, 1325. https://0-doi-org.brum.beds.ac.uk/10.3390/math8081325

AMA Style

Shang F, Liu Y, Shang F, Liu H, Kong L, Jiao L. A Unified Scalable Equivalent Formulation for Schatten Quasi-Norms. Mathematics. 2020; 8(8):1325. https://0-doi-org.brum.beds.ac.uk/10.3390/math8081325

Chicago/Turabian Style

Shang, Fanhua, Yuanyuan Liu, Fanjie Shang, Hongying Liu, Lin Kong, and Licheng Jiao. 2020. "A Unified Scalable Equivalent Formulation for Schatten Quasi-Norms" Mathematics 8, no. 8: 1325. https://0-doi-org.brum.beds.ac.uk/10.3390/math8081325

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop