s-Goodness for Low-Rank Matrix Recovery
Abstract
Low-rank matrix recovery (LMR) is a rank minimization problem subject to linear equality constraints, and it arises in many fields such as signal and image processing, statistics, computer vision, and system identification and control. This class of optimization problems is generally 𝒩𝒫 hard. A popular approach replaces the rank function with the nuclear norm of the matrix variable. In this paper, we extend and characterize the concept of s-goodness for a sensing matrix in sparse signal recovery (proposed by Juditsky and Nemirovski (Math Program, 2011)) to linear transformations in LMR. Using the two characteristic s-goodness constants, γs and , of a linear transformation, we derive necessary and sufficient conditions for a linear transformation to be s-good. Moreover, we establish the equivalence of s-goodness and the null space properties. Therefore, s-goodness is a necessary and sufficient condition for exact s-rank matrix recovery via the nuclear norm minimization.
1. Introduction
LMR problems have many applications and they appeared in the literature of a diverse set of fields including signal and image processing, statistics, computer vision, and system identification and control. For more details, see the recent paper [4]. LMR and NNM have been the focus of some recent research in the optimization community, see; for example, [4–15]. Although there are many papers dealing with algorithms for NNM such as interior-point methods, fixed point and Bregman iterative methods, and proximal point methods, there are fewer papers dealing with the conditions that guarantee the success of the low-rank matrix recovery via NNM. For instance, following the program laid out in the work of Candès and Tao in compressed sensing (CS, see, e.g., [16–18]), Recht et al. [4] provided a certain restricted isometry property (RIP) condition on the linear transformation which guarantees that the minimum nuclear norm solution is the minimum rank solution. Recht et al. [14, 19] gave the null space property (NSP) which characterizes a particular property of the null space of the linear transformation, which is also discussed by Oymak et al. [20, 21]. Note that NSP states a necessary and sufficient condition for exactly recovering the low-rank matrix via nuclear norm minimization. Recently, Chandrasekaran et al. [22] proposed that a fixed s-rank matrix X0 can be recovered if and only if the null space of 𝒜 does not intersect the tangent cone of the nuclear norm ball at X0.
In the setting of CS, there are other characterizations of the sensing matrix, under which ℓ1-norm minimization can be guaranteed to yield an optimal solution to SSR, in addition to RIP and null-space properties, see; for example, [23–26]. In particular, Juditsky and Nemirovski [24] established necessary and sufficient conditions for a Sensing matrix to be “s-good” to allow for exact ℓ1-recovery of sparse signals with s nonzero entries when no measurement noise is present. They also demonstrated that these characteristics, although difficult to evaluate, lead to verifiable sufficient conditions for exact SSR and to efficiently computable upper bounds on those s for which a given sensing matrix is s-good. Furthermore, they established instructive links between s-goodness and RIP in the CS context. One may wonder whether we can generalize the s-goodness concept to LMR and still maintain many of the nice properties as done in [24]. Here, we deal with this issue. Our approach is based on the singular value decomposition (SVD) of a matrix and the partition technique generalized from CS. In the next section, following Juditsky and Nemirovski’s terminology, we propose definitions of s-goodness and G-numbers, γs and , of a linear transformation in LMR and then we provide some basic properties of G-numbers. In Section 3, we characterize s-goodness of a linear transformation in LMR via G-numbers. We consider the connections between the s-goodness, NSP, and RIP in Section 4. We eventually obtain that δ2s < 0.472⇒𝒜 satisfying is s-good.
2. Definitions and Basic Properties
2.1. Definitions
We first go over some concepts related to s-goodness of the linear transformation in LMR (RMP). These are extensions of those given for SSR (CMP) in [24].
Definition 1. Let 𝒜 : ℝm×n → ℝp be a linear transformation and s ∈ {0,1, 2, …, r}. One says that 𝒜 is s-good, if for every s-rank matrix W ∈ ℝm×n, W is the unique optimal solution to the optimization problem
We denote by s*(𝒜) the largest integer s for which 𝒜 is s-good. Clearly, s*(𝒜)∈{0,1, …, r}. To characterize s-goodness we introduce two useful s-goodness constants: γs and . We call γs and G-numbers.
Definition 2. Let 𝒜 : ℝm×n → ℝp be a linear transformation, β ∈ [0, +∞] and s ∈ {0,1, 2, …, r}. Then we have the following.
(i) G-number γs(𝒜, β) is the infimum of γ ≥ 0 such that for every matrix X ∈ ℝm×n with singular value decomposition (i.e., s nonzero singular values, all equal to 1), there exists a vector y ∈ ℝp such that
(ii) G-number is the infimum of γ ≥ 0 such that for every matrix X ∈ ℝm×n with s nonzero singular values, all equal to 1, there exists a vector y ∈ ℝp such that 𝒜*y and X share the same orthogonal row and column spaces:
2.2. Basic Properties of G-Numbers
In order to characterize the s-goodness of a linear transformation 𝒜, we study the basic properties of G-numbers. We begin with the result that G-numbers γs(𝒜, β) and are convex nonincreasing functions of β.
Proposition 3. For every linear transformation 𝒜 and every s ∈ {0,1, …, r}, G-numbers γs(𝒜, β) and are convex nonincreasing functions of β ∈ [0, +∞].
Proof. We only need to demonstrate that the quantity γs(𝒜, β) is a convex nonincreasing function of β ∈ [0, +∞]. It is evident from the definition that γs(𝒜, β) is nonincreasing for given 𝒜, s. It remains to show that γs(𝒜, β) is a convex function of β. In other words, for every pair β1, β2 ∈ [0, +∞], we need to verify that
The following observation that G-numbers γs(𝒜, β), are nondecreasing in s is immediate.
Proposition 4. For every s′ ≤ s, one has , .
We further investigate the relationship between the G-numbers γs(𝒜, β) and .
Proposition 5. Let 𝒜 : ℝm×n → ℝp be a linear transformation, β ∈ [0, +∞], and s ∈ {0,1, 2, …, r}. Then one has
Proof. Let γ : = γs(𝒜, β) < 1. Then, for every matrix Z ∈ ℝm×n with s nonzero singular values, all equal to 1, there exists y ∈ ℝp, ∥y∥d ≤ β, such that 𝒜*y = Z + W, where ∥W∥ ≤ γ and W and Z have orthogonal row and column spaces. For a given pair Z, y as above, take . Then we have and
Claim 1. SJ contains the ∥·∥∞-ball of radius centered at the origin in ℝr.
Proof. Note that SJ is closed and convex. Moreover, SJ is the direct sum of its projections onto the pair of subspaces
Using the above claim, we conclude that for every J⊆{1,2, …, r} with cardinality s, there exists an x ∈ SJ such that , for all i ∈ J. From the definition of SJ, we obtain that there exists y ∈ ℝp with such that
To conclude the proof, we need to prove that the inequalities we established
We end this section with a simple argument which illustrates that for a given pair (𝒜, s), γs(𝒜, β) = γs(𝒜) and , for all β large enough.
Proposition 6. Let 𝒜 : ℝm×n → ℝp be a linear transformation and β ∈ [0, +∞]. Assume that for some ρ > 0, the image of the unit ∥·∥*-ball in ℝm×n under the mapping X ↦ 𝒜X contains the ball B = {x ∈ ℝp : ∥x∥1 ≤ ρ}. Then for every s ∈ {1,2, …, r}
Proof. Fix s ∈ {1,2, …, r}. We only need to show the first implication. Let γ : = γs(𝒜) < 1. Then for every matrix W ∈ ℝm×n with its SVD , there exists a vector y ∈ ℝp such that
3. s-Goodness and G-Numbers
We first give the following characterization result of s-goodness of a linear transformation 𝒜 via the G-number γs(𝒜), which explains the importance of γs(𝒜) in LMR.
Theorem 7. Let 𝒜 : ℝm×n → ℝp be a linear transformation, and s be an integer s ∈ {0,1, 2, …, r}. Then 𝒜 is s-good if and only if γs(𝒜) < 1.
Proof. Suppose 𝒜 is s-good. Let W ∈ ℝm×n be a matrix of rank s ∈ {1,2, …, r}. Without loss of generality, let be its SVD where Um×s ∈ ℝm×s, Vn×s ∈ ℝn×s are orthogonal matrices and Ws = Diag((σ1(W), …, σs(W)) T). By the definition of s-goodness of 𝒜, W is the unique solution to the optimization problem (6). Using the first-order optimality conditions, we obtain that there exists y ∈ ℝp such that the function fy(x) = ∥X∥* − yT[𝒜X − 𝒜W] attains its minimum value over X ∈ ℝm×n at X = W. So 0 ∈ ∂fy(W) or 𝒜*y ∈ ∂∥W∥*. Using the fact (see, e.g., [27])
- (i)
𝒜D = 0,
- (ii)
,
- (iii)
.
Clearly, for every t ∈ ℝ, the matrices Xt : = W + tD are feasible in (6). Note that
We next prove that 𝒜 is s-good if γs(𝒜) < 1. That is, we let W be an s-rank matrix and we show that W is the unique optimal solution to (6). Without loss of generality, let W be a matrix of rank s′ ≠ 0 and its SVD, where , are orthogonal matrices and . It follows from Proposition 4 that . By the definition of γs(𝒜), there exists y ∈ ℝp such that 𝒜*y = U Diag(σ(𝒜*y))VT, where , , and
For the G-number , we directly obtain the following equivalent theorem of s-goodness from Proposition 5 and Theorem 7.
Theorem 8. Let 𝒜 : ℝm×n → ℝp be a linear transformation, and s ∈ {1,2, …, r}. Then 𝒜 is s-good if and only if .
4. s-Goodness, NSP, and RIP
Proposition 9. For the linear transformation 𝒜, if and only if 𝒜 satisfies NSP.
Proof. We first give an equivalent representation of the G-number . We define a compact convex set first:
For X ∈ ℝm×n with 𝒜X = 0, let X = U Diag(σ(X))VT be its SVD. Then, we obtain the sum of the s largest singular values of X as
Proposition 10. Let 𝒜 : ℝm×n → ℝp be a linear transformation and s ∈ {0,1, 2, …, r}. For any nonsingular transformation T ∈ ℝp×p, .
Proof. It follows from the nonsingularity of T that {X : 𝒜X = 0} = {X : T𝒜X = 0}. Then, by the equivalent representation of the G-number in (59),
For the RIP constant δ2s, Oymak et al. [21] gave the current best bound on the restricted isometry constant δ2s < 0.472, where they proposed a general technique for translating results from SSR to LMR. Together with the above arguments, we immediately obtain the following theorem.
Theorem 11. δ2s < 0.472⇒𝒜 satisfying is s-good.
The above theorem says that s-goodness is a necessary and sufficient condition for recovering the low-rank solution exactly via nuclear norm minimization.
5. Conclusion
In this paper, we have shown that s-goodness of the linear transformation in LMR is a necessary and sufficient conditions for exact s-rank matrix recovery via the nuclear norm minimization, which is equivalent to the null space property. Our analysis is based on the two characteristic s-goodness constants, γs and , and the variational property of matrix norm in convex optimization. This shows that s-goodness is an elegant concept for low-rank matrix recovery, although γs and may not be easy to compute. Development of efficiently computable bounds on these quantities is left to future work. Even though we develop and use techniques based on optimization, convex analysis, and geometry, we do not provide explicit analogues to the results of Donoho [35] where necessary and sufficient conditions for vector recovery special case were derived based on the geometric notions of face preservation and neighborliness. The corresponding generalization to low-rank recovery is not known, currently the closest one being [22]. Moreover, it is also important to consider the semidefinite relaxation (SDR) for the rank minimization with the positive semidefinite constraint since the SDR convexifies nonconvex or discrete optimization problems by removing the rank-one constraint. Another future research topic is to extend the main results and the techniques in this paper to the SDR.
Acknowledgments
The authors thank two anonymous referees for their very useful comments. The work was supported in part by the National Basic Research Program of China (2010CB732501), the National Natural Science Foundation of China (11171018), a Discovery Grant from NSERC, a research grant from the University of Waterloo, ONR research Grant N00014-12-10049, and the Fundamental Research Funds for the Central Universities (2011JBM306, 2013JBM095).