ORIGINAL RESEARCH article

Front. Appl. Math. Stat., 28 May 2019

Sec. Mathematics of Computation and Data Science

Volume 5 - 2019 | https://doi.org/10.3389/fams.2019.00023

Simultaneous Structures in Convex Signal Recovery—Revisiting the Convex Combination of Norms

  • 1. Institute for Theoretical Physics, Heinrich Heine University Düsseldorf, Düsseldorf, Germany

  • 2. Institute of Theoretical Physics and Astrophysics, University of Gdańsk, Gdańsk, Poland

  • 3. Institut de Mathématiques de Jussieu-PRG, Sorbonne Université, Paris, France

  • 4. Department of Mathematics, Applied Mathematics and Statistics, Case Western Reserve University, Cleveland, OH, United States

  • 5. Communications and Information Theory Group, Technical University of Berlin, Berlin, Germany

Abstract

In compressed sensing one uses known structures of otherwise unknown signals to recover them from as few linear observations as possible. The structure comes in form of some compressibility including different notions of sparsity and low rankness. In many cases convex relaxations allow to efficiently solve the inverse problems using standard convex solvers at almost-optimal sampling rates. A standard practice to account for multiple simultaneous structures in convex optimization is to add further regularizers or constraints. From the compressed sensing perspective there is then the hope to also improve the sampling rate. Unfortunately, when taking simple combinations of regularizers, this seems not to be automatically the case as it has been shown for several examples in recent works. Here, we give an overview over ideas of combining multiple structures in convex programs by taking weighted sums and weighted maximums. We discuss explicitly cases where optimal weights are used reflecting an optimal tuning of the reconstruction. In particular, we extend known lower bounds on the number of required measurements to the optimally weighted maximum by using geometric arguments. As examples, we discuss simultaneously low rank and sparse matrices and notions of matrix norms (in the “square deal” sense) for regularizing for tensor products. We state an SDP formulation for numerically estimating the statistical dimensions and find a tensor case where the lower bound is roughly met up to a factor of two.

1. Introduction

The recovery of an unknown signal from a limited number of observations can be more efficient by exploiting compressibility and a priori known structure of the signal. This compressed sensing methodology has been applied to many fields in physics, applied math, and engineering. In the most common setting, the structure is given as sparsity in a known basis. To mention some more recent directions, block-, group-, and hierarchical sparsity, low-rankness in matrix or tensor recovery problems and also the generic concepts of atomic decomposition are important structures present in many estimation problems.

In most of these cases, convex relaxations render the inverse problems itself amenable to standard solvers at almost-optimal sampling rates and also show tractability from a theoretical perspective [1]. In one variant, one minimizes a regularizing function under the constraints given by the measurements. A good regularizing function, or just regularizer, gives a strong preference in the optimization toward the targeted structure. For instance, the ℓ1-norm can be used to regularize for sparsity and the nuclear norm for low rankness of matrices.

1.1. Problem Statement

Let us describe the compressed sensing setting in more detail. We consider a linear measurement map on a d-dimensional vector space V ≅ ℝd define by its output components

Throughout this work we assume the measurements to be Gaussian, i.e., iid., where {ej}j∈[d] is an orthonormal basis of V. Moreover, we consider a given signalx0V, which yields the (noiseless) measurement vector

We wish to reconstruct x0 from and y in a way that is computationally tractable. Following a standard compressed sensing approach, we consider a norm reg as regularizer. Specifically, we consider an outcome of the following convex optimization problem

as a candidate for a reconstruction of x0, where we will choose reg to be either the weighted sum or the weighted maximum of other norms (18). If computations related to reg are also computationally tractable, Equation (3) can be solved efficiently. We wish that xreg coincides with x0. Indeed, when the number of measurements m is large enough (compared to the dimension d) then, with high probability over the realization of , the signal is reconstructed, i.e., x0 = xreg. For instance, when md then is invertible with probability 1 and, hence, the constraint in (3) alone guarantees such a successful reconstruction.

However, when the signal is compressible, i.e., it can be described by fewer parameters than the dimension d, then one can hope for a reconstruction from fewer measurements by choosing a good regularizerreg. For the case of Gaussian measurements, a quantity called the statistical dimension gives the number of measurements sufficient [2] and necessary [3] (see also [4, Corollary 4]) for a successful reconstruction. Therefore, much of this work is focused on such statistical dimensions.

Now we consider the case where the signal has several structures simultaneously. Two important examples are (i) simultaneously low rank and sparse matrices and (ii) tensors with several low rank matrizations. In such cases one often knows good regularizers for the individual structures. In this work, we address the question of how well one can use convex combinations of the individual regularizers to regularize for the combined structure.

1.2. Related Work

It is a standard practice to account for multiple simultaneous structures in convex optimization by combining different regularizers or constraints. The hope is to improve the sampling rate, i.e., to recover x0 from fewer observations y. Unfortunately, when taking simple combinations of regularizers there are certain limitations on the improvement of sampling rate.

For the prominent example of entrywise sparse and low-rank matrices, Oymak et al. [5] showed that convex combinations of ℓ1– and nuclear norm cannot improve the scaling of the sampling rate m in convex recovery. Mu et al. [4] considered linear combinations of arbitrary norms and derived more explicit and simpler lower bounds on the sampling rate using elegant geometric arguments.

In particular, this analysis also covers certain approaches to tensor recovery. It should be noted that low-rank tensor reconstruction using few measurements is notoriously difficult. Initially, it was suggested to use linear combinations of nuclear norms as a regularizer [6], a setting where the lower bounds [4] apply. Therefore, the best guarantee for tractable reconstructions is still a non-optimal reduction to the matrix case [4].

If one gives up on having a convex reconstruction algorithm, non-convex quasi-norms can be minimized that lead to an almost optimal sampling rate [7]. This reconstruction is for tensors of small canonical rank (a.k.a. CP rank). Also for this rank another natural regularizer might be the so-called tensor nuclear norm. This is another semi-norm for which one can find tractable semidefinite programming relaxations based on so-called θ-bodies [8]. These norms provide promising candidates for (complexity theoretically) efficient and guaranteed reconstructions.

Also following this idea of atomic norm decompositions [2], a single regularizer was found by Richard et al. [9] that yields again optimal sampling rates at the price that the reconstruction is not give by a tractable convex program.

1.3. Contributions and Outline

We discuss further the idea of taking convex combinations of regularizers from a convex analysis viewpoint. Moreover, we focus on the scenario where the weights in a maximum and also a sum of norms can depend on the unknown object x0 itself, which reflects an optimal tuning of the convex program.

Based on tools established in Mu et al. [4], which may be not fully recognized in the community, we discuss simple convex combinations of regularizers. As already pointed out by Oymak et al. [5] an optimally weighted maximum of regularizers has the best sampling rate among such approaches. We follow this direction and discuss how sampling rates can be obtained in such setting. Specifically, we point out that the arguments leading to the lower bounds of Mu et al. [4] can also be used to obtain generic lower bounds for the maximum of norms, which implies similar bounds for the sum of norms (section 2). We also present an SDP characterization of dual norms for weighted sums and maxima and use them for numerically sampling the statistical dimension of descent cones, which correspond to the critical sampling rate [3].

In section 3 we discuss the prominent case of simultaneously sparse and low rank matrices. Here, our contributions are two-fold. We first show that the measurements satisfy a restricted isometry property (RIP) at a sampling rate that is essentially optimal for low rank matrices, which implies injectivity of the measurement map . Then, second, we provide numerical results showing that maxima of regularizers lead to recovery results that show an improvement over those obtained by the optimally weighted sum of norms above an intermediate sparsity level.

Then, in section 4 we discuss the regularization for rank-1 tensors using combinations of matrix norms (extending the “square deal” [4] idea). In particular, we suggest to consider the maximum of several nuclear norms of “balanced” or “square deal” matrizations for the reconstruction of tensor products. We point out that the maximum of regularizers can lead to an improved recovery, often better than expected.

It is the hope that this work will contribute to more a comprehensive understanding of convex approaches to simultaneous structures in important recovery problems.

2. Lower Bounds for Convex Recovery With Combined Regularizers

In this section we review the convex arguments used by Mu et al. [4] to establish lower bounds for the required number of measurements when using a linear combination of regularizers.

2.1. Setting and Preliminaries

For a positive integer m we use the notation [m] ≔ {1, 2, …, m}. The ℓp-norm of a vector x is denoted by ∥xp and Schatten p-norm of a matrix X is denoted by ∥Xp. For p = 2 these two norms coincide and are also called Frobenius norm or Hilbert-Schmidt norm. With slight abuse of notation, we generally denote the inner product norm of a Hilbert space by 2. For a cone S and a vector g in a vector space with norm we denote their induced distance by

The polar of a cone K is

For a set S we denote its convex hull by conv(S) and the cone generated by S by cone(S) ≔ {τx : xS, τ > 0}. For convex sets C1 and C2 one has

where “⊂” follows trivially from the definitions. In order to see “⊃” we write a conic combination z = ρx + σy ∈ cone(C1) + cone(C2) as . The Minkowski sum of two sets C1 and C2 is denoted by C1 + C2 ≔ {x + y : xC1, yC2}. It holds that

The subdifferential of a convex function f at base point x is denoted by ∂f(x). When f is a norm, f = , then the subdifferential is the set of vectors where Hölder's inequality is tight,

where ∥ y ∥° is the dual norm of defined by . The descent cone of a convex function f at point x is [10, Definition 2.4]

and it holds that [10, Fact 3.3]

where cl S denotes the closure of a set S. Let {fi}i∈[k] be proper convex functions such that the relative interiors of their domains have at least a point in common. Then

A function that is the point-wise maximum of convex functions {fi}i∈[m] has the subdifferential [11, Corollary D.4.3.2]

Hence, the generated cone is the Minkowski sum

The Lipschitz constant of a function f w.r.t. to a norm is the smallest constant L such that for all vectors x and x

Usually, is an ℓ2-norm, which also fits the Euclidean geometry of the circular cones defined below.

The statistical dimension of a convex cone K is given by (see e.g., [3, Proposition 3.1(4)])

where g ~ N(0, 𝟙) is a Gaussian vector. The statistical dimension satisfies [3, Proposition 3.1(8)]

for any cone KV in a vector space V ≅ ℝd. Now, let us consider a compressed sensing problem where we wish to recover a signal x0 from m fully Gaussian measurements using a convex regularizer f. For small m, a successful recovery is unlikely and for large m the recovery works with overwhelming probability. Between these two regions of m one typically observes a phase transition. This transition is centered at a value given by the statistical dimension of the descent cone δ((f, x0)) of f at x0 [3]. Thanks to Equation (10), this dimension can be calculated via the subdifferential of f,

By ≳ and ≲ we denote the usual greater or equal and less or equal relations up to a positive constant prefactor and ∝ if both holds simultaneously. Then we can summarize that the convex reconstruction (3) requires exactly a number of measurements m ≳ δ((f, x0)), which can be calculated via the last equation.

2.2. Combined Regularizers

We consider the convex combination and weighted maxima of norms (i), where i = 1, 2, …, k and set

where λ = (λ1, …, λk) ≥ 0 and μ = (μ1, …, μk) ≥ 0 are to be chosen later. Here, we assume that we have access to single norms of our original signal ∥x0(i) so that we can fine-tune the parameters λ and μ accordingly.

In Oymak et al. [5] lower bounds on the necessary number of measurements for the reconstruction (3) using general convex relaxations were derived. However, it has not been worked out explicitly what can be said in the case where one is allowed to choose the weights λ and μ dependending on the signal x0. For the sum norm λ,sum this case is covered by the lower bounds in Mu et al. [4]. Here, we will see that the arguments extend to optimally weighted max norms, i.e., μ,max with μ being optimally chosen for a given signal.

A description of the norms dual to those given by (18) is provided in section 2.4.1.

2.3. Lower Bounds on the Statistical Dimension

The statistical dimension of the descent cone is obtained as Gaussian squared distance from the cone generated by the subdifferential (17). Hence, it can be lower bounded by showing (i) that the subdifferential is contained in a small cone and (ii) bounding that cone [4]. A suitable choice for this small cone is the so-called circular cone.

2.3.1. Subdifferentials and Circular Cones

We use the following statements from Amelunxen et al. [3] and Mu et al. [4, section 3] which show that subdifferentials are contained in circular cones. More precisely, we say that the angle between vectors x, z ∈ ℝd is θ if 〈z, x〉 = cos(θ) ∥z2x2 and write in that case ∠(x, z) = θ. Then the circular cone with axis x and angle θ is defined as

Its statistical dimension has a simple upper bound in terms of its dimension: For all xV ≅ ℝd and some θ ∈ [0, π/2] [4, Lemma 5]

By the following argument this bound can be turned into a lower bound on the statistical dimension of descent cones and, hence, to the necessary number of measurements for signal reconstructions. The following two propositions summarize arguments made by Mu et al. [4].

Proposition 1 (Lower bound on descent cone statistical dimensions [4]). Let us consider a convex function f as a regularizer for the recovery of a signalin a d-dimensional space. If ∂f(x0) ⊂ circ(x0, θ) then

This statement is already implicitly contained in Mu et al. [4]. The arguments can be compactly summarized as follows.

Proof: With the polar of the descent cone (10), the assumption, and the statistical dimension of the polar cone (16) we obtain

Hence, with the bound on the statistical dimension of the circular cone (20),

     □

Recall from (14) a norm f is L-Lipschitz with respect to the ℓ2–norm on a (sub-)space V if

for all xV. Note that this implies for the dual norm f° that

for all xV.

Proposition 2 ([4, section 3]). Let f be a norm onV ≅ ℝdthat is L-Lipschitz with respect to the2-norm on V. Then

with .

For sake of self-containedness we summarize the proof of Mu et al. [4, section 3].

Proof: The subdifferential of a norm f on V ≅ ℝd with dual norm f° is given by

Then, thanks to Hölder's inequality , we have for any subgradient x ∈ ∂f(x0)

This bound directly implies the claim.      □

So together, Propositions 1 and 2 imply

where

is the f-rank of x0 (e.g., effective sparsity for f = 1 and effective matrix rank for f = 1).

2.3.2. Weighted Regularizers

A larger subdifferential leads to a smaller statistical dimension of the descent cone and, hence, to a reconstruction with a potentially smaller number of measurements in the optimization problems

and

with the norms (18). Having the statistical dimension (17) in mind, we set

which give the optimal number of measurements in a precise sense [3, Theorem II]. Note that it is clear that δsum(λ) is continuous in λ.

Now we fix x0 and consider adjusting the weights λi and μi depending on x0. We will see the geometric arguments from Mu et al. [4] extend to that case. Oymak et al. [5, Lemma 1] show that the max-norm μ,max with weights μi chosen as

has a better recovery performance than all other convex combinations of norms. With this choice the terms in the maximum (18) defining μ,max are all equal. Hence, from the subdifferential of a maximum of functions (12) it follows that this choice of weights is indeed optimal. Moreover, the optimally weighted max-norm leads to a better (smaller) statistical dimension for x0 than all sum-norms λ,sum and, hence, indeed to a better reconstruction performance:

Proposition 3 (Optimally weighted max-norm is better than any sum-norm). Consider a signal x0and the corresponding optimal weights μ* from (36). Then, for all possible weights λ ≽ 0 in the sum-norm, we have

In particular, for all λ ≽ 0.

Proof: Using (12) and (6) we obtain

with (11)

Then (7) concludes the proof.     □

Proposition 3 implies that if the max-norm minimization (32) with optimal weight (36) does not recover x0, then the sum-norm minimization (33) can also not recover x0 for any weight λ.

Now we will see that lower bounds on the number of measurements from Mu et al. [4, section 3] straightforwardly extend to the max-norm with optimal weights. These lower bound are obtained by deriving an inclusion into a circular cone. Combining Proposition 2 with (39) and noting that a Minkowski sum of circular cones (19) is just the largest circular cone yields the following:

Proposition 4 (Subdifferential contained in a circular cone). Let(signal) and(i)be norms satisfyingx(i)Lix2 for i ∈ [k] and for allxV. Then the subdifferential of

satisfies

In conjunction with Proposition 1 this yields the bound

Hence, upper bounds on the Lipschitz constants of the single norms yield a circular cone containing the subdifferential of the maximum, where the smaller the largest upper bound the smaller the circular cone. In terms of f-ranks it means that

Now, Mu et al. [4, Lemma 4] can be replaced by this slightly more general proposition and the main lower bound [4, Theorem 5] on the number of measurements m still holds. The factor 16 in [4, Corollary 4] can be replaced by an 8 (see updated Arxiv version [12] of [3]). These arguments [specifically, [12, (7.1) and (6.1)] with λ ≔ δ(C) − m] show the following for the statistical dimension δ of the descent cone. The probability of successful recovery for m ≤ δ is upper bounded by p for all . Conversely, the probability of unsuccessful recovery for m ≥ δ is upper bounded by q for any , so in particular, for any . Moreover, this yields the following implication of Amelunxen et al. [3]:

Theorem 5 (Lower bound). Let(i)be norms satisfyingx(i)Lix2for i ∈ [k] and for all xV. Fix(signal) and set

Consider the optimally weighted max-norm (41). Then, for allm ≤ κ, the probabilitythatx0is the unique minimizer of the reconstruction program (32) is bounded by

We will specify the results in more detail for the sparse and low-rank case in Corollary 8 and an example for the tensor case in Equation (100).

2.4. Estimating the Statistical Dimension Via Sampling and Semidefinite Programming

Often one can characterize the subdifferential of the regularizer by a semidefinite program (SDP). In this case, one can estimate the statistical dimension via sampling and solving such SDPs.

In more detail, in order to estimate the statistical dimension (17) for a norm f, we sample the Gaussian vector , calculate the distance using the SDP-formulation of ∂f(x0) and take the sample average in the end. In order to do so, we wish to also have an SDP characterization of the dual norm ∂f° of f, which provides a characterization of the subdifferential (8) of f.

In the case when the regularizer function f is either μ,max or λ,sum, where the single norms (i) have simple dual norms, we can indeed obtain such an SDP characterization of the dual norm f°.

2.4.1. Dual Norms

It will be convenient to have explicit formulae for the dual norms to the combined regularizers defined in section 2.2 [13].

Lemma 6 (Dual of the maximum/sum of norms). Let(i)be norms (i ∈ [k]) and denote byμ,max and λ,sumtheir weighted maximum and weighted sum as defined in (18). Then their dual norms are given by

For the proof of the lemma we use the notation from the convex analysis book by Rockafellar [14]. For an alternative more elementary proof see [15].

Let C ⊂ ℝd be a convex set. Then its support function is defined by

and its indicator function by

If C contains the origin, then we define its gauge function by

and the polar of C to be

[Note that while formally different, this definition is consistent with the polar of a cone introduced in (5).] The polar of a gauge function γ is denoted by γ°(y) := inf {λ ≥ 0 : 〈y, x〉 ≤ λγ(x)} and the convex conjugate of a function (Fenchel-Legendre dual) f by f*. The infimal convolution of convex functions f and g is defined by .

There are all sorts of rules [14] that we will use. Let f and g be lower semi-continuous convex functions and C1, C2 and B be convex sets where B contains the origin. Then f** = f, , B°° = c1 B, , (fg)* = f*+g*, , and .

Proof of Lemma 6: For the sake of better readability we focus on the case k = 2. The general case follows analogously.

In order to establish the statement of the lemma we will use the fact that the norms are the gauge functions of their unit balls, i.e., (i) = γBi with Bi := {x: ∥x(i) ≤ 1}. For the proofs, however, let Bi be closed convex sets containing the origin (they do not need to satisfy Bi = −Bi, as is the case for norms).

We begin with the polar of the sum (48). Note that, in general, the “polar body (K + L)° of a [Minkowski] sum of convex bodies has no plausible interpretation in terms of K°, L°” [16]. However, we can still find a useful expression for the dual norm:

so that

which implies the statement (48).

In order to establish the polar of the maximum (47) we start with the identity

to obtain

so that

which proves the identity (47).     □

2.4.2. Gaussian Distance as SDP

We were aiming to estimate the statistical dimension (17) by sampling over Gaussian vectors g and the corresponding SDP outcomes. For any vector g the distance to the cone generated by the subdifferential of a norm f is

For f = λ,sum we use its polar (48) to obtain

where one needs to note that an optimal feasible point of (60) also yields an optimal feasible point of (61) and vice versa. But this implies that

For the maximum of norms regularizer we again choose the optimal weights (36) to ensure that all norms are active, i.e.,

Then we use the subdifferential expression (13) for a point-wise maximum of functions to obtain

In the case that are ℓ-norms or spectral norms these programs can be written as SDPs and be solved by standard SDP solvers.

3. Simultaneously Sparse and Low-Rank Matrices

A class of structured signals that is important in many applications are matrices which are simultaneously sparse and of low rank. Such matrices occur in sparse phase retrieval1 [5, 17, 18], dictionary learning and sparse encoding [19], sparse matrix approximation [20], sparse PCA [21], bilinear compressed sensing problems like sparse blind deconvolution [2227] or, more general, sparse self-calibration [28]. For example, upcoming challenges in communication engineering and signal processing require efficient algorithms for such problems with theoretical guarantees [2931]. It is also well-known that recovery problems related to simultaneous structures like sparsity and low-rankness are at optimal rate often as hard as the classical planted/hidden clique problems, see for example [32] for further details and references.

3.1. Setting

We consider the d = n1 · n2–dimensional vector space of n1 × n2-matrices with components in the field 𝕂 being either ℝ or ℂ. The space V is equipped with the Hilbert-Schmidt inner product defined by

Our core problem is to recover structured matrices from m linear measurements of the form with a linear map . Hence, a single measurement is

It is clear that without further a-priori assumptions on the unknown matrix X we need md = n1 · n2 measurements to be able to successfully reconstruct X.

As additional structure, we consider subsets of V containing simultaneously low-rank and sparse matrices. However, there are different ways of combining low-rankness and sparsity. For example one could take matrices of rank r with different column and row sparsity, i.e., meaning that there are only a small number of non-zero rows and columns. Here, we consider a set having even more structure which is motivated by atomic decomposition [2]. Recall, that the rank of a matrix X can be defined as its “shortest description” as

and the vectors {xi} and {yi} can be required to be orthogonal, respectively. This characterization giving rise to the nuclear norm 1 as the corresponding atomic norm, see Chandrasekaran et al. [2] for a nice introduction to inverse problems from this viewpoint. Restricting the sets of feasible vectors {xi} and {yi} yields alternative formulations of rank. In the case of sparsity, one can formally ask for a description in terms of (s1, s2)-sparse atoms:

where ∥x0 denotes the number of non-zero elements of a vector x. The idea of the corresponding atomic norm [2] has been worked for the sparse setting by Richard et al. [9]. Note that, compared to (67), we do not require that and are orthogonal.

We say that a matrix is simultaneously (s1, s2)-sparse and of rankr if it is in the set

Our model differs to the joint-sparse setting as used in Lee et al. [22], since the vectors and may have individual sparse supports and need not to be orthogonal. Definition (69) fits more into the sparse model considered also in Fornasier et al. [33].

By ei we denote i-th canonical basis vector and define row-support supp1(X) and column-support supp2(X) of a matrix X as

Obviously, the matrices are then at most k1-column-sparse and k2-row-sparse (sometimes also called as joint-sparse), i.e.,

but not strictly –sparse2. Since we have sums of r different (s1, s2)-sparse matrices and there are at most r(s1 · s2) non-zero components. Note that a joint (s1, s2)-row and column sparse matrix has only s1 · s2 non-zero entries. Hence, considering this only from the perspective of sparse vectors, we expect that up to logarithmic terms recovery can be achieved from mr(s1 · s2) measurements. On the other hand, solely from a viewpoint of low-rankness, mr (n1 + n2) measurements also determine an n1 × n2-matrix of rank r. Combining both gives therefore mr min(s1s2, n1 + n2).

On the other hand, these matrices are determined by at most r(s1 + s2) non-zero numbers. Optimistically, we therefore hope that already mr(s1 + s2) sufficiently diverse observations are enough to infer on X which is substantially smaller and scales additive in s1 and s2. In the next part we will discuss that this intuitive parameter counting argument is indeed true in the low-rank regime . A generic low-dimensional embedding of this simultaneously sparse and low-rank structure into 𝕂m for mr(s1 + s2) via a Gaussian map is stably injective.

3.2. About RIP for Sparse and Low-Rank Matrices

Intuitively, (s1, s2)–sparse rank-r matrices can be stably identified from y if is almost-isometrically mapped into 𝕂m, i.e., distances between different matrices are preserved up to small error during the measurements process. Note that we have the inclusion

Since is linear it is therefore sufficient to ensure that norms are preserved for . We say that a map satisfies (r, s1, s2)-RIP with constant δ if

holds, where the supremum is taken over all and denotes the adjoint of (defined in the canonical way with respect to the Hilbert-Schmidt inner product). By we always denote the smallest value δ such that above condition holds. Equivalently, we have

A generic result, based on the ideas of Baraniuk et al. [34], Candes and Plan [35], and Recht et al. [36], has been presented already in Jung and Walk [29]. It shows that a random linear map which concentrates uniformly yields the RIP property (74) with overwhelming probability (exponential small outage) once the number of measurements are in the order of the metric entropy measuring the complexity of the structured set . In the case of simultaneous low rank and sparse matrices this quantity scales (up to logarithmic terms) additively in the sparsity, as desired. A version for rank-r matrices where and in (69) are orthonormal sets having joint-sparse supports has been sketched already in Lee et al. [22, Theorem III.7], i.e., with iid Gaussian entries acts almost isometrically in this case for with probability at least , c2 being an absolute constant. Another RIP perspective has been considered in Fornasier et al. [33] where the supremum in (73) is taken over effectively sparse ( and in (69) are now only well–approximated by sparse vectors) rank-r matrices X with (implying ). More precisely, for with Δ ∈ (0, 1) an operator with iid. centered sub-Gaussian entries acts almost-isometrically with probability at least 1 − 2 exp(−C′ Δm) at δ = ΔΓ2r and C′ is an absolute constant. Note that this probability is slightly weaker.

We provide now a condensed version of the generic statement in Jung and Walk [29]. More precisely, RIP (74) is satisfied with high probability for a random linear map that has uniformly sub-Gaussian marginals:

Theorem 7 (RIP for sub-Gaussian measurements). Letbe a random linear map which for given c > 0 and 0 < δ < 1 fulfillsuniformly for all. If

thensatisfies (r, s1, s2)-RIP with constantwith probability. Here, is an absolute constant andcis a constant depending only on δ and.

Clearly, standard Gaussian measurement maps fulfill the concentration assumption in the theorem. More general sub-Gaussian maps are included as well, see here also the discussion in Jung and Walk [29]. The proof steps are essentially well-known. For the sake of self-containedness we review the steps having our application in mind.

Proof: First, we construct a special ϵ-net for . By this we mean a set such that for each we have some such that ‖XR2 ≤ ϵ. Since the matrices are k1 := rs1 row-sparse and k2rs2 column-sparse there are

different combinations for the row support T1 ⊂ [n1] and column support T2 ⊂ [n2] with |T1| = k1 and |T2| = k2. For each of these canonical matrix subspaces supported on T1 × T2, we consider matrices of rank at most r. From Candes and Plan [35, Lemma 3] it is known that there exists an ϵ-net for k1 × k2 matrices of rank r of cardinality giving therefore

In other words, up to logarithmic factors, this quantity also reflects the intuitive parameter counting. The net construction also ensures that for each and close by net point R = R(X) we have | supp1(XR)| ≤ k1 and | supp2(XR)| ≤ k2. However, note that in non-trivial cases rank(RX) = 2r meaning that . Still, using a singular value decomposition one can find RX = X1 + X2 with 〈X1, X2〉 = 0 for some and . To show RIP, we define the constant

For some and close by net point and let us consider δ such that . Then,

Now we choose satisfying [ in (78) is compact]. For such an we also have

Requiring that the right hand side is bounded by δ and solving this inequality for A (assuming ϵ < 1) we find that indeed whenever ϵ ≤ δ/(2 + 2δ). In particular we can choose δ < 1 and we set ϵ = δ/4. Therefore, Equation (77) yields

Using the assumption −ecδ2m and the union bound we obtain RIP with probability at least

Thus, if we want to have RIP satisfied with probability for a given , i.e.,

it is sufficient to impose that mc″ δ−2log |R| for a some c″ > 0.     □

In essence the theorem shows that the intrinsic geometry of sparse and low-rank matrices is preserved in low-dimensional embeddings when choosing the dimension above a threshold. It states that, in the low-rank regime , for fixed δ this threshold for the RIP to hold scales indeed as r(s1 + s2). This additional low-rank restriction is an technical artifact due to suboptimal combining of covering number estimates. Indeed, upon revising the manuscript we found that the statement above may be improved by utilizing [33, Lemma 4.2] instead of (77) which yields a scaling of r(s1 + s2) without restrictions on r and with probability of at least . From the proof it follows also easily that for joint-sparse matrices where each of the sets and in (69) have also joint support as in Lee et al. [22], a sampling rate mr(s1 + s2) is sufficient anyway for all ranks r.

Intuitively, one should therefore be able reconstruct an unknown s1 × s2–sparse matrix of rank r from mr(s1 + s2) generic random measurements. This would indeed reflect the intuitive parameter counting argument. Unfortunately, as will be discussed next, so far, no algorithm is known that can achieve such a reconstruction for generic matrices.

3.3. Some More Details on Related Work

It is well-known that sufficiently small RIP constants imply successful convex recovery for sparse vectors [37] and low-rank matrices [36, 38], separately. An intuitive starting point for convex recovery of the elements from would therefore be the program:

which uses a weighted sum as a regularizer, where are noiseless measurements of the signal X0. Related approaches have been used also for applications including sparse phase retrieval and sparse blind deconvolution. Obviously, then the corresponding measurement map is different and depends on the particular application. The practical relevance of this convex formulation is that it always allows to use generic solvers and there is a rich theory available to analyze the performance for certain types measurement maps in terms of norms of the recovery error XX0. Intuitively, one might think that this amounts only to characterize the probability when the matrix A is robustly injective on feasible differences XX0, i.e., fulfills RIP or similar conditions. However, this is not enough as observed and worked out in Oymak et al. [5] and Jalali [39]. One of the famous no-go results in these works is that no extra reduction in the scaling of the sampling rate can be expected as compared to the best of recovering with respect to either the low-rank structure (μ1 = 0) or sparsity (μ1 = 0), separately. In other words, for any pair (μ1, μ1) the required sampling rate can not be better than the minimum one with single reguralizer norm (either μ1 = 0 or μ1 = 0). A difficult point in this discussion is what will happen if the program is optimally tuned, i.e., if μ1 = 1/‖X01 and μ1 = 1/‖X01. We have based our generic investigations given in section 2.3 on the considerably more simplified technique of Mu et al. [4] which also allows to obtain such results in more generality. An alternative convex approach is discussed [9] where the corresponding atomic norm [2] (called kq-norm) is used as a single regularizer. This leads to convex recovery at optimal sampling rate but the norm itself cannot be computed in a tractable manner, reflecting again the hardness of the problem itself. For certain restricted classes the hardness is not present and convex algorithms perform optimally, see for example [26] where signs in a particular basis are known a-priori.

Due to the limitations of tractable convex programs non-convex recovery approaches have been investigated intensively in the last years. In particular, the alternating and thresholding based algorithm “sparse power factorization”, as presented in Lee et al. [22, 23], can provably recover at optimal sampling rates when initialized optimally. However, this is again indeed the magic and difficult step since computing the optimal initialization is again computationally intractable. For a suboptimal but tractable initialization recovery can only be guaranteed for a considerably restricted set of very peaky signals. Relaxed conditions have been worked out recently [27] with the added benefit that the intrinsic balance between additivity and multiplicativity in sparsity is more explicitly established. Further alternating algorithms like [33] have been proposed with guaranteed local convergence and which have better empirical performance.

An interesting point has been discussed in Foucart et al. [40]. For simplicity, let n = n1 = n2 and s = s1 = s2. Assume that for given rank r and sparsity s the measurement map in (66) factorizes in the form where for i = 1,…, mrp and B ∈ ℝp×n are standard Gaussian matrices with ps log(en/s). In this case a possible reconstruction approach will factorize as well into two steps, (i) recovery of an intermediate matrix Y ∈ ℝp×n from the raw measurements y using nuclear norm minimization and (ii) recovery of the unknown matrix X0 from Y using the HiHTP algorithm (details see [41]). However, in the general case, hard-thresholding algorithms like HiHTP require computable and almost-exact (constants almost independent of s and n) head projections into (s, s)-sparse matrices. Positive-semidefiniteness is helpful in this respect [40] and, in particular, in the rank-one case this is relevant for sparse phase retrieval [31]. However, in the general case, to the best of the authors knowledge, no algorithm with tractable initialization has guaranteed global convergence for generic sparse and low-rank matrices so far.

3.4. The Lower Bound

In the following section we will further strengthen this “no-go” result for convex recovery. As already mentioned above, an issue which has not been discussed in sufficient depth is what can be said about optimally tuned convex programs and beyond convex combinations of multiple regularizers. For our simultaneously sparse and low rank matrices Theorem 5 yields the following.

Corollary 8 (Lower bound, sparse and low rank matrices). Letbe an (s1, s2)-sparse n1 × n2-matrix of rank at most r, andbe a Gaussian measurement operator. Then, for all, X0is the unique minimizer of

with probability at most

In words, even when optimally tuning convex algorithms and when using the intuitive best regularizer having the largest subdifferential, the required sampling rate still scales multiplicative in sparsity, i.e., it shows the same no-go behavior as the other (suboptimal) regularizer.

Proof: The Lipschitz constants of the ℓ1–norm and the nuclear norm (w.r.t. the Frobenius norm) are

respectively. Using that the matrix X0 is s := r(s1s2)–sparse yields . Hence,

is the expression in the minimum in (45) corresponding to the index “(i) = ℓ1” used for the ℓ1 norm. Using that X0 has rank at most r we obtain . Hence,

with is the expression in the minimum in (45) corresponding to the index “(i) = 1” used for the nuclear norm. Together,

and Theorem 5 establishes the corollary.     □

3.5. Numerical Experiments

We have numerically estimated the statistical dimension of the decent cones and have performed the actual reconstruction of simultaneously low rank and sparse matrices.

3.5.1. Gaussian Distance

In section 2.4.2 we showed that the Gaussian distance can be estimated numerically by sampling over (in this case) semidefinite programs (SDP) according to (62) and (64). When empirically averaging these results according to (17) one obtains an estimate of the statistical dimension and therefore the phase transition point for successful convex recovery. For the case of sparse and low-rank matrices with the weighted sum of nuclear norm and ℓ1–norm as regularizer the distance (62) becomes

A similar SDP can be obtained for the case of the maximum of these two regularizers. We solve both SDPs using the CVX toolbox in MATLAB (with SDPT3 as solver) for many realization of a Gaussian matrix G and then average those results. We show such results for the optimal weights in Figure 1 for where s = 4, 5, 15 and the size of the n × n matrices ranges in n ∈ [15, 40]. For s = 4, 5 the statistical dimension for the optimally weighted sum and the maximum are almost the same. However, for higher sparsity s = 15 there is a substantial difference, i.e., the optimally weighted sum of regularizers behaves worse than the maximum.

Figure 1

These results indeed suggest that the statistical dimension for optimally weighted maximum of regularizers is better than the sum of regularizers.

3.5.2. Convex Recovery

We numerically find the phase transition for the convex recovery of complex sparse and low-rank matrices using the sum and maximum of optimally weighted ℓ1 and nuclear norm. We also compare to the results obtained by convex recovery using only either the ℓ1–norm or the nuclear norm as regularizer and also provide and example for a non-convex algorithm.

The dimension of the matrices are n = n1 = n2 = 30, the sparsity range is s = s1 = s2 = 5, …, 20, and the rank is r = 1. For each parameter setup a matrix is drawn using a uniform distribution for supports of u and v of size s and iid. standard complex-normal distributed entries on those support. The measurement map itself also consists also of iid. complex-normal distributed entries. The reconstructed vector X is obtained using again the CVX toolbox in MATLAB with the SDPT3 solver and an reconstruction is marked to be successful exactly if

holds. Each (m, s)-bin in the phase transition plots contains 20 runs.

The results are shown in Figure 2. Figures 2A,B show the phase transition of only taking the ℓ1–norm and the nuclear norm as regularizer, respectively. The lower bound from Theorem 5 on the required number of measurements yield for those cases the sparsity s2 of X0 and n, respectively. Thus, only for very small values of s2 there is a clear advantage of ℓ1-regularization compared to the nuclear norm. The actual recovery rates scale roughly as 2s2 ln(n2 / s2) and 4rn, respectively.

Figure 2

However, combining both regularizers with optimal weights improves as shown in Figures 2C,D. Both combined approaches instantaneously balance between sparsity and rank. Moreover, there is a clear advantage of taking the maximum (Figure 2C) over of the sum (Figure 2D) of ℓ1– and nuclear norm. For sufficiently small sparsity the ℓ1–norm is more dominant and for higher sparsity than the nuclear norm determines the behavior of the phase transition. But only for the maximum of the regularizers the sampling rate saturates at approximately m = 130 due to rank(X0) = 1 (see Figure 2D).

We also mention that the maximum of regularizers improves only if it is optimally tuned, which is already indicated by the subdifferential of a maximum (12), where only the largest terms contribute. In contrast, reconstruction behavior of the sum of norms seems to more stable. This observations has also been mentioned in Jalali [39]. This feature motivates an empirical approach of guessing the weights from observations.

To sketch a greedy approach for guessing the weights we consider the following strategy. Ideally, we would like to choose λ1 = 1/∥X01 and λ1 = 1/∥X01 in the minimization of the objective function λ1X1 + λ1X1. Since, for Frobenius norm normalized X we have 1/∥X1 ≥ ∥X (similarly for the ℓ-norm) we choose for as initialization and for the iteration t = 1. After finding

we update and . The results obtained by this greedy approach after 3 iterations are shown in Figure 2E. Comparing this to the optimally weighted sum of regularizers in Figure 2C, we see that almost the same performance can be achieved with this iterative scheme.

Finally, there is indeed strong evidence that in many problems with simultaneous structures non-convex algorithms perform considerably better and faster then convex ones. Although we have focused in this work on better understand of convex recovery, let us also comment on other settings. For the sparse and low-rank setting there exists several very efficient and powerful algorithms, as examples we mention here sparse power factorization (SPF) [22] and ATLAS2, 1 [33]. In the noiseless setting SPF is known to clearly outperforms all convex algorithms (see Figure 2F). The numerical experiments in Fornasier et al. [33] suggests that in the noisy setting ATLAS2, 1 seems to be a better choice.

4. Special Low-Rank Tensors

Tensor recovery is an important and notoriously difficult problem, which can be seen as a generalization of low-rank matrix recovery. However, for tensors there are several notions of rank and corresponding tensor decompositions [42]. They include the higher order singular value decomposition (HOSVD), the tensor train (TT) decomposition (a.k.a. by matrix product states), the hierarchical Tucker (a.k.a. tree tensor network) decomposition, and the CP decomposition. For all these notions, the unit rank objects coincide and are given by tensor products.

Gandy et al. [6] suggested to use a sum of nuclear norms of different matrizations (see below) as a regularizer for the completion of 3-way tensors in image recovery problems. Mu et al. [4] showed that this approach leads to the same scaling in the number of required measurements as when one uses just one nuclear norm of one matrization as a regularizer. However, the prefactors are significantly different in these approaches. Moreover, Mu et al. suggested to analyze 4-way tensors, where the matrization can be chosen such that the matrices are close to being square matrices. In this case, the nuclear norm regularization yields an efficient reconstruction method with rigorous guarantees that has the so far best scaling in the number of measurements. For rank-1 tensors we will now suggest to use a maximum of nuclear norms of certain matrizations as a regularizer. While the no-go results [4] for an optimal scaling still hold, this leads to a significant improvement of prefactors.

4.1. Setting and Preliminaries

The effective rank of a matrix X is . Note that for matrices for which all non-zero singular values coincide, the rank coincides with the effective rank and for all other matrices the effective rank is smaller.

We consider the tensor spaces as a signal space and refer to the ni as local dimensions. The different matrix ranks of different matrizations are given as follows.

An index bipartition is

We call the b-matricization the canonical isomorphism , where nb = ∏ibni, i.e., the indices in b are joined together into the row index of a matrix and the indices in bc into the column index. It is performed by a reshape function in many numerics packages. The rank and effective of the b-matrization of X are denoted by rankb(X) and . The b-nuclear norm is given by the nuclear norm of the b-matricization of X.

Now, we consider ranks based on a set of index bipartitions

The corresponding (formal) rank is given by

Similarly, given a signal X0V the corresponding max-norm is given by

Note that for the case that X0 is a product we have

for all b ⊂ [L] and p ≥ 1. Hence, the reweighting in the optimal max-norm (97) is trivial in that case, i.e., it just yields an overall factor, which can be pulled out of the maximum.

Let us give more explicit examples for the set of bipartitions: defines the HOSVD rank and the tensor train rank. They also come along with a corresponding tensor decomposition. In other cases, accompanying tensor decompositions are not known. For instance, for k = 4 and it is clear that the tensors of (formal) rank (1, 1) are tensor products. The tensors of ranks (1, i) and (i, 1) are given by tensor products of two matrices, each of rank bounded by i. But in general, it is unclear what tensor decomposition corresponds to a -rank.

One interesting remark might be that there are several measures of entanglement in quantum physics, which measure the non-productness in case of quantum state vectors. The negativity [43] is such a measure. Now, for a non-trivial bipartition b and normalized tensor XV (i.e., ∥X2 = 1)

is the negativity [43] of the quantum state vector X w.r.t. the bipartition b, where Tb denotes the partial transposition w.r.t. b and vec(X) the vectorization of X, i.e., the [L]-matrization.

Theorem 5 applies to tensor recovery with the regularizer (97). We illustrate the lower bound for the special case of 4-way tensors (L = 4) with equal local dimensions ni = n and a regularizer norm given by . Then the critical number of measurements (45) in the lower bounds (46) and (43) is

If X0 is a tensor product, this becomes κ ≈ n2.

4.2. RIP for the HOSVD and TT Ranks

A similar statement as Theorem 7 has been proved for the HOSVD and TT rank for the case of sub-Gaussian measurements by Rauhut et al. [44, section 4]. They also showed that the RIP statements lead to a partial recovery guarantee for iterative hard thresholding algorithms. Having only suboptimal bounds for TT and HOSVD approximations has so far prevented proofs of full recovery guarantees.

It is unclear how RIP results could be extended to the “ranks” without an associated tensor decomposition and probably these ranks need to be better understood first.

4.3. Numerical Experiments

We sample the statistical dimension (17) numerically for L = 4 instances of the max-norm (97) and a unit rank signal X0; see Figure 3, where we have estimated the statistical dimension using the program (64) with the dual norms being spectral norms of the corresponding b-matrizations. The numerical experiments suggest that the actual reconstruction behavior of the -max-norm is close to twice the lower bound from Theorem 5, which is given by n2. The missing factor of two might be due to the following mismatch. In the argument with the circular cones we only considered tensors of unit b-rank whereas the actual descent cone contains tensors of b-rank 2 for some . This discrepancy should lead to the lower bound be too low by a factor of 2 which is compatible with the plots in Figure 3.

Figure 3

A similar experiment can be done for the related sum-norm from (18). This leads to similar statistical dimensions except for the tensor train bipartition, where the statistical dimension is significantly larger (~ 25%) for the sum-norm.

5. Conclusion and Outlook

We have investigated the problem of convex recovery of simultaneously structured objects from few random observations. We have revisited the idea of taking convex combinations of regularizers and have focused on the best among them, which is given by an optimally weighted maximum of the individual regularizers. We have extended lower bounds on the required number of measurements by Mu et al. [4] to this setting. The bounds are simpler and more explicit than those obtained by Oymak et al. [5] for simultaneously low rank and sparse matrices. They show that it is not possible to improve the scaling of the optimal sampling rate in convex recovery even if optimal tuning and the maximum of simultaneous regularizers are used, the latter giving the largest subdifferential. In more detail, we have obtained lower bounds for the number of measurements in the generic situation and applied this to the cases of (i) simultaneously sparse and low-rank matrices and (ii) certain tensor structures.

For these settings, we have compared the lower bounds to numerical experiments. In those experiments we have (i) demonstrated the actual recovery and (ii) estimated the statistical dimension that gives the actual value of the phase transition of the recovery rate. The latter can be achieved by sampling over certain SDPs. For tensors, we have observed that the lower bound can be quite tight up to a factor of 2.

The main question, whether or not one can derive strong rigorous recovery guarantees for efficient reconstruction algorithms in the case of simultaneous structures remains largely open. However, there are a few smaller questions that we would like to point out.

Numerically, we have observed that weights deviating from the optimal ones have a relatively small effect for the sum of norms as compared to the maximum of norms. Indeed, is a continuous function of μ whereas it appears to be non-continuous for μ,max due to (13). Of course it would be good to have tight upper bounds both regularizers. Perhaps, one can also find a useful interpolation between μ,sum and μ,max by using an ℓp norm of the vector containing the single norms . This interpolation would give the max-norm μ,max for p = ∞ and the sum-norm μ,sum for p = 1 and one could choose p depending on how accurately one knows the optimal weights μ*. Finally, perhaps one can modify an iterative non-convex procedure for solving the optimization problem we are using for the reconstructions such that one obtains recovery from fewer measurements.

Statements

Author contributions

All authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.

Acknowledgments

We thank Michał Horodecki, Omer Sakarya, David Gross, Ingo Roth, Dominik Stoeger, and Željka Stojanak for fruitful discussions.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Footnotes

1.^See the cited works for further references on the classical non-sparse phase retrieval problem.

2.^Assuming that k1rs1n1 and k2rs2n2.

References

  • 1.

    FoucartSRauhutH. A Mathematical Introduction to Compressive Sensing. New York, NY: Springer (2013).

  • 2.

    ChandrasekaranVRechtBParilloAWillskyAS. The convex geometry of linear inverse problems. Found Comput Math. (2012) 12:80549. 10.1007/s10208-012-9135-7

  • 3.

    AmelunxenDLotzMMcCoyMBTroppJA. Living on the edge: phase transitions in convex programs with random data. Inform Infer. (2014) 3:22494. 10.1093/imaiai/iau005

  • 4.

    MuCHuangBWrightJGoldfarbD. Square deal: lower bounds and improved relaxations for tensor recovery. In: Proceedings of the 31st International Conference on Machine Learning. Beijing (2013).

  • 5.

    OymakSJalaliAFazelMEldarYCHassibiB. Simultaneously structured models with application to sparse and low-rank matrices. IEEE Trans. Inf. Theory (2015) 61:2886908. 10.1109/TIT.2015.2401574

  • 6.

    GandySRechtBYamadaI. Tensor completion and low-n-rank tensor recovery via convex optimization. Inverse Probl. (2011) 27:025010. 10.1088/0266-5611/27/2/025010

  • 7.

    GhadermarzyNPlanYYilmazÖ. Near-optimal sample complexity for convex tensor completion. Inform Infer. (2017) iay019. 10.1093/imaiai/iay019

  • 8.

    RauhutHStojanacŽ. Tensor theta norms and low rank recovery. arXiv:1505.05175. (2015).

  • 9.

    RichardEObozinskiGRVertJP. Tight convex relaxations for sparse matrix factorization. In: GhahramaniZWellingMCortesCLawrenceNDWeinbergerKQ editors. Advances in Neural Information Processing Systems 27. Montréal, QC: Curran Associates, Inc. (2014). p. 328492.

  • 10.

    TroppJA. Convex recovery of a structured signal from independent random linear measurements. In: Pfander G, editor. Sampling Theory, a Renaissance. Applied and Numerical Harmonic Analysis.Cham: Birkhäuser (2014). p. 67101.

  • 11.

    Hiriart-UrrutyJBLemaréchalC. Fundamentals of Convex Analysis. Grundlehren text editions. Berlin; New York, NY: Springer (2001).

  • 12.

    AmelunxenDLotzMMcCoyMBTroppJA. Living on the edge: phase transitions in convex programs with random data. arXiv:1303.6672v2 (2013).

  • 13.

    DoanXVVavasisS. Finding approximately rank-one submatrices with the nuclear norm and ℓ1-Norm. SIAM J Optim. (2013) 23:250240. 10.1137/100814251

  • 14.

    RockafellarR. Convex Analysis.Princeton, NJ: Princeton University Press (1970). Second printing (1972).

  • 15.

    KlieschMSzarekSJJungP. Simultaneous structures in convex signal recovery - revisiting the convex combination of norms. arXiv:1904.07893. (2019).

  • 16.

    [Dataset] Dual of the Minkowski Sum,. Available onlie at: http://math.stackexchange.com/questions/1766775/dual-of-the-minkowski-sum (2016). (accessed February 15, 2017).

  • 17.

    LiXVoroninskiV. Sparse signal recovery from quadratic measurements via convex programming. SIAM J. Math. Anal. (2013) 45:301933. 10.1137/120893707

  • 18.

    JaganathanKOymakSHassibiB. Sparse phase retrieval: convex algorithms and limitations. In: 2013 IEEE International Symposium on Information Theory Proceedings (ISIT). Istanbul (2013). 10.1109/ISIT.2013.6620381

  • 19.

    RubinsteinRZibulevskyMEladM. Double sparsity: learning sparse dictionaries for sparse signal approximation. IEEE Trans. Signal Process. (2010) 58:155364. 10.1109/TSP.2009.2036477

  • 20.

    SmolaAJSchölkopfB. Sparse greedy matrix approximation for machine learning. In: Proceedings of the Seventeenth International Conference on Machine Learning. San Francisco, CA (2000). p. 9118.

  • 21.

    JohnstoneIMLuAY. On consistency and sparsity for principal components analysis in high dimensions. J Am Stat Assoc. (2009) 104:68293. 10.1198/jasa.2009.0121

  • 22.

    LeeKWuYBreslerY. Near optimal compressed sensing of sparse rank-one matrices via sparse power factorization. IEEE Trans Inform Theor. (2013) 64:1666-98. 10.1109/TIT.2017.2784479

  • 23.

    LeeKLiYJungeMBreslerY. Blind recovery of sparse signals from subsampled convolution. IEEE Trans Inform Theor. (2017) 63:80221. 10.1109/TIT.2016.2636204

  • 24.

    FlinthA. Sparse blind deconvolution and demixing through ℓ_1, 2-minimization. Adv Comput Math. (2018) 44:121. 10.1007/s10444-017-9533-0

  • 25.

    LeeKLiYJungeMBreslerY. Stability in blind deconvolution of sparse signals and reconstruction by alternating minimization. In: 2015 International Conference on Sampling Theory and Applications, SampTA 2015. Washington, DC (2015). p. 15862. 10.1109/SAMPTA.2015.7148871

  • 26.

    AghasiAAhmedAHandP. BranchHull: convex bilinear inversion from the entrywise product of signals with known signs. Appl Comput Harmon Anal. (2017). 10.1016/j.acha.2019.03.002

  • 27.

    GeppertJKrahmerFStögerD. Sparse power factorization: balancing peakiness and sample complexity. In: 2018 IEEE Statistical Signal Processing Workshop (SSP). Freiburg (2018).

  • 28.

    LingSStrohmerT. Self-calibration and biconvex compressive sensing. Inverse Probl. (2015) 31:115002. 10.1088/0266-5611/31/11/115002

  • 29.

    JungPWalkP. Sparse model uncertainties in compressed sensing with application to convolutions and sporadic communication. In: Boche H, Calderbank R, Kutyniok G, Vybiral J, editors. Compressed Sensing and its Applications. (Cham: Springer) (2015). p. 129. 10.1007/978-3-319-16042-9_10

  • 30.

    WunderGBocheHStrohmerTJungP. Sparse signal processing concepts for efficient 5g system design. IEEE Access. (2015) 3:195208. 10.1109/ACCESS.2015.2407194

  • 31.

    RothIKlieschMWunderGEisertJ. Reliable recovery of hierarchically sparse signals and application in machine-type communications. arXiv:1612.07806. (2016).

  • 32.

    BerthetQRigolletP. Complexity theoretic lower bounds for sparse principal component detection. JMLR (2013) 30:121. 10.1109/ICEEI.2011.6021515

  • 33.

    FornasierMMalyJNaumovaV. Sparse PCA from inaccurate and incomplete measurements. arXiv:1801.06240. (2018).

  • 34.

    BaraniukRDavenportMDeVoreRWakinM. A simple proof of the restricted isometry property for random matrices. Construct Approx. (2008) 28:25363. 10.1007/s00365-007-9003-x

  • 35.

    CandesEPlanY. Tight oracle inequalities for low-rank matrix recovery from a minimal number of noisy random measurements. IEEE Trans Inf Theory. (2011) 57:2342359. 10.1109/TIT.2011.2111771

  • 36.

    RechtBFazelMParriloP. Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization. SIAM Rev. (2010) 52:471501. 10.1137/070697835

  • 37.

    CandesEJTaoT. Decoding by linear programming. IEEE Trans Inf Theory. (2005) 51:420315. 10.1109/TIT.2005.858979

  • 38.

    GrossD. Recovering low-rank matrices from few coefficients in any basis. IEEE Trans Inf Theory. (2011) 57:154866. 10.1109/TIT.2011.2104999

  • 39.

    JalaliA. Convex Optimization Algorithms and Statistical Bounds for Learning Structured Models. Ph.D. thesis, University of Washington, Seattle, WA (2016).

  • 40.

    FoucartSGribonvalRJacquesLRauhutH. Jointly low-rank and bisparse recovery: Questions and partial answers. ArXiv-preprint. (2019) arxiv:1902.04731.

  • 41.

    IwenMViswanathanAWangY. Robust sparse phase retrieval made easy. Applied and Computational Harmonic Analysis42 (2017) 135142. 10.1016/j.acha.2015.06.007

  • 42.

    KoldaTBaderB. Tensor decompositions and applications. SIAM Rev. (2009) 51:455500. 10.1137/07070111X.

  • 43.

    VidalGWernerRF. Computable measure of entanglement. Phys Rev A. (2002) 65:032314. 10.1103/PhysRevA.65.032314

  • 44.

    RauhutHSchneiderRStojanacZ. Low rank tensor recovery via iterative hard thresholding. Linear Algebra and its Applications (2016) 523:22062. 10.1016/j.laa.2017.02.028

Summary

Keywords

compressed sensing, low rank, sparse, matrix, tensor, reconstruction, statistical dimension, convex relaxation

Citation

Kliesch M, Szarek SJ and Jung P (2019) Simultaneous Structures in Convex Signal Recovery—Revisiting the Convex Combination of Norms. Front. Appl. Math. Stat. 5:23. doi: 10.3389/fams.2019.00023

Received

23 November 2018

Accepted

16 April 2019

Published

28 May 2019

Volume

5 - 2019

Edited by

Jean-Luc Bouchot, Beijing Institute of Technology, China

Reviewed by

Valeriya Naumova, Simula Research Laboratory, Norway; Xin Guo, Hong Kong Polytechnic University, Hong Kong

Updates

Copyright

*Correspondence: Martin Kliesch Peter Jung

This article was submitted to Mathematics of Computation and Data Science, a section of the journal Frontiers in Applied Mathematics and Statistics

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics