6
An ADMM algorithm for matrix completion of partially known state covariances Fu Lin, Mihailo R. Jovanovi´ c, and Tryphon T. Georgiou Abstract—We study the inverse problem of reproducing partially known second-order statistics of a linear time invariant system by the least number of possible input disturbance chan- nels. This can be formulated as a rank minimization problem, and for its solution, we employ a convex relaxation based on the nuclear norm. The resulting optimization problem can be cast as a semi-definite program and solved efficiently using general- purpose solvers for small- and medium-size problems. In this paper, we focus on issues and techniques that are pertinent to large-scale systems. We bring in a re-parameterization which transforms the problem into a form suitable for the alternating direction method of multipliers. Furthermore, we show that each iteration of this algorithm amounts to solving a system of linear equations, an eigenvalue decomposition, and a singular value thresholding. An illustrative example is provided to demonstrate the effectiveness of the developed approach. Index Terms— Alternating direction method of multipliers, convex optimization, low-rank approximation, nuclear norm regularization, singular value thresholding, state covariances, structured matrix completion problems. I. I NTRODUCTION We consider the problem of reproducing partially known second-order statistics using a linear time invariant (LTI) system with a minimum number of input channels. The mo- tivation for this inverse problem stems from applications in which it is desired to reproduce state covariances generated in experiments or simulations using linear systems driven by stochastic disturbances. This completion of state covariances is closely related to well-known matrix completion prob- lems [1]–[4] with a Lyapunov equality constraint imposed by the underlying linear dynamics. The search for the least number of possible input dis- turbance channels gives rise to a rank minimization prob- lem. We employ a convex relaxation based on the nuclear norm [5], [6]. The resulting optimization problem can be cast as a semi-definite program (SDP) and solved efficiently using general-purpose SDP solvers for problems of small and medium size. To deal with large problems, we rely on a re-parameterization that brings the optimization problem into a form suitable for the alternating direction method of multipliers (ADMM). Furthermore, we show that each ADMM iteration amounts to solving a system of linear Supported in part by the University of Minnesota Initiative for Renewable Energy and the Environment under Early Career Award RC-0014-11, the NSF under Grant 1027696, and the AFOSR under Grant FA9550-12-1-0319. F. Lin, M. R. Jovanovi´ c, and T. T. Georgiou are with the De- partment of Electrical and Computer Engineering, University of Min- nesota, Minneapolis, MN 55455. E-mails: [email protected], [email protected], [email protected]. equations, an eigenvalue decomposition, and a singular value thresholding. The paper is organized as follows. We formulate the minimum rank covariance completion problem and make connections to well-known matrix completion problems in Section II. We relax the rank using the nuclear norm and provide a re-parameterization suitable for large problems in Section III. We then cast the constrained convex optimization problem into a form suitable for ADMM in Section IV. We give an illustrative example to demonstrate the effectiveness of the developed approach in Section V. We summarize our contributions and discuss future directions in Section VI. II. PROBLEM FORMULATION Consider a stochastically forced LTI system ˙ x = Ax + Bu (1) where A C n×n , B C n×m , and u(t) C m is a stationary zero-mean stochastic process. For a Hurwitz matrix A and a controllable pair (A, B), it was shown in [7] that the steady- state covariance X := lim t →∞ E (x(t)x * (t)), with E (·) denoting the expectation operator, satisfies a con- straint imposed by the underlying linear dynamics AX + XA * = - (BH * + HB * ). (2) More specifically, a positive semi-definite matrix X is quali- fied to be the state covariance of the linear system (1) if and only if equation (2) is solvable in terms of H C n×m . Our objective is to identify input statistics from partially available state covariances. The motivation for this inverse problem stems from applications in which it is desired to reproduce second-order statistics (generated in experiments or simulations) using an LTI system driven by stochastic disturbances. For a given set of N constraints on a state covariance matrix trace (T i X)= g i , i =1,...,N, (3) we are interested in finding B, H, and X = X * 0 that satisfy the Lyapunov equation (2) and the constraint (3). Here, problem data is specified by (A, T i , g i ) where matrices T i relate X to sampled statistics g i . The structure of the matrix B determines the way distur- bance inputs enter into the state equation; in particular, the

An ADMM algorithm for matrix completion of partially known ...fulin/papers/CDC13_matrix_completion.pdf · it can be shown [5] that the matrix completion problem (MC) is equivalent

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: An ADMM algorithm for matrix completion of partially known ...fulin/papers/CDC13_matrix_completion.pdf · it can be shown [5] that the matrix completion problem (MC) is equivalent

An ADMM algorithm for matrix completionof partially known state covariances

Fu Lin, Mihailo R. Jovanovic, and Tryphon T. Georgiou

Abstract— We study the inverse problem of reproducingpartially known second-order statistics of a linear time invariantsystem by the least number of possible input disturbance chan-nels. This can be formulated as a rank minimization problem,and for its solution, we employ a convex relaxation based on thenuclear norm. The resulting optimization problem can be castas a semi-definite program and solved efficiently using general-purpose solvers for small- and medium-size problems. In thispaper, we focus on issues and techniques that are pertinent tolarge-scale systems. We bring in a re-parameterization whichtransforms the problem into a form suitable for the alternatingdirection method of multipliers. Furthermore, we show thateach iteration of this algorithm amounts to solving a system oflinear equations, an eigenvalue decomposition, and a singularvalue thresholding. An illustrative example is provided todemonstrate the effectiveness of the developed approach.

Index Terms— Alternating direction method of multipliers,convex optimization, low-rank approximation, nuclear normregularization, singular value thresholding, state covariances,structured matrix completion problems.

I. INTRODUCTION

We consider the problem of reproducing partially knownsecond-order statistics using a linear time invariant (LTI)system with a minimum number of input channels. The mo-tivation for this inverse problem stems from applications inwhich it is desired to reproduce state covariances generatedin experiments or simulations using linear systems driven bystochastic disturbances. This completion of state covariancesis closely related to well-known matrix completion prob-lems [1]–[4] with a Lyapunov equality constraint imposedby the underlying linear dynamics.

The search for the least number of possible input dis-turbance channels gives rise to a rank minimization prob-lem. We employ a convex relaxation based on the nuclearnorm [5], [6]. The resulting optimization problem can becast as a semi-definite program (SDP) and solved efficientlyusing general-purpose SDP solvers for problems of smalland medium size. To deal with large problems, we rely ona re-parameterization that brings the optimization probleminto a form suitable for the alternating direction methodof multipliers (ADMM). Furthermore, we show that eachADMM iteration amounts to solving a system of linear

Supported in part by the University of Minnesota Initiative for RenewableEnergy and the Environment under Early Career Award RC-0014-11, theNSF under Grant 1027696, and the AFOSR under Grant FA9550-12-1-0319.

F. Lin, M. R. Jovanovic, and T. T. Georgiou are with the De-partment of Electrical and Computer Engineering, University of Min-nesota, Minneapolis, MN 55455. E-mails: [email protected], [email protected],[email protected].

equations, an eigenvalue decomposition, and a singular valuethresholding.

The paper is organized as follows. We formulate theminimum rank covariance completion problem and makeconnections to well-known matrix completion problems inSection II. We relax the rank using the nuclear norm andprovide a re-parameterization suitable for large problems inSection III. We then cast the constrained convex optimizationproblem into a form suitable for ADMM in Section IV. Wegive an illustrative example to demonstrate the effectivenessof the developed approach in Section V. We summarize ourcontributions and discuss future directions in Section VI.

II. PROBLEM FORMULATION

Consider a stochastically forced LTI system

x = Ax + B u (1)

where A ∈ Cn×n, B ∈ Cn×m, and u(t) ∈ Cm is a stationaryzero-mean stochastic process. For a Hurwitz matrix A and acontrollable pair (A,B), it was shown in [7] that the steady-state covariance

X := limt→∞

E(x(t)x∗(t)),

with E(·) denoting the expectation operator, satisfies a con-straint imposed by the underlying linear dynamics

AX + XA∗ = − (BH∗ +HB∗). (2)

More specifically, a positive semi-definite matrix X is quali-fied to be the state covariance of the linear system (1) if andonly if equation (2) is solvable in terms of H ∈ Cn×m.

Our objective is to identify input statistics from partiallyavailable state covariances. The motivation for this inverseproblem stems from applications in which it is desired toreproduce second-order statistics (generated in experimentsor simulations) using an LTI system driven by stochasticdisturbances.

For a given set of N constraints on a state covariancematrix

trace (TiX) = gi, i = 1, . . . , N, (3)

we are interested in finding B, H , and X = X∗ 0 thatsatisfy the Lyapunov equation (2) and the constraint (3).Here, problem data is specified by (A, Ti, gi) where matricesTi relate X to sampled statistics gi.

The structure of the matrix B determines the way distur-bance inputs enter into the state equation; in particular, the

Page 2: An ADMM algorithm for matrix completion of partially known ...fulin/papers/CDC13_matrix_completion.pdf · it can be shown [5] that the matrix completion problem (MC) is equivalent

dimension of B dictates the number of input channels. It isthus of interest to seek a linear model with a small numberof inputs that is capable of reproducing available data. In thiscase matrix B also becomes a design parameter. However,since the product of B and H in (2) results in a difficultnon-convex equality constraint, we introduce new variable

Q := BH∗ + HB∗.

Since the rank of Q bounds the rank of BH∗, and hence thenumber of input channels [8], we seek a matrix Q with theminimum rank. Therefore, we consider the following rankminimization problem

minimizeQ,X

rank (Q)

subject to AX + XA∗ + Q = 0

trace (TiX) = gi, i = 1, . . . , N

X 0.

(4)

Note that the constraint set is convex because it is theintersection of the positive semi-definite cone and the linearsubspace determined by the Lyapunov equation and traceconstraints. On the other hand, the rank in the objectivefunction is the source of non-convexity.

A. Connections to matrix completion problems

Matrix completion problems have received considerableattention in recent years due to emerging applications, ex-citing theoretical developments, and efficient optimizationalgorithms [1]–[4], [6], [9]–[13]. It is therefore instructiveto provide connections and highlight differences between (4)and the positive semi-definite matrix completion problem.

Consider the problem of recovering a positive semi-definite matrix from a sample of its entries. This problemcan be expressed as

find X 0

subject to Xjl = Mjl, (j, l) ∈ Ω

where Ω is the index set of known entries Mjl. The problemof finding a minimum rank matrix that is consistent with theobserved data can be formulated as

minimizeX

rank (X)

subject to Xjl = Mjl, (j, l) ∈ Ω

X 0.

(5)

Note that the rank minimization problem (4) with

A = − (1/2)I, Ti = el eTj

where ej is the jth element of the canonical basis of Rn,simplifies to the standard matrix completion problem (5). Inother words, the rank minimization problem (4) encompassesthe following structured covariance completion problem

minimizeQ,X

rank (Q)

subject to AX + XA∗ + Q = 0

Xjl = Mjl, (j, l) ∈ Ω

X 0.

(6)

It is worth mentioning that problem (4) also accommodatesthe output covariance completion problem. Suppose thatinstead of having access to some of the entries of the statecovariance matrix, we have a sample of entries of the outputcovariance matrix

(CXC∗)jl = Mjl, (j, l) ∈ Ω

with C ∈ Cp×n denoting the output matrix. Since theselinear constraints can be written as

trace(C∗ele

Tj CX

)= Mjl, (j, l) ∈ Ω,

we conclude that the structured output covariance comple-tion problem

minimizeQ,X

rank (Q)

subject to AX + XA∗ + Q = 0

(CXC∗)jl = Mjl, (j, l) ∈ Ω

X 0

(7)

is also a special case of (4) with

Ti = C∗eleTj C, gi = Mjl.

III. NUCLEAR NORM RELAXATION

In this section, we relax the rank with the nuclear norm toobtain a convex optimization problem. For the relaxed con-vex problem, we provide an SDP formulation. Furthermore,we parameterize the constraint set and bring the relaxedformulation into a form that is well-suited to large problems.

The nuclear norm, i.e., the sum of the singular values ofa matrix

‖Q‖∗ :=

n∑i=1

σi(Q)

provides an effective surrogate for rank minimization prob-lems [5], [6]. In contrast to the non-convex rank function, thenuclear norm is convex and it is the best convex approxima-tion of rank function over the set of matrices with spectralnorm no greater than one [6]. Thus, relaxation of rank tonuclear norm in (4) yields a convex optimization problem

minimizeQ,X

‖Q‖∗subject to AX + XA∗ + Q = 0

trace (TiX) = gi, i = 1, . . . , N

X 0.

(MC)

A. SDP formulation

For a Hermitian matrix Q, the nuclear norm of Q can bewritten as

‖Q‖∗ =

n∑i=1

|λi(Q)|.

Therefore, by splitting Q into positive and negative semi-definite parts

Q = Q+ − Q−, Q+ 0, Q− 0

Page 3: An ADMM algorithm for matrix completion of partially known ...fulin/papers/CDC13_matrix_completion.pdf · it can be shown [5] that the matrix completion problem (MC) is equivalent

it can be shown [5] that the matrix completion problem (MC)is equivalent to the following semi-definite program

minimizeQ+, Q−, X

trace (Q+) + trace (Q−)

subject to AX + XA∗ + Q+ − Q− = 0

trace (TiX) = gi, i = 1, . . . , N

X 0, Q+ 0, Q− 0.

When the number of states is n . 50, this SDP canbe solved using general-purpose SDP solvers. Customizedinterior point method for unconstrained nuclear norm mini-mization [9] can deal with matrices with a few hundreds ofrows and columns. However, the positive semi-definite con-straint and the Lyapunov equality constraint prevent us fromemploying this approach. We next provide an alternativeproblem formulation that is well-suited to large problems.

B. Null space parameterization

We begin by parameterizing the null space of the linearoperator determined by trace constraints. This null spaceparameterization will allow us to eliminate trace constraintsand the Lyapunov equation from (MC).

Since the set of trace constraints

trace (TiX) = gi, i = 1, . . . , N

defines a linear mapping

T (X) = g (8)

from the Hermitian matrix X to the vector g :=[ g1 · · · gN ]

T , it follows that all solutions to (8) can beparameterized using the null space of the linear operator T .Specifically,

X = X0 +

q∑j=1

zjXj (9)

where X0 = X∗0 is a particular solution of (8), zj’s arereal coefficients, and Xj = X∗j ’s are linearly independentmatrices that satisfy

trace (TiXj) = 0, i = 1, . . . , N, j = 1, . . . , q.

Here, q is the dimension of the null space of T .

Similarly, we parameterize Q by substituting (9) into theLyapunov equation in (MC)

Q = − (AX + XA∗) = Q0 +

q∑j=1

zjQj

where

Q0 := − (AX0 + X0A∗), Qj := − (AXj + XjA

∗).

Therefore, the above presented null space parameterizationcan be used to recast the state covariance completion prob-

lem (MC) into

minimizeQ,X, z

‖Q‖∗

subject to Q = Q0 +

q∑j=1

zjQj

X = X0 +

q∑j=1

zjXj

X 0.

(10)

Here, Qj and Xj are problem data and Q = Q∗, X = X∗,and z are the optimization variables. Note that the vector zcouples linear constraints in (10). This particular structurewill be exploited in Section IV.

Remark 1: One approach to compute a solution X0 of (8)is to solve the under-determined system of linear equations

[ vec (T1) · · · vec (TN ) ]T

vec (X) = g

where vec (X) denotes the vectorization of the upper trian-gular part (including the main diagonal) of the Hermitianmatrix X . The basis elements Xj’s of the null space of Tcan be computed via the singular value decomposition of T .

IV. ALTERNATING DIRECTION METHOD OF MULTIPLIERS

We next employ the alternating direction method of mul-tipliers (ADMM) to solve (10). This method has been em-ployed effectively in low-rank matrix recovery [11], sparsecovariance selection [14], image denoising and magneticresonance imaging [15], sparse feedback synthesis [16], andmany other applications [17]. When applied to (10), wewill demonstrate that each ADMM iteration requires aneigenvalue decomposition of X , a singular value thresholdingof Q, and the solution to a system of linear equations for z.

Note that problem (10) is not in a form suitable forADMM since it has three independent variables Q, X , zand two coupling linear constraints. To put (10) into the so-called two-block form suitable for ADMM, let us introduce

f(Q,X) := ‖Q‖∗ + φ(X), h(z) ≡ 0

where φ(X) is the indicator function

φ(X) =

0, X 0∞, otherwise.

Since f is a convex separable function of Q and X andsince h is a trivially convex function of z, the followingequivalent formulation to (10) is in the standard form suitablefor ADMM [17]

minimizeX,Q, z

f(Q,X) + h(z)

subject to[Q(z)X (z)

]−[QX

]= 0

(11)

where

Q(z) := Q0 +

q∑j=1

zjQj , X (z) := X0 +

q∑j=1

zjXj .

We next form the augmented Lagrangian associated with

Page 4: An ADMM algorithm for matrix completion of partially known ...fulin/papers/CDC13_matrix_completion.pdf · it can be shown [5] that the matrix completion problem (MC) is equivalent

the constrained problem (11)

Lρ(Q,X, z, Y, Z) = ‖Q‖∗ + φ(X) + 〈Y,Q(z) − Q〉+

〈Z,X (z) − X〉 +ρ

2‖Q(z) − Q‖2F +

ρ

2‖X (z) − X‖2F

where Y = Y ∗ and Z = Z∗ are Lagrange multipliers, ρ is apositive scalar, 〈·, ·〉 is the inner product of two matrices, and‖ · ‖F is the Frobenius norm. The ADMM algorithm uses asequence of iterations

(Q,X)k+1 := arg minQ,X

Lρ(Q,X, zk, Y k, Zk) (12a)

zk+1 := arg minz

Lρ(Qk+1, Xk+1, z, Y k, Zk) (12b)[Y k+1

Zk+1

]:=

[Y k

Zk

]+ ρ

[Q(zk+1) − Qk+1

X (zk+1) − Xk+1

](12c)

to find the solution of (11). These iterations terminate whenthe primal and dual residuals are sufficiently small

primal residues:

‖Q(zk+1)−Qk+1‖F ≤ ε1

‖X (zk+1)−Xk+1‖F ≤ ε2

dual residues:

‖Qk+1 −Qk‖F ≤ ε3

‖Xk+1 −Xk‖F ≤ ε4

‖zk+1 − zk‖2 ≤ ε5

where εi’s are determined by problem data; see [17, Section3.3] for details.

A. Solution to the (Q,X)-minimization problem

For fixed zk, Y k, Zk, the minimization of the aug-mented Lagrangian Lρ with respect to (Q,X) amounts to

minimizeQ,X

‖Q‖∗ + φ(X) +ρ

2‖Q − (Q(zk) +

1

ρY k)‖2F

2‖X − (X (zk) +

1

ρZk)‖2F .

Separability of the objective function with respect to Q andX results in the following two uncoupled problems

minimizeQ

‖Q‖∗ + (ρ/2)‖Q − (Q(zk) + (1/ρ)Y k)‖2F(13)

and

minimizeX

φ(X) + (ρ/2)‖X − (X (zk) + (1/ρ)Zk)‖2F .(14)

The solution to the Q-minimization problem (13) is givenby the singular value thresholding operator [18]. Specifi-cally, we compute the singular value decomposition of thesymmetric matrix

Q(zk) + (1/ρ)Y k = U ΣU∗

and apply the soft-thresholding operator to the resultingsingular values

S1/ρ(Σ) = diag(σi − 1/ρ)+

where a+ := maxa, 0. Then, the solution to the Q-minimization problem (13) is determined by [18]

Qk+1 = U S1/ρ(Σ)U∗.

On the other hand, the solution of the X-minimizationproblem (14) is determined by the projection of X (zk) +(1/ρ)Zk onto the positive semi-definite cone. Specifically,we compute the eigenvalue decomposition of the matrix

X (zk) + (1/ρ)Zk = V ΛV ∗

and truncate the negative eigenvalues

Λ+ = diag(λi)+

to obtain the solution

Xk+1 = V Λ+ V∗.

B. Solution to the z-minimization problem

Problem (12b) is a convex quadratic program

minimizez

ρ

2‖Q(z) − Qk+1‖2F +

ρ

2‖X (z) − Xk+1‖2F +

〈Y k,Q(z) − Qk+1〉 + 〈Zk,X (z) − Xk+1〉

that can be rewritten compactly as

minimizez

(ρ/2) zTHz + (bk)T z.

Here, the Hessian H and the gradient bk are given by

Hij = trace (QiQj + XiXj)

bki = trace (Y kQi + ZkXi + ρQi(Q0 −Qk+1) +

ρXi(X0 −Xk+1)).

Therefore, this z-minimization problem (12b) amounts tosolving the system of linear equations

ρHz + bk = 0

whose solution can be computed efficiently using Choleskyfactorization and back-solve operations. Note that the Hes-sian matrix H is independent of ADMM iterations. This im-plies that we only need to compute the Cholesky factorizationof H once, and then save the Cholesky factor for the back-solve operations in subsequent z-minimization steps.

C. Computational complexity

Since the Q-minimization problem amounts to a singularvalue decomposition, it costs O(n3) operations, where n isthe number of states. Similarly, since the X-minimizationproblem amounts to an eigenvalue decomposition, this stepalso requires O(n3) operations. On the other hand, theCholesky factorization of H takes O(q3) operations and thesubsequent z-minimization step takes O(q2) operations forthe back-solve operations, where

q := n(n+ 1)/2 − N

Page 5: An ADMM algorithm for matrix completion of partially known ...fulin/papers/CDC13_matrix_completion.pdf · it can be shown [5] that the matrix completion problem (MC) is equivalent

is the dimension of the null space. Therefore, the totalcomputational cost is determined by

O(maxq3, kn3, kq2)

where k is the number of ADMM iterations.

Without exploiting sparsity in basis representation, thememory requirements are O(q2) for the Cholesky factor ofH and O(qn2) for the basis of the null space. Note that forthe structured covariance completion problem (6), the basiselements are sparse matrices that contain only single nonzeroentry.

V. AN EXAMPLE

Consider a mass-spring-damper system with 50 massessubject to disturbances that are outputs of a low-pass filterdriven by white noise

low-pass filter: xf = −xf + d

u = xf

mass-spring-damper system: x = Ax + B u

where d is a white stochastic disturbance with zero-meanand unit variance. The state and the input matrices of themass-spring-damper system are

A =

[O I−K − 2K

], B =

[OI

]where O and I are zero and identity matrices and K ∈R50×50 is a symmetric tridiagonal Toeplitz matrix with 2on the main diagonal and −1 on the first upper- and lower-subdiagonal.

The steady-state covariance Σ of the cascade connectionis determined by the solution of the Lyapunov equation

AΣ + ΣA∗ + BB∗ = 0 (15)

whereA =

[A BO − I

], B =

[BI

]and

Σ =

[Σxx Σxxf

Σxfx Σxfxf

].

We partition the state covariance Σxx of the mass-spring-damper system into 2 × 2 block matrices compatible withthe position and velocity of masses

Σxx =

[Σpp ΣpvΣvp Σvv

].

The diagonals of Σpp, Σvv , and Σpv are assumed to be theavailable data and hence

g =

diag (Σpp)diag (Σvv)diag (Σpv)

.We are interested in completing the state covariance X that

are consistent with the diagonals of Σpp, Σvv , and Σpv . Tothis end, we solve the covariance completion problem (MC).

(a) Xpp (b) Xvv

Fig. 1: The matrices Xpp and Xvv resulting from thecovariance completion problem (MC) for the mass-spring-damper system. Sampled statistics in (MC) are determinedby the diagonals of the matrices Σpp, Σvv , and Σpv .

(a) Σpp (b) Σvv

Fig. 2: Position and velocity covariances, Σpp and Σvv ,resulting from the solution Σ to the Lyapunov equation (15)for the mass-spring-damper system.

Figure 1 shows the completed position and velocity covari-ances Xpp and Xvv . We observe their close correspondenceto the true covariance matrices Σpp and Σvv shown in Fig. 2.In this example, it turns out that both Σpv and Xpv are zeromatrices.

We next examine the singular values of Q that results fromsolving covariance completion problem (MC). As shown inFig. 3, there are seven relatively large singular values andthe rest of singular values are much smaller. On the otherhand, Fig. 4 demonstrates that there is no clear-cut in thesingular values of Q with

Q = − (AΣxx + ΣxxA∗).

The approximately low-rank feature of the solution Q il-lustrates the utility of the covariance matrix completionproblem (MC).

VI. CONCLUDING REMARKS

We study the inverse problem of reproducing partiallyknown second-order statistics using an LTI system with aminimum number of input channels. We formulate a rankminimization problem that encompasses the positive semi-definite matrix completion problem. We employ the nuclearnorm relaxation and show that the resulting optimizationproblem can be formulated as an SDP. Furthermore, weprovide a null space parameterization that allows us to cast

Page 6: An ADMM algorithm for matrix completion of partially known ...fulin/papers/CDC13_matrix_completion.pdf · it can be shown [5] that the matrix completion problem (MC) is equivalent

0 20 40 60 80 1000

0.2

0.4

0.6

0.8

1

1.2

i

σi(Q

)

Fig. 3: Singular values of the approximately low-rank solu-tion Q to the covariance matrix completion problem (MC)for the mass-spring-damper system.

0 20 40 60 80 1000

0.2

0.4

0.6

0.8

1

1.2

i

σi

( Q)

Fig. 4: Singular values of Q = −(AΣxx + ΣxxA∗) for the

mass-spring-damper system.

the problem into a form suitable for the ADMM algorithm.We show that each ADMM iteration amounts to comput-ing an eigenvalue and a singular value decomposition andsolving a system of linear equations. An illustrative exampleis provided to demonstrate the utility of the developedapproach.

A. Future directions

Suppose that Q is decomposed into Q = S + S∗. It isshown in [8] that the number of positive and negative eigen-values of Q is bounded by the rank of S. In particular, themaximum of the number of positive and negative eigenvaluesof Q coincides with the rank of S; see [8]. In view of this, itis also of interest to consider the following rank minimization

problem

minimizeQ+, Q−, X

max rank (Q+) , rank (Q−)

subject to AX + XA∗ + Q+ − Q− = 0

trace (TiX) = gi, i = 1, . . . , N

X 0, Q+ 0, Q− 0.

We intend to develop efficient algorithms for the convexrelaxation of this problem in future work.

REFERENCES

[1] E. J. Candes and B. Recht, “Exact matrix completion via convexoptimization,” Foundations of Computational mathematics, vol. 9,no. 6, pp. 717–772, 2009.

[2] E. J. Candes and Y. Plan, “Matrix completion with noise,” Proceedingsof the IEEE, vol. 98, no. 6, pp. 925–936, 2010.

[3] E. J. Candes and T. Tao, “The power of convex relaxation: Near-optimal matrix completion,” IEEE Trans. Information Theory, vol. 56,no. 5, pp. 2053–2080, 2010.

[4] R. H. Keshavan, A. Montanari, and S. Oh, “Matrix completion froma few entries,” IEEE Trans. Information Theory, vol. 56, no. 6, pp.2980–2998, 2010.

[5] M. Fazel, “Matrix rank minimization with applications,” Ph.D. disser-tation, Stanford University, March 2002.

[6] B. Recht, M. Fazel, and P. A. Parrilo, “Guaranteed minimum-ranksolutions of linear matrix equations via nuclear norm minimization,”SIAM Review, vol. 52, no. 3, pp. 471–501, 2010.

[7] T. T. Georgiou, “The structure of state covariances and its relationto the power spectrum of the input,” IEEE Trans. Automat. Control,vol. 47, no. 7, pp. 1056–1066, 2002.

[8] Y. Chen, M. R. Jovanovic, and T. T. Georgiou, “State covariances andthe matrix completion problem,” in Proceedings of the 52nd IEEEConference on Decision and Control, Florence, Italy, 2013.

[9] Z. Liu and L. Vandenberghe, “Interior-point method for nuclear normapproximation with application to system identification,” SIAM J.Matrix Anal. & Appl., vol. 31, no. 3, pp. 1235–1256, 2009.

[10] K.-C. Toh and S. Yun, “An accelerated proximal gradient algorithmfor nuclear norm regularized linear least squares problems,” Pacific J.Optimization, vol. 6, pp. 615–640, 2010.

[11] M. Tao and X. Yuan, “Recovering low-rank and sparse componentsof matrices from incomplete and noisy observations,” SIAM J. Optim.,vol. 21, no. 1, pp. 57–81, 2011.

[12] S. Ma, D. Goldfarb, and L. Chen, “Fixed point and Bregman iterativemethods for matrix rank minimization,” Mathematical Programming,vol. 128, no. 1, pp. 321–353, 2011.

[13] D. Goldfarb and S. Ma, “Convergence of fixed-point continuation algo-rithms for matrix rank minimization,” Foundations of ComputationalMathematics, vol. 11, no. 2, pp. 183–210, 2011.

[14] X. Yuan, “Alternating direction method of multipliers for covarianceselection models,” Journal of Scientific Computing, vol. 51, pp. 261–273, 2012.

[15] T. Goldstein and S. Osher, “The split Bregman method for `1-regularized problems,” SIAM J. Imaging Sci., vol. 2, no. 2, pp. 323–343, 2009.

[16] F. Lin, M. Fardad, and M. R. Jovanovic, “Design of optimal sparsefeedback gains via the alternating direction method of multipliers,”IEEE Trans. Automat. Control, vol. 58, no. 9, pp. 2426–2431, 2013.

[17] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein, “Distributedoptimization and statistical learning via the alternating directionmethod of multipliers,” Foundations and Trends in Machine Learning,vol. 3, no. 1, pp. 1–124, 2011.

[18] J.-F. Cai, E. J. Candes, and Z. Shen, “A singular value thresholdingalgorithm for matrix completion,” SIAM J. Optim., vol. 20, no. 4, pp.1956–1982, 2010.