14
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012 151 On Distributed Convex Optimization Under Inequality and Equality Constraints Minghui Zhu, Member, IEEE, and Sonia Martínez, Senior Member, IEEE Abstract—We consider a general multi-agent convex optimiza- tion problem where the agents are to collectively minimize a global objective function subject to a global inequality constraint, a global equality constraint, and a global constraint set. The objective func- tion is defined by a sum of local objective functions, while the global constraint set is produced by the intersection of local constraint sets. In particular, we study two cases: one where the equality con- straint is absent, and the other where the local constraint sets are identical. We devise two distributed primal-dual subgradient algo- rithms based on the characterization of the primal-dual optimal solutions as the saddle points of the Lagrangian and penalty func- tions. These algorithms can be implemented over networks with dynamically changing topologies but satisfying a standard connec- tivity property, and allow the agents to asymptotically agree on optimal solutions and optimal values of the optimization problem under the Slater’s condition. Index Terms—Cooperative control, distributed optimization, multi-agent systems. I. INTRODUCTION R ECENT advances in sensing, communication and com- putation technologies are challenging the way in which control mechanisms are designed for their efficient exploitation in a coordinated manner. This has motivated a wealth of al- gorithms for information processing, cooperative control, and optimization of large-scale networked multi-agent systems per- forming a variety of tasks. Due to a lack of a centralized au- thority, the proposed algorithms aim to be executed by indi- vidual agents through local actions, with the main feature of being robust to dynamic changes of network topologies. In this paper, we consider a general multi-agent optimiza- tion problem where the goal is to minimize a global objective function, given as a sum of local objective functions, subject to global constraints, which include an inequality constraint, an equality constraint and a (state) constraint set. Each local objec- tive function is convex and only known to one particular agent. On the other hand, the inequality (respectively, equality) con- straint is given by a convex (respectively, affine) function and known by all agents. Each node has its own convex constraint set, and the global constraint set is defined as their intersection. This problem is motivated by others in distributed estimation [21], [28], distributed source localization [25], network utility maximization [13], optimal flow control in power systems [23], Manuscript received October 26, 2009; revised July 22, 2010; accepted April 25, 2011. Date of publication September 15, 2011; date of current version December 29, 2011. This work was supported by the NSF CAREER Award CMS-0643673 and NSF IIS-0712746. Recommended by Associate Editor F. Dabbene. The authors are with the Department of Mechanical and Aerospace Engi- neering, University of California at San Diego, La Jolla CA 92093 USA (e-mail: [email protected]; [email protected]). Digital Object Identifier 10.1109/TAC.2011.2167817 [30] and optimal shape changes of mobile robots [9]. An im- portant feature of the problem is that the objective and (or) con- straint functions depend upon a global decision vector. This re- quires the design of distributed algorithms where, on the one hand, agents can align their decisions through a local informa- tion exchange and, on the other hand, the common decisions will coincide with an optimal solution and the optimal value. Literature Review: In [2], the authors develop a general framework for parallel and distributed computation over a set of processors. Consensus problems, a class of canonical problems on networked multi-agent systems, have been in- tensively studied since then. A necessarily incomplete list of references includes [22] tackling continuous-time consensus, [2], [10], [16] investigating discrete-time versions, and [15] where asynchronous implementation of consensus algorithms is discussed. The papers [5], [6], [12], [29] treat randomized consensus via gossip communication, achieving consensus through quantized information, consensus over random graphs and average consensus through memoryless erasure broadcast channels, respectively. The convergence rate of consensus al- gorithms is discussed, e.g., in [24], and the author in [7] derives conditions to achieve different consensus values. Applications of consensus algorithm to cooperative control are discussed in [27]. In robotics and control communities, convex optimization has been exploited to design algorithms coordinating mobile robots. In [8], in order to increase the connectivity of a multi-agent system, a distributed supergradient-based algorithm is proposed to maximize the second smallest eigenvalue of the Laplacian matrix of the state-dependent proximity graph. In [9], optimal shape changes of mobile robots are achieved through second- order cone programming techniques. The recent papers [18], [20] are the most relevant to our work. In [18], the authors solve a multi-agent unconstrained convex optimization problem through a novel combination of average consensus algorithms with subgradient methods. More recently, the paper [20] further takes local constraint sets into account. To deal with these constraints, the authors in [20] present an exten- sion of their distributed subgradient algorithm, by projecting the original algorithm onto the local constraint sets. Two cases are solved in [20]: the first assumes that the network topologies can dynamically change and satisfy a periodic strong connectivity assumption, but then the local constraint sets are identical; the second requires that the communication graphs are (fixed and) complete and then the local constraint sets can be different. An- other related paper is [11] where a special case of [20], the net- work topology is fixed and all the local constraint sets are iden- tical, is addressed. Statement of Contributions: Building on the work [20], this paper further incorporates global inequality and equality con- 0018-9286/$26.00 © 2011 IEEE

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012 151

On Distributed Convex Optimization UnderInequality and Equality Constraints

Minghui Zhu, Member, IEEE, and Sonia Martínez, Senior Member, IEEE

Abstract—We consider a general multi-agent convex optimiza-tion problem where the agents are to collectively minimize a globalobjective function subject to a global inequality constraint, a globalequality constraint, and a global constraint set. The objective func-tion is defined by a sum of local objective functions, while the globalconstraint set is produced by the intersection of local constraintsets. In particular, we study two cases: one where the equality con-straint is absent, and the other where the local constraint sets areidentical. We devise two distributed primal-dual subgradient algo-rithms based on the characterization of the primal-dual optimalsolutions as the saddle points of the Lagrangian and penalty func-tions. These algorithms can be implemented over networks withdynamically changing topologies but satisfying a standard connec-tivity property, and allow the agents to asymptotically agree onoptimal solutions and optimal values of the optimization problemunder the Slater’s condition.

Index Terms—Cooperative control, distributed optimization,multi-agent systems.

I. INTRODUCTION

R ECENT advances in sensing, communication and com-putation technologies are challenging the way in which

control mechanisms are designed for their efficient exploitationin a coordinated manner. This has motivated a wealth of al-gorithms for information processing, cooperative control, andoptimization of large-scale networked multi-agent systems per-forming a variety of tasks. Due to a lack of a centralized au-thority, the proposed algorithms aim to be executed by indi-vidual agents through local actions, with the main feature ofbeing robust to dynamic changes of network topologies.

In this paper, we consider a general multi-agent optimiza-tion problem where the goal is to minimize a global objectivefunction, given as a sum of local objective functions, subjectto global constraints, which include an inequality constraint, anequality constraint and a (state) constraint set. Each local objec-tive function is convex and only known to one particular agent.On the other hand, the inequality (respectively, equality) con-straint is given by a convex (respectively, affine) function andknown by all agents. Each node has its own convex constraintset, and the global constraint set is defined as their intersection.This problem is motivated by others in distributed estimation[21], [28], distributed source localization [25], network utilitymaximization [13], optimal flow control in power systems [23],

Manuscript received October 26, 2009; revised July 22, 2010; accepted April25, 2011. Date of publication September 15, 2011; date of current versionDecember 29, 2011. This work was supported by the NSF CAREER AwardCMS-0643673 and NSF IIS-0712746. Recommended by Associate Editor F.Dabbene.

The authors are with the Department of Mechanical and Aerospace Engi-neering, University of California at San Diego, La Jolla CA 92093 USA (e-mail:[email protected]; [email protected]).

Digital Object Identifier 10.1109/TAC.2011.2167817

[30] and optimal shape changes of mobile robots [9]. An im-portant feature of the problem is that the objective and (or) con-straint functions depend upon a global decision vector. This re-quires the design of distributed algorithms where, on the onehand, agents can align their decisions through a local informa-tion exchange and, on the other hand, the common decisionswill coincide with an optimal solution and the optimal value.

Literature Review: In [2], the authors develop a generalframework for parallel and distributed computation over aset of processors. Consensus problems, a class of canonicalproblems on networked multi-agent systems, have been in-tensively studied since then. A necessarily incomplete list ofreferences includes [22] tackling continuous-time consensus,[2], [10], [16] investigating discrete-time versions, and [15]where asynchronous implementation of consensus algorithmsis discussed. The papers [5], [6], [12], [29] treat randomizedconsensus via gossip communication, achieving consensusthrough quantized information, consensus over random graphsand average consensus through memoryless erasure broadcastchannels, respectively. The convergence rate of consensus al-gorithms is discussed, e.g., in [24], and the author in [7] derivesconditions to achieve different consensus values. Applicationsof consensus algorithm to cooperative control are discussed in[27].

In robotics and control communities, convex optimization hasbeen exploited to design algorithms coordinating mobile robots.In [8], in order to increase the connectivity of a multi-agentsystem, a distributed supergradient-based algorithm is proposedto maximize the second smallest eigenvalue of the Laplacianmatrix of the state-dependent proximity graph. In [9], optimalshape changes of mobile robots are achieved through second-order cone programming techniques.

The recent papers [18], [20] are the most relevant to our work.In [18], the authors solve a multi-agent unconstrained convexoptimization problem through a novel combination of averageconsensus algorithms with subgradient methods. More recently,the paper [20] further takes local constraint sets into account. Todeal with these constraints, the authors in [20] present an exten-sion of their distributed subgradient algorithm, by projecting theoriginal algorithm onto the local constraint sets. Two cases aresolved in [20]: the first assumes that the network topologies candynamically change and satisfy a periodic strong connectivityassumption, but then the local constraint sets are identical; thesecond requires that the communication graphs are (fixed and)complete and then the local constraint sets can be different. An-other related paper is [11] where a special case of [20], the net-work topology is fixed and all the local constraint sets are iden-tical, is addressed.

Statement of Contributions: Building on the work [20], thispaper further incorporates global inequality and equality con-

0018-9286/$26.00 © 2011 IEEE

Page 2: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

straints. More precisely, we study two cases: one in which theequality constraint is absent, and the other in which the localconstraint sets are identical. For the first case, we adopt a La-grangian relaxation approach, define a Lagrangian dual problemand devise the distributed Lagrangian primal-dual subgradientalgorithm (DLPDS, for short) based on the characterization ofthe primal-dual optimal solutions as the saddle points of the La-grangian function. The DLPDS algorithm involves each agentupdating its estimates of the saddle points via a combinationof an average consensus step, a subgradient (or supgradient)step and a primal (or dual) projection step onto its local con-straint set (or a compact set containing the dual optimal set). TheDLPDS algorithm is shown to asymptotically converge to a pairof primal-dual optimal solutions under the Slater’s conditionand the periodic strong connectivity assumption. Furthermore,each agent asymptotically agrees on the optimal value by imple-menting a dynamic average consensus algorithm developed in[31], which allows a multi-agent system to track time-varyingaverage values.

For the second case, to dispense with the additional equalityconstraint, we adopt a penalty relaxation approach, whiledefining a penalty dual problem and devising the distributedpenalty primal-dual subgradient algorithm (DPPDS, for short).Unlike the first case, the dual optimal set of the second casemay not be bounded, and thus the dual projection steps arenot involved in the DPPDS algorithm. It renders that dualestimates and thus (primal) subgradients may not be uniformlybounded. This challenge is addressed by a more careful choiceof step-sizes. We show that the DPPDS algorithm asymptoti-cally converges to a primal optimal solution and the optimalvalue under the Slater’s condition and the periodic strongconnectivity assumption.

II. PROBLEM FORMULATION AND ASSUMPTIONS

A. Problem Formulation

Consider a network of agents labeled bythat can only interact with each other through local communica-tion. The objective of the multi-agent group is to cooperativelysolve the following optimization problem:

(1)

where is the convex objective function of agent ,is the compact and convex constraint set of agent ,

and is a global decision vector. Assume that and areonly known by agent , and probably different. The function

is known to all the agents with each component, for , being convex. The inequalityis understood component-wise; i.e., , for all

, and represents a global inequality constraint. Thefunction , defined as with

, represents a global equality constraint, and is known toall the agents. We denote , ,and . We assumethat the set of feasible points is non-empty; i.e., .Since is compact and is closed, then we can deduce that

is compact. The convexity of implies that of andthus is continuous. In this way, the optimal value of the

problem (1) is finite and , the set of primal optimal points,is non-empty. Throughout this paper, we suppose the followingSlater’s condition holds:

Assumption 2.1 (Slater’s Condition): There exists a vectorsuch that and . And there exists a

relative interior point of such that where is arelative interior point of ; i.e., and there exists an opensphere centered at such that withbeing the affine hull of .

In this paper, we will study two particular cases of problem(1): one in which the global equality constraint is notincluded, and the other in which all the local constraint sets areidentical. For the case where the constraint is absent,the Slater’s condition 2.1 reduces to the existence of a vector

such that .

B. Network Model

We will consider that the multi-agent network oper-ates synchronously. The topology of the network at time

will be represented by a directed weighted graphwhere is

the adjacency matrix with being the weight assignedto the edge and is the set of edgeswith non-zero weights . The in-neighbors of node at time

are denoted by .We here make the following assumptions on the networkcommunication graphs, which are standard in the analysis ofaverage consensus algorithms; e.g., see [22], [24], and dis-tributed optimization in [18], [20].

Assumption 2.2 (Non-Degeneracy): There exists a constantsuch that , and , for , satisfies

, for all .Assumption 2.3 (Balanced Communication)1: It holds that

for all and , andfor all and .

Assumption 2.4 (Periodical Strong Connectivity): There is apositive integer such that, for all , the directed graph

is strongly connected.

C. Notion and Notations

The notion of saddle point plays a critical role in our paper.Definition 2.1 (Saddle Point): Consider a function

where and are non-empty subsets of and .A pair of vectors is called a saddle point of

over if hold for all.

Remark 2.1: Equivalently, is a saddle point ofover if and only if , and

.In this paper, we do not assume the differentiability of and. At the points where the function is not differentiable, the

subgradient plays the role of the gradient. For a given convexfunction and a point , a subgradient of thefunction at is a vector such that the followingsubgradient inequality holds for any

.Similarly, for a given concave function and a

point , a supgradient of the function at is a vector

1It is also referred to as double stochasticity.

Page 3: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

ZHU AND MARTÍNEZ: ON DISTRIBUTED CONVEX OPTIMIZATION UNDER INEQUALITY AND EQUALITY CONSTRAINTS 153

such that the following supgradient inequalityholds for any .

Given a set , we denote by its convex hull. We letthe function denote the projection operatoronto the non-negative orthant in . For any vector , wedenote , while is the 2-norm in theEuclidean space.

III. CASE (I): ABSENCE OF EQUALITY CONSTRAINT

In this section, we study the case of problem (1) where theequality constraint is absent; i.e.

(2)

We first provide some preliminaries, including a Lagrangiansaddle-point characterization of the problem (2) and finding asuperset containing its Lagrangian dual optimal set. We thenpresent the distributed Lagrangian primal-dual subgradient al-gorithm and summarize the convergence properties.

A. Preliminaries

We here introduce some preliminary results which are essen-tial to the development of the distributed Lagrangian primal-dual subgradient algorithm.

1) A Lagrangian Saddle-Point Characterization: Firstly, theproblem (2) is equivalent to

with associated Lagrangian dual problem given by

Here, the Lagrangian dual function, , is defined as, where is the La-

grangian function . We denote theLagrangian dual optimal value of the Lagrangian dual problemby and the set of Lagrangian dual optimal points by . Asis well-known, under the Slater’s condition 2.1, the property ofstrong duality holds; i.e., , and . The followingtheorem is a standard result on Lagrangian duality stating thatthe primal and Lagrangian dual optimal solutions can be char-acterized as the saddle points of the Lagrangian function.

Theorem 3.1 (Lagrangian Saddle-Point Theorem [3]): Thepair of is a saddle point of the Lagrangianfunction over if and only if it is a pair of primaland Lagrangian dual optimal solutions and the following La-grangian minimax equality holds:

This following lemma presents some preliminary analysis ofsaddle points.

Lemma 3.1 (Preliminary Results on Saddle Points): Letbe any superset of .

(a) If is a saddle point of over , thenis also a saddle point of over .

(b) There is at least one saddle point of over .

(c) If is a saddle point of over , thenand is Lagrangian dual optimal.

Proof:(a) It just follows from the definition of saddle point of over

.(b) Observe that

Since the Slater’s condition 2.1 implies zero duality gap,the Lagrangian minimax equality holds. From Theorem3.1 it follows that the set of saddle points of over

is the Cartesian product . Recall thatand are non-empty, so we can guarantee the existenceof the saddle point of over . Then by (a), wehave that (b) holds.

(c) Pick any saddle point of over .Since the Slater’s condition 2.1 holds, from Theorem3.1 one can deduce that is a pair of primal andLagrangian dual optimal solutions. This implies that

From Theorem 3.1, we have . Hence,. On the other hand, we pick any

saddle point of over . Then for alland , it holds that .By Theorem 3.1, then . Recall ,and thus we have . Since

and , we have. Combining the above two relations gives that

. From Remark 2.1 wesee that . Since

, then and thusis a Lagrangian dual optimal solution.

Remark 3.1: Despite that (c) holds, the reverse of (a) maynot be true in general. In particular, may be infeasible; i.e.,

for some .2) An Upper Estimate of the Lagrangian Dual Optimal Set:

In what follows, we will find a compact superset of . To dothat, we define the following primal problem for each agent :

Due to the fact that is compact and the are continuous,the primal optimal value of each agent’s primal problem isfinite and the set of its primal optimal solutions is non-empty.The associated dual problem is given by

Here, the dual function is defined by, where is the Lagrangian

function of agent and given by .The corresponding dual optimal value is denoted by . In thisway, is decomposed into a sum of local Lagrangian functions;i.e., .

Page 4: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

154 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

Define now the set-valued map by. Additionally, define a

function by . Ob-serve that if is a Slater vector, then . The followinglemma is a direct result of Lemma 1 in [17].

Lemma 3.2 (Boundedness of Dual Solution Sets): The setis bounded for any , and, in particular, for any

Slater vector , it holds that.

Notice that . Picking anySlater vector , and letting in Lemma 3.2gives that

(3)

Define the function by. By the property of weak du-

ality, it holds that and thus for any. Since , thus for any

. With this observation, we pick any and thefollowing set is well-defined:

for some . Observe that for all

(4)

Since , it follows from (3) and (4) that

Hence, we have for all . To simplifythe notations, we will use the shorthands and

. It is easy to see that and are convexand compact.

3) Convexity of : For each , we define the func-tion as . Note that isconvex since it is a nonnegative weighted sum of convex func-tions. For each , we define the functionas . It is easy to check that is a con-cave (actually affine) function. Then the Lagrangian functionis the sum of a collection of convex-concave local functions.This property motivates us to significantly extend primal-dualsubgradient methods in [1], [19] to the networked multi-agentscenario.

B. Distributed Lagrangian Primal-Dual SubgradientAlgorithm

Here, we introduce the Distributed Lagrangian Primal-DualSubgradient Algorithm (DLPDS, for short) to find a saddle pointof the Lagrangian function over and the optimal value.This saddle point will coincide with a pair of primal and La-grangian dual optimal solutions which is not always the case;see Remark 3.1.

Through the algorithm, at each time , each agent main-tains the estimate of of the saddle point of theLagrangian function over and the estimate of of

. To produce (respectively, ), agent takesa convex combination (respectively, ) of its estimate

(respectively, ) with the estimates of its neighboringagents at time , makes a subgradient (respectively, supgradient)step to minimize (respectively, maximize) the local Lagrangianfunction , and takes a primal (respectively, dual) projectiononto the local constraint (respectively, ). Furthermore,agent generates the estimate by taking a convex com-bination of its estimate with those of its neighborsat time and taking one step to track the variation of the localobjective function . More precisely, the DLPDS algorithm isdescribed as follows:

Initially, each agent picks a common Slater vector and acommon . Then agent computes through amax-consensus algorithm, and computes the set with some

. Furthermore, agent chooses any initial state, , and .At every , each agent generates ,

and according to the following rules:

where (respectively, ) is the projection operator ontothe set (respectively, ), the scalars are non-neg-ative weights and the scalars are step-sizes2. Weuse the shorthands , and

.The following theorem summarizes the convergence proper-

ties of the DLPDS algorithm.Theorem 3.2: (Convergence Properties of the DLPDS Algo-

rithm): Consider the optimization problem (2). Let the non-de-generacy assumption 2.2, the balanced communication assump-tion 2.3 and the periodic strong connectivity assumptions 2.4hold. Consider the sequences of , andof the distributed Lagrangian primal-dual subgradient algorithmwith the step-sizes satisfying ,

, and . Then, thereis a pair of primal and Lagrangian dual optimal solutions

such thatand for all . Moreover,

for all .Remark 3.2: For a convex-concave function, continuous-time

gradient-based methods are proved in [1] to converge globallytowards the saddle-point. Recently, [19] presents (discrete-time)

2Each agent � executes the update law of � ��� for � � �.

Page 5: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

ZHU AND MARTÍNEZ: ON DISTRIBUTED CONVEX OPTIMIZATION UNDER INEQUALITY AND EQUALITY CONSTRAINTS 155

primal-dual subgradient methods which relax the differentia-bility in [1] and further incorporate state constraints. Themethod in [1] is adopted by [14] and [26] to study a distributedoptimization problem on fixed graphs where objective functionsare separable.

The DLPDS algorithm is a generalization of primal-dual sub-gradient methods in [19] to the networked multi-agent scenario.It is also an extension of the distributed projected subgradient al-gorithm in [20] to solve multi-agent convex optimization prob-lems with inequality constraints. Additionally, the DLPDS algo-rithm enables agents to find the optimal value. Furthermore, theDLPDS algorithm objective is that of reaching a saddle point ofthe Lagrangian function in contrast to achieving a (primal) op-timal solution in [20].

IV. CASE (II): IDENTICAL LOCAL CONSTRAINT SETS

In last section, we study the case where the equality constraintis absent in problem (1). In this section, we turn our attentionto another case of problem (1) where is taken intoaccount but we require that local constraint sets are identical;i.e., for all . We first adopt a penalty relax-ation and provide a penalty saddle-point characterization of theprimal problem (1) with . We then introduce the dis-tributed penalty primal-dual subgradient algorithm, followed byits convergence properties.

A. Preliminaries

Some preliminary results are developed in this section, andthese results are essential to the design of the distributed penaltyprimal-dual subgradient algorithm.

1) A Penalty Saddle-Point Characterization: The primalproblem (1) with is equivalent to the following:

(5)

with associated penalty dual problem given by

(6)

Here, the penalty dual function, ,is defined by , where

is the penalty function givenby . We de-note the penalty dual optimal value by and the set of penaltydual optimal solutions by . We define the penalty function

for each agent as follows:. In this way, we

have that . As proven in thenext lemma, the Slater’s condition 2.1 ensures zero duality gapand the existence of penalty dual optimal solutions.

Lemma 4.1: (Strong Duality and Non-Emptyness of thePenalty Dual Optimal Set): The values of and coincide,and is non-empty.

Proof: Consider the auxiliary Lagrangian functiongiven by

, with the associated dual problem defined by

(7)

Here, the dual function, , is definedby . The dual optimal valueof problem (7) is denoted by and the set of dual optimalsolutions is denoted . Since is convex, and , for

, are convex, is finite and the Slater’s con-dition 2.1 holds, it follows from Proposition 5.3.5 in [3] that

and . We now proceed to characterize and. Pick any . Since , then

(8)

On the other hand, pick any . Then is feasible,i.e., , and . It implies that

holds for anyand , and thus

. Therefore, we have .To prove , we pick any . From (8) and

, it follows that and thus .The following is a slight extension of Theorem 3.1 to penalty

functions.Theorem 4.1: (Penalty Saddle-Point Theorem): The pair of

is a saddle point of the penalty function overif and only if it is a pair of primal and penalty dual

optimal solutions and the following penalty minimax equalityholds:

Proof: The proof is analogous to that of Proposition 6.2.4in [4], and thus omitted here. For the sake of completeness, weprovide the details in the enlarged version [32].

2) Convexity of : Since is convex and is convexand non-decreasing, thus is convex in for each

. Denote . Since is convexand is an affine mapping, then is convex in

for each .We denote . For each , we define

the function as . Notethat is convex in by using the fact that a nonnegativeweighted sum of convex functions is convex. For each ,we define the function as

. It is easy to check that is concave (actuallyaffine) in . Then the penalty function is the sum ofconvex-concave local functions.

Remark 4.1: The Lagrangian relaxation does not fit to ourapproach here since the Lagrangian function is not convex inby allowing entries to be negative.

B. Distributed Penalty Primal-Dual Subgradient Algorithm

We now proceed to devise the Distributed Penalty Primal-Dual Subgradient Algorithm (DPPDS, for short), that is basedon the penalty saddle-point theorem 4.1, to find the optimalvalue and a primal optimal solution to the primal problem (1)

Page 6: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

156 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

with . The main steps of the DPPDS algorithm are de-scribed as follow.

Initially, agent chooses any initial state ,, , and . At every time

, each agent computes the following convex combinations:

and generates , , andaccording to the following rules:

where is the projection operator onto the set , the scalarsare non-negative weights and the positive scalars

are step-sizes3. The vector defined below is a subgradientof at where

Given a step-size sequence , we define the sum overby and assume that:

Assumption 4.1 (Step-Size Assumption): The step-sizes satisfy , ,

, and ,, .

The following theorem characterizes the convergence of theDPPDS algorithm where a optimal solution and the optimalvalue are asymptotically achieved.

Theorem 4.2 (Convergence Properties of the DPPDS Al-gorithm): Consider the problem (1) with . Let thenon-degeneracy assumption 2.2, the balanced communicationassumption 2.3 and the periodic strong connectivity assumption2.4 hold. Consider the sequences of andof the distributed penalty primal-dual subgradient algorithmwhere the step-sizes satisfy Assumption 4.1. Thenthere exists a primal optimal solution such that

for all . Furthermore, wehave for all .

3Each agent � executes the update law of � ��� for � � �.

Remark 4.2: As the primal-dual subgradient algorithm in[1], [19], the DPPDS algorithm produces a pair of primal anddual estimates at each step. Main differences include: firstly,the DPPDS algorithm extends the primal-dual subgradient al-gorithm in [19] to the multi-agent scenario; secondly, it furthertakes the equality constraint into account. The presence of theequality constraint can make unbounded. Therefore, unlikethe DLPDS algorithm, the DPPDS algorithm does not involvethe dual projection steps onto compact sets. This may causethe subgradient not to be uniformly bounded, while theboundedness of subgradients is a standard assumption in theanalysis of subgradient methods, e.g., see [3], [4], [17]–[20].This difficulty will be addressed by a more careful choice of thestep-size policy; i.e, Assumption 4.1, which is stronger than themore standard diminishing step-size scheme, e.g., in the DLPDSalgorithm and [20]. We require this condition in order to prove,in the absence of the boundedness of , the existence ofa number of limits and summability of expansions toward The-orem 4.2. Thirdly, the DPPDS algorithm adopts the penalty re-laxation instead of the Lagrangian relaxation in [19].

Remark 4.3: Observe that , and(due to the fact that is convex). Furthermore,

is a supgradient of ;i.e., the following penalty supgradient inequality holds for any

and :

(9)

In addition, a step-size sequence that satisfies the step-size as-sumption 4.1 is the harmonic series .The verification can be found in the extended version [32].

V. CONVERGENCE ANALYSIS

We next provide the proofs for Theorem 3.2 and 4.2. Westart our analysis by providing some useful properties of the se-quences weighted by .

Lemma 5.1: (Convergence Properties of Weighted Se-quences): Let . Consider the sequence definedby where ,

and .(a) If , then .(b) If , then .

Proof:(a) For any , there exists such that

for all . Then the following holds for all :

Take the limit on in the above estimate and we have. Since is arbitrary, then

.

Page 7: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

ZHU AND MARTÍNEZ: ON DISTRIBUTED CONVEX OPTIMIZATION UNDER INEQUALITY AND EQUALITY CONSTRAINTS 157

(b) For any , there exists such thatfor all . Then we have

Take the limit on in the above estimate and we have. Since is arbitrary, then

.

Proofs of Theorem 3.2

We now proceed to show Theorem 3.2. To do that, we firstrewrite the DLPDS algorithm into the following form:

where and are projection errors described by

and is the local inputwhich allows agent to track the variation of the local objectivefunction .

Lemma 5.2 (Lipschitz Continuity of ): Consider and. Then there are and such that

and for each pair of and. Furthermore, for each , the

function is Lipschitz continuous with Lipschitz constantover , and for each , the func-

tion is Lipschitz continuous with Lipschitz constant over.

Proof: The readers can find the proof in [32].The following lemma provides a basic iteration relation used

in the convergence proof for the DLPDS algorithm. Many sim-ilar relations have been employed to analyze the subgradientmethods in, e.g., [18], [20], [28].

Lemma 5.3 (Basic Iteration Relation): Let the balanced com-munication assumption 2.3 and the periodic strong connectivityassumption 2.4 hold. For any , any and all ,the following estimates hold:

(10)

(11)

Proof: By Lemma 1 in [20] with ,and , we have that for all

(12)

One can show (11) by substituting the following Lagrangiansupgradient inequality into (12):

Similarly, the equality (10) can be shown by using the followingLagrangian subgradient inequality:

.The following lemma shows that the consensus is asymptot-

ically reached.Lemma 5.4 (Achieving Consensus): Let the non-de-

generacy assumption 2.2, the balanced communicationassumption 2.3 and the periodic strong connectivity as-sumption 2.4 hold. Consider the sequences of ,

and of the DLPDS algorithm with thestep-size sequence satisfying .Then there exist and such that

,for all , and for all ,

.Proof: Observe that and

. Then it follows from Lemma 5.2 that. From Lemma 5.3 it follows that:

(13)

Page 8: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

158 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

Notice that ,and are bounded. Since is continuous, then

and are bounded. Since, the last two terms on the right-hand

side of (13) converge to zero as . One can verify thatexists for any .

On the other hand, taking limits on both sides of (10) weobtain and there-fore we deduce that for all .It follows from [31] thatfor all , . Combining this with the property that

exists for any , we deduce thatthere exists such that forall . Since and is closed, it implies that

for all and thus . Similarly, one can showthat there is such thatfor all .

Since and is continuous,then . It follows from [31] that

for all , .From Lemma 5.4, we know that the sequences of and

of the DLPDS algorithm asymptotically agree on tosome point in and some point in , respectively. Denote by

the set of such limit points. Denote by the averageof primal and dual estimates and

, respectively. The following lemmafurther characterizes that the points in are saddle points of theLagrangian function over .

Lemma 5.5 (Saddle-Point Characterization of ): Each pointin is a saddle point of over .

Proof: Denote by the maximum deviation of primalestimates . Notice that

. Denote by the maximum deviationof dual estimates . Sim-ilarly, we have . We will show thislemma by contradiction. Suppose that there iswhich is not a saddle point of over . At least one ofthe following holds:

(14)

(15)

Suppose first that (14) holds. Then, there exists such that. Consider the sequences of

and which converge respectively to and definedabove. The estimate (10) leads to

where ,,

, ,, and .

It follows from the Lipschitz continuity property of ; seeLemma 5.2, that:

Since , ,, and , then

all , , , , converge to zero as. Then there exists such that for all , it

holds that

Similarly, we have that for all , it holds that

Since and and, are bounded, the above estimate yields a

contradiction by taking sufficiently large. In other words, (14)cannot hold. Following a parallel argument, one can show that(15) cannot hold either. This ensures that each isa saddle point of over .

The combination of (c) in Lemmas 3.1 and Lemma 5.5 givesthat, for each , we have that and

is Lagrangian dual optimal. We still need to verify that isa primal optimal solution. We are now in the position to showTheorem 3.2 based on two claims.

Proofs of Theorem 3.2:Claim 1: Each point is a point in .

Proof: The Lagrangian dual optimality of followsfrom (c) in Lemma 3.1 and Lemma 5.5. To characterize theprimal optimality of , we define an auxiliary sequence

by which is aweighted version of the average of primal estimates. Since

, it follows from Lemma 5.1 (b) that.

Page 9: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

ZHU AND MARTÍNEZ: ON DISTRIBUTED CONVEX OPTIMIZATION UNDER INEQUALITY AND EQUALITY CONSTRAINTS 159

Since is a saddle point of over , thenfor any ; i.e., the following re-

lation holds for any :

(16)

Choose whereis given in the definition of . Then and

implying . Lettingin (16) gives that .

Since , we have . On the other hand, wechoose and then . Letting in (16)gives that and thus . Thecombination of the above two estimates guarantees the propertyof .

We now proceed to show by contradiction. As-sume that does not hold. Denote

and .Then . Since is continuous and converges to ,there exists such that for alland all . Since converges to , without lossof generality, we say that forall . Choose such that for and

for . Since and, thus . Furthermore, ,

then . Equating to and letting inthe estimate (12), the following holds for :

(17)

Summing (17) over with , dividing byon both sides, and using

, we obtain

(18)

Since , are bounded and, then the limit of the first term on the right hand side

of (18) is zero as . Since ,then the limit of the second term is zero as .Since and , thus

.Then it follows from Lemma 5.1 (b) that then the limit

of the third term is zero as . Then we have. Recall that ,

and . Then we reach a contradiction, implyingthat .

Since and , then is a feasible solutionand thus . On the other hand, since is a convexcombination of and is convex, thus wehave the following estimate:

(19)

Recall the following convergence properties:

It follows from Lemma 5.1 (b) that . Therefore, wehave , and thus is a primal optimal point.

Claim 2: It holds that .Proof: The following can be proven by induction on for

a fixed :

(20)

Let in (20) and recall that initial statefor all . Then we have

. The combination of this relation withgives the desired result.

Proofs of Theorem 4.2

In order to analyze the DPPDS algorithm, we first rewrite itinto the following form:

where is projection error described by

and , ,are some local in-

puts. Denote by the maximum deviations of dual estimates

Page 10: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

160 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

and .Before showing Lemma 5.6, we present some useful facts.Since is compact, and , and are continuous, thereexist , , such that for all , it holds that

for all , and .Since is a compact set and , , are convex,then it follows from Proposition 5.4.2 in [3] that there exist

, , such that for all , it holds that,

and . Denote by the av-erages of primal and dual estimates ,

and .Lemma 5.6 (Diminishing and Summable Properties): Sup-

pose the balanced communication assumption 2.3 and the step-size assumption 4.1 hold.

(a) It holds that ,,

, and the sequences of, and

are summable.(b) The sequences ,

, ,and

are summable.Proof: (a) Notice that

where in the last equality we use the balanced communicationassumption 2.3. Recall that . This implies that thefollowing inequalities hold for all :

From here, then we deduce the following recursive estimate on. Repeatedly

applying the above estimates yields that

(21)

Similar arguments can be employed to show that

(22)

Since and ,then we know that and

. Notice that the followingestimate on holds:

(23)

Recall that ,and . Then the result of

follows. By (21), we have

It follows from the step-size assumption 4.1 that. Similarly, one can show that

. By using (21), (22) and (23), wehave the following estimate:

Then the summability of , andverifies that of .

(b) Consider the dynamics of which is in the sameform as the distributed projected subgradient algorithm in[20]. Recall that is uniformly bounded. Thenfollowing from Lemma 7.1 in Appendix withand , we have the summability of

. Thenis summable by using the following set of inequalities:

(24)

where we use . Similarly, it holds that

. We now consider theevolution of . Recall that . By Lemma 1 in [20]with , and , we havethe following:

and thus . With this re-lation, from Lemma 7.1 with and , thefollowing holds for some and :

(25)

Page 11: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

ZHU AND MARTÍNEZ: ON DISTRIBUTED CONVEX OPTIMIZATION UNDER INEQUALITY AND EQUALITY CONSTRAINTS 161

Multiplying both sides of (25) by and using (23),we obtain

Notice that the above inequalities hold for all . Then byemploying the relation of and regroupingsimilar terms, we obtain

Part (a) gives that is summable. Combiningthis fact with , thenwe can say that the first term on the right-hand side in theabove estimate is summable. It is easy to check that thesecond term is also summable. It follows from Part (a) that

andis summable. Then

Lemma 7 in [20] withensures that the third term is summable. Therefore,

the summability of isguaranteed. Following the same lines in (24), one can showthe summability of . Analo-gously, the sequences of and

are summable.Remark 5.1: In Lemma 5.6, the assumption of all local con-

straint sets being identical is utilized to find an upper bound ofthe convergence rate of to zero. This propertyis crucial to establish the summability of expansions pertainingto in part (b).

The following lemma provides basic iteration relations of theDPPDS algorithm.

Lemma 5.7 (Basic Iteration Relation): The following esti-mates hold for any and :

(26)

(27)

Proof: The proof is similar to that of Lemma 5.3.Lemma 5.8 (Achieving Consensus): Let us suppose that the

non-degeneracy assumption 2.2, the balanced communicationassumption 2.3 and the periodical strong connectivity assump-tion 2.4 hold. Consider the sequences of , ,

and of the distributed penalty primal-dualsubgradient algorithm with the step-size sequenceand the associated satisfyingand . Then there existssuch that for all .Furthermore, ,

andfor all , .

Proof: The proof is analogous to Lemma 5.4 and providedin the enlarged version of [32].

We now proceed to show Theorem 4.2 based on five claims.Proof of Theorem 4.2:

Claim 1: For any and, the sequences of

andare summable.

Proof: Observe that

(28)

By using the summability of andin Part (b) of Lemma 5.6, we have that

and thusare summable. Similarly, the

following estimates hold:

Then the property of inPart (b) of Lemma 5.6 implies the summability of the se-quenceand thus the sequence

.

Page 12: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

162 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

Claim 2: Denote the weighted version of as. The following

property holds: .Proof: Summing (26) over and replacing by

leads to

(29)

The summability of in Part (b) of Lemma 5.6implies that the right-hand side of (29) is finite as , andthus

(30)

Pick any . It follows from Theorem 4.1 thatis a saddle point of over . Notice

that . Combining thisrelation, Claim 1 and (30) renders that

and thus . On the other hand,implies that .

Similarly, by using (27) with , , and Claim 1,we have . We reach the desiredrelation.

Claim 3: Denote by.

Denote the weighted version of as. The

following property holds: .

Proof: Notice that

(31)

By using the boundedness of subdifferentials and the primal es-timates, it follows from (31) that:

(32)

Then it follows from (b) in Lemma 5.6 thatis summable. Notice that

, and thus. The desired result

immediately follows from Claim 2.Claim 4: The limit point in Lemma 5.8 is primal optimal.

Proof: Let . By thebalanced communication assumption 2.3, we obtain

This implies that the sequence is non-decreasing in. Observe that is lower bounded by zero. In this

way, we distinguish the following two cases:Case 1: The sequence is upper

bounded. Then is convergent in .Recall that forall , . This implies that there exists

such thatfor all . Observe that

. Thus,we have , im-plying that . Since

for all , then, and thus .

Case 2: The sequence is not upper bounded. Sinceis non-decreasing, then . It follows

from Claim 3 and (a) in Lemma 5.1 that it is impossible that

Page 13: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

ZHU AND MARTÍNEZ: ON DISTRIBUTED CONVEX OPTIMIZATION UNDER INEQUALITY AND EQUALITY CONSTRAINTS 163

. Assume that . Thenwe have the following:

(33)

Notice that the right-hand side of (33) diverges and so does. Then we reach a contradiction, implying

that .In both cases, we have for any . By

utilizing similar arguments, we can further prove that. Since , then is feasible and thus . On the

other hand, since is a convex com-bination of and , thenClaim 3 and (b) in Lemma 5.1 implies that

Hence, we have and thus .Claim 5: It holds that .

Proof: The proof follows the same lines in Claim 2 of The-orem 3.2, and thus omitted here.

VI. CONCLUSION

We have studied a general multi-agent optimization problemwhere the agents aim to minimize a sum of local objective func-tions subject to a global inequality constraint, a global equalityconstraint and a global constraint set defined as the intersectionof local constraint sets. We have considered two cases: the firstone in the absence of the equality constraint and the second onewith identical local constraint sets. To address these cases, wehave introduced two distributed subgradient algorithms whichare based on primal-dual methods. These two algorithms wereshown to asymptotically converge to primal solutions and op-timal values.

APPENDIX

Consider the distributed projected subgradient algorithm pro-posed in [20]: . Denote by

. The following is aslight modification of Lemma 8 and its proof in [20].

Lemma 7.1: Let Assumption 2.2, 2.3 and 2.4 hold. Supposeis a closed and convex set. Then there exist and

such that

Suppose is uniformly bounded for each, and , then we have

.

REFERENCES

[1] K. J. Arrow, L. Hurwicz, and H. Uzawa, Studies in Linear and Non-linear Programming. Stanford, CA: Stanford Univ. Press, 1958.

[2] D. P. Bertsekas and J. N. Tsitsiklis, Parallel and Distributed Compu-tation: Numerical Methods. Belmont, MA: Athena Scientific, 1997.

[3] D. P. Bertsekas, Convex Optimization Theory. Belmont, MA: An-thena Scietific, 2009.

[4] D. P. Bertsekas, A. Nedic, and A. Ozdaglar, Convex Analysis and Op-timization. Belmont, MA: Anthena Scietific, 2003.

[5] S. Boyd, A. Ghosh, B. Prabhakar, and D. Shah, “Randomized gossip al-gorithms,” IEEE Trans. Inform. Theory, vol. 52, no. 6, pp. 2508–2530,Sep. 2006.

[6] R. Carli, G. Como, P. Frasca, and F. Garin, “Distributed averagingon digital erasure networks,” Automatica, vol. 47, no. 1, pp. 115–121,2011.

[7] J. Cortés, “Distributed algorithms for reaching consensus on generalfunctions,” Automatica, vol. 44, no. 3, pp. 726–737, 2008.

[8] M. C. DeGennaro and A. Jadbabaie, “Decentralized control of connec-tivity for multi-agent systems,” in Proc. IEEE Conf. Decision Control,San Diego, CA, Dec. 2006, pp. 3947–3952.

[9] J. Derenick and J. Spletzer, “Convex optimization strategies for coor-dinating large-scale robot formations,” IEEE Trans. Robot., vol. 23, no.6, pp. 1252–1259, Jun. 2007.

[10] A. Jadbabaie, J. Lin, and A. S. Morse, “Coordination of groups ofmobile autonomous agents using nearest neighbor rules,” IEEE Trans.Autom. Control, vol. 48, no. 6, pp. 988–1001, Jun. 2003.

[11] B. Johansson, T. Keviczky, M. Johansson, and K. H. Johansson, “Sub-gradient methods and consensus algorithms for solving convex opti-mization problems,” in Proc. IEEE Conf. Decision Control, Cancun,Mexico, Dec. 2008, pp. 4185–4190.

[12] A. Kashyap, T. Basar, and R. Srikant, “Quantized consensus,” Auto-matica, vol. 43, no. 7, pp. 1192–1203, 2007.

[13] F. P. Kelly, A. Maulloo, and D. Tan, “Rate control in communicationnetworks: Shadow prices, proportional fairness and stability,” J. Oper.Res. Soc., vol. 49, no. 3, pp. 237–252, 1998.

[14] P. Martin and M. Egerstedt, “Optimization of multi-agent motion pro-grams with applications to robotic marionettes,” in Proc. Hybrid Syst.:Computat. Control, Apr. 2009, pp. 262–275.

[15] M. Mehyar, D. Spanos, J. Pongsajapan, S. H. Low, and R. M. Murray,“Asynchronous distributed averaging on communication networks,”IEEE/ACM Trans. Netw., vol. 15, no. 3, pp. 512–520, 2007.

[16] L. Moreau, “Stability of multiagent systems with time-dependent com-munication links,” IEEE Trans. Autom. Control, vol. 50, no. 2, pp.169–182, Feb. 2005.

[17] A. Nedic and A. Ozdaglar, “Approximate primal solutions and rateanalysis for dual subgradient methods,” SIAM J. Optim., vol. 19, no.4, pp. 1757–1780, 2009.

[18] A. Nedic and A. Ozdaglar, “Distributed subgradient methods for multi-agent optimization,” IEEE Trans. Autom. Control, vol. 54, no. 1, pp.48–61, Jan. 2009.

[19] A. Nedic and A. Ozdaglar, “Subgradient methods for saddle-pointproblems,” J. Optim. Theory Appl., vol. 142, no. 1, pp. 205–228, 2009.

[20] A. Nedic, A. Ozdaglar, and P. A. Parrilo, “Constrained consensus andoptimization in multi-agent networks,” IEEE Trans. Autom. Control,vol. 55, no. 4, pp. 922–938, Apr. 2010.

[21] R. D. Nowak, “Distributed EM algorithms for density estimation andclustering in sensor networks,” IEEE Trans. Signal Processing, vol. 51,no. 8, pp. 2245–2253, Aug. 2003.

[22] R. Olfati-Saber, J. A. Fax, and R. M. Murray, “Consensus and coop-eration in networked multi-agent systems,” Proc. IEEE, vol. 95, no. 1,pp. 215–233, Jan. 2007.

[23] A. Oliveira, S. Soares, and L. Nepomuce, “Optimal active power dis-patch combining network flow and interior point approaches,” IEEETrans. Power Syst., vol. 18, no. 4, pp. 1235–1240, Nov. 2003.

[24] A. Olshevsky and J. N. Tsitsiklis, “Convergence speed in distributedconsensus and averaging,” SIAM J. Control Optim., vol. 48, no. 1, pp.33–55, 2009.

[25] M. G. Rabbat and R. D. Nowak, “Decentralized source localization andtracking,” in Proc. IEEE Int. Conf. Acoust., Speech Signal Processing,May 2004, pp. 921–924.

[26] A. Rantzer, “Using game theory for distributed control engineering,” inProc. 3rd World Congress Game Theory Soc. (Games’08),, 2008, [CDROM].

[27] W. Ren and R. Beard, Distributed Consensus in Multi-Vehicle Coop-erative Control. London, U.K.: Springer-Verlag, 2008.

Page 14: IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1 ...php.scripts.psu.edu/muz16/pdf/MZ-SM-TAC12.pdf · 152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

164 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 1, JANUARY 2012

[28] S. S. Ram, A. Nedic, and V. V. Veeravalli, “Distributed and recursiveparameter estimation in parametrized linear state-space models,” IEEETrans. Autom. Control, vol. 55, no. 2, pp. 488–492, Feb. 2010.

[29] A. Tahbaz-Salehi and A. Jadbabaie, “Consensus over random net-works,” IEEE Trans. Autom. Control, vol. 53, no. 3, pp. 791–795, Mar.2008.

[30] H. Wei, H. Sasaki, J. Kubokawa, and R. Yokoyama, “An interior pointnonlinear programming for optimal power flow problems with a noveldata structure,” IEEE Trans. Power Syst., vol. 13, no. 3, pp. 870–877,Aug. 1998.

[31] M. Zhu and S. Martínez, “Discrete-time dynamic average consensus,”Automatica, vol. 46, no. 2, pp. 322–329, 2010.

[32] M. Zhu and S. Martínez, “On Distributed Convex Optimization UnderInequality and Equality Constraints via Primal-Dual SubgradientMethods. Extended Version,” Tech. Rep., 2010 [Online]. Available:http://arxiv.org/abs/1001.2612

Minghui Zhu (M’11) received the B.E. degree (withhonors) in mechanical engineering from ZhejiangUniversity, Hangzhou, China, in 2004, the M.Phil.in mechanical engineering from The Chinese Uni-versity of Hong Kong, Hong Kong, China, in 2006,and is currently pursuing the Ph.D. degree in theMechanical and Aerospace Engineering Department,University of California, San Diego (UC San Diego),La Jolla.

His main research interests include nonlinear con-trol, and distributed control, optimization and infor-

mation processing in networked multi-agent systems.Mr. Zhu received the Powell Fellowship from 2007 to 2010 and the Back

Fellowship from 2007 to 2008 at UC San Diego.

Sonia Martínez (SM’08) received the Ph.D. degreein engineering mathematics from the UniversidadCarlos III de Madrid, Spain, in 2002.

She is an Associate Professor at the Mechanicaland Aerospace Engineering Department, Universityof California at San Diego, La Jolla. Following ayear as a Visiting Assistant Professor of AppliedMathematics at the Technical University of Cat-alonia, Spain, she obtained a Postdoctoral FulbrightFellowship and held positions as a Visiting Re-searcher at UIUC and UCSB. After this, she was an

Assistant Professor at MAE—UCSD during 2006–2010. Her main researchinterests include nonlinear control theory, cooperative control and networkedcontrol systems. In particular, her work has focused on the modeling andcontrol of robotic sensor networks, the development of distributed coordinationalgorithms for groups of autonomous vehicles, and the geometric control ofmechanical systems.

Dr. Martinez received the Best Student Paper Award at the 2002 IEEE Con-ference on Decision and Control, the NSF CAREER Award in 2007, and the2008 Control Systems Magazine Outstanding Paper Award.