Non Convex 1

Embed Size (px)

Citation preview

  • 8/12/2019 Non Convex 1

    1/61

    Advances in

    Mechanics and Mathematics

    Volume III

    Dedicated to Gilbert Strang on the Occasion

    of His 70th Birthday

    Edited by

    David Y. Gao & Hanif D. SheraliVirginia Polytechnic Institute & State UniversityBlacksburg, VA 24061, USAE-mails: [email protected], [email protected]

  • 8/12/2019 Non Convex 1

    2/61

    Chapter 5

    Nonconvex Optimization forCommunication Networks

    Mung Chiang

    Summary. Nonlinear convex optimization has provided both an insightfulmodeling language and a powerful solution tool to the analysis and design ofcommunication systems over the last decade. A main challenge today is onnonconvex problems in these applications. This chapter presents an overviewon some of the important nonconvex optimization problems in communi-cation networks. Four typical applications are covered: Internet congestioncontrol through nonconcave network utility maximization, wireless networkpower control through geometric and sigmoidal programming, DSL spectrummanagement through distributed nonconvex optimization, and Internet intra-domain routing through nonconvex, nonsmooth optimization. A variety ofnonconvex optimization techniques are showcased: sum-of-squares program-ming through successive SDP relaxation, signomial programming throughsuccessive GP relaxation, leveraging specific structures in these engineeringproblems for efficient and distributed heuristics, and changing the underlyingprotocol to enable a different problem formulation in the first place. Collec-

    tively, they illustrate three alternatives of tackling nonconvex optimizationfor communication networks: going through nonconvexity, around non-convexity, and above nonconvexity.

    Key words: Digital subscriber line, duality, geometric programming, Inter-net, network utility maximization, nonconvex optimization, power control,routing, semidefinite programming, sum of squares, TCP/IP, wireless net-work

    Mung ChiangElectrical Engineering Department, Princeton University, Princeton, NJ 08544, U.S.A.e-mail: [email protected]

    137

  • 8/12/2019 Non Convex 1

    3/61

  • 8/12/2019 Non Convex 1

    4/61

    5 Nonconvex Optimization for Communication Networks 139

    This chapter overviews the latest results in recent publications about thefirst two topics, with a particular focus on showing the connections betweenthe engineering intuitions about important problems in communication net-works and the state-of-the-art algorithms in nonconvex optimization theory.

    Most of the results surveyed here were obtained in 20052006, and the prob-lems driven by fundamental issues in the Internet, wireless, and broadbandaccess networks. As this chapter illustrates, even after much progress madein recent years, there are still many challenging mysteries to be resolved onthese important nonconvex optimization problems.

    1 2 3

    Fig. 5.1 Three major types of approaches when tackling nonconvex optimization problemsin communication networks: Go (1) through, (2) around, or (3) above nonconvexity.

    It is interesting to point out that, as illustrated in Figure 5.1, there are atleast three very different approaches to tackle the difficult issue of noncon-vexity.

    Go through nonconvexity. In this approach, we try to solve the difficultnonconvex problem; for example, we may use successive convex relaxations(e.g., sum-of-squares, signomial programming), utilize special structures inthe problem (e.g., difference of convex functions, generalized quasiconcav-ity), or leverage smarter branch and bound methods.

    Go around nonconvexity. In this approach, we try to avoid solving theconvex problem; for example, we may discover a change of variables thatturns the seemingly nonconvex problem into a convex one, determine con-ditions under which the problem is convex or the KKT point is unique, ormake approximations to make the problem convex.

    Go above nonconvexity. In this approach, we try to reformulate thenonconvex problem in the first place to make it more solvable or ap-proximately solvable. We observe that optimization problem formulationsare induced by some underlying assumptions on what the network archi-tectures and protocols should look like. By changing these assumptions, a

    diff

    erent, much easier-to-solve or easier-to-approximate formulations mayresult. We refer to this approach asdesign for optimizability, which is con-cerned with redrawing architectures to make the resulting optimization

  • 8/12/2019 Non Convex 1

    5/61

    140 Mung Chiang

    problem easier to solve. This approach of changing a hard problem intoan easier one is in contrast to optimization, which tries to solve a given,possibly difficult, problem.

    The four topics chosen in this chapter span a range of application contexts and

    tasks in communication networks. The sources of difficulty in these nonconvexoptimization problems are summarized in Table 5.1, together with the keyideas in solving them and the type of approaches used. For more detailsbeyond this brief overview chapter, please refer to the related publications[29, 19, 14, 35, 7, 70, 71] by the author and coworkers and the referencestherein.

    Table 5.1 Summary of four nonconvex optimization problems in this chapter

    Section Application Task Di fficulty Solution Approach

    5.2 Internet Congestion Nonconcave U Sum of Throughcontrol squares

    5.3 Wireless Power Posynomial Geometric Aroundcontrol ratio program

    5.4 DSL Spectrum Posynomial Problem Aroundmanagement ratio structure

    5.5 Internet Routing Nonconvex Approximation Aboveconstraint

    5.2 Internet Congestion Control

    5.2.1 Introduction

    Basic Network Utility Maximization

    Since the publication of the seminal paper [37] by Kelly, Maulloo, and Tan in1998, the framework of network utility maximization (NUM) has found manyapplications in network rate allocation algorithms and Internet congestioncontrol protocols (e.g., surveyed in [45, 60]). It has also led to a systematicunderstanding of the entire network protocol stack in the unifying frameworkof layering as optimization decomposition (e.g., surveyed in [13, 49, 44]).By allowing nonlinear concave utility objective functions, NUM substantiallyexpands the scope of the classical LP-based network flow problems.

    Consider a communication network withLlinks, each with a fixed capacityofc

    lbps, and Ssources (i.e., end-users), each transmitting at a source rate

    ofxs bps. Each source s emits one flow, using a fixed set L(s) of links in itspath, and has a utility functionUs(xs). Each link l is shared by a set S(l) of

  • 8/12/2019 Non Convex 1

    6/61

    5 Nonconvex Optimization for Communication Networks 141

    sources. Network utility maximization, in its basic version, is the followingproblem of maximizing the total utility of the network

    Ps Us(xs), over the

    source rates x, subject to linear flow constraintsP

    s:lL(s) xs cl for alllinksl:

    maximizeP

    s Us(xs)

    subject toP

    sS(l) xs cl, l,

    x0,

    (5.1)

    where the variables are x RS.There are many nice properties of the basic NUM model due to several

    simplifying assumptions of the utility functions and flow constraints, whichprovide the mathematical tractability of problem (5.1) but also limit its ap-plicability. In particular, the utility functions {Us} are often assumed to beincreasing and strictly concave functions.

    Assuming thatUs

    (xs

    ) becomes concave for large enoughxs

    is reasonable,because the law of diminishing marginal utility eventually will be effective.However,Us may not be concave throughout its domain. In his seminal pa-per in 1995, Shenker [57] differentiated inelastic network traffic from elastictraffic. Utility functions for elastic traffic were modeled as strictly concavefunctions. Although inelastic flows with nonconcave utility functions repre-sent important applications in practice, they have received little attentionand rate allocation among them has only a limited mathematical foundation.There have been three recent publications [41, 29, 19] (see also earlier workin [69, 42, 43] related to the approach in [41]) on this topic.

    In this section, we investigate the extension of the basic NUM to max-imization of nonconcave utilities, as in the approach of [19]. We provide acentralized algorithm for offline analysis and establishment of a performance

    benchmark for nonconcave utility maximization when the utility function is apolynomial or signomial. Based on the semialgebraic approach to polynomialoptimization, we employ convex sum-of-squares (SOS) relaxations solved bya sequence of semidefinite programs (SDP), to obtain increasingly tighter up-per bounds on total achievable utility for polynomial utilities. Surprisingly,in all our experiments, a very low-order and often a minimal-order relaxationyields not just a bound on attainable network utility, but the globally maxi-mized network utility. When the bound is exact, which can be proved usinga sufficient test, we can also recover a globally optimal rate allocation.

    Canonical Distributed Algorithm

    A reason that the assumption of a utility functions concavity is upheld inmany papers on NUM is that it leads to three highly desirable mathematicalproperties of the basic NUM:

  • 8/12/2019 Non Convex 1

    7/61

    142 Mung Chiang

    It is a convex optimization problem, therefore the global minimum can becomputed (at least in centralized algorithms) in worst-case polynomial-time complexity [4].

    Strong duality holds for (5.1) and its Lagrange dual problem. A zero du-

    ality gap enables a dual approach to solve (5.1). Minimization of a separable objective function over linear constraints can

    be conducted by distributed algorithms based on the dual approach.

    Indeed, the basic NUM (5.1) is such a nice optimization problem thatits theoretical and computational properties have been well studied since the1960s in the field of monotropic programming (e.g., as summarized in [54]).For network rate allocation problems, a dual-decomposition-based distributedalgorithm has been widely studied (e.g., in [37, 45]), and is summarized below.

    Zero duality gap for (5.1) states that solving the Lagrange dual problem isequivalent to solving the primal problem (5.1). The Lagrange dual problemis readily derived. We first form the Lagrangian of (5.1):

    L(x, ) =Xs

    Us(xs) +Xl

    lcl X

    sS(l)

    xs ,

    wherel 0 is the Lagrange multiplier (can be interpreted as the link con-gestion price) associated with the linear flow constraint on link l. Additivityof total utility and linearity of flow constraints lead to a Lagrangian dualdecomposition into individual source terms:

    L(x, ) =Xs

    Us(xs)

    XlL(s)

    l

    xs

    +

    Xl

    cll

    = Xs

    Ls(xs, s) +X

    l

    cll,

    where s =P

    lL(s) l. For each source s, Ls(xs, s) = Us(xs)

    sxs onlydepends on local xs and the link prices l on those links used by sources.

    The Lagrange dual function g() is defined as the maximizedL(x, ) overx. This net utility maximization obviously can be conducted distributivelyby each source, as long as the aggregate link price s =

    PlL(s) lis available

    to source s, where source s maximizes a strictly concave function Ls(xs, s)

    over xs for a given s:

    xs(s) = argmax [Us(xs)

    sxs] , s. (5.2)

    The Lagrange dual problem is

    minimize g() =L(x(), )subject to 0,

    (5.3)

  • 8/12/2019 Non Convex 1

    8/61

  • 8/12/2019 Non Convex 1

    9/61

    144 Mung Chiang

    0 2 4 6 8 10 120

    0.5

    1

    1.5

    2

    2.5

    3

    x

    U(x

    )

    Fig. 5.2 Some examples of utility functions Us(xs): it can be concave or sigmoidal asshown in the graph, or any general nonconcave function. If the bottleneck link capacityused by the source is small enough, that is, if the dotted vertical line is pushed to the left,a sigmoidal utility function effectively becomes a convex utility function.

    rithm still converges to the globally optimal solution. However, these condi-tions may not hold in many cases. These two approaches illustrate the choicebetween admission control and capacity planning to deal with nonconvexity(see also the discussion in [36]). But neither approach provides a theoreticallypolynomial-time and practically efficient algorithm (distributed or central-ized) for nonconcave utility maximization.

    In [19], using a family of convex semidefinite programming (SDP) relax-ations based on the sum-of-squares (SOS) relaxation and the positivstellen-satz theorem in real algebraic geometry, we apply a centralized computationalmethod to bound the total network utility in polynomial time. A surprising

    result is that for all the examples we have tried, wherever we could verifythe result, the tightest possible bound (i.e., the globally optimal solution)of NUM with nonconcave utilities is computed with a very low-order relax-ation. This efficient numerical method for offline analysis also provides thebenchmark for distributed heuristics.

    These three different approaches: proposing distributed but suboptimalheuristics (for sigmoidal utilities) in [41], determining optimality conditionsfor the canonical distributed algorithm to converge globally (for all nonlinearutilities) in [29], and proposing an efficient but centralized method to computethe global optimum (for a wide class of utilities that can be transformed intopolynomial utilities) in [19] (and this section), are complementary in thestudy of distributed rate allocation by nonconcave NUM.

  • 8/12/2019 Non Convex 1

    10/61

    5 Nonconvex Optimization for Communication Networks 145

    5.2.2 Global Maximization of Nonconcave NetworkUtility

    Sum-of-Squares Method

    We would like to bound the maximum network utility by in polynomialtime and search for a tight bound. Had there been no link capacity con-straints, maximizing a polynomial is already an NP-hard problem, but canbe relaxed into an SDP [58]. This is because testing if the following bound-ing inequality holds p(x), where p(x) is a polynomial of degree d in nvariables, is equivalent to testing the positivity of p(x), which can berelaxed into testing ifp(x) can be written as a sum of squares (SOS):

    p(x) =Pr

    i=1 qi(x)2 for some polynomials qi, where the degree of qi is less

    than or equal to d/2. This is referred to as the SOS relaxation. If a poly-nomial can be written as a sum of squares, it must be nonnegative, but notvice versa. Conditions under which this relaxation is tight have been studiedsince Hilbert. Determining if a sum of squares decomposition exists can beformulated as an SDP feasibility problem, thus polynomial-time solvable.

    Constrained nonconcave NUM can be relaxed by a generalization of theLagrange duality theory, which involves nonlinear combinations of the con-straints instead of linear combinations in the standard duality theory. Thekey result is the positivstellensatz, due to Stengle [62], in real algebraic geom-etry, which states that for a system of polynomial inequalities, either thereexists a solution in Rn or there exists a polynomial which is a certificatethat no solution exists. This infeasibility certificate has recently been shownto be also computable by an SDP of sufficient size [51, 50], a process thatis referred to as the sum-of-squares method and automated by the softwareSOSTOOLS [52] initiated by Parrilo in 2000. For a complete theory and manyapplications of SOS methods, see [51] and references therein.

    Furthermore, the bound itself can become an optimization variable in theSDP and can be directly minimized. A nested family of SDP relaxations, eachindexed by the degree of the certificate polynomial, is guaranteed to producethe exact global maximum. Of course, given the problem is NP-hard, it isnot surprising that the worst-case degree of certificate (thus the number ofSDP relaxations needed) is exponential in the number of variables. What isinteresting is the observation that in applying SOSTOOLS to nonconcaveutility maximization, a very low-order, often the minimum-order relaxationalready produces the globally optimal solution.

    Application of SOS Method to Nonconcave NUM

    Using sum-of-squares and the positivstellensatz, we set up the following prob-lem whose objective value converges to the optimal value of problem (5.1),

  • 8/12/2019 Non Convex 1

    11/61

    146 Mung Chiang

    where {Ui} are now general polynomials, as the degree of the polynomialsinvolved is increased.

    minimize subject to

    Ps Us(xs)

    Pl l(x)(cl

    PsS(l) xs)

    P

    j,k jk(x)(cj P

    sS(j) xs)(ck P

    sS(k) xs)

    . . . 12...n(x)(c1 P

    sS(1) xs) . . . (cn P

    sS(n) xs)

    is SOS,l(x), jk(x), . . . , 12...n(x) are SOS.

    (5.5)

    The optimization variables are and all of the coefficients in polynomialsl(x),jk(x),. . .,12...n(x). Note thatx is not an optimization variable; theconstraints hold for all x, therefore imposing constraints on the coefficients.This formulation uses Schmudgens representation of positive polynomialsover compact sets [56].

    Let D be the degree of the expression in the first constraint in (5.5). We

    refer to problem (5.5) as the SOS relaxation of order D for the constrainedNUM. For a fixedD, the problem can be solved via SDP. As D is increased,the expression includes more terms, the corresponding SDP becomes larger,and the relaxation gives tighter bounds. An important property of this nestedfamily of relaxations is guaranteed convergence of the bound to the globalmaximum.

    Regarding the choice of degree D for each level of relaxation, clearly apolynomial of odd degree cannot be SOS, so we need to consider only thecases where the expression has even degree. Therefore, the degree of the firstnontrivial relaxation is the largest even number greater than or equal todegree

    Ps Us(xs), and the degree is increased by 2 for the next level.

    A key question now becomes: how do we find out, after solving an SOSrelaxation, if the bound happens to be exact? Fortunately, there is a suffi-

    cient test that can reveal this, using the properties of the SDP and its dualsolution. In [31, 39], a parallel set of relaxations, equivalent to the SOS ones,is developed in the dual framework. The dual of checking the nonnegativityof a polynomial over a semialgebraic set turns out to be finding a sequence ofmoments that represent a probability measure with support in that set. Tobe a valid set of moments, the sequence should form a positive semidefinitemoment matrix. Then, each level of relaxation fixes the size of this matrix(i.e., considers moments up a certain order) and therefore solves an SDP.This is equivalent to fixing the order of the polynomials appearing in SOSrelaxations. The sufficient rank test checks a rank condition on this momentmatrix and recovers (one or several) optimal x, as discussed in [31].

    In summary, we have the following algorithm for centralized computa-tion of a globally optimal rate allocation to nonconcave utility maximization,

    where the utility functions can be written as or converted into polynomials.

  • 8/12/2019 Non Convex 1

    12/61

    5 Nonconvex Optimization for Communication Networks 147

    Algorithm 1. Sum-of-squares for nonconcave utility maximization.

    1. Formulate the relaxed problem (5.5) for a given degree D.2. Use SDP to solve the Dth order relaxation, which can be conducted

    using SOSTOOLS [52].

    3. If the resulting dual SDP solution satisfies the sufficient rank condition,theDth-order optimizer(D) is the globally optimal network utility, and acorrespondingx can be obtained.1

    4. IncreaseDtoD+2, that is, the next higher-order relaxation, and repeat.

    In the following section, we give examples of the application of SOS re-laxation to the nonconcave NUM. We also apply the above sufficient test tocheck if the bound is exact, and if so, we recover the optimum rate allocationx that achieve this tightest bound.

    5.2.3 Numerical Examples and Sigmoidal Utilities

    Polynomial Utility Examples

    First, consider quadratic utilities (i.e., Us(xs) = x2s) as a simple case to start

    with (this can be useful, for example, when the bottleneck link capacity limitssources to their convex region of a sigmoidal utility). We present examplesthat are typical, in our experience, of the performance of the relaxations.

    x1

    x2 x3

    c1 c2

    Fig. 5.3 Network topology for Example 5.1.

    Example 5.1. A small illustrative example. Consider the simple 2-link, 3-user network shown in Figure 5.3, with c = [1, 2]. The optimization problemis

    maximizeP

    s x2s

    subject tox1+ x2 1x1+ x3 2x1, x2, x3 0.

    (5.6)

    1 Otherwise, (D) may still be the globally optimal network utility but is only provablyan upper bound.

  • 8/12/2019 Non Convex 1

    13/61

    148 Mung Chiang

    The first level relaxation with D = 2 is

    minimize subject to (x2

    1

    + x2

    2

    + x2

    3

    ) 1(x1 x2+ 1) 2(x1x3+ 2) 3x1 4x2 5x3 6(x1 x2+ 1)(x1 x3+ 2) 7x1(x1 x2+ 1) 8x2(x1x2+ 1) 9x3(x1 x2+ 1) 10x1(x1 x3+ 2)11x2(x1 x3+ 2) 12x3(x1 x3+ 2)13x1x2 14x1x3 15x2x3 is SOS,i 0, i= 1, . . . , 15.

    (5.7)

    The first constraint above can be written as xTQx for x= [1, x1, x2, x3]T

    and an appropriate Q. For example, the (1,1) entry which is the constantterm reads 1 22 26, the (2,1) entry, coefficient ofx1, reads 1+2 3+ 36 7 210, and so on. The expression is SOS if and only ifQ 0. The optimal is 5, which is achieved by, for example, 1 = 1,2 = 2,

    3 = 1, 8 = 1, 10 = 1, 12 = 1, 13 = 1, 14 = 2 and the rest of the iequal to zero. Using the sufficient test (or, in this example, by inspection) wefind the optimal rates x0 = [0, 1, 2].

    In this example, many of the i could be chosen to be zero. This meansnot all product terms appearing in (5.7) are needed in constructing the SOSpolynomial. Such information is valuable from the decentralization point ofview, and can help determine to what extent our bound can be calculated ina distributed manner. This is a challenging topic for future work.

    c1 c2

    c3

    c4

    c5

    c6

    c7

    Fig. 5.4 Network topology for Example 5.2.

    Example 5.2. Larger tree topology. As a larger example, consider the net-work shown in Figure 5.4 with seven links. There are nine users, with thefollowing routing table that lists the links on each users path.

    x1 x2 x3 x4 x5 x6 x7 x8 x91,2 1,2,4 2,3 4,5 2,4 6,5,7 5,6 7 5

  • 8/12/2019 Non Convex 1

    14/61

    5 Nonconvex Optimization for Communication Networks 149

    Forc = [5, 10, 4, 3, 7, 3, 5], we obtain the bound = 116 with D = 2,which turns out to be globally optimal, and the globally optimal rate vectorcan be recovered:x0 = [5, 0, 4, 0, 1, 0, 0, 5, 7]. In this example, exhaustivesearch is too computationally intensive, and the sufficient condition test plays

    an important role in proving the bound is exact and in recovering x0.

    c1

    c2

    c3

    c4

    c5

    c6

    Fig. 5.5 Network topology for Example 5.3.

    Example 5.3. Large m-hop ring topology. Consider a ring network with nnodes,n users, andn links where each users flow starts from a node and goesclockwise through the next m links, as shown in Figure 5.5 for n = 6,m = 2.As a large example, with n = 25, m = 2, and capacities chosen randomlyfor a uniform distribution on [0, 10], using relaxation of order D = 2 weobtain the exact bound = 321.11 and recover an optimal rate allocation.Forn = 30,m = 2, and capacities randomly chosen from [0, 15], it turns outthatD = 2 relaxation yields the exact bound 816 .95 and a globally optimal

    rate allocation.

    Sigmoidal Utility Examples

    Now consider sigmoidal utilities in a standard form:

    Us(xs) = 1

    1 + e(asxs+bs),

    where{as, bs} are constant integers. Even though these sigmoidal functionsare not polynomials, we show the problem can be cast as one with polynomialcost and constraints, with a change of variables.

    Example 5.4. Sigmoidal utility. Consider the simple 2-link, 3-user exampleshown in Figure 5.3 for as = 1 and bs= 5.

    The NUM problem is to

  • 8/12/2019 Non Convex 1

    15/61

    150 Mung Chiang

    maximizeP

    s1

    1+e(xs5)

    subject tox1+ x2 c1x1+ x3 c2x 0.

    (5.8)

    Letys = 1/

    1 + e(xs5)

    , then xs = log((1/ys) 1) + 5. Substitutingfor x1, x2 in the first constraint, arranging terms and taking exponentials,then multiplying the sides by y1y2 (note that y1, y2 > 0), we get

    (1 y1)(1 y2) e(10c1)y1y2,

    which is polynomial in the new variables y. This applies to all capac-ity constraints, and the nonnegativity constraints for xs translate to ys 1/

    1 + e5

    . Therefore the whole problem can be written in polynomial form,and SOS methods apply. This transformation renders the problem polynomialfor general sigmoidal utility functions, with any as andbs.

    We present some numerical results, using a small illustrative example. Here

    SOS relaxations of order 4 (D = 4) were used. For c1 = 4, c2 = 8, we fi

    nd = 1.228, which turns out to be a global optimum, with x0 = [0, 4, 8]as the optimal rate vector. For c1 = 9, c2 = 10, we find = 1.982 andx0 = [0, 9, 10]. Now place a weight of 2 on y1, and the other ys have weightone; we obtain = 1.982 and x0 = [9, 0, 1].

    In general, ifas6= 1 for somes, however, the degree of the polynomials inthe transformed problem may be very high. If we write the general problemas

    maximizeP

    s1

    1+e(asxs+bs)

    subject toP

    sS(l) xs cl, l,

    x 0,

    (5.9)

    each capacity constraint after transformation will be

    Qs(1 ys)rlsk6=sak

    exp(Qs as(cl+

    Ps rls/asbs))

    Qs y

    rlsQ

    k6=saks ,

    whererls= 1 ifl L(s) and equals 0 otherwise. Because the product of theas appears in the exponents, as > 1 significantly increases the degree of thepolynomials appearing in the problem and hence the dimension of the SDPin the SOS method.

    It is therefore also useful to consider alternative representations of sig-moidal functions such as the following rational function:

    Us(xs) = xnsa + xns

    ,

    where the inflection point is x0 = ((a(n 1)) / (n + 1))1/n and the slope at

    the inflection point is Us(x0) = ((n 1) /4n) ((n + 1) / (a(n 1)))1/n. Let

  • 8/12/2019 Non Convex 1

    16/61

    5 Nonconvex Optimization for Communication Networks 151

    ys = Us(xs); the NUM problem in this case is equivalent to

    maximizeP

    s yssubject toxns ysx

    ns ays= 0P

    sS(l) xs cl, lx 0

    (5.10)

    which again can be accommodated in the SOS method and be solved byAlgorithm 1.

    The benefit of this choice of utility function is that the largest degree ofthe polynomials in the problem is n+ 1, therefore growing linearly with n.The disadvantage compared to the exponential form for sigmoidal functionsis that the location of the inflection point and the slope at that point cannotbe set independently.

    5.2.4 Alternative Representations for Convex

    Relaxations to Nonconcave NUM

    The SOS relaxation we used in the last two sections is based on Schm udgensrepresentation for positive polynomials over compact sets described by otherpolynomials. We now briefly discuss two other representations of relevanceto the NUM, that are interesting from both theoretical (e.g., interpretation)and computational points of view.

    LP Relaxation

    Exploiting linearity of the constraints in NUM and with the additional as-

    sumption of nonempty interior for the feasible set (which holds for NUM),we can use Handelmans representation [30] and refine the positivstellen-satz condition to obtain the following convex relaxation of nonconcave NUMproblem.

    maximizesubject to

    P

    s Us(xs) =XNL

    LYl=1

    (cl P

    sS(l) xs)l , x

    0, ,

    (5.11)

    where the optimization variables are and, and denotes an ordered set

    of integers {l}.Fixing D whereP

    l l D, and equating the coefficients on the twosides of the equality in (5.11), yields a linear program (LP). There are no

  • 8/12/2019 Non Convex 1

    17/61

    152 Mung Chiang

    SOS terms, therefore no semidefiniteness conditions. As before, increasingthe degreeD gives higher-order relaxations and a tighter bound.

    We provide a pricing interpretation for problem (5.11). First, normalizeeach capacity constraint as 1 ul(x) 0, where ul(x) =PsS(l) xs/cl. Wecan interpret ul(x) as link usage, or the probability that link l is used atany given point in time. Then, in (5.11), we have terms linear in u such asl(1ul(x)), in which l has a similar interpretation as in concave NUM, asthe price of using linkl. We also have product terms such asjk(1uj(x))(1uk(x)), wherejkuj(x)uk(x) indicates the probability of simultaneous usageof links j and k, for links whose usage probabilities are independent (e.g.,they do not share any flows). Products of more terms can be interpretedsimilarly.

    Although the above price interpretation is not complete and does not jus-tify all the terms appearing in (5.11) (e.g., powers of the constraints, productterms for links with shared flows), it does provide some useful intuition: thisrelaxation results in a pricing scheme that provides better incentives for theusers to observe the constraints, by giving an additional reward (because the

    corresponding term adds positively to the utility) for simultaneously keepingtwo links free. Such incentive helps tighten the upper bound and eventuallyachieve a feasible (and optimal) allocation.

    This relaxation is computationally attractive because we need to solve anLPs instead of the previous SDPs at each level. However, significantly morelevels may be required [40].

    Relaxation with No Product Terms

    Putinar [53] showed that a polynomial positive over a compact set2 can berepresented as an SOS-combination of the constraints. This yields the follow-ing convex relaxation for nonconcave NUM problem.

    maximize subject to

    P

    s Us(xs) =PL

    l=1 l(x)(cl P

    sS(l) xs), x

    (x) is SOS,

    (5.12)

    where the optimization variables are the coefficients in l(x). Similar to theSOS relaxation (5.5), fixing the order D of the expression in (5.12) results inan SDP. This relaxation has the nice property that no product terms appear:the relaxation becomes exact with a high enough D without the need ofproduct terms. However, this degree might be much higher than what theprevious SOS method requires.

    2 With an extra assumption that always holds for linear constraints as in NUM problems.

  • 8/12/2019 Non Convex 1

    18/61

    5 Nonconvex Optimization for Communication Networks 153

    5.2.5 Concluding Remarks and Future Directions

    We consider the NUM problem in the presence of inelastic flows, that is, flowswith nonconcave utilities. Despite its practical importance, this problem has

    not been studied widely, mainly due to the fact it is a nonconvex problem.There has been no effective mechanism, centralized or distributed, to com-pute the globally optimal rate allocation for nonconcave utility maximizationproblems in networks. This limitation has made performance assessment anddesign of networks that include inelastic flows very difficult.

    In one of the recent works on this topic [19], we employed convex SOS re-laxations, solved by a sequence of SDPs, to obtain high-quality, increasinglytighter upper bounds on total achievable utility. In practice, the performanceof our SOSTOOLS-based algorithm was surprisingly good, and bounds ob-tained using a polynomial-time (and indeed a low-order and often minimal-order) relaxation were found to be exact, achieving the global optimum ofnonconcave NUM problems. Furthermore, a dual-based sufficient test, if suc-cessful, detects the exactness of the bound, in which case the optimal rateallocation can also be recovered. This performance of the proposed algorithmbrings up a fundamental question on whether there is any particular propertyor structure in nonconcave NUM that makes it especially suitable for SOSrelaxations.

    We further examined the use of two more specialized polynomial repre-sentations, one that uses products of constraints with constant multipliers,resulting in LP relaxations; and at the other end of spectrum, one that usesa linear combination of constraints with SOS multipliers. We expect these re-laxations to give higher-order certificates, thus their potential computationalbenefits need to be examined further. We also show they admit economicsinterpretations (e.g., prices, incentives) that provide some insight on how theSOS relaxations work in the framework of link congestion pricing for the

    simultaneous usage of multiple links.An important research issue to be further investigated is decentralizationmethods for rate allocation among sources with nonconcave utilities. Theproposed algorithm here is not easy to decentralize, given the products of theconstraints or polynomial multipliers that destroy the separable structure ofthe problem. However, when relaxations become exact, the sparsity patternof the coefficients can provide information about partially decentralized com-putation of optimal rates. For example, if after solving the NUM offline, weobtain an exact bound, then if the coefficient of the cross-termxixj turns outto be zero, it means users i and j do not need to communicate to each otherto find their optimal rates. An interesting next step in this area of research isto investigate a distributed version of the proposed algorithm through limitedmessage passing among clusters of network nodes and links.

  • 8/12/2019 Non Convex 1

    19/61

  • 8/12/2019 Non Convex 1

    20/61

    5 Nonconvex Optimization for Communication Networks 155

    with no efficient and global solution methods. In this case, we present aheuristic that is provably convergent and empirically almost always computesthe globally optimal power allocation by solving a sequence of GPs throughthe approach of successive convex approximations.

    The GP approach reveals the hidden convexity structure, which impliesefficient solution methods and the global optimality of any local optimum inpower control problems with nonlinear objective functions. It clearly differ-entiates the tractable formulations in a high-SIR regime from the intractableones in a low-SIR regime. Power control by GP is applicable to formulationsin both cellular networks with single-hop transmission between mobile usersand base stations, and ad hoc networks with multihop transmission amongthe nodes, as illustrated through several numerical examples in this section.Traditionally, GP is solved by centralized computation through the highlyefficient interior point methods. In this section we present a new result onhow GP can be solved distributively with message passing, which has inde-pendent value to general maximization of coupled objective, and applies it topower control problems with a further reduction of message-passing overhead

    by leveraging the specific structures of power control problems.More generally, the technique of nonlinear change of variables, including

    the log change of variables, to reveal hidden convexity in optimization for-mulations has recently become quite popular in the communication networkresearch community.

    5.3.2 Geometric Programming

    GP is a class of nonlinear, nonconvex optimization problems with many usefultheoretical and computational properties. It was invented in 1967 by Duffin,Peterson, and Zener [17], and much of the development by the early 1980s was

    summarized in [1]. Because a GP can be turned into a convex optimizationproblem, a local optimum is also a global optimum, the Lagrange duality gapis zero under mild conditions, and a global optimum can be computed veryefficiently. Numerical efficiency holds both in theory and in practice: interiorpoint methods applied to GP have provably polynomial-time complexity [48],and are very fast in practice with high-quality software downloadable fromthe Internet (e.g., the MOSEK package). Convexity and duality propertiesof GP are well understood, and large-scale, robust numerical solvers for GPare available. Furthermore, special structures in GP and its Lagrange dualproblem lead to distributed algorithms, physical interpretations, and compu-tational acceleration beyond the generic results for convex optimization. Adetailed tutorial of GP and comprehensive survey of its recent applicationsto communication systems and to circuit design can be found in [11] and [3],respectively. This section contains a brief introduction of GP terminology.

  • 8/12/2019 Non Convex 1

    21/61

    156 Mung Chiang

    There are two equivalent forms of GP: standard form and convex form.Thefirst is a constrained optimization of a type of function called posynomial,and the second form is obtained from the first through a logarithmic changeof variable.

    We first define a monomial as a function f :Rn++ R:

    f(x) = dxa(1)

    1 xa(2)

    2 . . . xa(n)

    n ,

    where the multiplicative constantd 0 and the exponential constantsa(j) R, j = 1, 2, . . . , n. A sum of monomials, indexed by k below, is called aposynomial:

    f(x) =KXk=1

    dkxa(1)k

    1 xa(2)k

    2 . . . xa(n)kn ,

    wheredk 0, k = 1, 2, . . . , K , and a(j)k R, j = 1, 2, . . . , n, k = 1, 2, . . . , K .

    For example, 2x1 x0.52 + 3x1x

    1003 is a posynomial in x, x1 x2 is not a

    posynomial, andx1/x2 is a monomial, thus also a posynomial.

    Minimizing a posynomial subject to posynomial upper bound inequalityconstraints and monomial equality constraints is called GP in standard form:

    minimize f0(x)subject tofi(x) 1, i= 1, 2, . . . , m ,

    hl(x) = 1, l= 1, 2, . . . , M ,(5.13)

    where fi, i= 0, 1, . . . , m, are posynomials: fi(x) =PKi

    k=1 dikxa(1)ik

    1 xa(2)ik

    2 . . . xa(n)ikn ,

    andhl, l = 1, 2, . . . , M , are monomials: hl(x) = dlxa(1)l

    1 xa(2)l

    2 . . . xa(n)ln .

    GP in standard form is not a convex optimization problem, because posy-nomials are not convex functions. However, with a logarithmic change of thevariables and multiplicative constants: yi = log xi, bik = log dik, bl = log dl,

    and a logarithmic change of the functions values, we can turn it into thefollowing equivalent problem in y.

    minimize p0(y) = logPK0

    k=1exp(aT0ky + b0k)

    subject topi(y) = logPKi

    k=1exp(aTiky + bik) 0, i= 1, 2, . . . , m ,

    ql(y) = aTl y + bl= 0, l= 1, 2, . . . , M .

    (5.14)

    This is referred to as GP in convex form, which is a convex optimizationproblem because it can be verified that the log-sum-exp function is convex[4].

    In summary, GP is a nonlinear, nonconvex optimization problem that canbe transformed into a nonlinear convex problem. GP in standard form canbe used to formulate network resource allocation problems with nonlinear

    objectives under nonlinear QoS constraints. The basic idea is that resourcesare often allocated proportional to some parameters, and when resource allo-

  • 8/12/2019 Non Convex 1

    22/61

    5 Nonconvex Optimization for Communication Networks 157

    05

    10

    0

    5

    10

    0

    20

    40

    60

    80

    100

    120

    Y

    X

    Function

    02

    4

    02

    40.5

    1

    1.5

    2

    2.5

    3

    3.5

    4

    4.5

    5

    AB

    Function

    Fig. 5.6 A bivariate posynomial before (left graph) and after (right graph) the log trans-formation. A nonconvex function is turned into a convex one.

    cations are optimized over these parameters, we are maximizing an invertedposynomial subject to lower bounds on other inverted posynomials, whichare equivalent to GP in standard form.

    SP/GP, SOS/SDP

    Note that, although the posynomial seems to be a nonconvex function, it be-comes a convex function after the log transformation, as shown in an examplein Figure 5.6. Compared to the (constrained or unconstrained) minimizationof a polynomial, the minimization of a posynomial in GP relaxes the integerconstraint on the exponential constants but imposes a positivity constraint onthe multiplicative constants and variables. There is a sharp contrast betweenthese two problems: polynomial minimization is NP-hard, but GP can beturned into convex optimization with provably polynomial-time algorithmsfor a global optimum.

    In an extension of GP called signomial programming discussed later in thissection, the restriction of nonnegative multiplicative constants is removed.This results in a general class of nonlinear and truly nonconvex problemsthat is simultaneously a generalization of GP and polynomial minimization

    over the positive quadrant, as summarized in the comparison Table 5.2.

  • 8/12/2019 Non Convex 1

    23/61

    158 Mung Chiang

    Table 5.2 Comparison of GP, constrained polynomial minimization over the positivequadrant (PMoP), and signomial programming (SP). All three types of problems minimizea sum of monomials subject to upper bound inequality constraints on sums of monomials,

    but have different definitions of monomial:c

    Qjx

    a(j)

    j , as shown in the table. GP is knownto be polynomial-time solvable, but PMoP and SP are not.

    GP PMoP SP

    c R+ R R

    a(j) R Z+ R

    xj R++ R++ R++

    The objective function of signomial programming can be formulated asminimizing a ratio between two posynomials, which is not a posynomial (be-cause posynomials are closed under positive multiplication and addition butnot division). As shown in Figure 5.7, a ratio between two posynomials is anonconvex function both before and after the log transformation. Although itdoes not seem likely that signomial programming can be turned into a convexoptimization problem, there are heuristics to solve it through a sequence ofGP relaxations. However, due to the absence of algebraic structures foundin polynomials, such methods for signomial programming currently lack atheoretical foundation of convergence to global optimality. This is in contrastto the sum-of-squares method [51], which uses a nested family of SDP relax-

    0

    5

    10

    0

    5

    10

    20

    0

    20

    40

    60

    XY

    Function

    0

    1

    2

    3

    0

    1

    2

    3

    2

    2.5

    3

    3.5

    AB

    Function

    Fig. 5.7 Ratio between two bivariate posynomials before (left graph) and after (rightgraph) the log transformation. It is a nonconvex function in both cases.

  • 8/12/2019 Non Convex 1

    24/61

    5 Nonconvex Optimization for Communication Networks 159

    ations to solve constrained polynomial minimization problems as explainedin the last section.

    5.3.3 Power Control by Geometric Programming:

    Convex Case

    Various schemes for power control, centralized or distributed, have been ex-tensively studied since the 1990s based on different transmission models andapplication needs (e.g., in [2, 26, 47, 55, 63, 72]). This section summarizesthe new approach of formulating power control problems through GP. Thekey advantage is that globally optimal power allocations can be efficientlycomputed for a variety of nonlinear systemwide objectives and user QoSconstraints, even when these nonlinear problems appear to be nonconvexoptimization.

    Basic Model

    Consider a wireless (cellular or multihop) network with n logical transmit-ter/receiver pairs. Transmit powers are denoted as P1, . . . , P n. In the cellularuplink case, all logical receivers may reside in the same physical receiver,that is, the base station. In the multihop case, because the transmission en-vironment can be different on the links comprising an end-to-end path, powercontrol schemes must consider each link along a flows path.

    Under Rayleigh fading, the power received from transmitter j at receiveri is given by GijFijPj where Gij 0 represents the path gain (it may alsoencompass antenna gain and coding gain) that is often modeled as propor-tional to d

    ij

    , where dij denotes distance, is the power fall-off factor, andFijmodel Rayleigh fading and are independent and exponentially distributedwith unit mean. The distribution of the received power from transmitter jat receiver i is then exponential with mean valueE [GijFijPj ] = GijPj . TheSIR for the receiver on logical link i is:

    SIRi= PiGiiFiiPNj6=i PjGijFij+ ni

    (5.15)

    whereni is the noise power for receiver i.The constellation size Mused by a link can be closely approximated for

    MQAM modulations as follows.M= 1+(1/ (ln(2BER)))SIR, where BERis the bit error rate and 1, 2 are constants that depend on the modulationtype. DefiningK= 1/ (ln(2BER)) leads to an expression of the data rateRion theith link as a function of the SIR:Ri= (1/T)log2(1+ KSIRi), whichcan be approximated as

  • 8/12/2019 Non Convex 1

    25/61

    160 Mung Chiang

    Ri = 1

    T log2(KSIRi) (5.16)

    when KSIR is much larger than 1. This approximation is reasonable eitherwhen the signal level is much higher than the interference level or, in CDMA

    systems, when the spreading gain is large. For notational simplicity in therest of this section, we redefineGiiasKtimes the originalGii, thus absorbingconstant Kinto the definition of SIR.

    The aggregate data rate for the system can then be written as

    Rsystem=Xi

    Ri = 1

    T log2

    "Yi

    SIRi

    #.

    So in the high SIR regime, aggregate data rate maximization is equivalentto maximizing a product of SIR. The system throughput is the aggregatedata rate supportable by the system given a set of users with specified QoSrequirements.

    Outage probability is another important QoS parameter for reliable com-munication in wireless networks. A channel outage is declared and packetslost when the received SIR falls below a given threshold SIRth, often com-puted from the BER requirement. Most systems are interference-dominatedand the thermal noise is relatively small, thus the ith link outage probabilityis

    Po,i = Prob{SIRi SIRth}

    =Prob{GiiFiiPi SIRthXj6=i

    GijFijPj}.

    The outage probability can be expressed as [38]

    Po,i= 1Yj6=i

    1

    1 + SIRthGijPjGiiPi

    ,

    which means that the upper bound Po,i Po,i,maxcan be written as an upperbound on a posynomial in P:

    Yj6=i

    1 +

    SIRthGijPjGiiPi

    1

    1 Po,i,max. (5.17)

    Cellular Wireless Networks

    We first present how GP-based power control applies to cellular wireless

    networks with one-hop transmission from N users to a base station. Theseresults extend the scope of power control by the classical solution in CDMA

  • 8/12/2019 Non Convex 1

    26/61

    5 Nonconvex Optimization for Communication Networks 161

    systems that equalizes SIRs, and those by the iterative algorithms (e.g., in[2, 26, 47]) that minimize total power (a linear objective function) subject toSIR constraints.

    We start the discussion on the suite of power control problem formula-

    tions with a simple objective function and basic constraints. The followingconstrained problem of maximizing the SIR of a particular user i is a GP.

    maximize Ri(P)subject toRi(P) Ri,min, i,

    Pi1Gi1 = Pi2Gi2,0 Pi Pi,max, i.

    The first constraint, equivalent to SIRi SIRi,min, sets a floor on the SIRof other users and protects these users from user i increasing her transmitpower excessively. The second constraint reflects the classical power controlcriterion in solving the nearfar problem in CDMA systems: the expectedreceived power from one transmitter i1 must equal that from another i2. The

    third constraint is regulatory or system limitations on transmit powers. Allconstraints can be verified to be inequality upper bounds on posynomials intransmit power vector P.

    Alternatively, we can use GP to maximize the minimum rate among allusers. The maxmin fairness objective

    maximizeP mini{Ri}

    can be accommodated in GP-based power control because it can be turnedinto equivalently maximizing an auxiliary variable t such that SIRi(P) exp(t), i, which has a posynomial objective and constraints in ( P, t).

    Example 5.5. A small illustrative example. A simple system comprised offive users is used for a numerical example. The five users are spaced at dis-tancesd of 1, 5, 10, 15, and 20 units from the base station. The power fall-offfactor= 4. Each user has a maximum power constraint ofPmax= 0.5 mW.The noise power is 0.5 W for all users. The SIR of all users, other thanthe user we are optimizing for, must be greater than a common thresholdSIR level . In different experiments, is varied to observe the effect on theoptimized users SIR. This is done independently for the near user at d = 1,a medium distance user at d = 15, and the far user at d = 20. The resultsare plotted in Figure 5.8.

    Several interesting effects are illustrated. First, when the required thresh-old SIR in the constraints is sufficiently high, there is no feasible power con-trol solution. At moderate threshold SIR, as is decreased, the optimizedSIR initially increases rapidly. This is because it is allowed to increase its

    own power by the sum of the power reductions in the four other users, andthe noise is relatively insignificant. At low threshold SIR, the noise becomesmore significant and the power tradeoff from the other users less significant,

  • 8/12/2019 Non Convex 1

    27/61

    162 Mung Chiang

    5 0 5 1020

    15

    10

    5

    0

    5

    10

    15

    20

    Threshold SIR (dB)

    OptimizedSIR(dB)

    Optimized SIR vs. Threshold SIR

    nearmediumfar

    Fig. 5.8 Constrained optimization of power control in a cellular network (Example 5.5).

    so the curve starts to bend over. Eventually, the optimized user reaches itsupper bound on power and cannot utilize the excess power allowed by thelower threshold SIR for other users. This is exhibited by the transition froma sharp bend in the curve to a much shallower sloped curve.

    We now proceed to show that GP can also be applied to the problemformulations with an overall system objective of total system throughput,under both user data rate constraints and outage probability constraints.

    The following constrained problem of maximizing system throughput is aGP.

    maximize Rsystem(P)

    subject toRi(P)

    Ri,min,

    i,Po,i(P) Po,i,max, i,0 Pi Pi,max, i

    (5.18)

    where the optimization variables are the transmit powers P. The objectiveis equivalent to minimizing the posynomial

    QiISRi, where ISR is 1/SIR.

    Each ISR is a posynomial in P and the product of posynomials is again aposynomial. The first constraint is from the data rate demand Ri,minby eachuser. The second constraint represents the outage probability upper boundsPo,i,max. These inequality constraints put upper bounds on posynomials ofP, as can be readily verified through (5.16) and (5.17). Thus (5.18) is indeeda GP, and efficiently solvable for global optimality.

    There are several obvious variations of problem (5.18) that can be solvedby GP; for example, we can lower bound Rsystemas a constraint and maximize

    Rifor a particular user i, or have a total powerP

    i Piconstraint or objectivefunction.

  • 8/12/2019 Non Convex 1

    28/61

    5 Nonconvex Optimization for Communication Networks 163

    Table 5.3 Suite of power control optimization solvable by GP

    Objective Function Constraints

    (A) Max Ri (a)Ri Ri,min(specific user) (rate constraint)

    (B) Max miniRi (b)Pi1Gi1= Pi2Gi2(worst-case user) (nearfar constraint)

    (C) MaxP

    iRi (c)P

    iRi Rsystem,min(total throughput) (sum rate constraint)

    (D) MaxP

    iwiRi (d)Po,i Po,i,max(weighted rate sum) (outage probability constraint)

    (E) MinP

    i Pi (e) 0 Pi Pi,max(total power) (power constraint)

    The objective function to be maximized can also be generalized to aweighted sum of data rates,

    Pi wiRi, where w 0 is a given weight vector.

    This is still a GP because maximizing

    Piwi logSIRi is equivalent to maxi-

    mizing logQiSIR

    wi

    i , which is in turn equivalent to minimizingQ

    iISRwi

    i . Now

    use auxiliary variables{ti}, and minimizeQ

    i twii over the original constraints

    in (5.18) plus the additional constraints ISRi ti for all i. This is readilyverified to be a GP in (x, t), and is equivalent to the original problem.

    Generalizing the above discussions and observing that high-SIR assump-tion is needed for GP formulation only when there are sums of log(1 + SIR)in the optimization problem, we have the following summary.

    Proposition 5.1. In the high-SIR regime, any combination of objectives(A)(E) and constraints(a)(e) in Table5.3 (pick any one of the objectivesand any subset of the constraints) is a power control optimization problemthat can be solved by GP, that is, can be transformed into a convex optimiza-

    tion with efficient algorithms to compute the globally optimal power vector.

    When objectives (C)(D) or constraints (c)(d) do not appear, the powercontrol optimization problem can be solved by GP in any SIR regime.

    In addition to efficient computation of the globally optimal power allo-cation with nonlinear objectives and constraints, GP can also be used foradmission control based on feasibility study described in [11], and for deter-mining which QoS constraint is a performance bottleneck, that is, met tightlyat the optimal power allocation.3

    3 This is because most GP solution algorithms solve both the primal GP and its Lagrangedual problem, and by the complementary slackness condition, a resource constraint is tightat optimal power allocation when the corresponding optimal dual variable is nonzero.

  • 8/12/2019 Non Convex 1

    29/61

    164 Mung Chiang

    Extensions

    In wireless multihop networks, system throughput may be measured either byend-to-end transport layer utilities or by link layer aggregate throughput. GP

    application to the first approach has appeared in [10], and those to the secondapproach in [11]. Furthermore, delay and buffer overflow properties can alsobe accommodated in the constraints or objective function of GP-based powercontrol.

    5.3.4 Power Control by Geometric Programming:

    Nonconvex Case

    If we maximize the total throughput Rsystemin the medium to low SIR case(i.e., when SIR is not much larger than 0 dB), the approximation of log(1 +SIR) as log SIR does not hold. Unlike SIR, which is an inverted posynomial,

    1+SIRis not an inverted posynomial. Instead, 1/ (1 + SIR) is a ratio betweentwo posynomials:

    f(P)

    g(P) =

    Pj6=i GijPj+ niPjGijPj+ ni

    . (5.19)

    Minimizing, or upper bounding, a ratio between two posynomials be-longs to a truly nonconvex class of problems known as complementary GP[1, 11] that is in general an NP-hard problem. An equivalent generalizationof GP is signomial programming [1, 11]: minimizing a signomial subject toupper bound inequality constraints on signomials, where a signomial s(x)is a sum of monomials, possibly with negative multiplicative coefficients:s(x) =

    PNi=1 cigi(x) where c R

    N andgi(x) are monomials.4

    Successive Convex Approximation Method

    Consider the following nonconvex problem,

    minimize f0(x)subject tofi(x) 1, i= 1, 2, . . . , m ,

    (5.20)

    wheref0 is convex without loss of generality,5 but thefi(x)s, iare noncon-

    vex. Because directly solving this problem is NP-hard, we want to solve it by

    4 An SP can always be converted into a complementary GP, because an inequality in SP,which can be written as fi1(x) fi2(x) 1, where fi1, fi2 are posynomials, is equivalentto an inequality fi1(x)/ (1 + fi2(x)) 1 in complementary GP.

    5 Iff0 is nonconvex, we can move the objective function to the constraint by introducingauxiliary scalar variable t and writing minimize t subject to the additional constraintf0(x) t 0.

  • 8/12/2019 Non Convex 1

    30/61

    5 Nonconvex Optimization for Communication Networks 165

    a series of approximations fi(x) fi(x),x, each of which can be optimallysolved in an easy way. It is known [46] that if the approximations satisfy thefollowing three properties, then the solutions of this series of approximationsconverge to a point satisfying the necessary optimality KarushKuhnTucker

    (KKT) conditions of the original problem.(1)fi(x) fi(x) for all x.(2)fi(x0) = fi(x0) where x0 is the optimal solution of the approximated

    problem in the previous iteration.(3) fi(x0) = fi(x0).

    The following algorithm describes the generic successive approximationapproach. Given a method to approximate fi(x) with fi(x) , i, around somepoint of interest x0, the following algorithm provides the output of a vectorthat satisfies the KKT conditions of the original problem.

    Algorithm 2. Successive approximation to a nonconvex problem.

    1. Choose an initial feasible pointx(0) and setk = 1.

    2. Form an approximated problem of (5.20) based on the previous pointx(k1).

    3. Solve the kth approximated problem to obtain x(k).4. Increment k and go to step 2 until convergence to a stationary point.

    Single condensation method. Complementary GPs involve upper boundson the ratio of posynomials as in (5.19); they can be turned into GPs byapproximating the denominator of the ratio of posynomials, g(x), with amonomial g(x), but leaving the numerator f(x) as a posynomial.

    The following basic result can be readily proved using the arithmetic-meangeometric-mean inequality.

    Lemma 5.1. Letg (x) =

    Pi ui(x) be a posynomial. Then

    g(x) g(x) =Yi

    ui(x)

    i

    i. (5.21)

    If, in addition,i = ui(x0)/g(x0), i,for anyfixed positivex0, theng(x0) =g(x0), andg(x) is the best local monomial approximation to g (x) nearx0 inthe sense offirst-order Taylor approximation.

    The above lemma easily leads to the following

    Proposition 5.2. The approximation of a ratio of posynomials f(x)/g(x)withf(x)/g(x) whereg(x) is the monomial approximation ofg(x) using thearithmetic-geometric mean approximation of Lemma 5.1 satisfies the threeconditions for the convergence of the successive approximation method.

    Double condensation method. Another choice of approximation is to makea double monomial approximation for both the denominator and numeratorin (5.19). However, in order to satisfy the three conditions for the convergence

  • 8/12/2019 Non Convex 1

    31/61

    166 Mung Chiang

    of the successive approximation method, a monomial approximation for thenumerator f(x) should satisfy f(x) f(x).

    Applications to Power Control

    Figure 5.9 shows a block diagram of the approach of GP-based power controlfor a general SIR regime [64]. In the high SIR regime, we need to solve onlyone GP. In the medium to low SIR regimes, we solve truly nonconvex powercontrol problems that cannot be turned into convex formulation through aseries of GPs.

    (High SIR)OriginalProblem

    Solve1 GP

    (Medium

    toLow SIR)

    OriginalProblem SPComplementary

    GP (Condensed) Solve1 GP

    -

    - - -6

    Fig. 5.9 GP-based power control in different SIR regimes.

    GP-based power control problems in the medium to low SIR regimes be-come SP (or, equivalently, complementary GP), which can be solved by thesingle or double condensation method. We focus on the single condensationmethod here. Consider a representative problem formulation of maximizingtotal system throughput in a cellular wireless network subject to user rate

    and outage probability constraints in problem (5.18), which can be explicitlywritten out as

    minimizeQN

    i=11

    1+SIRisubject to (2TRi,min 1) 1

    SIRi 1, i= 1, . . . , N ,

    (SIRth)N1(1 Po,i,max)

    QNj6=i

    GijPjGiiPi

    1, i= 1, . . . , N ,

    Pi(Pi,max)1 1, i= 1, . . . , N .

    (5.22)

    All the constraints are posynomials. However, the objective is not a posyno-mial, but a ratio between two posynomials as in (5.19). This power controlproblem can be solved by the condensation method by solving a series ofGPs. Specifically, we have the following single-condensation algorithm.

    Algorithm 3. Single condensation GP power control.

    1. Evaluate the denominator posynomial of the objective function in (5.22)with the given P.

  • 8/12/2019 Non Convex 1

    32/61

    5 Nonconvex Optimization for Communication Networks 167

    0 50 100 150 200 250 300 350 400 450 5005050

    5100

    5150

    5200

    5250

    5300

    Experiment index

    Totalsystem

    throughputachiev

    ed

    ptimized total system throughput

    Fig. 5.10 Maximized total system throughput achieved by the (single) condensationmethod for 500 different initial feasible vectors (Example 5.6). Each point represents adifferent experiment with a different initial power vector.

    2. Compute for each term i in this posynomial,

    i=value ofith term in posynomial

    value of posynomial .

    3. Condense the denominator posynomial of the (5.22) objective functioninto a monomial using (5.21) with weights i.

    4. Solve the resulting GP using an interior point method.5. Go to step 1 using P of step 4.6. Terminate the kth loop if

    P(k) P(k1) where is the errortolerance for exit condition.

    As condensing the objective in the above problem gives us an underesti-mate of the objective value, each GP in the condensation iteration loop triesto improve the accuracy of the approximation to a particular minimum inthe original feasible region. All three conditions for convergence are satisfied,and the algorithm is convergent. Empirically through extensive numerical ex-periments, we observe that it almost always computes the globally optimalpower allocation.

    Example 5.6. Single condensation example. We consider a wireless cellularnetwork with three users. LetT = 106 s,Gii= 1.5, and generateGij ,i 6=j ,

    as independent random variables uniformly distributed between 0 and 0.3.Threshold SIR is SIRth = 10 dB, and minimal data rate requirements are100 kbps, 600 kbps, and 1000 kbps for logical links 1, 2, and 3, respectively.

  • 8/12/2019 Non Convex 1

    33/61

    168 Mung Chiang

    Maximal outage probabilities are 0.01 for all links, and maximal transmitpowers are 3 mW, 4 mW, and 5 mW for links 1, 2, and 3, respectively. Foreach instance of problem (5.22), we pick a random initial feasible power vec-tor P uniformly between 0 and Pmax. Figure 5.10 compares the maximized

    total network throughput achieved over 500 sets of experiments with differ-ent initial vectors. With the (single) condensation method, SP converges todifferent optima over the entire set of experiments, achieving (or coming veryclose to) the global optimum at 5290 bps 96% of the time and a local op-timum at 5060 bps 4% of the time. The average number of GP iterationsrequired by the condensation method over the same set of experiments is 15if an extremely tight exit condition is picked for SP condensation iteration:= 1 1010. This average can be substantially reduced by using a larger ;for example, increasing to 1 102 requires on average only 4 GPs.

    We have thus far discussed a power control problem (5.22) where theobjective function needs to be condensed. The method is also applicable ifsome constraint functions are signomials and need to be condensed [14, 35].

    5.3.5 Distributed Algorithm

    A limitation for GP-based power control in ad hoc networks (without basestations) is the need for centralized computation (e.g., by interior point meth-ods). The GP formulations of power control problems can also be solved bya new method of distributed algorithm for GP. The basic idea is that eachuser solves its own local optimization problem and the coupling among usersis taken care of by message passing among the users. Interestingly, the spe-cial structure of coupling for the problem at hand (all coupling among thelogical links can be lumped together using interference terms) allows one to

    further reduce the amount of message passing among the users. Specifically,we use a dual decomposition method to decompose a GP into smaller sub-problems whose solutions are jointly and iteratively coordinated by the useof dual variables. The key step is to introduce auxiliary variables and to addextra equality constraints, thus transferring the coupling in the objective tocoupling in the constraints, which can be solved by introducing consistencypricing (in contrast to congestion pricing). We illustrate this idea throughan unconstrained GP followed by an application of the technique to powercontrol.

    Distributed Algorithm for GP

    Suppose we have the following unconstrained standard form GP inx 0,

    minimizeP

    i fi(xi, {xj}jI(i)), (5.23)

  • 8/12/2019 Non Convex 1

    34/61

    5 Nonconvex Optimization for Communication Networks 169

    where xi denotes the local variable of the ith user, {xj}jI(i) denotes thecoupled variables from other users, andfiis either a monomial or posynomial.Making a change of variableyi= log xi,i, in the original problem, we obtain

    minimizeP

    i fi(eyi

    , {eyj

    }jI(i)).

    We now rewrite the problem by introducing auxiliary variables yij for thecoupled arguments and additional equality constraints to enforce consistency:

    minimizeP

    i fi(eyi , {eyij}jI(i))

    subject toyij =yj , j I(i), i. (5.24)

    Each ith user controls the local variables (yi, {yij}jI(i)). Next, the Lagran-gian of (5.24) is formed as

    L({yi}, {yij}; {ij}) =Xi

    fi(eyi , {eyij}jI(i)) +

    Xi

    XjI(i)

    ij(yj yij)

    =Xi

    Li(yi, {yij}; {ij}),

    where

    Li(yi, {yij}; {ij}) = fi(eyi , {eyij}jI(i)) +

    Xj:iI(j)

    ji

    yi

    XjI(i)

    ijyij .

    (5.25)The minimization of the Lagrangian with respect to the primal variables({yi}, {yij}) can be done simultaneously and distributively by each user inparallel. In the more general case where the original problem (5.23) is con-strained, the additional constraints can be included in the minimization ateachLi.

    In addition, the following master Lagrange dual problem has to be solvedto obtain the optimal dual variables or consistency prices {ij},

    max{ij}

    g({ij}), (5.26)

    whereg({ij}) =

    Xi

    minyi,{yij}

    Li(yi, {yij}; {ij}).

    Note that the transformed primal problem (5.24) is convex with zero dualitygap; hence the Lagrange dual problem indeed solves the original standardGP problem. A simple way to solve the maximization in (5.26) is with thefollowing subgradient update for the consistency prices,

    ij(t + 1) =ij(t) + (t)(yj(t) yij(t)). (5.27)

  • 8/12/2019 Non Convex 1

    35/61

    170 Mung Chiang

    Appropriate choice of the stepsize (t)> 0, for example, (t) = 0/tfor someconstant 0> 0, leads to convergence of the dual algorithm.

    Summarizing, the ith user has to: (i) minimize the function Li in (5.25)involving only local variables, upon receiving the updated dual variables

    {ji, j : i I(j)}, and (ii) update the local consistency prices {ij , j I(i)}with (5.27), and broadcast the updated prices to the coupled users.

    Applications to Power Control

    As an illustrative example, we maximize the total system throughput in thehigh SIR regime with constraints local to each user. If we directly appliedthe distributed approach described in the last section, the resulting algorithmwould require knowledge by each user of the interfering channels and inter-fering transmit powers, which would translate into a large amount of messagepassing. To obtain a practical distributed solution, we can leverage the struc-tures of power control problems at hand, and instead keep a local copy of

    each of the effective received powersPRij =GijPj . Again using problem (5.18)as an example formulation and assuming high SIR, we can write the problemas (after the log change of variable)

    minimizeP

    i log

    G1ii exp(Pi)P

    j6=i exp(PRij ) +

    2

    subject to PRij =Gij+ Pj ,

    (5.28)

    Constraints are local to each user, for example, (a), (d), and (e) in Table 5.3.The partial Lagrangian is

    L=

    Xilog

    G1ii exp(Pi)

    Xj6=i

    exp(PRij ) + 2

    +Xi

    Xj6=i

    ij

    PRij

    Gij+ Pj

    , (5.29)

    and the localith Lagrangian function in (5.29) is distributed to the ith user,from which the dual decomposition method can be used to determine the op-timal power allocation P. The distributed power control algorithm is sum-marized as follows.

    Algorithm 4. Distributed power allocation update to maximize Rsystem.

    At each iteration t:1. The ith user receives the term

    Pj6=i ji(t)

    involving the dual vari-

    ables from the interfering users by message-passing and minimizes the fol-

    lowing local Lagrangian with respect to Pi(t),n

    PRij (t)oj

    subject to the local

    constraints.

  • 8/12/2019 Non Convex 1

    36/61

    5 Nonconvex Optimization for Communication Networks 171

    Li

    Pi(t),

    nPRij (t)

    oj

    ; {ij(t)}j

    = log

    G1ii exp(Pi(t))

    Xj6=i

    exp(PRij (t)) + 2

    +Xj6=i

    ijPRij (t)

    Xj6=i

    ji(t)

    Pi(t).

    2. The ith user estimates the effective received power from each of theinterfering users PRij (t) = GijPj(t) for j 6=i, updates the dual variable by

    ij(t + 1) =ij(t) + (0/t)

    PRij (t) log GijPj(t)

    , (5.30)

    and then broadcasts them by message passing to all interfering users in thesystem.

    Example 5.7. Distributed GP power control.We apply the distributed algo-rithm to solve the above power control problem for three logical links withGij = 0.2,i 6=j, Gii= 1, i, maximal transmit powers of 6 mW, 7 mW, and7 mW for links 1, 2, and 3 respectively. Figure 5.11 shows the convergence ofthe dual objective function towards the globally optimal total throughput ofthe network. Figure 5.12 shows the convergence of the two auxiliary variablesin links 1 and 3 towards the optimal solutions.

    0 50 100 150 2000.4

    0.6

    0.8

    1

    1.2

    1.4

    1.6

    1.8

    2

    2.2x 10

    4

    Iteration

    Dual objective function

    Fig. 5.11 Convergence of the dual objective function through distributed algorithm (Ex-ample 5.7).

  • 8/12/2019 Non Convex 1

    37/61

  • 8/12/2019 Non Convex 1

    38/61

    5 Nonconvex Optimization for Communication Networks 173

    5.4 DSL Spectrum Management

    5.4.1 Introduction

    Digital subscriber line (DSL) technologies transform traditional voice-bandcopper channels into high-bandwidth data pipes, which are currently capableof delivering data rates up to several Mbps per twisted-pair over a distanceof about 10 kft. The major obstacle for performance improvement in todaysDSL systems (e.g., ADSL and VDSL) is crosstalk, which is the interferencegenerated between different lines in the same binder. The crosstalk is typically1020 dB larger than the background noise, and direct crosstalk cancellation(e.g., [6, 27]) may not be feasible in many cases due to complexity issues oras a result of unbundling. To mitigate the detriments caused by crosstalk,static spectrum managementwhich mandates spectrum mask or flat powerbackoffacross all frequencies (i.e., tones) has been implemented in the currentsystem.

    Dynamic spectrum management (DSM) techniques, on the other hand,can significantly improve data rates over the current practice of static spec-trum management. Within the current capability of the DSL modems, eachmodem has the capability to shape its own power spectrum density (PSD)across different tones, but can only treat crosstalk as background noise (i.e.,no signal level coordination, such as vector transmission or iterative decod-ing, is allowed), and each modem is inherently a single-inputsingle-outputcommunication system. The objective would be to optimize the PSD of allusers on all tones (i.e., continuous power loading or discrete bit loading), suchthat they are compatible with each other and the system performance (e.g.,weighted rate sum as discussed below) is maximized.

    Compared to power control in wireless networks treated in the last sec-tion, the channel gains are not time-varying in DSL systems, but the prob-

    lem dimension increases tremendously because there are many tones (orfrequency carriers) over which transmission takes place. Nonconvexity stillremains a major technical challenge, and high SIR approximation in gen-eral cannot be made. However, utilizing the specific structures of the prob-lem (e.g., the interference channel gain values), an efficient and distributedheuristic is shown to perform close to the optimum in many realistic DSLnetwork scenarios.

    Following [7], this section presents a new algorithm for spectrum manage-ment in frequency selective interference channels for DSL, called autonomousspectrum balancing (ASB). It is the first DSL spectrum management algo-rithm that satisfies all of the following requirements for performance andcomplexity. It is autonomous (distributed algorithm across the users withoutexplicit information exchange) with linear-complexity, while provably conver-

    gent, and comes close to the globally optimal rate region in practice. ASB

  • 8/12/2019 Non Convex 1

    39/61

    174 Mung Chiang

    overcomes the bottlenecks in the state-of-the-art algorithms in DSM, includ-ing IW, OSB, and ISB summarized below.

    Let Kbe the number of tones and N the number of users (lines). Theiterative waterfilling(IW) algorithm [74] is among one of the first DSM algo-

    rithms proposed. In IW, each user views any crosstalk experienced as additiveGaussian noise, and seeks to maximize its data rate by waterfilling over theaggregated noise plus interference. No information exchange is needed amongusers, and all the actions are completely autonomous. IW leads to a greatperformance increase over the static approach, and enjoys a low complexitythat is linear in N. However, the greedy nature of IW leads to a performancefar from optimal in the nearfar scenarios such as mixed CO/RT deploymentand upstream VDSL.

    To address this, an optimal spectrum balancing(OSB) algorithm [9] hasbeen proposed, which finds the best possible spectrum management solutionunder the current capabilities of the DSL modems. OSB avoids the selfish be-haviors of individual users by aiming at the maximization of a total weightedsum of user rates, which corresponds to a boundary point of the achievable

    rate region. On the other hand, OSB has a high computational complex-ity that is exponential in N, which quickly leads to intractability when Nis larger than 6. Moreover, it is a completely centralized algorithm where aspectrum management center at the central office needs to know the globalinformation(i.e., all the noise PSDs and crosstalk channel gains in the samebinder) to perform the algorithm.

    As an improvement to the OSB algorithm, an iterative spectrum balancing(ISB) algorithm [8] has been proposed, which is based on a weighted sumrate maximization similar to OSB. Different from OSB, ISB performs theoptimization iteratively through users, which leads to a quadratic complexityin N. Close to optimal performance can be achieved by the ISB algorithmin most cases. However, each user still needs to know the global informationas in OSB, thus ISB is still a centralized algorithm and is considered to be

    impractical in many cases.This section presents the ASB algorithm [7], which attains near-optimal

    performance in an implementable way. The basic idea is to use the conceptof a reference line to mimic a typical victim line in the current binder.By setting the power spectrum level to protect the reference line, a goodbalance between selfish and global maximizations can be achieved. The ASBalgorithm enjoys a linear complexity in N and K, and can be implementedin a completely autonomous way. We prove the convergence of ASB underboth sequential and parallel updates.

    Table 5.4 compares various aspects of different DSM algorithms. Utilizingthe structures of the DSL problem, in particular, the lack of channel variationand user mobility, is the key to provide a linear complexity, distributed, con-vergent, and almost optimal solution to this coupled, nonconvex optimizationproblem.

  • 8/12/2019 Non Convex 1

    40/61

    5 Nonconvex Optimization for Communication Networks 175

    Table 5.4 Comparison of different DSM algorithms

    Algorithm Operation Complexity Performance Reference

    IW Autonomous O (KN) Suboptimal [74]

    OSB Centralized O KeN

    Optimal [9]ISB Centralized O

    KN

    2 Near optimal [8]ASB Autonomous O (KN) Near optimal [7]

    5.4.2 System Model

    Using the notation as in [9, 8], we consider a DSL bundle withN ={1, . . . , N }modems (i.e., lines, users) and K = {1, . . . , K } tones. Assume discrete mul-titone (DMT) modulation is employed by all modems; transmission can bemodeled independently on each tone as

    yk = Hkxk+ zk.

    The vector xk ={xnk , n N} contains transmitted signals on tone k , where

    xnk is the signal transmitted onto line n at tone k. yk and zk have similarstructures.yk is the vector of received signals on tone k. zk is the vector ofadditive noise on tone k and contains thermal noise, alien crosstalk, single-carrier modems, radio frequency interference, and so on.Hk = [h

    n,mk ]n,mN is

    theNNchannel transfer matrix on tone k, wherehn,mk is the channel fromTX m to RX n on tone k. The diagonal elements ofHk contain the direct-channels whereas the off-diagonal elements contain the crosstalk channels. We

    denote the transmit power spectrum density (PSD) snk =En|xnk |

    2o

    . In the

    last sections notation for single-carrier systems, we would have snk =Pn, k.For convenience we denote the vector containing the PSD of user n on alltones assn ={snk , k K}. We denote the DMT symbol rate as fs.

    Assume that each modem treats interference from other modems as noise.When the number of interfering modems is large, the interference can bewell approximated by a Gaussian distribution. Under this assumption theachievable bit loading of user n on tone k is

    bnk = log

    1 +

    1

    snkPm6=n

    n,mk s

    mk +

    nk

    !, (5.31)

    wheren,mk =|hn,mk |

    2/ |hn,nk |

    2is the normalized crosstalk channel gain, and

    nk is the noise power density normalized by the direct channel gain |hn,nk |

    2.

    Here denotes the SINR-gap to capacity, which is a function of the de-sired BER, coding gain, and noise margin [61]. Without loss of generality, we

    assume= 1. The data rate on line n is thus

  • 8/12/2019 Non Convex 1

    41/61

    176 Mung Chiang

    Rn =fsXkK

    bnk . (5.32)

    Each modemn is typically subject to a total power constraint Pn, due to thelimitations on each modems analogue frontend.X

    kK

    snk Pn. (5.33)

    5.4.3 Spectrum Management Problem Formulation

    One way to define the spectrum management problem is start with the fol-lowing optimization problem.

    maximize R1

    subject toRn Rn,target, n >1PkK snk Pn,n.

    (5.34)

    Here Rn,target is the target rate constraint of user n. In other words, wetry to maximize the achievable rate of user 1, under the condition that allother users achieve their target rates Rn,target. The mutual interference in(5.31) causes Problem (5.34) to be coupled across users on each tone, andthe individual total power constraint causes Problem (5.34) to be coupledacross tones as well.

    Moreover, the objective function in Problem (5.34) is nonconvex due tothe coupling of interference, and the convexity of the rate region cannot beguaranteed in general.

    However, it has been shown in [75] that the duality gap between the dualgap of Problem (5.34) goes to zero when the number of tones K gets large

    (e.g., for VDSL), thus Problem (5.34) can be solved by the dual decompositionmethod, which brings the complexity as a function of K down to linear.Moreover, a frequency-sharing property ensures the rate region is convexwith large enough K, and each boundary point of the boundary point of therate region can be achieved by a weighted rate maximization as (following[9]),

    maximize R1 +P

    n>1 wnRn

    subject toP

    kK snk P

    n, n N, (5.35)

    such that the nonnegative weight coefficientwn is adjusted to ensure that thetarget rate constraint of user n is met. Without loss of generality, here wedefinew1 = 1. By changing the rate constraintsRn,target for usersn >1 (orequivalently, changing the weight coefficients, wn forn >1), every boundary

    point of the convex rate region can be traced.

  • 8/12/2019 Non Convex 1

    42/61

    5 Nonconvex Optimization for Communication Networks 177

    We observe that at the optimal solutions of (5 .34), each user chooses aPSD level that leads to a good balance of maximization of its own rateand minimization of the damages it causes to the other users. To accuratelycalculate the latter, the user needs to know the global information of the noise

    PSDs and crosstalk channel gains. However, if we aim at a less aggressiveobjective and only require each user give enough protection to the otherusers in the binder while maximizing her own rate, then global informationmay not be needed. Indeed, we can introduce the concept of a referenceline, a virtual line that represents a typical victim in the current binder.Then instead of solving (5.34), each user tries to maximize the achievabledata rate on the reference line, subject to its own data rate and total powerconstraint. Define the rate of the reference line to user n as

    Rn,ref =XkK

    bnk =XkK

    log

    1 +

    sknks

    nk + k

    .

    The coefficients {sk, k,nk , k, n} are parameters of the reference line and

    can be obtained from field measurement. They represent the conditions of atypical victim user in an interference channel (here a binder of DSL lines),and are known to the users a priori. They can be further updated on a muchslower time scale through channel measurement data. User n then wants tosolve the following problem local to itself,

    maximize Rn,ref

    subject toRn Rn,target,PkK s

    nk P

    n.(5.36)

    By using Lagrangian relaxation on the rate target constraint in Problem(5.36) with a weight coefficient (dual variable) wn, the relaxed version of(5.36) is

    maximize wn

    Rn

    + Rn,ref

    subject toP

    kK snk P

    n. (5.37)

    The weight coefficientwn needs to be adjusted to enforce the rate constraint.

    5.4.4 ASB Algorithms

    We first introduce the basic version of the ASB algorithm (ASB-I), where eachusern chooses the PSD sn to solve (5.36) ,and updates the weight coefficientwn to enforce the target rate constraint. Then we introduce a variation of theASB algorithm (ASB-II) that enjoys even lower computational complexityand provable convergence.

  • 8/12/2019 Non Convex 1

    43/61

    178 Mung Chiang

    ASB-I

    For each usern, replacing the original optimization (5.36) with the Lagrangedual problem

    maxn0,

    PkK

    snkPn

    XkK

    maxsnk

    Jnk

    wn, n, snk , snk

    , (5.38)

    whereJnk

    wn, n, snk , snk

    = wnbnk + b

    nk

    nsnk . (5.39)

    By introducing the dual variable n, we decouple (5.36) into several smallersubproblems, one for each tone. And defineJnk as user ns objective functionon tonek . The optimal PSD that maximizes Jnk for given w

    n and n is

    sn,Ik

    wn, n, snk

    = arg maxsnk[0,Pn]

    Jnk

    wn, n, snk , snk

    , (5.40)

    which can be found by solving the first-order condition,

    Jnk

    wn, n, snk , snk

    /snk = 0,

    which leads to

    wn

    sn,Ik +P

    m6=n n,mk s

    mk +

    nk

    nk sk

    sk+ nksn,Ik + k

    nks

    n,Ik + k

    n = 0.(5.41)

    Note that (5.41) can be simplified into a cubic equation that has threesolutions. The optimal PSD can be found by substituting these three solutionsback to the objective function Jnk

    wn, n, snk , s

    nk

    , as well as checking the

    boundary solutions snk = 0 andsnk =P

    n, and picking the one that yields thelargest value ofJnk .

    The user then updates

    n

    to enforce the power constraint, and updateswn to enforce the target rate constraint. The complete algorithm is given asfollows, where and w are small stepsizes for updating

    n andwn.

    Algorithm 5. Autonomous Spectrum Balancing.

    repeatfor each user n = 1, . . . , N

    repeatfor each tonek = 1, . . . , K , find

    sn,Ik = arg maxsnk0 Jnk

    n =h

    n + P

    ksn,Ik P

    ni+

    ;

    wn = [wn + w(Rn,target Pkbnk)]

    +;

    until convergenceend

    until convergence

  • 8/12/2019 Non Convex 1

    44/61

    5 Nonconvex Optimization for Communication Networks 179

    ASB-II with Frequency-Selective Waterfilling

    To obtain the optimal PSD in ASB-I (for fixedn andwn), we have to solvethe roots of a cubic equation. To reduce the computational complexity and

    gain more insights of the solution structure, we assume that the referenceline operates in the high SIR regime whenever it is active: If sk > 0, thensk k

    n,mk s

    nk for any feasible s

    nk , n N, and k K. This assumption

    is motivated by our observations on optimal solutions in the DSL type ofinterference channels. It means that the reference PSD is much larger thanthe reference noise, which is in turn much larger than the interference fromuser n. Then on any tone k K = {k| sk > 0, k K}, the reference linesachievable rate is

    log

    1 +

    sknks

    nk + k

    log

    skk

    nksnk

    k,

    and user ns objective function on tonek can be approximated by

    Jn,II,1k

    wn, n, snk , snk

    = wnbnk

    nksnkk

    nsnk + log

    skk

    .

    The corresponding optimal PSD is

    sn,II,1k

    wn, n, snk

    =

    wn

    n + nk/kXm6=n

    n,mk smk

    nk

    +

    . (5.42)

    This is a waterfilling type of solution and is intuitively satisfying: the PSDshould be smaller when the power constraint is tighter (i.e., n is larger), orthe interference coefficient to the reference line nk is higher, or the noise levelon the reference line k is smaller, or there is more interference plus noiseP

    m6=n n,mk smk + nk on the current tone. It is different from the conventional

    waterfilling in that the water level in each tone is not only determined by thedual variables wn and n, but also by the parameters of the reference line,nk/k.

    On the other hand, on any tone where the reference line is inactive, thatis,k KC= {k| sk = 0, k K}, the objective function is

    Jn,II,2k

    wn, n, snk , snk

    = wnbnk

    nsnk ,

    and the corresponding optimal PSD is

    sn,II,2k wn, n, snk =

    wn

    n X

    m6=n

    n,mk smk

    nk

    +

    . (5.43)

    This is the same solution as the iterative waterfilling.

  • 8/12/2019 Non Convex 1

    45/61

    180 Mung Chiang

    The choice of optimal PSD in ASB-II can be summarized as the following.

    sn,IIk wn, n, snk =

    wn

    n+nk/k

    Pm6=n

    n,mk s

    mk

    nk

    +, k K,

    wnn

    Pm6=n

    n,m

    k sm

    k n

    k+

    , k KC

    .(5.44)

    This is essentially a waterfill