Contract Theory, by Alex Young

Embed Size (px)

Citation preview

  • 8/10/2019 Contract Theory, by Alex Young

    1/86

    Contract Theory

    Alex Young

    December 19, 2012

    Contents

    1 Preface 4

    2 Hidden Information, Screening 4

    2.1 The Simple Economics of Adverse Selection . . . . . . . . . . . . . . . . . . 4

    2.1.1 First-Best Outcome: Perfect Price Discrimination . . . . . . . . . . . 6

    2.1.1.1 Comparison of type-specific contracts with each other. . . . 6

    2.1.1.2 In case its been too long since youve last done a maximiza-

    tion problem . . . . . . . . . . . . . . . . . . . . . . . . . . 7

    2.1.1.3 Mas-Colell et al.[1995]. . . . . . . . . . . . . . . . . . . . . 8

    2.1.1.4 Just what is the single crossing condition, exactly? . . . . . 10

    2.1.2 Adverse Selection, Linear Pricing, and Simple Two-Part Tariffs . . . 10

    2.1.2.1 Linear Pricing . . . . . . . . . . . . . . . . . . . . . . . . . 10

    2.1.2.2 Single Two-Part Tariff . . . . . . . . . . . . . . . . . . . . . 122.1.2.3 Did you forget the envelope theorem?. . . . . . . . . . . . . 12

    2.1.3 Second-Best Outcome: Optimal Nonlinear Pricing Algorithm. . . . . 14

    2.1.3.1 Comparison with the First Best . . . . . . . . . . . . . . . . 17

    2.2 Application: Credit Rationing . . . . . . . . . . . . . . . . . . . . . . . . . . 19

    1

  • 8/10/2019 Contract Theory, by Alex Young

    2/86

    2.2.1 Solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20

    2.3 A Continuum of Types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

    2.3.1 The Implementation Problem . . . . . . . . . . . . . . . . . . . . . . 23

    2.3.2 The Optimization Problems Algorithm. . . . . . . . . . . . . . . . . 25

    2.3.2.1 How do you get the local incentive compability constraint? . 25

    2.3.2.2 Doesnt that look familiar? . . . . . . . . . . . . . . . . . . 30

    2.4 Application: Costly State Verification. . . . . . . . . . . . . . . . . . . . . . 31

    3 Hidden Information, Signaling 34

    3.1 Spences Model of Education as a Signal . . . . . . . . . . . . . . . . . . . . 34

    3.1.1 Did you forget the beer quiche game? . . . . . . . . . . . . . . . . . . 36

    3.1.2 Refinements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42

    3.1.2.1 Cho and Kreps Intuitive Criterion . . . . . . . . . . . . . . 42

    3.1.2.2 Maskin and Tiroles Informed Principal Problem . . . . . . 45

    4 The Principal-Agent ProblemMas-Colell et al. [1995] 46

    4.1 Comment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46

    4.2 Hidden Actions (Moral Hazard) . . . . . . . . . . . . . . . . . . . . . . . . . 47

    4.2.1 The Optimal Contract when Effort is Observable . . . . . . . . . . . 48

    4.2.1.1 A risk-averse owner. . . . . . . . . . . . . . . . . . . . . . . 50

    4.2.2 The Optimal Contract when Effort is Not Observable . . . . . . . . . 51

    4.2.2.1 A risk-neutral manager. . . . . . . . . . . . . . . . . . . . . 52

    4.2.2.2 A risk-averse manager . . . . . . . . . . . . . . . . . . . . . 53

    4.2.3 Comparing the first and second-best wage contracts . . . . . . . . . . 554.2.3.1 An additional demonstration of the implications of risk aversion 55

    4.2.4 The Value of Information . . . . . . . . . . . . . . . . . . . . . . . . 56

    4.3 Linear Contracts, Exponential Utility, and Normally Distributed Performance

    (LEN)[Bolton and Dewatripont,2004] . . . . . . . . . . . . . . . . . . . . . 58

    2

  • 8/10/2019 Contract Theory, by Alex Young

    3/86

    4.3.1 Did you forget the properties of the log-normal distribution? . . . . . 60

    4.4 Moral Hazard in Teams[Bolton and Dewatripont,2004]. . . . . . . . . . . . 60

    4.4.1 Unobservable Individual Outputs . . . . . . . . . . . . . . . . . . . . 61

    4.4.2 Observable Individual Outputs . . . . . . . . . . . . . . . . . . . . . 64

    4.5 Combining Moral Hazard and Adverse Selection . . . . . . . . . . . . . . . . 68

    4.5.1 Optimal Contract, Moral Hazard Only . . . . . . . . . . . . . . . . . 70

    4.5.2 Optimal Contract, Adverse Selection Only . . . . . . . . . . . . . . . 71

    4.5.3 Optimal Contract, Moral Hazard and Adverse Selection. . . . . . . . 74

    5 Applications 77

    5.1 Application:Antle and Fellingham[1995]. . . . . . . . . . . . . . . . . . . . 77

    5.1.1 The Model. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77

    5.1.2 Solution of Benchmark Case . . . . . . . . . . . . . . . . . . . . . . . 79

    5.1.2.1 How? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79

    5.2 Prendergast [1993] . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80

    5.2.1 Simplified Representation . . . . . . . . . . . . . . . . . . . . . . . . 80

    3

  • 8/10/2019 Contract Theory, by Alex Young

    4/86

    1 Preface

    These notes are mostly follow Bolton and Dewatripont [2004] andMas-Colell et al. [1995].

    Textbooks always omit steps in derivations due to space constraints. Moreover, its a good

    exercise for the reader to fill in the blanks. Since I habitually forget what I have done,

    why not keep a record of all those aha! moments I had while trodding through the tedious

    algebra? Why not comment on what the algebra means in English?

    I thank Qi Chen for his teaching, Zeqiong Huang and Thomas Steffen for studying together,

    and of course my father for many helpful Skype conversations.

    2 Hidden Information, Screening

    2.1 The Simple Economics of Adverse Selection

    Consider a transaction between a buyer and a seller, where the seller does not know the

    buyers exact valuation but sets the terms of the contract. The buyers preferences are

    represented by the utility function

    u(q , T , ) =

    q0

    P(x, ) dx T

    whereqis the number of units purchased,Tis the total amount paid to the seller, and P(x, )

    is the inverse demand curve of a buyer with preference characteristics . For tractability,

    consider the following functional form for the buyers preferences:

    u(q , T , ) =v(q) T

    4

  • 8/10/2019 Contract Theory, by Alex Young

    5/86

    wherev(0) = 0, v(q)>0, and v (q)< 0 q. That is,

    u(0, 0, ) =v(0) 0 = 0

    u(q , T , )

    q =v (q)>0

    2u(q , T , )

    q2 =v (q)0. His profit from sellingq

    units against a sum of moneyT is given by

    = T cq

    The question is, What is the best (i.e. the profit maximizing) pair (T, q) that the seller will

    be able to induce the buyer to choose? The answer depends on the information the seller

    has on the buyers preferences. For simplification, consider the case where there are only two

    types of buyers: {L, H}, with H > L. The consumer is of typeL with probability

    [0, 1] and of type Hwith probability (1 ). The probabilitycan also be interpreted

    as the proportion of consumers of type L.

    5

  • 8/10/2019 Contract Theory, by Alex Young

    6/86

    2.1.1 First-Best Outcome: Perfect Price Discrimination

    Suppose instead that the seller knows the buyers exact valuation. The seller can then

    treat each type of buyer separately and offer type-specific contracts, that is, (Ti, qi) for type

    i (i = H, L). Assume the buyer obtains a payoff of u if she does not take the offer. The

    seller will try to maximize his profits subject to inducing the buyer to accept the proposed

    contract:

    maxTi,qi

    Ti cqi

    subject to

    iv(qi) Ti u (IR)

    The constraint means that the buyer will participate only if doing so does not make him

    worse off than not participating (i.e. individual rationality). The solution to the problem is

    the contract (qi, Ti) such that

    iv(qi) =c

    iv(qi) = Ti+ u

    2.1.1.1 Comparison of type-specific contracts with each other We can relate the

    quantities and prices of one type to the other.

    Hv(qH) =Lv(qL)

    SinceH> L, it must be that v(qH)< v

    (qL). Since v() 0, qH> qL. It makes intuitive

    6

  • 8/10/2019 Contract Theory, by Alex Young

    7/86

    sense that the high type will consume more than the low type. What about price?

    Hv(qH) TH=Lv(qL) TL

    Hv(qH) Lv(qL) = TH TL

    Since H > L and v(qH) > v(qL) because qH > qL and v() > 0, the left hand side is

    positive, and so is the right hand side. Therefore, TH> TL. In the first best, the seller leaves

    no rents to either buyer type. Hence, along with the higher consumption for the high type

    comes a higher price. This will play a role in the next section when the seller cannot observe

    the buyers type. We will see that the seller cannot use these bundles in such a situation to

    maximize profit.

    2.1.1.2 In case its been too long since youve last done a maximization problem

    To see this, first determine whether the IR constraint will hold as an equality in the optimal

    solution. The answer is yes if either the partial derivative of the profit function with

    respect to Ti or the partial derivative with respect to qi is positive, for then the optimal

    pair could not lie below the IR constraint since profit could be made greater by increasing

    whichever argument had positive marginal profit [Graham, 2005].

    Ti= 1> 0

    qi= c

  • 8/10/2019 Contract Theory, by Alex Young

    8/86

    press the constraint in terms ofTi, substitute into the profit function, and take the FOC:

    {(iv(qi) u) cqi}

    qi=iv

    (qi) c

    = 0

    iv(qi) =c

    Thus, the seller finds it optimal to maximize his total surplus by setting a quantity such

    that marginal utility equals marginal cost, and then setting payment to appropriate the full

    surplus and leave no rent to the buyer above u. Without adverse selection, the total profit

    of the seller is thus

    (TL cqL) + (1 )(TH cqH)

    2.1.1.3 Mas-Colell et al. [1995] Mas-Colell et al. [1995] discuss hidden information

    in the context of the principal-agent problem. That is, postcontractual uncertainty takes

    the form of hidden information, not hidden action, and the problem is one of inducing

    truthtelling, not effort. The unobservable in the model is the random realization of the

    managers disutility from effort.

    The managers Bernoulli utility function over wages and effort depends on a state of nature

    that is realized after the contract is signed and that only the manager observes. We assume

    that Rand that

    u(w,e,) =v(w g(e, ))

    8

  • 8/10/2019 Contract Theory, by Alex Young

    9/86

    g(e, ) measures disutility of effort. It has the following properties:

    ge(e, )>0 (Disutility of effort is increasing in effort)

    gee(e, )>0 (Marginal disutility of effort is increasing in effort)

    g(e, )

  • 8/10/2019 Contract Theory, by Alex Young

    10/86

    Now then,

    (eL) ge(eL, H)>

    (eL) ge(eL, L) (Single crossing condition)

    = 0 (FOC)

    = (eH) ge(eH, H) (Equal across states)

    (eL) ge(eL, H)>

    (eH) ge(eH, H)

    eL< eH (fe(ei, i)

  • 8/10/2019 Contract Theory, by Alex Young

    11/86

    The buyers net surplus can be written as follows:

    Si(P) iv(qi) P qi

    =iv[Di(P)] P[Di(P)]

    D(P) DL(P) + (1 )DH(P)

    S(P) SL(P) + (1 )SH(P)

    With linear pricing, the sellers problem is the classic monopoly pricing problem; he solves

    maxP

    = maxP

    T cq

    = maxP

    P D(P) cD(P)

    = maxP

    (P c)D(P)

    with the monopoly price given by

    d{(P c)D(P)}dP

    =P D(P) +D(P) cD(P)

    = 0

    P =c D(P)

    D(P)

    In this solution, we have both positive rents for the buyers (S(P)> 0) and inefficiently low

    consumption (i.e. iv(q) = P > c). The seller can do better by moving away from linear

    pricing only if buyers cannot make arbitrage profits by trading in a secondary market.

    Otherwise, buyers would buy at the minimum average price and then resell in the secondary

    market if they do not want to consume everything they have purchased.

    11

  • 8/10/2019 Contract Theory, by Alex Young

    12/86

    2.1.2.2 Single Two-Part Tariff We assume that there is only one buyer and that is

    a probability measure; thus, there are no arbitrage opportunities to the buyer. Therefore,

    a single two-part tariff (Z, P), where P is the unit price and Z the fixed fee, will improve

    upon linear pricing for the seller. P, the minimum fixed fee the seller will set is given by

    Z=SL(P). A type-Hbuyer will always purchaseq >0 when charged T(q) =SL(P) + P q,

    sinceH> L. If the seller sets Z=SL(P), he also solves

    maxP

    T(q) = maxP

    SL(P) + (P c)q

    = maxP

    SL(P) + (P c)D(P)

    d{SL(P) + (P c)D(P)}

    dP =SL(P) + (P c)D

    (P) +D(P)

    = 0

    P =c D(P) +SL(P)

    D(P)

    By the envelope theorem,

    Lv(qL) PqL= Lv(DL(P

    )) PDL(P)

    = maxP

    SL(P)

    Lv(DL(P))D

    L(P) (P D

    L(P) +DL(P)) =S

    L(P)

    (Lv(DL(P)) P)

    =0 by FOC

    DL(P) DL(P) =SL(P)

    DL(P) =SL(P)

    so that D(P) +SL(P)> 0; in addition,D(P)< 0 P > c.

    2.1.2.3 Did you forget the envelope theorem? Recall that a property of the indirect

    utility function is that it is strictly increasing in w, income[Jehle and Reny,2001]. We prove

    12

  • 8/10/2019 Contract Theory, by Alex Young

    13/86

    this using the Envelope Theorem.

    Proof. To solve

    v(p, w) = maxxRn

    +

    u(x)

    subject to

    p x= w

    form the Lagrangian:

    L(x, ) =u(x) +(w p x)

    L(x, )

    xi =

    u(x)

    xi > 0

    pi = 0

    >0

    By the Envelope Theorem,

    v(p, w)

    w =

    L(x, w)

    w = >0

    Alternatively, instead of using the envelope theorem, consider a graphical argument:

    Use the Fundamental Theorem of Calculus:

    SL(P) = d

    dPSL(P)

    =

    d

    dP aP DL(t) dt=

    d

    dP

    Pa

    DL(t) dt

    = DL(P)

    where a is where the demand curve intersects the (horizontal) P axis. Thus, if the seller

    13

  • 8/10/2019 Contract Theory, by Alex Young

    14/86

    Figure 1: An inverse demand curve.

    decides to serve both types of customers, the first-best outcome cannot be ahcieved and

    underconsumption remains relative to the first-best outcome. Alternatively, an optimal

    single two-part tariff contract is always preferred by the seller to an optimal linear pricing

    contract because the seller can always raise his profits by setting Z=SL(Pm).

    2.1.3 Second-Best Outcome: Optimal Nonlinear Pricing Algorithm

    Since the seller does not observe the buyers type, he is forced to offer a set of choices

    independent of her type. The seller solves

    maxT(q)

    [T(qL) cqL] + (1 )[T(qH) cqH]

    subject to

    qi = arg maxq

    iv(q) T(q) for i = L, H (IC)

    iv(qi) T(qi) 0 for i = L, H (IR)

    The problem as-is involves optimization over a schedule T(q) under constraints that them-

    selves involve optimization problems. To simplify the problem, follow these steps:

    14

  • 8/10/2019 Contract Theory, by Alex Young

    15/86

    1. Use the revelation principle.

    2. Observe that the participation constraint of the high type will not bind at the opti-

    mum.

    Hv(qH) T(qH) Hv(qL) T(qL) (ICH)

    > Lv(qL) T(qL) (H> L)

    0 (IRL)

    That is, the high type will earn an informational rent.

    3. Solve the relaxed problem without the incentive constraint that is satisfied at the first-

    best optimum.

    The first-best problems outcome is not incentive compatible because the Hbuyer will

    prefer to choose (qL, TL) rather than his own first-best allocation because it gives him

    a strictly positive surplus:

    Hv(qL) TL> Lv(qL) TL (H> L)

    = 0 (No rent in first-best)

    (H L)v(qL)>0

    Instead, type L will not find it attractive to raise his consumption to the level qH:

    Hv(qH) T(qH) = 0 (No rent in first-best)

    > Lv(qH) T(qH) (H> L)

    15

  • 8/10/2019 Contract Theory, by Alex Young

    16/86

    Therefore, we omit the (ICL) constraint. That is, we omit the constraint for the type

    that does not want to deviate; the constraint for the type that does will bind.

    4. Observe that the two remaining constraints of the relaxed problem will bind at the

    optimum.

    5. Eliminate TL and THfrom the maximand using the two binding constraints, perform

    the unconstrained optimization, and check that (ICL) is indeed satisfied.

    How do you get the second FOC (w.r.t qL) in Step 5?

    [Lv(qL) c] (1 )(H L)v

    (qL) =v(qL)(L (1 )(H L)) c

    = 0

    v(qL) = c

    L (1 )(H L)

    =

    c

    L

    1 (1 )(H L)

    L

    Lv(qL) = c

    1 (1 )(H L)

    L

    16

  • 8/10/2019 Contract Theory, by Alex Young

    17/86

    2.1.3.1 Comparison with the First Best From the previous analysis, we see that

    Lv(qSL) =

    c

    1 1

    H L

    L > c

    =Lv(qFL ) (No rent in first best)

    Lv(qSL)> Lv

    (qFL )

    v(qSL)> v(qFL ) (L> 0)

    If we assume v () qSL andv

    ()>0,

    L[v(qFL ) v(q

    SL)]> 0

    ThereforeTFL > TSL . So in the second best, the low type has a lower quantity and a lower

    price compared to the first best.

    17

  • 8/10/2019 Contract Theory, by Alex Young

    18/86

    What about the high type? We saw previously that

    Hv(qSH) =c

    Marginal benefit also equaled marginal cost in the first best:

    Hv(qFH) =c

    Sincev ()< 0, qFH=qSH; that is, the optimal quantity for the high type is unchanged in the

    second best relative to the first best. What about the price?

    Hv(qSH) T

    SH=Hv(q

    SL) T

    SL (ICH)

    > Lv(qSL) T

    SL (H> L)

    = 0 (IRL)

    =Hv(qFH) T

    FH (No rent in first best)

    Hv(qSH) T

    SH> Hv(q

    FH) T

    FH

    TFH TSH>0 (q

    SH=q

    FH)

    TFH > TSH

    So in the second best, the high type gets the same quantity as he gets in the first best at a

    lower price. Let us link this result with the result for the low type, who gets a lower quantity

    at a lower price in the second best compared to the first best. The seller designs the bundle

    such that it is not worth it for the high type to pick the low types bundle. He does so in

    two ways:

    1. He gives the high type a good deal by offering him his optimal quantity at a lower

    price (relative to the first best).

    18

  • 8/10/2019 Contract Theory, by Alex Young

    19/86

    2. He makes the low types bundle less attractive by reducing the quantity.

    2.2 Application: Credit Rationing

    Consider a population of risk-neutral borrowers who each own a project that requires an

    initial outlay of I = 1 and yields a random return X, where X {R, 0}. Let p [0, 1]

    denote the probability that X=R. A borrower can of two types i= s, r, where sis safe

    and r is risky. The borrower of type i has a project with return characteristics (pi, Ri).

    We assume

    piRi= m >1 (A1)

    ps> pr and Rs< Rr (A2)

    A bank can offer to finance the initial outlay in exchange for a future repayment. Assume

    that there is a single bank and excess demand for funds: the bank has a total amount max{, 1 }, so that available funds are sufficient to avoid crowding out either type

    of borrower completely.

    Call (xi, Di) a contract that offers financing with probability xi and repayment Di. The

    bank solves

    maxxi,Di

    [xs(psDs 1) + (1 )xr(prDr 1)]

    19

  • 8/10/2019 Contract Theory, by Alex Young

    20/86

    subject to

    0 xi 1 i = s, r (Regularity)

    Di Ri i = s, r (IR)

    xipi(Ri Di) xjpi(Ri Dj)i, j=s, r (IC)

    xs+ (1 )xr (Supply)

    2.2.1 Solution

    1. We have already applied the revelation principle.

    2. The risky type has an incentive to mimic the safe type. Therefore, his (IR) con-

    straint does not bind:

    RrxrDr xsDs

    xr xs(ICR)

    >xrDr xsDr

    xr xs(Dr > Ds)

    =Dr

    and the (IR) of the safe type does.

    3. The first-best problems outcome is not incentive compatible because the risky type

    will prefer to choose (xs, Ds) rather than his own first-best allocation because it gives

    him a strictly positive surplus (i.e. an informational rent):

    xspr(Rr Ds) xsps(Rs Ds) (pr ps)

    = 0 (No rent in first-best)

    Rr > Ds

    20

  • 8/10/2019 Contract Theory, by Alex Young

    21/86

    Instead, the safe type will not find it attractive to choose (xr, Dr):

    xrpr(Rr Dr) = 0 (No rent in first-best)

    >xrps(Rs Dr) (pr < ps)

    Therefore, we omit the incentive constraint of the safe type.

    4. The banks problem is now to solve

    maxxi,Di

    [xs(ps(Ds 1) + (1 )xr(prDr 1)]

    subject to

    Ds= Rs (IRS)

    xr(Rr Dr) =xs(Rr Rs) (ICR)

    5. As both of these constraints bind, the problem becomes

    maxxs,xr

    {xs(psRs 1) + (1 )[xr(prRr 1) xspr(Rr Rs)]}

    = maxxs,xr

    {xs(m 1) + (1 )[xr(m 1) xspr(Rr Rs)]}

    subject to

    0 xs xr 1

    xs+ (1 )xr

    The FOC with respect to xr is (1 )(m 1) 0, therefore xr = 1; there is no

    rationing of risky borrowers.

    21

  • 8/10/2019 Contract Theory, by Alex Young

    22/86

    The FOC with respect to xs is(m 1) (1 )pr(Rr Rs). Thus,

    xs =0, if(m 1) (1 )p

    r(Rr Rs)< 0 (1 )

    , otherwise

    2.3 A Continuum of Types

    Suppose that is distributed according to the p.d.ff() with c.d.fF() on an interval [,].

    With the revelation principle, the seller solves

    maxq(),T()

    E[()] = maxq(),T()

    E[T() cq()]

    = maxq(),T()

    [T() cq()]f() d

    subject to

    v[q()] T() 0 [,] (IR)

    v[q()] T() v[q()] T() , [,] (IC)

    If all (IC)s hold, the (IR)s can be replaced by

    v[q()] T() 0 (IR)

    Consider decomposing the sellers problem an implementation problem (whichq() are IC?)

    and anoptimization problem(among all feasibleq(), which one is the arg maxq() E[()]?)

    22

  • 8/10/2019 Contract Theory, by Alex Young

    23/86

    2.3.1 The Implementation Problem

    We want to show that if the buyers utility function satisfies the single-crossing condition

    [ u/qu/T]> 0

    then the set of (IC)s in the sellers problem is equivalent to the following:

    1. Monotonicity: dq()

    d 0.

    2. Local incentive compatibility: v[q()]dq()

    d =T() [,].

    Proof. We first show that (IC) satisfaction implies monotonicity and local incentive compat-

    ibility. If consumptionq() and paymentT() are differentiable, the following conditions for

    the buyers problem are satisfied at = :

    u(q(), T(), )

    =

    [v(q()) T()]

    =v [q()]dq()

    d

    = 0 (FOC)

    v[q()](dq()

    d)2 +v[q()]

    d2q()

    d2 T() 0 (SOC)

    The FOC is the local incentive compatibility constraint. If we differentiate it with respect

    to, we obtain

    v[q()]dq()

    d

    +v[q()](dq()

    d

    )2 +v[q()]d2q()

    d2

    T() 0= 0

    v[q()]dq()

    d 0

    dq()

    d 0

    We next show that monotonicity and local incentive compatibility imply that all the buyers

    23

  • 8/10/2019 Contract Theory, by Alex Young

    24/86

    (IC)s hold. Suppose not. Then, there exists at least one type = such that

    v[q()] T()< v[q()] T()

    0 , thenx [,],

    v

    [q(x)]< xv

    [q(x)]

    v[q(x)] T(x)< xv[q(x)] T(x)

    [v[q(x)]dq(x)

    dx T(x)] dx xv[q(x)]

    v[q(x)] T(x)> xv[q(x)] T(x)

    [v[q(x)]dq(x)

    dx T(x)] dx >

    [xv[q(x)]dq(x)

    dx T(x)] dx

    = 0

    But < [v[q(x)]

    dq(x)

    dx T(x)] dx

  • 8/10/2019 Contract Theory, by Alex Young

    25/86

  • 8/10/2019 Contract Theory, by Alex Young

    26/86

    Proof.

    v[q()] T() v[q()] T() (Type )

    v[q()] T() v[q()] T() (Type)

    (v[q()] v[q()]) T() T() (From first ineq.)

    (v[q()] v[q()]) T() T() (From second ineq.)

    (v[q()] v[q()]) T() T() (v[q()] v[q()])

    lim

    (v[q()] v[q()])

    lim

    T() T()

    lim

    (v[q()] v[q()])

    v[q()]dq()

    d =T()

    by the Sandwich Theorem.

    The standard procedure for solving this is to first ignore the monotonicity constraint and

    solve the problem with only the (IR) and LIC constraints. Define

    W() v[q()] T() = max

    {v[q()] T()}

    By the envelope theorem,

    dW()

    d

    =

    =

    W()

    =

    =v[q()]

    26

  • 8/10/2019 Contract Theory, by Alex Young

    27/86

    or, integrating,

    dW(x)

    dx dx=

    v[q(x)] dx

    W() W() =

    v[q(x)] dx

    W() =

    v[q(x)] dx+W()

    At the optimum, the participation constraint of the lowest type is binding. SinceT() =

    v[q()] W(), we can rewrite the sellers profits as

    =

    [T() cq()] f() d

    =

    [(v[q()] W()) cq()] f() d

    =

    v[q()]

    v[q(x)] dx cq()

    f() d

    =

    {v[q()] cq()} f() d

    v[q(x)] dxf() d

    Solve the second integral through integration by parts:

    u=

    v[q(x)] dx, dv= f() d

    du= v[q()] d, v= F()

    27

  • 8/10/2019 Contract Theory, by Alex Young

    28/86

  • 8/10/2019 Contract Theory, by Alex Young

    29/86

    From this equation, we make two observations:

    1. There is underconsumption < since first-best efficiency requires v[q()] =c.

    2. We can obtain an expression for the price-cost margin:

    P[q()] T()

    =v [q()]

    =c+1 F()

    f() v[q()]

    =c+1 F()

    f()

    P[q()]

    P cP

    =1 F()f()

    Recall that we have ignored the monotonicity constraint in the second equality. But we have

    to check that the optimal solution defined by (2.31) and the local incentive compatibility

    constraint satisfies the monotonicity constraint. In general, that depends on the form of

    the buyers utility function and/or on the form of the density function f(). A sufficient

    condition for the monotonicity constraint to be satisfied is that the hazard rate is increasing

    in :

    d

    dh()

    d

    d

    f()

    1 F()

    >0

    Letting

    g() 1

    h()

    = 1 F()

    f()

    29

  • 8/10/2019 Contract Theory, by Alex Young

    30/86

    the FOCs can be rewritten as

    {g()}v[q()] =c

    Implicitly differentiating this equation with respect to then yields

    g()v[q()]dq

    d+v[q()]g() = 0

    dq

    d=

    g()v[q()]

    v[q()]g()

    Since v() is concave and g()> 0 , dq/d 0 ifg()> 0. Hence, a sufficient condition

    for g ()> 0 is that 1/h() is decreasing in .

    2.3.2.2 Doesnt that look familiar? Well, it should. Recall the optimal auction design

    problem[Krishna, 2010]. The seller maximizes

    iN

    mi(0) +iN

    X

    xi

    1 Fi(xi)

    fi(xi)

    Qi(x)f(x) dx

    subject to

    mi(0) 0 (IR)

    dqi(xi)

    dxi 0 (IC-1)

    Ui(xi) =Ui(0) +

    xi0

    qi(ti) dti (IC-2)

    But (IC-1) is just a monotonicity condition. And (IC-2) is a local incentive compatibility

    constraint. Moreover, the design problem is regularif the virtual valuation is an increasing

    function ofxi, and a sufficient condition for regularity is that the hazard rate function, i(xi),

    30

  • 8/10/2019 Contract Theory, by Alex Young

    31/86

    is increasing:

    i(xi) xi 1

    i(xi)

    =xi 1 Fi(xi)fi(xi)

    2.4 Application: Costly State Verification

    Consider a financial contracting problem involving two risk-neutral agents:

    1. an entrepreneur with an investment project but no investment funds, and

    2. a financier with unlimited funds.

    The fixed investment requires a setup cost of I > 0 at t = 0 and generates random cash

    flows at t = 1 of [0, +), with p.d.f f(). The entrepreneur observes realized cash

    flows and can credibly disclose to the investor only by incurring a certification cost K >0.

    The contract-design problem is to specify in advance which cash-flow realizations should be

    certified.

    The set of contracts from which the contracting parties can choose follows: the contract

    specifies whether an audit should take place following the realization of cash flows and what

    fraction of realized cash flows the entrepreneur should pay back to the investor. Therevela-

    tion principlereduces the set of relevant contracts.

    Thus, following the realization of , the entrepreneur truthfully reveals , and the contract

    specifies the probability p() [0, 1] of certifying (auditing) cash flows. When there is no

    certification, the contract can only specify a repaymentr(). With certification, the contract

    specifies repayment contingent on both the announced cash flow and the true, certified

    cash flow : r(, ). ra() denotes the repayment for audited cash flows.

    31

  • 8/10/2019 Contract Theory, by Alex Young

    32/86

    A central result of the CSV approach to financial contracting is that standard debt contracts

    may be optimal financial contracts. But it is driven by fairly strong assumptions. For

    example, financial contracts are assumed to specify only deterministic certification policies:

    , p() {0, 1}

    That is, commitment to random audits is not feasible. How do we then show that a standard

    debt contract is an optimal financial contract?

    Assuming that the project has positive NPV, the optimal contracting problem reduces to

    minimizing expected audit costs, subject to meeting the entrepreneurs incentive constraints

    for truthful revelation, and the financiers participation:

    minp(),r(),ra()

    K

    +0

    p()f() d

    subject to

    +0

    p()[ra() K]f() d+

    +0

    [1 p()]r()f() d I (IR)

    ra(1) r(2)1=2 : p(1) = 1 and p(2) = 0 (IC-1)

    r(1) =r(2) =r 1=2: p(1) = 0 =p(2) (IC-2)

    ra() : p() = 1 (Limited wealth (A))

    r() : p() = 0 (Limited wealth (B))

    The incentive constraints imply that for any two cash-flow realizations that do not require

    certification, the repayment to the financer has to be the same: r(1) = r(2) = r. Oth-

    erwise, the entrepreneur could lie about realized cash flows and announce the one with the

    32

  • 8/10/2019 Contract Theory, by Alex Young

    33/86

    lower repayment. Similarly, a cash flow realization that requires certification should not en-

    tail a higher repayment than the repayment for a cash-flow realization that does not involve

    certification: ra(1) r(2). Otherwise, the entrepreneur could lie about the realization of

    1 and thus make a smaller repayment.

    The only dimensions along which the financial contract can be varied are the certification

    subset and the size of the repayment in the certification subset. Characterizing the optimal

    contract follows:

    1. Any feasible contract must include the cash-flow realization = 0 in the audit subset.

    Otherwise, the entrepreneur could always claim to have a cash flow of zero, thereby

    avoiding an audit as well as any repayments to the financier.

    2. Any contract that minimizes expected audit costs must be such that for any cash-flow

    realization in the audit subset,ra() = min{, r}.

    3. Any contract with a disconnected audit subset [0, ] [, ] (with < ) would be

    inefficient, since then an obvious improvement is available by shifting to a connected

    subset with the same probability mass.

    We conclude that the uniquely optimal financial contract, which minimizes expected audit

    costs, is such that

    1. there is a single connected audit region a = [0, ], with < ;

    2. over this region, audit repayments arera() =; and

    3. for cash flows >, which are not audited, the repayment isr = . The unique cutoff

    is given by the solution to the (IR) constraint

    0

    ( K)f() d+ [1 F()]r= I

    33

  • 8/10/2019 Contract Theory, by Alex Young

    34/86

    and expected audit costs are F()K.

    3 Hidden Information, Signaling

    3.1 Spences Model of Education as a Signal

    A workers productivity can be eitherrHorrL, withrH> rL> 0. Leti Pr(r= ri) [0, 1]

    be the firms prior belief that r = ri. Workers are willing to work at any wagew >0, and

    firms are willing to hire any worker at a wage less than the workers expected productivity.

    A worker of type i = L, Hcan gete years of education at costc(e) =iebefore entering the

    labor market. The key assumption is that H< L; the marginal cost of education is lower

    for high-productivity workers.

    In a first stage, the worker chooses education, and in a second stage, the wage is determined

    through bargaining. We assume that in the bargaining phase, the worker has all the bar-

    gaining power.

    Suppose that the workers productivity is perfectly observable. Then each worker type

    chooses a level of education ei = 0. Since the workers productivity is known, the highest

    offer the firm is willing to accept is wi = ri regardless of the level of education of the worker,

    and we have the first-best solution.

    Now suppose that productivity is not observable. Then the first-best solution can no longer

    be an equilibrium outcome. We specify precisely the game played by the worker and firm,

    as well as the notion of equilibrium:

    1. The worker chooses a level of education, possibly randomly, to maximize his expected

    return. pi(e) Pr(e|i) denotes the probability that the worker/principal of type i

    34

  • 8/10/2019 Contract Theory, by Alex Young

    35/86

    chooses education level e.

    2. The outcome is entirely driven by how the agents beliefs have been affected by the

    observation of the principals education level. (i|e) Pr(i|e) denotes the agents

    revised beliefs about productivity upon observing e. Then the equilibrium wage in the

    second stage is given by

    w(e) =(H|e)rH+ (L|e)rL

    where (L|e) = 1 (H|e). This is the maximum wage the firm is willing to pay

    given its updated beliefs.

    Imposing the minimum consistency requirements on the agents conditional beliefs leads to

    the definition of a perfect Bayesian equilibrium: a set of (possibly mixed) strategies

    {pi(e)} for the principals types and conditional beliefs (i|e) for the agent such that

    1. All education levels observed with positive probability in equilibrium must maximize

    workers expected payoff: e :pi(e)> 0,

    e arg maxe

    {(H|e)rH+ (L|e)rL ie}

    2. Firms posterior beliefs conditional on equilibrium education levels must satisfy Bayes

    rule:

    (i|e) = Pr(i|e)

    = Pr(e|i)Pr(i)

    Pr(e|H) + Pr(e|L) (Familiar now?)

    = pi(e)i2

    i=1 ipi(e)

    whenever pi(e)> 0 for at least one type.

    35

  • 8/10/2019 Contract Theory, by Alex Young

    36/86

    3. Posterior beliefs are otherwise not restricted: if pi(e) = 0 f or i = L, H (so that2i=1 ipie) = 0, and Bayes rule gives no prediction for posterior beliefs), then (i|e)

    can take any value in [0, 1].

    4. Firms pay workers their expected productivity:

    w(e) =(H|e)rH+ (L|e)rL

    Thus, an equilibrium outcome is a situation where the agents beliefs about what action

    each type of principal takes are correct (i.e. the action believed to be chosen by a given

    type is indeed the action chosen by that type) and where, given the agents updated beliefs(following the action), each type of principal is acting optimally.

    When a PBE is taken to be the outcome, essentially the only restriction imposed is that in

    equilibrium, the agents beliefs are consistent with the agents knowledge of the optimizing

    behavior of the principal. To solve for a PBE, one typically proceeds as follows:

    1. Using ones basic understanding of how the game works, one guesses (i.e. conjectures)

    conditional beliefs(i|e) for the agent.

    2. Then one determines the principals best response pi(e) given these beliefs.

    3. Finally one checks whether the beliefs (i|e) are consistent with the principals opti-

    mizing behavior.

    In signaling games, the difficulty is usually not to find aPBE. Rather, the problem is

    that there exist too many PBEs.

    3.1.1 Did you forget the beer quiche game?

    Lets use the beer quiche game as an example of how to solve for a PBE. As a student,

    I found my professors instruction to be frustrating at first. Conjecture an equilibrium and

    36

  • 8/10/2019 Contract Theory, by Alex Young

    37/86

    Figure 2: Image from Professor Meghan Busse.

    then verify it? What the hell is that?! Im supposed to know the answer before I solve for

    it?

    No, you are not supposed to know the answer from the onset. But that leads to another

    frustration during an exam: you have limited time, but several of your conjectures may

    turn out not to be equilibria. Well, hopefully when you write a paper, youre not under a

    ninety-minute time constraint, so lets proceed with the steps.

    There are two players in the game: the sender and the receiver. The sender is one individual

    who can be of two types: brave or coward. If the sender is the brave type, he prefers beer

    to quiche; if he is the coward type, he prefers quiche to beer. The receiver prefers to fight

    the coward type but not the brave type.

    1. Let us first conjecture a separating equilibrium: if brave, then beer; if coward, then

    quiche. This is the senders strategy.

    2. We can show by Bayes Rule that such a conjecture allows the receiver to perfectly

    37

  • 8/10/2019 Contract Theory, by Alex Young

    38/86

    distinguish between the types:

    Pr(tb | b) = Pr(b | tb)Pr(tb)

    Pr(b | tb)Pr(tb) + Pr(b | tc)Pr(tc)

    = 1 0.81 0.8 + 0 0.2

    = 1

    By the same procedure, we see that Pr(tb | q) = 0, Pr(tc | q) = 1, and Pr(tc | b) = 0.

    Hence, the receiver will not fight if he sees beer but will fight if he sees quiche. This is

    the receiversresponse.

    3. Knowing that the receiver will respond that way, will the sender still stick to the original

    strategy? No. The brave type gets his preferred breakfast without a fight, but the

    coward type will have to fight if he chooses quiche. If instead he lied and pretended

    to be the brave type, then he would receive a higher payoff due to not fighting, even

    though he prefers quiche to beer. So our conjectured separating equilibrium is not an

    equilibrium.

    It is easy to see that the other candidate separating equilibrium, if brave, then quiche; if

    coward, then beer, makes no sense. The coward type will again deviate, as doing so would

    yield the best payoff for him: he gets his preferred breakfast without a fight. What about

    pooling equilibria? Lets follow the steps again:

    1. Conjecture always beer.

    2. We can show by Bayes Rule that such a conjecture implies that the receiver cannot

    38

  • 8/10/2019 Contract Theory, by Alex Young

    39/86

  • 8/10/2019 Contract Theory, by Alex Young

    40/86

    To prevent the coward type from deviating, the response must be fight if quiche.

    Accordingly, >1 >1/2.

    3. Knowing that the receiver will respond that way, will the sender still stick to the orig-

    inal strategy? Yes. The brave type will not deviate as he receives his best payoff in

    equilibrium. The coward type might have deviated if he could have gotten the payoff

    of 3, but he cannot as the receiver will fight if he sees quiche. Thus, the coward type

    will not deviate either.

    There is another candidate pooling equilibrium, but we will save that for later.

    We conjecture the following PBE. If the observed education level is high, it is likely that

    the worker is type H. Even if a type L worker could obtain a wage w = rH by acquiring

    education, he would not be willing to choose a level of education above e, where e is given

    by1

    w c= rH Le

    =rL

    Thus a candidate is to set (H|e) = 1 e e and (H|e) = 0 e < e. If the principal

    optimizes against these beliefs, then he chooses e= 0 when he is of type L and e when he

    is of type H.

    There are many other PBEs, which can be classified into three categories:

    1. Separating PBEsThe signal chosen by the principal identifies the principals type

    exactly.

    1That is, e is a hurdle. Any level of education higher than e means that the worker is better off notpursuing any education at all: e >e, rH Le < rL.

    40

  • 8/10/2019 Contract Theory, by Alex Young

    41/86

    2. Pooling PBEsThe observed signal reveals no additional information about the prin-

    cipals type.

    3. Semiseparating PBEsSome, but not all, information about the principals tpye is

    obtained from the observation of the signal.

    Aseparating equilibriumis a PBE where each type of principal chooses a different signal

    in equilibrium: eH = eL : (H|eH) = 1, (L|eL) = 1, and wi = ri. The set of separating

    equilibrium levels of education is given by

    Ss=

    (eH, eL) :eL= 0 & eH

    rH rL

    L,rH rL

    H

    because incentive compatibility requires

    rH LeH rL

    rH HeH rL

    Therefore

    rH rLL

    eHrH rL

    H

    eH

    rH rL

    L,rH rL

    H

    A pooling equilibrium is a PBE where each type of principal chooses the same signal in

    equilibrium: eH=eL: (H|eH) =H, (L|eL) =L, andw(eH) =w(eL) =LrL+HrH

    r. The set of pooling equilibrium levels of education is

    Sp=

    (eH, eL) :eL= eH=ep & ep

    0,

    LrL+HrH rLL

    41

  • 8/10/2019 Contract Theory, by Alex Young

    42/86

    because incentive compatibility requires

    r Lep rL

    Therefore

    r rLL

    =LrL+HrH rL

    L

    ep

    0

    ep

    0,LrL+HrH rL

    L

    Finally, there also exists a set of semiseparating equilibria, where at least one type of principal

    is mixing between two signals, one of which is also chosen with positive probability by the

    other type of principal. The key requirement here is that, for mixing to be optimal, the type

    of principal who is mixing has to be indifferent between the two signals that are played with

    positive probability.

    3.1.2 Refinements

    A large body of research in game theory has attempted to develop criteria based on theo-

    retical considerations alone for selecting a particular subset of equilibria. The basic idea is

    to enrich the specification of the game by introducing restrictions on the set of allowable

    off-equilibrium beliefs using the observation that some types of players are more likely to

    choose some deviations than others (i.e. who wants to deviate?)

    3.1.2.1 Cho and Kreps Intuitive Criterion The basic observation this selection cri-

    terion builds on is that most deviations from equilibrium play can never be in the interest

    of some principal types. Beliefs conditional on off-equilibrium actions, therefore, ought to

    42

  • 8/10/2019 Contract Theory, by Alex Young

    43/86

    reflect that these actions are more likely to be chosen by some types than others. That is,

    beliefs conditional on off-equilibrium actions must be restricted to reflect that only some

    types are ever likely to choose these actions.

    Formally (i.e. use this in papers, not workshops), the Cho-Kreps intuitive criterion

    follows: Letui = wi (ei) iei denote the equilibrium payoff of type i. Then, (j|e) = 0,

    for e = (ei; ej), whenever rH je < uj and rH ie ui (i= L, H; i =j).

    In a language comprehensible to mortals, the intuitive criterion states that when a deviation

    isdominatedfor one type of player but not the other, the deviation should not be attributed

    to the player for which it is dominated. By dominated, one means that the player is getting

    a worsepayoff than his equilibrium payoff for anybelief of the uninformed party following

    the deviation.

    Perhaps an example will help illuminate or remind you about the intuitive criterion. Recall

    the beer quiche game:

    1. Lets finish with the final conjecture: always quiche.

    2. By the same procedure, the receiver does not learn the senders type since both types

    43

  • 8/10/2019 Contract Theory, by Alex Young

    44/86

    choose the same message. Fighting yields

    0 0.8 + 1 0.2 = 0.2

    Not fighting yields

    1 0.8 + 0 0.2 = 0.8

    Therefore, the receiver will choose to not fight if quiche. We again must specify

    off-equilibrium probabilities. Let denote Pr(tb | b), with Pr(tc | b) = 1 . Fighting

    yields

    0 + 1 (1 ) = 1

    Not fighting yields

    1 + 0 (1 ) =

    To prevent the brave type from deviating, the response must be fight if beer. Ac-

    cordingly, 1 >

  • 8/10/2019 Contract Theory, by Alex Young

    45/86

    to beer from quiche came from the coward type.

    That makes no sense. The coward type would NEVER deviate to beer; he already

    received his best payoff in equilibrium. The probability that the sender is the coward

    type given that the receiver sees beer should not be greater than 1/2; it should be

    zero.

    Therefore, while always quiche is a perfect Bayesian equilibrium, it does not survive

    refinement via the intuitive criterion. Note that in the always beer case, the brave

    type had no reason to deviate, and accordingly, the probability that that the sender

    was the brave type given that the receiver saw quiche was less than 1/2. Hence,

    always beer survives the intuitive criterion.

    The intuitive criterion predicts implausible equilibrium outcomes in some situations. The

    least-cost separating equilibrium in which eL = 0 and eH = (rH rL)/L is the same for

    all i >0; i= L, H. Suppose now that L is arbitrarily small (L= 0). Then it seems

    excessive to payc(eH; H) =H(rHrL)/Ljust to raise the waige by a commensurately small

    amount [w= rH (1 )rH rL= (rH rL) 0]. Then the pooling equilibrium where

    no education costs are incurred seems more plausible as it is Pareto-dominant. Moreover,

    note that with no adverse selection (i.e. L = 0), eL = eH = 0 in equilibrium, so the

    pooling equilibrium without education is the limit of this complete-information case, not the

    Cho-Kreps equilibrium.

    3.1.2.2 Maskin and Tiroles Informed Principal Problem The problem of multi-

    plicity of PBEs is reduced when the timing of the game is changed so as to let the principal

    offer the agent a contingent contract before the choice of the signal. Consider the linear

    model of education specified previously, but invert the stages of contracting and education

    choice. That is,

    45

  • 8/10/2019 Contract Theory, by Alex Young

    46/86

    1. now the worker offers his employer a contract before undertaking education.

    2. This contract then specifies a wage schedule contingent on the level of education chosen

    by the worker after signing the contract.

    Let {we} denote the contingent wage schedule specified by the contract. There are two

    different cases to consider:

    1. r = HrH+ LrL rH H(rH rL)/L. In this case a type Hworker is better off

    in the least-cost separating equilibrium than in the efficient pooling equilibrium.

    2. r > rH H(rH rL)/L. Here the typeHworker is better off in the efficient pooling

    equilibrium.

    4 The Principal-Agent ProblemMas-Colell et al.[1995]

    4.1 Comment

    One thing about contract theory (and perhaps teaching game theory more generally) is that

    some people do certain topics way, way, way better than others. We can see this in textbooks.

    When I took Micro 1, I found the consumer choice, classical demand theory, and production

    chapters to be unreadable; I considered it a testament to the readability ofJehle and Reny

    [2001] that I could read the sections in that book and still do the assigned problems in

    Mas-Colell et al. [1995]2. But Chapter 14 in Mas-Colell et al. is amazingly readable. The

    notes accordingly follow Mas-Colell et al. for the basic treatment instead of Bolton and

    Dewatripont[2004].2Of course by do I meant that I stared at the page for half an hour, took a stab at how to begin, and

    then consulted the solutions manual.

    46

  • 8/10/2019 Contract Theory, by Alex Young

    47/86

    4.2 Hidden Actions (Moral Hazard)

    The owner of a firm (the principal) wishes to hire a manager (the agent) for a project. The

    projects profits are affected in part by the managers actions. How should the principal

    design the managers compensation in a way that indirectly gives him (the manager) the

    incentive to take the desired action when the managers actions are unobservable?

    Let denote the projects observable profits, and let e E R denote the managers

    action choice. For the nonobservability of managerial effort to matter, e must not be per-

    fectly deducible from observing . We thus assume that although the projects profits are

    affected by e, they are not fully determined by it. Profit can take values in [, ] and

    it is stochastically related to e in a manner described by f( | e) with f( | e) > 0 for all

    e Eand all [, ]. Put another way, any realization ofcan arise following any givene.

    We focus on the case in which the manager has only two possible effort choices, eH and

    eL. eH leads to a higher profit level for the firm than eL but entails greater difficulty for

    the manager. That is, we assume that the distribution of conditional on eH first-order

    stochastically dominates the distribution conditional on eL:

    F( | eH) F( | eL) f( | eH) d >

    f( | eL) d

    The manager is an expected utility maximizer with a Bernoulli utility funciton u(w, e) over

    47

  • 8/10/2019 Contract Theory, by Alex Young

    48/86

    his wage and effort level with the following properties:

    u(w, e) =v(w) g(e)

    v

    (w)> 0 (Prefers more income to less)

    v(w) 0 (Weakly risk averse)

    g(eH)> g(eL) (Dislikes a high level of effort)

    The owner receives the projects profits less any wage payments made to the manager. We

    assume that the owner is risk neutral; his objective is to maximize his expected return.

    4.2.1 The Optimal Contract when Effort is Observable

    The owner chooses a take it or leave it contract to offer the manager. The optimal contract

    for the owner then solves the following problem:

    maxe{eL,eH},w()

    ( w())f( | e) d

    s.t.

    v(w())f( | e) d g(e) u

    We write the Lagrangian corresponding to a discrete version of the model and take the FOC:

    L =i

    (i wi)pi,e+

    i

    v(wi)pi,e g(e) u

    L

    wi= pi,e+(v

    (wi)pi,e)

    = 0

    = 1

    v(wi)

    = 1

    v(w()) (Borch rule)

    48

  • 8/10/2019 Contract Theory, by Alex Young

    49/86

    Exercise 14.B.1 asks us whether the reservation utility constraint must be binding in an

    optimal contract. We can see this a few ways. First, from the previous derivation, since

    v() > 0, >0. This implies that the constraint must bind. Second, if the constraint did

    not bind, then the owner could reduce the wage payment while still satisfying the constraint

    andmaking himself strictly better off. But that would mean that such a wage payment was

    not optimal.

    If the manager is strictly risk averse (i.e. v (w)

  • 8/10/2019 Contract Theory, by Alex Young

    50/86

    utility level:

    v(we) g(e) = u

    v(we) = u+g(e)

    we =v1(u+g(e))

    g(eH)> g(eL)

    we,H> we,L

    Thus, in the principal-agent model with observable managerial effort, an optimal contractspecifies that the manager choose e that maximizes

    f( | e) d v1(u+g(e))

    and pays the manager a fixed wage

    w

    =v1

    (u+g(e

    ))

    4.2.1.1 A risk-averse owner Exercise 14.B.2 asks us to derive the FOC characterizing

    the optimal compensation scheme for a two-effort-level hidden action (i.e. effort unobserv-

    able) model with a strictly risk averse principal. Let V() denote the owners utility function

    over wealth. As before, the owner chooses a take it or leave it contract to offer the manager.

    50

  • 8/10/2019 Contract Theory, by Alex Young

    51/86

    The optimal contract for the owner then solves the following problem:

    maxe{eL,eH},w()

    V( w())f( | e) d

    s.t.

    v(w())f( | e) d g(e) u (Individual Rationality)

    e arg maxe

    v(w())f( | e) d g(e) (Incentive Compatibility)

    Note that the IC constraint can be written as

    v(w())f( | eH) d g(eH)

    v(w())f( | eL) d g(eL)

    We write the Lagrangian corresponding to a discrete version of the model and take the FOC:

    L =i

    V(i wi)pi,eH+

    i

    v(wi)pi,eH g(e) u

    +

    i

    v(wi)pi,eH g(eH) i

    v(wi)pi,eL+ g(eL)

    L

    wi= pi,eHV

    (i wi) +(v(wi)pi,eH) +(v

    (wi)pi,eH v(wi)pi,eL)

    = 0

    +

    1

    pi,eLpi,eH

    =

    V(i wi)

    v(wi)

    +

    1

    f( | eL)

    f( | eH)

    =

    V( w())

    v(w()) (Borch rule)

    4.2.2 The Optimal Contract when Effort is Not Observable

    When effort is not observable, the only way to get the manager to take the desired action is to

    relate his pay to the realization of profits, which we assumed was random. Nonobservability

    of effort can thus lead to inefficiencies.

    51

  • 8/10/2019 Contract Theory, by Alex Young

    52/86

    4.2.2.1 A risk-neutral manager When the manager has no concern about risk bearing,

    the owner can still achieve the same outcome as when effort is observable. Suppose that the

    owner offers a compensation schedule of the form

    w() = (Selling the project to the manager)

    If the manager accepts this contract, he chooses e to maximize his expected utility

    w()f( | e) d g(e) =

    f( | e) d g(e)

    But whateverethat maximizes the above expression is also theethat maximizes the owners

    profit given observable effort and a risk-neutral manager3:

    f( | e) d g(e) u

    Thus, w() = induces the first-best effort level e. The manager is willing to accept

    it as long as it gives him an expected utility of at least u; that is, if

    f( | e) d g(e) u

    Let be the level at which the above inequality holds with equality. Note that the owners

    payoff givenw() = is . We rearrange

    f( | e) d g(e) = u

    f( | e) d g(e) u=

    But that is the maximized value of the owners expected profit given a risk-neutral manager

    and w() = . Both the owner and the manager get the same payoff as when effort is

    3The and the u are both constants that drop out in the FOC with respect to e.

    52

  • 8/10/2019 Contract Theory, by Alex Young

    53/86

    observable.

    4.2.2.2 A risk-averse manager Now, incentives for high effort can be provided only

    at the cost of having the manager face risk. The principal solves

    maxe{eL,eH},w()

    ( w())f( | e) d

    s.t.

    v(w())f( | e) d g(e) u (Individual Rationality)

    e arg maxe

    v(w())f( | e) d g(e) (Incentive Compatibility)

    As before, note that the IC constraint can be written as

    v(w())f( | eH) d g(eH)

    v(w())f( | eL) d g(eL)

    We write the Lagrangian corresponding to a discrete version of the model and take the FOC:

    L =

    i(i wi)pi,eH+

    i

    v(wi)pi,eH g(e) u

    +

    i

    v(wi)pi,eH g(eH) i

    v(wi)pi,eL+g(eL)

    L

    wi= pi,eH+(v

    (wi)pi,eH) +(v(wi)pi,eH v

    (wi)pi,eL)

    = 0

    +

    1

    pi,eLpi,eH

    =

    1

    v(wi)

    +

    1 f( | eL)

    f( | eH)

    =

    1

    v(w()) (Borch rule)

    Note that we can draw comparisons between the first-best and the second-best. We letw

    53

  • 8/10/2019 Contract Theory, by Alex Young

    54/86

    be the fixed wage payment such that

    1

    v(

    w)

    =

    Iff( | eL)/f( | eH)< 1, then

    1

    v(w())=+

    1

    f( | eL)

    f( | eH)

    =

    1

    v(w)+

    1 f( | eL)

    f( | eH)

    =

    1

    v(

    w)

    +

    1

    f( | eL)

    f( | eH)

    >0

    1

    v(w())>

    1

    v(w)

    v(w())< v(w)

    w()>w (v() 0)By the same procedure, we conclude that iff( | eL)/f( | eH)>1, then

    w()

  • 8/10/2019 Contract Theory, by Alex Young

    55/86

    4.2.3 Comparing the first and second-best wage contracts

    Given the variability introduced into the managers compensation in the second-best scenario,

    the expected value of the managers wage payment must be strictly greater than the fixed

    wage payment in the first-best. Why? The manager must be assured an expected utility

    level of uor else he will not participate. The owner subsequently compensates him through

    a higher average wage payment for any risk he bears.

    v(E[w() | eH])> E[v(w()) | eH] (v() 0)

    =

    v(w())f( | eH) d

    = u+g(eH) (IR constraint)

    =v(weH)

    E[w() | eH]> weH

    (v()>0)

    To reiterate, from the owners perspective, if he wishes to induce the manager to perform

    eHwhen effort is unobservable, on average, he must pay more than in the case where effort

    is observable. Nonobservability of effort can thus lead to an inefficiently low level of effort

    being implemented.

    4.2.3.1 An additional demonstration of the implications of risk aversion This

    comes fromBolton and Dewatripont [2004]. Take a standard moral hazard problem where

    the principal considers offering the agent a lottery rather than a fixed payment wi if output

    qi is observed. Specifically, the agent would receive wi,j with probability pi,j 0, with

    j = 1, 2, . . . , mandm

    j=1

    pi,j = 1

    Assume the agents utility function isu(w) (a). Show that such a randomizing incentive

    55

  • 8/10/2019 Contract Theory, by Alex Young

    56/86

    scheme cannot be optimal if the principal is risk neutral and the agent is strictly risk averse.

    Proof. Fix an output qi, i {L, H}.

    v(Qwi,1+ (1 Q)wi,2)> Qv(wi,1) + (1 Q)v(wi,2)

    =v(wi)

    Qwi,1+ (1 Q)wi,2> wi

    Therefore, since there is another wage payment for a given output that makes the agent

    indifferent but at less cost to the principal, the lottery is not optimal. This leads us to the

    next section, which is in fact where I found out how to solve this problem.

    4.2.4 The Value of Information

    Suppose that another statistical signal of effort, y, is available to the owner in addition to

    the realization of profits (i.e. the outcome), and that the joint density of and y given e

    (action / effort) is f(, y | e). In principle, the agents wage could depend on both output

    and this additional signal. When should that be the case? That is, when should the optimal

    compensation function depend ony?

    We can derive the Borch rule given that w is a function of and y :

    +

    1

    f(, y | eL)

    f(, y | eH)

    =

    1

    v(w(, y))

    Suppose that y is simply a noisy random variable that is unrelated to e (cf. the lottery

    56

  • 8/10/2019 Contract Theory, by Alex Young

    57/86

    problem inBolton and Dewatripont[2004]). In that case,

    f(, y | e) =f1( | e)f2(y)

    But then

    +

    1

    f(, y | eL)

    f(, y | eH)

    =

    1

    v(w(, y))

    +

    1

    f1( | eL)f2(y)

    f1( | eH)f2(y)

    =

    1

    v(w(, y))

    +

    1

    f1( | eL)

    f1( | eH)

    =

    1

    v(w(, y))

    Thus, the optimal compensation package is independent of y. Why? Suppose that the

    principal does offer a contract that has wage payments depend ony. But ally does is induce

    additional uncertainty in the agents wage that is unrelated to effort; that is, the agent bears

    additional risk for no reason. If the principal instead offers, for each outcome, the certain

    payment w() such that

    v(w()) = E[v(w(, y)) | ]

    then the agent gets exactly the same expected utility under w() as under w(, y) for any

    level of effort. But the agent faces less risk under the certain payment w() per outcome

    contract. Thus, the expected wage payments ar elower and the principal is better off under

    that contract than the one that depends on y :

    v(E[w(, y) | ])> E[v(w(, y)) | ]

    =v(w())

    E[w(, y) | ]> w() (v()>0)

    57

  • 8/10/2019 Contract Theory, by Alex Young

    58/86

    4.3 Linear Contracts, Exponential Utility, and Normally Distributed

    Performance (LEN) [Bolton and Dewatripont,2004]

    Performance is assumed to be equal to effort plus noise:

    q= a+

    where is normally distributed with zero mean and variance 2. The principal is assumed

    to be risk neuteral. The agent has CARA risk preferences represented by

    u(w, a) = exp([w (a)])

    where w is the amount of monetary compensation and > 0 is the agents coefficient of

    absolute risk aversion ( = u/u from Arrow and Pratt). Effort cost is measured in

    monetary units. For simplicity, the cost-of-effort function is assumed to be quadratic:

    (a) =1

    2ca2

    Suppose that the principal and agent can write only linear contracts of the form

    w= t+sq

    wheret is the fixed compensation level and s is the variable, performance-related component

    of compensation. The principals problem is then to solve

    maxa,t,s

    E(q w)

    58

  • 8/10/2019 Contract Theory, by Alex Young

    59/86

  • 8/10/2019 Contract Theory, by Alex Young

    60/86

    subject to

    t+sa 1

    2ca2

    1

    2s22 =w (IR)

    Plugging int and taking the FOC with respect to s, we have

    s= 1

    1 +c2

    Effort and the variable compensation component goes down as c, , and 2 increases.

    4.3.1 Did you forget the properties of the log-normal distribution?

    All you really need to remember, thanks to Wikipedia, is that ifx N(, 2), then exp(x)

    is distributed lognormally with mean

    exp

    E[x] +

    1

    2Var[x]

    = exp

    +

    1

    22

    4.4 Moral Hazard in Teams [Bolton and Dewatripont,2004]

    A risk-neutral principal has a contract with n 2 agents, each of whom has the following

    utility function separable in income and effort:

    ui(wi) i(ai)

    ui()>0

    ui () 0

    i()>0

    i() 0

    60

  • 8/10/2019 Contract Theory, by Alex Young

    61/86

    The agents actions produce outputs:

    qn1= (q1, . . . , q n)

    F(q| a) (Joint CDF)

    an1= (a1, . . . , an) (Vector of actions taken by the agents)

    We consider two cases:

    1. Moral hazard in a team- the output produced by the agents is a single aggregate, Q,

    with conditional distribution F(Q | a).

    2. Each agent produces an individual random output qi, which may be imperfectly cor-

    related with the other agents outputs.

    4.4.1 Unobservable Individual Outputs

    n 2 agents form a partnership to produce some deterministic (scalar) aggregate output

    Q= Q(a)

    by supplying a vector of individual hidden actions a. We assume that the output function

    has the following properties:

    Q

    ai>0 (Strictly increasing)

    2Q

    a2i

  • 8/10/2019 Contract Theory, by Alex Young

    62/86

    We suppose further that all agents are risk neutral:

    ui(w) =w

    A partnership is a vector of output-contingent compensations for each agent, w(Q) =

    [w1(Q), . . . , wn(Q)], such thatn

    i=1

    wi(Q) =Q (Budget balance)

    When effort levels are unobservable and no measure of individual performance is available,

    rewarding an agent for raising aggregate output means rewarding her higher effort and also

    the higher effort of other agents. Thus, there is a free-riding problem.

    We know that a risk-neutral individual agent would supply the first-best level of effort, given

    that all the other agents supply the first-best, if he is compensated with the full marginal

    return of his effort. The level of effort that maximizes social welfare in the first best is

    defined by

    a = arg maxa

    Q(a)

    ni=1

    i(ai)

    From the FOC, at the optimum, we have

    Q(a)

    ai i(a

    i ) = 0 i

    Fix an arbitrary partnership contract w(Q). Each agent i independently chooses ai to

    maximize his own utility given the other agents actions, ai. Thus, agent i solves

    maxai

    wi(Q(ai, ai)) i(ai)

    62

  • 8/10/2019 Contract Theory, by Alex Young

    63/86

    From the FOC, we have

    dwi(Q(ai, ai))

    dQ

    Q(ai, ai)

    ai i(ai) = 0

    It seems that we may achieve a Nash equilibrium where all agents supply first-best effort

    levels ifeach agent gets a compensation contract wi[Q(ai, ai)] providing him with the full

    marginal return from his effort, when all other agents also supply the first-best effort level:

    dwi(Q(ai, ai))

    dQ = 1

    But that implies

    dwi(Q(ai, ai)) = dQ

    wi(Q(ai, ai)) =Q+k

    Sincen

    i=1(Q+k) =nQ+nk > Q

    we cannot implement such a contract while satisfying budget balance. What can we do?

    Introduce a budget breaker into the organization. A third party agrees to sign a contract

    with all n agents offering to pay each of them wi(Q) = Q. There then exists a Nash

    equilibrium where all agents supply their first-best action and where the budget breaker

    pays nQ(a) in equilibrium.

    Proof. The proof is due to S. Viswanathan. If all other agents pickai

    and agent i picks

    ai > ai , he gets nothing more even though he does produce more. Due to costly action,

    i() >0, this is suboptimal. If instead agent i picks ai < ai , then he gets nothing. (And

    neither does anyone else?) Therefore, a is Nash.

    63

  • 8/10/2019 Contract Theory, by Alex Young

    64/86

    4.4.2 Observable Individual Outputs

    Previously, the problem was one of eliciting cooperation / avoiding free riding. We now turn

    to the problem of harnessing competition between agents. When does an optimal linear

    incnetive scheme for agent i also take int oaccount agent j s performance? We consider the

    setting in which both agents have CARA risk preferences, normally distributed individual

    outputs, and incentive contracts linear in output.

    We have two identical agents, each producing an individual output qi by supplying effortai:

    q1= a1+1+2

    q2= a2+2+1

    1,2 N(0, 2) (Independently and normally distributed)

    When = 0, the outputs are correlated. Hence, it should be optimal to base an individual

    agents compensation on both output realizations, as both provide information about an

    individual agents action choice.

    The principal is risk neutral, and both agents have CARA risk preferences as follows:

    u(w, a) = exp( [w (a)])

    ais the agents effort

    is the coefficient of absolute risk aversion

    (a) =12

    ca2

    64

  • 8/10/2019 Contract Theory, by Alex Young

    65/86

    We restrict attention to linear incentive contracts of the form

    w1= z1+v1q1+u1q2

    w2= z2+v2q2+u2q1

    The principals problem is symmetric. Therefore, we illustrate his maximization problem in

    relation with agent 1:

    maxa1,z1,v1,u1

    E(q1 w1)

    s.t. E [ exp { [w1 (a)]}] u(w) (IR)

    a1 arg maxa

    E [ exp { [w1 (a)]}] (IC)

    We can rewrite the agents maximization problem:

    E [ exp { [w1 (a)]}] = E [ exp { [z1+v1q1+u1q2 (a)]}]

    = E [ exp { [z1+v1(a1+1+2) +u1(a2+2+1) (a)]}]

    E [ exp {w1(a)}]= exp

    E (w1(a)) +1

    2Var (w1(a))

    = exp

    z1+v1a1+u1a2

    1

    2ca21

    +

    1

    22Var [v11+v12+u12+u11]

    = exp

    z1+v1a1+u1a2

    1

    2ca21

    +1

    22Var [(v1+u1)1+ (u1+v1)2]

    = exp

    z1+v1a1+u1a2 1

    2ca21

    +1

    222

    (v1+u1)

    2 + (u1+v1)2

    = exp

    z1+v1a1+u1a2

    1

    2ca21

    1

    22

    (v1+u1)

    2 + (u1+v1)2

    65

  • 8/10/2019 Contract Theory, by Alex Young

    66/86

    Solving the above is equivalent to solving5

    maxa1

    z1+v1a1+u1a2

    1

    2ca21

    1

    22

    (v1+u1)

    2 + (u1+v1)2

    From the FOC, we have

    v1 ca1 = 0

    a1 =v1

    c

    We plug this back in:

    z1+v1

    v1c

    +u1v2c

    1

    2cv1

    c

    2

    1

    22

    (v1+u1)

    2 + (u1+v1)2

    =

    z1+

    v 21c

    +u1v2

    c

    1

    2

    v21c

    1

    22

    (v1+u1)

    2 + (u1+v1)2

    =

    z1+

    1

    2

    v21c

    +u1v2

    c

    1

    22

    (v1+u1)

    2 + (u1+v1)2

    =w (Certainty equivalent, reservation utility)

    The principals problem can now be written as

    maxa1,z1,v1,u1

    E(q1 w1) = maxz1,v1,u1

    E (a1+1+2 [z1+v1q1+u1q2])

    = maxz1,v1,u1

    E

    v1c

    z1+

    v21c

    +u1v2

    c

    s.t.

    z1+

    1

    2

    v21c

    +u1v2

    c

    1

    22

    (v1+u1)

    2 + (u1+v1)2

    = w

    5

    Draw the graph of exp(x) if you ever forget / are too drunk to remember.

    66

  • 8/10/2019 Contract Theory, by Alex Young

    67/86

    Since the constraint binds, we can rewrite it as an unconstrained problem:

    maxz1,v1,u1

    E

    v1c

    1

    2

    v21c

    u1v2

    c +

    1

    22

    (v1+u1)

    2 + (u1+v1)2

    +

    v 21c

    +u1v2

    c

    =max

    v1,u1Ev1

    c 1

    2v

    2

    1

    c 1

    22 (v1+u1)2 + (u1+v1)2

    We solve this sequentially:

    1. For a given v1,u1 is determined to minimize risk.

    2. v1 is then set optimally to trade off risk sharing and incentives.

    We minimize the variance with respect to u1:

    u1

    1

    22

    (v1+u1)

    2 + (u1+v1)2

    =1

    22 [2(v1+u1)+ 2(u1+v1)]

    = 0

    u1 = 22 + 1

    v1We postpone the interpretation ofu1. For now, we substitute u1 in and take the FOC with

    respect to v1:

    maxv1,u1

    E

    v1c

    1

    2

    v21c

    1

    22

    v1+

    2

    2 + 1

    v1

    2+

    2

    2 + 1

    v1

    +v1

    2

    =maxv1,u1

    Ev1c 12 v21c 122v21 1 222 + 12

    +1 22 + 12

    2=max

    v1,u1E

    v1c

    1

    2

    v21c

    1

    22v21

    1 2

    2 + 1

    2(1 +2)

    =maxv1,u1

    E

    v1c

    1

    2

    v21c

    1

    22v21

    (1 2)2

    1 +2

    67

  • 8/10/2019 Contract Theory, by Alex Young

    68/86

    We get

    1

    c

    1

    cv1

    2 (1 2)2

    1 +2 v1 = 0

    1 v1 c2 (1

    2

    )

    2

    1 +2 v1 = 0

    v1 = 1

    1 +c2(1 2)2

    1 +2

    = 1 +2

    1 +2 +c2(1 2)2

    >0 (1 1)

    When= 0, v1 reduces to1

    1 +c2

    the expression for the variable wage component in the LEN case. When = 0, the correlation

    in the two agents outputs can reduce the overall risk exposure of any individual agent and

    thus enable the provision of stronger incentives to both agents. To see this, consider the

    extremes where = 1: v1 = 1. That is, by filtering out a common shock, the optimal

    incentive scheme can almost eliminate each agents exposure to risk and approximate first-

    best incentives by letting v1 tend to 1 (i.e. full marginal return). Equilibrium effort is

    higher because of lower risk exposure.

    4.5 Combining Moral Hazard and Adverse Selection

    Consider the problem of a risk-neutral seller of a firm interacting with a risk-neutral buyer.

    The buyer can generate an uncertain revenue stream by running the firm: X {0, R}. The

    buyer may be more or less able at running the firm, and ability translates into a higher

    or lower probability of getting the outcome R > 0. In addition, the buyer can raise the

    probability of gettingRby working harder.

    68

  • 8/10/2019 Contract Theory, by Alex Young

    69/86

    Let denote ability, and suppose that {L, H} with L < H. The seller does not

    know the buyers type, and his prior is that the buyer has ability H (L) with probability

    (1 ). e denotes effort, and a buyer with ability who supplies effort e generates R with

    a probabilitye at a private cost of

    (e) 1

    2ce2

    To summarize, the setup is a model with

    1. two states (or outcomes),

    2. two types, and

    3. a continuum of actions.

    The sellers problem is to offer a menu of contracts (ti, ri) to maximize his expected re-

    turn from the sale, where ti denotes an up-front cash payment and ri a repayment to be

    paid from future revenues generated by the buyer. We assume limited liability: X ri 0.6

    The buyers payoff under such a contract takes the form:

    iei(R ri) ti 1

    2ce2i

    The FOC gives the choice of effort that maximizes his payoff:

    i(R ri) cei= 0

    ei=i(R ri)

    c

    Bolton and Dewatripont [2004]note that the optimal ei is independent ofti and decreasing

    6That is, the repayment cannot exceed the realization of revenue.

    69

  • 8/10/2019 Contract Theory, by Alex Young

    70/86

    in ri7. Plugging that in to the payoff produces the maximized payoff:

    ii(R ri)

    c (R ri) ti

    1

    2c

    i(R ri)

    c

    2=1

    22i (R ri)2

    c ti

    The seller solves

    max(ti,ri)

    {[tH+ HeHrH] + (1 ) [tL+Le

    LrL]}

    =max(ti,ri)

    tH+ H

    H(R rH)

    c rH

    + (1 )

    tL+L

    L(R rL)

    c rL

    =max

    (ti,ri)tH+ 2H(R rH)

    c rH + (1 ) tL+ 2L(R rL)

    c rL

    subject to

    1

    2

    2i (R ri)

    2

    c

    ti

    1

    2

    2i (R rj)

    2

    c

    tj (IC)

    1

    2

    2i (R ri)

    2

    c

    ti 0 (IR)

    4.5.1 Optimal Contract, Moral Hazard Only

    When there is no adverse selection problem that is, the seller can observe the buyers

    ability / type he can offer ability-specific contracts:

    max{ti,ri}

    ti+

    2i (R ri)ric

    subject to

    1

    2c [i(R ri)]

    2

    ti 0 (IR)7Its obvious, but think of an extreme case when ri = R. In that case, the seller demands everything.

    Would the buyer work in such a case?

    70

  • 8/10/2019 Contract Theory, by Alex Young

    71/86

    Since the participation constraint binds, the seller solves

    max{ri}

    1

    2c[i(R ri)]

    2 +2i (R ri)ri

    c

    We take the FOC:

    1

    c[i(R ri)](i) +

    1

    c2i [(R ri) ri] = 0

    R+ri+R 2ri = 0

    ri = 0

    ri = 0

    This implies that

    ti = 1

    2c[i(R ri)]

    2

    = 1

    2c[iR

    2]

    That is, the seller sells the firm for an upfront cash payment and does not maintain anyadditional financial participation / future ownership.

    4.5.2 Optimal Contract, Adverse Selection Only

    We now suppose that the buyers effort level is fixed at some levelebut that the seller cannotobserve the buyers ability.

    max{ti,ri}

    {(tH+HerH) + (1 )(tL+LerL)}subject to

    ie(R ri) ti ie(R rj) tj (IC)ie(R ri) ti 1

    2ce2 0 (IR)

    71

  • 8/10/2019 Contract Theory, by Alex Young

    72/86

    Bolton and Dewatripont[2004] claim that the problem has a simple solution:

    ri= R

    ti=

    1

    2ce2And indeed it does8. How? Lets follow the algorithm from before.

    1. A less-productive agent always has to be paid more than a more-productive agent.

    Hence, the more-productive agent has the incentive to deviate / lie.

    2. Thus, IR-L and IC-H bind.

    3. We solve for tL and tH:

    Le(R rL) tL 12

    ce2 = 0 (IR-L)tL= Le(R rL) 1

    2ce2

    H

    e(R rH) tH=H

    e(R rL) tL (IC-H)

    tH=He(R rH) He(R rL) +Le(R rL) 1

    2

    ce24. We plug these in to the maximization problem to simplify it:

    max{ri}

    He(R rH) He(R rL) +Le(R rL) 1

    2ce2 +HerH

    +(1)

    L

    e(R rL)

    1

    2c

    e2 +L

    erL

    8I thank Zeqiong Huang for alerting me to the corner solution.

    72

  • 8/10/2019 Contract Theory, by Alex Young

    73/86

    5. We take the FOC with respect to rH:

    H

    e+H

    e= H

    e(1 )

    >0

    6. We do the same with respect to rL:

    (He Le) =(H L)e>0

    7. What does it mean when both FOCs are strictly positive? It means that we have a

    corner solution: keep increasing bothri until they cant be increased any more. This

    happens whenri = R.

    Intuitively, the solutions to the two subcases have different interpretations. When there is

    moral hazard but no adverse selection, the seller literally sells the firm to the buyer. He

    accepts an upfront payment and maintains no additional ownership. This is reminiscent of

    how moral hazard can be avoided even when effort is unobservable and can only be imper-

    fectly deduced from output so long as the agent is risk-neutral, for then the principal can

    simply sell the project to the agent.

    When there is adverse selection but no moral hazard, the seller does not actually sell anything

    to the buyer. Rather, he employs the buyer. In exchange for a payment note thatti

  • 8/10/2019 Contract Theory, by Alex Young

    74/86

    4.5.3 Optimal Contract, Moral Hazard and Adverse Selection

    The problem becomes, unsurprisingly, a combination of the two subcases we have considered,

    and it is none other than the problem we set up at the beginning:

    max(ti,ri)

    tH+

    2H(R rH)

    c rH

    + (1 )

    tL+

    2L(R rL)

    c rL

    subject to

    1

    2

    2i (R ri)

    2

    c

    ti

    1

    2

    2i (R rj)

    2

    c

    tj (IC)

    1

    2

    2i (R ri)

    2

    c

    ti 0 (IR)

    We first comment on why the previous solutions are not optimal when both problems exist.

    Recall that when there is no adverse selection but there is moral hazard, there are no future

    repayments, but there is a fixed upfront cost that is type-specific:

    tH=2HR

    2

    2c

    >2LR

    2

    2c

    =tL

    Given that rH=rL= 0, if the seller offered the moral hazard only menu in a dual-problem

    setup, the high type would never choose (rH, tH). Why should he? If he chooses (rL, tL), he

    can pay less for the same thing.

    Now, recall that when there is no moral hazard but there is adverse selection, the seller paysthe buyer upfront, who in turn pays the seller everything in the future. But the ti upfront

    payment to the buyer we derived is the same for both types:

    tH=tL= ce22

    74

  • 8/10/2019 Contract Theory, by Alex Young

    75/86

  • 8/10/2019 Contract Theory, by Alex Young

    76/86

    We take the FOC with respect to rH:

    1

    2

    1

    c2H2(R rH) +

    1

    c2H(R 2rH) = 0

    (R rH) +R 2rH= 0

    rH= 0

    rH= 0

    Just as in the second-best adverse selection case, we have the so-called efficiency at the

    top. We take the FOC with respect to rL:

    1

    2

    1

    c2H2(R rL)(1) +

    1

    2

    1

    c2L2(R rL)(1)

    +(1)

    1

    2

    1

    c2L2(R rL)(1) +

    1

    c2L(R 2rL)

    =(2H(R rL)

    2L(R rL)) + (1 )(

    2L(R 2rL)

    2L(R rL))

    =(2H 2L)(R rL) + (1 )(R 2rL R+rL)

    2L

    =(2H 2L)R (2H 2L)rL+ (1 )(rL)2L

    (2H 2L)R=rL[(

    2H

    2L) + (1 )

    2L]

    rL=

    (2H

    2L)

    (2H 2L) + (1 )

    2L

    R

    >0

    From that we can derive tSL, where the Sdenotes second-best:

    tSL= [L(R rSL)]

    2

    2c

    < [L(R)]

    2

    2c

    =tFL

    76

  • 8/10/2019 Contract Theory, by Alex Young

    77/86

  • 8/10/2019 Contract Theory, by Alex Young

    78/86

    transferred from the owner to the manager, who consumes cash above what he puts into

    production. The managers utility for cash transferred,y; and production requirement, x,

    with cost per dollar c is therefore

    U(y, x; c) =y cx

    The owner designs a menu of contracts, {x(c), y(c)} c C, such that the manager selects

    (x(c), y(c)) when he knows the per dollar cost is c. The owners objective in choosing among

    alternative budgets is to maximize expected profits:

    10

    [x(c) y(c)] f(c) dc

    There are five constraints on the owners choice of schemes.

    1. The manager must receive an expected utility at least as high as his next-best alterna-

    tive, with denoting the managers opportunity cost of participation and normalized

    to zero.

    2. The contract must respect the managers lack of cash.

    3. The contract must induce a manager who knows the cost is ci to select (x(c), y(c)).

    4. To ensure a solution exists, we must assume the production of cash flows is bounded.

    5. Cash produced must not be negative.

    Therefore, the owners problem is

    max(x(c),y(c))

    10

    [x(c) y(c)] f(c) dc

    78

  • 8/10/2019 Contract Theory, by Alex Young

    79/86

    subject to

    10

    [y(c) cx(c)] f(c) dc = 0 (IR-1)

    y(c) cx(c) 0c C (IR-2)

    y(c) cx(c) y(c) cx(c) c, c C (IC)

    x(c) xmax c C (Existence)

    x(c) 0c C (Nonnegativity)

    5.1.2 Solution of Benchmark Case

    Antle and Eppen [1985]showed that the optimal amounts of production and resources trans-

    ferred are given by a simple hurdle strategy: if the reported cost is above c, nothing is

    produced and no resources are given; if the reported cost is at or below c,xmax is produced

    and cxmax is given.

    5.1.2.1 How? If the problem is simplified from a continuum of types to two, following

    the algorithm outlined above, the owners problem is equivalent to

    maxx

    (1 c F(c)

    f(c))x

    Hence, by inspection[Antle and Eppen,1985],

    x = xmax, if (1 c

    F(c)

    f(c))> 0

    0, otherwise

    The solution involves a trade-off between productive efficiency and distributional conse-

    quences. Lowering (raising) the hurdle cost gives up (increases) valuable production. But it

    79

  • 8/10/2019 Contract Theory, by Alex Young

    80/86

    allows the owner to capture (give) more of the surplus to the manager in the form of reduced

    (excess) resources.

    The optimal policy balances these factors. Consider a candidate for the critical hurdle cost,

    ci. If the hurdle is lowered tock1, production with an expected gross revenue to the owner

    ofpkxmax is lost. But the expected resources allocated decline:

    (k

    i=1

    pickxmax =

    k1i=1

    pickxmax +pkckx

    max) (Original)

    k1

    i=1pick1x

    max (New)

    =k1i=1

    pi(ck ck1)xmax +pkckx

    max (Difference)

    Likewise, if the hurdle is raised to ck+1, production with an expected gross revenue to the

    owner ofpk+1xmax is gained. But the expected resources allocated increase:

    (k+1

    i=1pick+1x

    max =k

    i=1pick+1x

    max +pk+1ck+1xmax) (New)

    k

    i=1

    pickxmax (Original)

    =k

    i=1

    pi(ck+1 ck)xmax +pk+1ck+1x

    max (Difference)

    5.2 Prendergast[1993]

    5.2.1 Simplified Representation

    The following is a simplification of Prendergast [1993] due to Qi Chen. A risk-neutral

    manager-owner of a firm chooses a binary decisionx {0, 1} to maximize profits. Choosing

    the correct decision results in an additional profit of to the firm. We let {0, 1} denote

    the optimal decision, and the ex ante prior places equal probabilities on each decision being

    80

  • 8/10/2019 Contract Theory, by Alex Young

    81/86

    optimal.

    The manager privately observes a signalm regarding the true state where

    Pr[m= ] =m>1

    2

    The manager also employs a risk-neutral worker who privately observes another independent

    signal w regarding the true state where

    Pr[w =] =w >1

    2

    In addition, the worker privately observes a signal as to what the manager has observed

    where

    Pr[= m] = >1

    2

    Thus, the worker has information about the managers observation, which is indirectly in-

    formative about the true state, and direct information about the true state. Note that

    information about the managers observation can easily be motivated as some guess about

    what the managers taste / preferences are.

    The manager offers wage contracts of the following form: the worker is paid a fixed wage

    (normalized to zero) plus a non-negative bonusif his reported optimal decisionw matchesthe managers signalm and zero otherwise.

    If the worker exerts effort at cost c, then w =w; and if he does not, then w =w < w.

    In addition, we assume that w > w > m, such that in a first-best world, the manager

    would always ignore his own information and choose x = w regardless of the workers effort.

    81

  • 8/10/2019 Contract Theory, by Alex Young

    82/86

    We assume that the workers effort is socially efficient:

    (w) (w)> c

    Lastly, we assume that the probability of the agent getting paid conditional on working ex-

    ceeds the probability of getting paid if the agent reports based on what he thinks the manager

    wants to hear, which is in turn greater than the probability of getting paid conditional on