information economics - cpb-us-w2.wpmucdn.com · 4. 70s and 80s: –rst phase of applications of...

88
Dirk Bergemann Department of Economics Yale University Information Economics March 2009 Springer-Verlag Berlin Heidelberg NewYork London Paris Tokyo Hong Kong Barcelona Budapest

Upload: others

Post on 29-Oct-2019

5 views

Category:

Documents


0 download

TRANSCRIPT

Dirk BergemannDepartment of EconomicsYale University

Information Economics

March 2009

Springer-VerlagBerlin Heidelberg NewYorkLondon Paris TokyoHongKong BarcelonaBudapest

Contents

Part I. Introduction

1. Game Theory : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 91.1 Game theory and parlor games - a brief history . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91.2 Game theory in microeconomics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10

2. Information Economics : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 11

3. Akerlof�s Lemon Model : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 133.1 Basic Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133.2 Extensions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

4. Wolinsky�s Price Signal�s Quality : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 174.1 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 174.2 Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18

Part II. Static Games of Incomplete Information

5. Harsanyi�s insight : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 21

6. De�nition of a Bayesian Game : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 236.1 Global Games . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

6.1.1 Incomplete Information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 256.2 Common Knowledge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26

7. Puri�cation : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 27

8. Global Games : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 298.0.1 Investment Game . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

Part III. Dynamic Games of Incomplete Information

9. Job Market Signalling : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 339.1 Pure Strategy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 339.2 Perfect Bayesian Equilibrium . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 339.3 Equilibrium Domination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 369.4 Informed Principal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38

9.4.1 Maskin and Tirole�s informed principal problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38

4 Contents

10. Spence-Mirrlees Single Crossing Condition : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 4110.0.2 Separating Condition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41

10.1 Supermodular . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4210.2 Supermodular and Single Crossing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4310.3 Signalling versus Disclosure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4310.4 Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44

Part IV. Moral Hazard10.5 Introduction and Basics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4710.6 Binary Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47

10.6.1 First Best . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4810.6.2 Second Best . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

10.7 General Model with �nite outcomes and actions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4910.7.1 Optimal Contract . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5010.7.2 Monotone Likelihood Ratio . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5110.7.3 Convexity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51

10.8 Information and Contract . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5310.8.1 Informativeness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5310.8.2 Additional Signals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54

10.9 Linear contracts with normally distributed performance and exponential utility . . . . . . . . . . . 5410.9.1 Certainty Equivalent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5510.9.2 Rewriting Incentive and Participation Constraints . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55

10.10Readings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56

Part V. Mechanism Design

11. Introduction : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 59

12. Adverse selection: Mechanism Design with One Agent : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 6112.1 Monopolistic Price Discrimination with Binary Types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61

12.1.1 First Best . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6212.1.2 Second Best: Asymmetric information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62

12.2 Continuous type model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6412.2.1 Information Rent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6412.2.2 Utilities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6512.2.3 Incentive Compatibility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

12.3 Optimal Contracts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6712.3.1 Optimality Conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6812.3.2 Pointwise Optimization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68

13. Mechanism Design Problem with Many Agents : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 7113.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7113.2 Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7113.3 Mechanism as a Game . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7213.4 Second Price Sealed Bid Auction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74

14. Dominant Strategy Equilibrium : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 77

Contents 5

15. Bayesian Equilibrium : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 7915.1 First Price Auctions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7915.2 Optimal Auctions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81

15.2.1 Revenue Equivalence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8115.2.2 Optimal Auction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82

15.3 Additional Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8415.3.1 Procurement Bidding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8415.3.2 Bilateral Trade . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8415.3.3 Bilateral Trade: Myerson and Sattherthwaite . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8415.3.4 Readings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84

16. E¢ ciency : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 8516.1 First Best . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8516.2 Second Best . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86

17. Social Choice : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 8717.1 Social Welfare Functional . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8717.2 Social Choice Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87

Part I

Introduction

1. Game Theory

Game theory is the study of multi-person decision problems. The focus of game theory is interdependence,situations in which an entire group of people is a¤ected by the choices made by every individual within thatgroup. As such they appear frequently in economics. Models and situations of trading processes (auction,bargaining) involve game theory, labor and �nancial markets. There are multi-agent decision problemswithin an organization, many person may compete for a promotion, several divisions compete for investmentcapital. In international economics countries choose tari¤s and trade policies, in macroeconomics, the FRBattempts to control prices.Why game theory and economics? In competitive environments, large populations interact. How-

ever, the competitive assumption allows us to analyze that interaction without detailed analysis of strategicinteraction. This gives us a very powerful theory and also lies behind the remarkable property that e¢ cientallocations can be decentralized through markets.In many economic settings, the competitive assumption does not makes sense and strategic issues must

addressed directly. Rather than come up with a menu of di¤erent theories to deal with non-competitiveeconomic environments, it is useful to come up with an encompassing theory of strategic interaction (gametheory) and then see how various non-competitive economic environments �t into that theory. Thus this sec-tion of the course will provide a self-contained introduction to game theory that simultaneously introducessome key ideas from the theory of imperfect competition.

1. What will each individual guess about the other choices?2. What action will each person take?3. What is the outcome of these actions?

In addition we may ask

1. Does it make a di¤erence if the group interacts more than once?2. What if each individual is uncertain about the characteristics of the other players?

Three basic distinctions may be made at the outset

1. non-cooperative vs. cooperative games2. strategic (or normal form) games and extensive (form) games3. games with perfect or imperfect information

In all game theoretic models, the basic entity is a player. In noncooperative games the individual playerand her actions are the primitives of the model, whereas in cooperative games coalition of players and theirjoint actions are the primitives.

1.1 Game theory and parlor games - a brief history

1. 20s and 30s: precursorsa) E. Zermelo (1913) chess, the game has a solution, solution concept: backwards induction

10 1. Game Theory

b) E. Borel (1913) mixed strategies, conjecture of non-existence2. 40s and 50s: core conceptual development

a) J. v. Neumann (1928) existence in of zero-sum gamesb) J. v. Neumann / O. Morgenstern (1944) Theory of Games and Economic Behavior: Axiomaticexpected utility theory, Zero-sum games, cooperative game theory

c) J. Nash (1950) Nonzero sum games3. 60s: two crucial ingredients for future development: credibility (subgame perfection) and incompleteinformationa) R. Selten (1965,75) dynamic games, subgame perfect equilibriumb) J. Harsanyi (1967/68) games of incomplete information

4. 70s and 80s: �rst phase of applications of game theory (and information economics) in applied �eldsof economics

5. 90s and on: real integration of game theory insights in empirical work and institutional design

� For more on the history of game theory, see Aumann�s entry on �Game Theory� in the New PalgraveDictionary of Economics.

1.2 Game theory in microeconomics

1. decision theory (single agent)2. game theory (few agents)3. general equilibrium theory (many agents)

2. Information Economics

We shall start with two very simple, but classic models and results to demonstrate (i) how asymmetry ofinformation changes classical e¢ ciency results of markets and (ii) how asymmetry of information changesclassical arguments about the role of prices and the equilibrium process.Economics of information examines the role of information in economic relationship. It is therefore an

investigation into the role of imperfect and incomplete information. In the presence of imperfect information,learning, Bayesian or not, becomes important, and in consequence dynamic models are prevalent. In thisclass we will focus mostly on incomplete or asymmetric information.The nature of information economics as a �eld is perhaps best understood when contrasted with the

standard general equilibrium theory. Information consists of a set of tools rather than a single methodology.Furthermore, the choice of tools is very issue driven. Frequently we will make use of the following tools:

� small number of participants� institutions may be represented by constraints� noncooperative (Bayesian) game theory� simple assumptions on bargaining: Principal-Agent paradigm

We refer to the Principal-Agent paradigm as a setting where one agent, called the Principal, can make allthe contract o¤ers, and hence has (almost) all bargaining power and a second agent, called the Agent, canonly choose whether to accept or reject the o¤er by the principal. With this general structure the interactivedecision problem can often by simpli�ed to a constrained optimization problem, where the Principal hasan objective function, and the Agent simply represents constraints on the Principal�s objective function.Two recurrent constraints will be the participation constraint and the incentive constraint.We refer to the Principal-Agent paradigm as a setting where one agent, called the Principal, can make all

the contract o¤ers, and hence has (almost) all bargaining power and a second agent, called the Agent, canonly choose whether to accept or reject the o¤er by the principal. With this general structure the interactivedecision problem can often by simpli�ed to a constrained optimization problem, where the Principal hasan objective function, and the Agent simply represents constraints on the Principal�s objective function.Two recurrent constraints will be the participation constraint and the incentive constraint.The following scheme organizes most of the models prevalent in information economics according to

two criteria: (i) whether the informed or an uniformed agent has the initiative (makes the �rst move, o¤ersthe initial contract, arrangement; and (ii) whether the uniformed agent is uncertain about the action orthe type (information) of the informed agent.

hidden action hidden informationinformed agent signalling

uninformed agent moral hazard adverse selection

3. Akerlof�s Lemon Model

3.1 Basic Model

This section is based on ?. Suppose an object with value v s U [0; 1] is o¤ered by the seller. The valuationsare

us = �sv

and

ub = �bv

with �b > �s, and hence trading is always pareto-optimal. But trade has to be voluntary. We then ask isthere a price at which trade occurs. Suppose then at a price p trade would occur. What properties wouldthe price have to induce trade. The seller sells if

�sv � p

and thus by selling the object he signals that

v � p

�s(3.1)

The buyer buys the object if

�bE [v] � p (3.2)

and as he knows that (3.1) has to hold, he can form a conditional expectation, that

�bE [v jp ] � p, �bp

2�s� p (3.3)

Thus for the sale to occur,

�b � 2�s. (3.4)

Thus unless, the tastes di¤er substantially, the market breaks down completely:

� market mechanism in which a lower prices increases sales fails to work as lowering the price decreasesthe average quality, lower price is �bad news�.

� market may not disappear but display lower volume of transaction than socially optimal.

14 3. Akerlof�s Lemon Model

3.2 Extensions

In class, we made some informed guesses how the volume of trade may depend on the distribution ofprivate information among the sellers. In particular, we ventured the claim that as the amount of privateinformation held by the sellers decreases, the possibilities for trade should increase. We made a secondobservation, namely that in the example we studied, for a given constellation of preferences by buyer andseller, represented by �b and �s, trade would either occur with positive probability for all prices or it wouldnot occur at all. We then argued that this result may be due to the speci�c density in the example, butmay not hold in general. We now address both issues with a uniform density with varying support andconstant mean:

v~U�1

2� "; 1

2+ "

�with " 2

�0; 12�. We can formalize the amount of private information by the variance of f (�) and how it

a¤ects the e¢ ciency of the trade.We may redo the analysis above where the seller signals by selling the object at a �xed price p that

v � p

�s.

The buyer buys the object if

�bE [v jp ] � p,

The expected value is now given by

E [v jp ] =12 � "+

p�s

2

and hence for sale to occur

�b

12 � "+

p�s

2� p (3.5)

and inequality prevails if, provided that " 2 [0; 12 ) and 2�s > �b,

p � 1

2�b�s

1� 2"2�s � �b

Consider next the e¢ ciency issue. All types of sellers are willing to sell if

p = �s

�1

2+ "

�(3.6)

in which case the expected conditional value for the buyer is

1

2�b � �s

�1

2+ "

�or equivalently

�b � �s (1 + 2") (3.7)

Thus as the amount of private information, measured by ", decreases, the ine¢ ciencies in trade decrease.Notice that we derived a condition in terms of the preferences such that all sellers wish to sell. However

3.2 Extensions 15

even if condition (3.7) is not met, we can in general support some trade. This is indicated already in theinequality (3.5):

�b

12 � "+

p�s

2� p (3.8)

as the lhs increases slower in p, then the rhs of the inequality above. Thus there might be lower prices,which will not induce all sellers to show up at the market, yet allow some trade. We show how this can arisenext. For a positive mass of seller willing to sell the price may vary between p 2 (�s

�12 � "

�; �s�12 + "

�]. If

we then insert the price as a function of x 2 (�"; "], we �nd

�b

12 � "+

�s( 12+x)�s

2� �s

�1

2+ x

�or

�b1� "+ x

2� �s

�1

2+ x

�For a given �b; �s and ", we may then solve the (in-)quality for x, to �nd the (maximal) price p where thevolume of trade is maximized, or:

x =�b (1� ")� �s2�s � �b

:

Thus x is increasing in �b and decreasing in ";con�rming now in terms of the volume of trade our intuitionabout private information and e¢ ciency. In fact, we can verify that for a given " 2 [0; 12 ), the volume oftrade is positive, but as "! 1

2 becomes arbitrarily small and converges as expected to zero for �b < 2�s.

4. Wolinsky�s Price Signal�s Quality

This section is based on ?. Suppose a continuum of identical consumers. They have preferences

u = �v � p

and the monopolist can provide low and high quality: v = f0; 1g at cost 0 < c0 < c1. The monopolistselects price and quality simultaneously. Assume that � > c1 so that it is socially e¢ cient to produce thehigh quality good. Assume that the consumers do not observe quality before purchasing. It is clear that anequilibrium in which the monopolist sells and provides high quality cannot exist.Suppose now that some consumers are informed about the quality of the product, say a fraction �.

Observe �rst that if the informed consumers are buying then the uniformed consumer are buying as well.When is the seller better o¤ to sell to both segments of the market:

p� c1 � (1� �) (p� c0)

or

�p � c1 � (1� �) c0. (4.1)

We can then make two observations:

� high quality is supplied only if price is su¢ ciently high, �high price can signal high quality�.� a higher fraction, �, of informed consumers favors e¢ ciency as it prevents the monopolist from cuttingquality

� the informational externality favors government intervention as individuals only take private bene�t andcost into account.

4.1 Conclusion

We considered �hidden information�or �hidden action�models and suggested how asymmetry in informa-tion may reduce or end trade completely. In contrast to the canonical model of goods, which we may call�search goods�, where we can assert the quality by inspection, we considered �experience goods� (?, ?),where the quality can only be ascertained after the purchase. The situation is only further acerbated with�credence goods�(?).In both models, there was room for a third party, government or other institution, to induce pareto

improvement. In either case, and improvement in the symmetry of information lead to an improvementin the e¢ ciency of the resulting allocation. This suggest that we may look for optimal or equilibriumarrangements to reduce the asymmetry in information, either through:

� costly signalling� optimal contracting to avoid moral hazard, or� optimal information extraction through a menu of contract (i.e. mechanism design).

18 4. Wolinsky�s Price Signal�s Quality

4.2 Reading

The lecture is based on Chapter 1 in ?, Chapter 2 in ?, and Chapter 5 in ?.

Part II

Static Games of Incomplete Information

5. Harsanyi�s insight

Harsanyi�s insights is illustrated by the following example.

Example: Suppose payo¤s of a two player two action game are either:

H TH 1,1 0,0T 0,1 1,0

or

H TH 1,0 0,1T 0,0 1,1

i.e. either player II has dominant strategy to play H or a dominant strategy to play T . Suppose thatII knows his own payo¤s but player I thinks there is probability � that payo¤s are given by the �rstmatrix, probability 1 � � that they are given by the second matrix. Say that player II is of type1 if payo¤s are given by the �rst matrix, type 2 if payo¤s are given by the second matrix. Clearlyequilibrium must have: II plays H if type 1, T if type 2; I plays H if � > 1

2 , T if � <12 . But how to

analyze this problem in general?

All payo¤ uncertainty can be captured by a single move of �nature�at the beginning of the game. See�gure 14.As with repeated games, we are embedding simple normal form games in a bigger game.

6. De�nition of a Bayesian Game

A (�nite) static incomplete information game consists of

� Players 1; :::; I� Actions sets A1; :::; AI� Sets of types T1; :::; TI� A probability distribution over types p 2 � (T ), where T = T1 � :::� TI� Payo¤ functions g1; :::; gI , each gi : A� T ! R

Interpretation: Nature chooses a pro�le of players types, t � (t1; :::; tI) 2 T according to probabilitydistribution p (:). Each player i observes his own type ti and chooses an action ai. Now player i receivespayo¤s gi (a; t).

A strategy is a mapping si : Ti ! Ai. Write Si for the set of such strategies, and let S = S1 � ::� SI .Player i�s payo¤ function of the incomplete information game, ui : S ! R, is

ui (s) =Xt2T

p (t) gi (s (t) ; t)

where s = (s1; :::; sI) and s (t) = (s1 (t1) ; :::; sI (tI)).

(Old) De�nition: Strategy pro�le s� is a pure strategy Nash equilibrium if

ui�s�i ; s

��i�� ui

�si; s

��i�for all si 2 Si and i = 1; ::; I

This can be re-written as:

Pt2T

p (t) gi��s�i (ti) ; s

��i (t�i)

�; t��Pt2T

p (t) gi��si (ti) ; s

��i (t�i)

�; t�

for all si 2 Si and i = 1; ::; I

Writing p (ti) =Pt0�i

p�ti; t

0�i�and p (t�ijti) � p(ti;t�i)

p(ti), this can be re-written as:

Pt�i2T�i

p (t�ijti) gi��s�i (ti) ; s

��i (t�i)

�; t��

Pt�i2T�i

p (t�ijti) gi��ai; s

��i (t�i)

�; t�

for all ti 2 Ti, ai 2 Ai and i = 1; :::; I

Example: Duopoly

�2 (q1; q2; cL) = [(a� q1 � q2)� cL] q2�2 (q1; q2; cH) = [(a� q1 � q2)� cH ] q2

�1 (q1; q2) = [(a� q1 � q2)� c] q1T1 = fcg, T2 = fcL; cHg, p (cH) = �. A strategy for player 1 is a quantity q�1 . A strategy for player 2 isa function q�2 : T2 ! R, i.e. we must solve for three numbers q�1 ; q�2 (cH) ; q�2 (cL).

24 6. De�nition of a Bayesian Game

Assume interior solution. We must have:

q�2 (cH) = argmaxq2

[(a� q�1 � q2)� cH ] q2

i.e. q�2 (cH) =1

2(a� q�1 � cH)

Similarly q�2 (cL) =1

2(a� q�1 � cL)

We must also have:

q�1 = argmaxq1

� [(a� q1 � q�2 (cH))� c] q1 + (1� �) [(a� q1 � q�2 (cL))� c] q1

i.e. q�1 = argmaxq1

[(a� q1 � �q�2 (cH)� (1� �) q�2 (cL))� c] q1

i.e. q�1 =1

2(a� c� �q�2 (cH)� (1� �) q�2 (cL))

Solution:

q�2 (cH) =a� 2cH + c

3+1� �6

(cH � cL)

q�2 (cL) =a� 2cL + c

3� �6(cH � cL)

q�1 =a� 2c+ �cH + (1� �) cL

3

[lecture 11]

6.1 Global Games

Consider a two player game, each agent must choose an action, "invest" or "not invest". The state �represents "fundamentals" of the economy and the payo¤ is given by

I NI �; � � � 1; 0N 0; � � 1 0; 0

Complete Information. Complete information: there is "common knowledge" of payo¤s, including � :

1. If 0 � � � 1, the game has multiple Nash equilibria:a) An equilibrium where all players invest (expecting all others to invest)b) An equilibrium where no one invests (since they expect no one else to invest)

2. If � is strictly greater than 1, everyone has a dominant strategy to invest3. If � is strictly less than 0, everyone has a dominant strategy to not invest

6.1 Global Games 25

6.1.1 Incomplete Information

Suppose there is not common knowledge of payo¤, suppose � is not common knowledge. We have

� � U [�M;M ]

for large M > 0, but each player only observe a signal xi

xi = � +1

z"i; "i � U

��12;1

2

�and so z represent the accuracy of the signal xi.Suppose the agents pursue a cut-o¤ strategy:

si (xi) =

�I if xi > xN if xi < x

Then the best response condition for investment is

E [� jxi ] Pr (sj = I jxi ) + (1� Pr (sj = I jxi ))E [� � 1 jxi ] � 0 (6.1)

Posteriors and Beliefs. We have

� = xi �1

z"i;

and hence

gi (� jxi ) =

8<: 0 if �i < xi � 12z

z if xi � 12z < � < xi +

12z

0 if �i > xi +12z

Hence, simple enough:

E [� jxi ] = xi:

Now

Pr (sj = I jxi ) = Pr (xj � x jxi ) = Pr�� +

1

z"j � x jxi

�= Pr

�xi �

1

z"i +

1

z"j � x

�which we could compute easily by the uniformity of the error term "i and "j :

Pr

�1

z("j � "i) � x� xi

�. (6.2)

But to compute the critical threshold, we do not have to do this, since we only need to evaluate at xi = x :

xi Pr (sj = I jxi ) + (1� Pr (sj = I jxi )) (xi � 1) = 0()

xi Pr

�1

z("j � "i) � x� xi

�+

�1� Pr

�1

z("j � "i) � x� xi

��(xi � 1) = 0()

xPr

�1

z("j � "i) � x� x

�+

�1� Pr

�1

z("j � "i) � x� x

��(x� 1) = 0()

xPr

�1

z("j � "i) � 0

�+

�1� Pr

�1

z("j � "i) � 0

��(x� 1) = 0()

x1

2+1

2(x� 1) = 0()

x =1

2;

and thus we have a complete description of the equilibrium, and in fact it is now a unique equilibrium.

26 6. De�nition of a Bayesian Game

6.2 Common Knowledge

A state space with ! 2 Knowledge is represented by a partition of , possibly di¤erent across players hi and hi (!) is the

element of the partition that contains !.The common prior is p (!)The posterior is

p (! jhi ) =p (!)P

!2hi p (!0)

We say that player i knows event E at ! if

hi (!) � E

and hence "player i knows E"

Ki (E) = f! jhi (!) � E g

The event "everyone knows E" is

KI (E) =

(!

�����[i2Ihi (!) � E

)

and hence

K2I (E) =

(!

�����[i2Ihi (!) � KI (E)

),

and so on:

KnI (E) =

(!

�����[i2Ihi (!) � Kn�1

I (E)

).

De�nition 6.2.1. Event E is common knowledge at ! if ! 2 K1I (E)

7. Puri�cation

N SN 1,-1 0,0S 0,0 1,-1

Suppose I has some private value to going N , " � U [0; x]; II has some private value to going N , � � U [0; x]New game:

N SN 1+",-1+� ",0S 0,� 1,-1

Best Responses: If I attaches probability q to II going N , then I�s best response is N if " + q � 1 � q i.e." � 1�2q. Similarly, if II attaches prob. p to I going N , then II�s best response is N if ��p � � (1� p)i.e. � � 2p� 1.

Thus players 1 and 2 must each have a threshold value of " and �, respectively, above which they chooseN . We may label the thresholds so that the ex ante probability of choosing N is p and q, respectively:

s1 (") =

�N , if " � (1� p)xS, if " < (1� p)x

s2 (�) =

�N , if � � (1� q)xS, if � < (1� q)x

Now s1 is a best response to s2 if and only if (1� p)x = 1 � 2q; s2 is a best response to s1 if and only if(1� q)x = 2p� 1. Thus�

�x 2�2 x

��pq

�=

�1� x1 + x

�.

So �pq

�=

��x 2�2 x

��1�1� x1 + x

�=

1

�x2 � 4

�x �2�2 �x

��1� x1 + x

�=

�1x2 + 4

�x �2�2 �x

��1� x1 + x

�=

2+x+x2

4+x2

2�x+x24+x2

!

!�1212

�, as x! 0

28 7. Puri�cation

We say that the mixed strategy equilibrium of the complete information game is �puri�ed�by addingthe incomplete information.Parameterize a complete information game by the payo¤ functions, g = (g1; :::; gI).Let player i have a type �i = (�

ai )a2A 2 R#A; let player i�s type be independently drawn according to

smooth density fi (�) with bounded support, say [�1; 1]#A. Let player i�s payo¤ in the �"-perturbed game�(this is an incomplete information game) be

egi (a; �) = gi (a) + "�aiTheorem (Harsanyi): Fix a generic complete information game g and any Nash equilibrium of g. Then

there exists a sequence of pure strategy equilibria of the "-perturbed with the property that the probabilitydistribution over actions of each player is converging to his strategy in the original Nash equilibrium.

8. Global Games

8.0.1 Investment Game

The state � is drawn uniformly from the interval [ M,M] for largeM > 0Each player observes a signal si = q + 1x #i , where #i U1

2 , 12andx measures the accuracy of the signalStephen

Part III

Dynamic Games of Incomplete Information

9. Job Market Signalling

Consider the following game �rst analyzed by ?. Nature chooses a workers type (productivity) a 2 f1; 2g.The worker has productivity a with probability p (a). For notational convenience, we de�ne

p , p (a = 2) :

The worker can choose an educational level e 2 R+. The worker chooses the �rm The workers utility is

u (w; e; a) = w � e

a.

The objective function of the �rm is given by

minX

a2f1;2g

(a� w)2 p (a je )

We observe that

@2u (w; e; a)

@e@a=1

a2> 0

and thus type and strategic variable are complements (supermodular).

9.1 Pure Strategy

A pure strategy for the worker is a function

be : f1; 2g ! R+

A pure strategy for the �rm is a function

w : R+ ! R+

where w (e) is the wage o¤ered to a worker of educational level e.

9.2 Perfect Bayesian Equilibrium

The novelty in signalling games is that the uninformed party gets a chance to update its prior belief onthe basis of a signal sent by the informed party. The updated prior belief is the posterior belief, dependingon e and denoted by p (a je ). The posterior belief is a mapping

p : R+ ! [0; 1]

34 9. Job Market Signalling

In sequential or extensive form games we required that the strategy are sequential rational, or time consis-tent. We now impose a similar consistency requirement on the posterior beliefs by imposing the Bayes�rulewhenever possible. Since p (a je ) is posterior belief, and hence a probability function, it is required that:

8e; 9p (a je ) ; s.th. p (a je ) � 0; andX

a2f1;2g

p (a je ) = 1: (9.1)

Moreover, when the �rm can apply Bayes law, it does so, or

if 9a s.th. at e� (a) = e; then p (a je ) = p (a)Pfa0je�(a)=eg p (a

0): (9.2)

We refer to educational choice which are selected by some worker-types in equilibrium, or if 9a s.th. ate� (a) = e, as �on-the-equilibrium path� and educational choices s.th. @a with e� (a) = e as �o¤-the-equilibrium-path�.As before it will be sometimes easier to refer to

p (e) , p (a = 2 je )

and hence

p (1 je ) = 1� p (e) :

De�nition 9.2.1 (PBE). A pure strategy Perfect Bayesian Equilibrium is a set of strategies fe� (a) ; w� (e)gand posterior beliefs p (e) such that:

1. 8e; 9p� (a je ) ; s.th. p� (a je ) � 0; andP

�2f1;2g p� (a je ) = 1;

2. 8i; w�i (e) =P

a p� (a je ) a;

3. 8a; e� (a) 2 argmax�w� (e)� e

a

4. if 9a s.th. at e� (a) = e; then:

p (a je ) = p (a)Pfa0je�(a)=eg p (a

0):

De�nition 9.2.2 (Separating PBE). A pure strategy PBE is a separating equilibrium if

a 6= a0 ) e (a) 6= e (a0) .

De�nition 9.2.3 (Pooling PBE). A pure strategy PBE is a pooling equilibrium if

8a; a0 ) e (a) = e (a0) .

[Lecture 3]

Theorem 9.2.1.

1. A pooling equilibrium exists for all education levels e = e (1) = e (2) 2 [0; p] :2. A separating equilibrium exists for all e (1) = 0 and e (2) 2 [1; 2].

Proof. (1) We �rst construct a pooling equilibrium. For a pooling equilibrium to exist it must satisfy thefollowing incentive compatibility constraints

8e; 1 + p (e�)� e� � 1 + p (e)� e (9.3)

and

9.2 Perfect Bayesian Equilibrium 35

8e; 1 + p (e�)� e�

2� 1 + p (e)� e

2(9.4)

Consider �rst downward deviations, i.e. e < e�, then (9.3) requires that

p (e�)� p (e) � e� � e: (9.5)

Then consider upward deviations, i.e. e > e�, then (9.4) requires that

p (e)� p (e�) � 1

2(e� e�) : (9.6)

We can then ask for what levels can both inequalities be satis�ed. Clearly both inequalities are easiest tosatisfy if

e < e� ) p (e) = 0

e > e� ) p (e) = 0;

which leaves us with

e < e� ) p � e� � e (9.7)

e > e� ) �p � 1

2(e� e�) :

We may then rewrite the inequalities in (9.7) as

e < e� ) e� � p+ e (9.8)

e > e� ) e� � p+ 2e

As the inequality has to hold for all e, the asserting that 0 � p � e� holds, follows immediately.(2) Consider then a separating equilibrium. It must satisfy the following incentive compatibility con-

straints

8e; 1 + p (e�1)� e�1 � 1 + p (e)� e

and

8e; 1 + p (e�2)�e�22� 1 + p (e)� e

2: (9.9)

As along the equilibrium path, the �rms must apply Bayes law, we can rewrite the equations as

8e; 1� e�1 � 1 + p (e)� e

and

8e; 2� e�2

2� 1 + p (e)� e

2:

Consider �rst the low productivity type. It must be that e�1 = 0. This leaves us with

p (e) � e: (9.10)

But if e = e�2 is supposed to be part of a separating equilibrium, then p (e�2) = 1. Thus it follows further

from (9.10) that e�2 � 1, for otherwise we would not satisfy 1 = p (e�2) � e�2. Finally, we want to determinean upper bound for e�2. AS we cannot require the high ability worker to produce too much e¤ort otherwisehe would mimic the lower ability, we can rewrite (9.9) to obtain:

36 9. Job Market Signalling

1� p (e) � 1

2(e�2 � e) ;

which is easiest to satisfy if

p (e) = 0;

and hence

8e; 2 + e � e�2

which implies that:

e�2 � 2.

which completes the proof.

The theorem above only de�nes the range of educational choices which can be supported as an equi-librium but is not a complete equilibrium description as we have not speci�ed the beliefs in detail. Thejob market signalling model suggests a series of further questions and issues. There were multiple equilib-ria, reducing the predictive ability of the model and we may look at di¤erent approaches to reduced themultiplicity:

� re�ned equilibrium notion� di¤erent model: informed principal

The signal was modelled as a costly action. We may then ask for conditions:

� when do costly signals matter for Pareto improvements (or simply separation): Spence-Mirrlees singlecrossing conditions

� when do costless signals matter: cheap-talk games.

In the model, education could also be interpreted as an act of disclosure of information through theveri�cation of the private information by a third party. It may therefore be of some interest to analyze thepossibilities of voluntary information disclosure.

9.3 Equilibrium Domination

The construction of some of the equilibria relied on rather arbitrary assumptions about the beliefs of the�rms for educational choice levels o¤ the equilibrium path. Next we re�ne our argument to obtain somelogical restrictions on the beliefs. The restrictions will not be bases on Bayes law, but on the plausibilityof deviations as strategic choices by the agents. The notions to be de�ned follow ?.

De�nition 9.3.1 (Equilibrium Domination). Given a PBE, the message e is equilibrium-dominatedfor type a, if for all possible assessments p (a je ) (or simply p(e)):

w� (e� (a))� e� (a)

a> w (e)� e

a

or equivalently

1 + p� (e� (a))� e� (a)

a> 1 + p (e)� e

a

9.3 Equilibrium Domination 37

De�nition 9.3.2 (Intuitive Criterion). If the information set following e is o¤ the equilibrium pathand e is equilibrium dominated for type a, then, if possible,

p (a je ) = 0: (9.11)

Remark 9.3.1. The quali�er, if possible, in the de�nition above refers to the fact that p (a je ) has to be awell-de�ned probability, and thus if the e¤ort e is equilibrium dominated for all a, then the requirementimposed by (9.11) is vacuous.

De�nition 9.3.3. A PBE where all beliefs o¤ the equilibrium path satisfy the intuitive criterion is said tosatisfy the intuitive criterion.

Theorem 9.3.1 (Uniqueness). The unique Perfect Bayesian equilibrium outcome which satis�es theintuitive criterion is given by

fe�1 = 0; e�2 = 1; w� (0) = 1; w� (1) = 2g : (9.12)

The beliefs are required to satisfy

p� (e) =

8<: 0; for e = 02 [0; e] ; for 0 < e < 11; for e � 1

Remark 9.3.2. The equilibrium outcome is referred to as the least-cost separating equilibrium (or ? equi-librium)

Proof. We �rst show that there can�t be any pooling equilibria satisfying the intuitive criterion, and thenproceed to show that only one of the separating equilibria survives the test of the intuitive criterion.Suppose 0 � e� � p. Consider e 2 (e� + (1� p) ; e� + 2 (1� p)). Any message e in the interval is

equilibrium dominated for the low productivity worker as

1 + p� e� > 2� e;

but any message in the interval is not equilibrium dominated for the high productivity worker, as

1 + p� e

2

�< 2� e

2

and thus for e = e� + (1� p), we have

p < 1� (1� p)2

, 1

2p <

1

2;

which certainly holds. Thus if (e; p (e)) satis�es the intuitive criterion, we must have p (e) = 1 for e 2(e� + (1� p) ; e� + 2 (1� p)), but then pooling is not an equilibrium anymore as the high productivityworker has a pro�table deviation with any e 2 (e� + (1� p) ; e� + 2 (1� p)).Consider now the separating equilibria. For e�2 > 1, any e 2 (1; e�2) is equilibrium dominated for the low

ability worker as

1 > 2� e;

but is not equilibrium dominated for the high ability worker, as

2� e�2

2< 2� e

2.

It follows that p (e) = 1 for all e 2 (1; e�2). But then e�2 > 1, cannot be supported as an equilibrium as thehigh ability worker has a pro�table deviation by lowering his educational level to some e 2 (1; e�2), and stillreceive his full productivity in terms of wages. It remains e�2 = 1 as the unique PBE satisfying the intuitivecriterion.

38 9. Job Market Signalling

Criticism. Suppose p! 1:

Remark 9.3.3. Add the Kreps critique of equilibrium domination.

[Lecture 4]

9.4 Informed Principal

As plausible as the Cho-Kreps intuitive criterion may be, it does seem to predict implausible equilibriumoutcomes in some situations. For example, suppose that, p, the prior probability that a worker of typea = 2 is present, is arbitrarily large, (p �! 1). In that case, it seems a rather high cost to pay to incur aneducation cost of

c(e = 1) =1

2

just to be able to raise the wage by a commensurately small amount �w = 2 � (1 + p) �! 0 as p ! 1.Indeed, in that case the pooling equilibrium where no education costs are incurred seems a more plausibleoutcome. This particular case should serve as a useful warning not to rely too blindly on selection criteria(such as Cho-Kreps�intuitive criterion) to single out particular PBEs.

9.4.1 Maskin and Tirole�s informed principal problem

Interestingly, much of the problem of multiplicity of PBEs disappears when the timing of the game ischanged to letting agent and principal sign a contract before the choice of signal. This is one importantlesson to be drawn from ?.To see this, consider the model of education as speci�ed above and invert the stages of contracting

and education choice. That is, now the worker signs a contract with his/her employer before undertakingeducation. This contract then speci�es a wage schedule contingent on the level of education chosen by theworker after signing the contract.Let fw(e)g denote the contingent wage schedule speci�ed by the contract.Consider the problem for the high productivity worker. Suppose he would like to make an o¤er by

which he can separate himself from a low ability worker

maxfe;w(e)g

nw (e)� e

2

osubject to

w (e1)� e1 � w (e2)� e2 (IC1)

and

w (e2)�e22� w (e1)�

e12

(IC2)

and

a1 � w (e1) � 0 (IR1)

a2 � w (e2) � 0 (IR2)

Thus to make incentive compatibility as easy as possible he suggests e1 = 0 and w (e1 = 0) = 1. Asw (e2) = 2, it follows that after setting e2 = 1, he indeed maximizes his payo¤.

9.4 Informed Principal 39

Suppose instead he would like to o¤er a pooling contract. Then he would suggest

maxe;w

nw � e

2

o1 + p� w � 0 (IR1)

which would yield w = 1 + p for all e.This suggest that there are two di¤erent cases to consider:

1. 1 + p � 2� 12 : a high productivity worker is better o¤ in the �least cost�separating equilibrium than

in the e¢ cient pooling equilibrium.2. 1 + p > 2� 1

2 : a high productivity worker is better o¤ in the e¢ cient pooling equilibrium.

It is easy to verify that in the case where p � 12 , the high productivity worker cannot do better than

o¤ering the separating contract, nor can the low productivity worker. More precisely, the high productivityworker strictly prefers this contract over any contract resulting in pooling or any contract with more costlyseparation. As for the low productivity worker, he has everything to lose by o¤ering another contract whichwould identify himself.In the alternative case where p > 1

2 , the unique equilibrium contract is the one where:

w�(e) = 1 + p for all e � 0.

Again, if the �rm accepts this contract, both types of workers choose an education level of zero. Thus, onaverage the �rm breaks even by accepting this contract, provided that it is as (or more) likely to originatefrom a high productivity worker than a low productivity worker. Now, a high productivity worker strictlyprefers to o¤er this contract over any other separating contract. Similarly, a low productivity worker haseverything to lose from o¤ering another contract and thus identifying himself. Thus, in this case again thisis the unique contract o¤er made by the workers.

10. Spence-Mirrlees Single Crossing Condition

10.0.2 Separating Condition

Consider now a general model with a continuum of types:

A � R+

and an arbitrary number of signals

E � R+

with a general quasilinear utility function

u (t; e; a) = t+ v (e; a)

where we recall that the utility function used in Spence model was given by:

u (t; e; a) = t� e

a

We now want ask when is it possible in general to sustain a separating equilibrium for all n agents, suchthat

a 6= a0 ) e 6= e0

Suppose we can support a separating equilibrium for all types, then we must be able to satisfy for a0 > aand without loss of generality e0 > e:

t+ v (e; a) � t0 + v (e0; a), t� t0 � v (e0; a)� v (e; a) (10.1)

and

t0 + v (e0; a0) � t+ v (e; a0), t� t0 � v (e0; a0)� v (e; a0) ; (10.2)

where t , t (e) and t0 , t (e0), and by combining the two inequalities, we �nd

v (e0; a)� v (e; a) � v (e0; a0)� v (e; a0) ; (10.3)

or similarly, again recall that e < e0

v (e; a0)� v (e; a) � v (e0; a0)� v (e0; a) : (10.4)

We then would like to know what are su¢ cient conditions on v (�; �) such that for every a there exists esuch that (10.3) can hold.

42 10. Spence-Mirrlees Single Crossing Condition

10.1 Supermodular

De�nition 10.1.1. A function v : E � A ! R has increasing di¤erences in (e; a) if for any a0 > a,v(e; a0)� v(e; a) is nondecreasing in e.

Note that if v has increasing di¤erences in (e; a), it has increasing di¤erences in (a; e). Alternatively, wesay the function v is supermodular in (e; a).If v is su¢ ciently smooth, then v is supermodular in (e; a) if and only if @2v=@e@a � 0.As the transfers t; t0 could either be determined exogenously, as in the Spence signalling model through

market clearing conditions, or endogenously, as in optimally chosen by the mechanism designer, we want toask when we can make a separation incentive compatible. We shall not consider here additional constraintssuch as a participation constraints. We shall merely assume that t (a), i.e. the transfer that the agent of typea gets, t (a), conditional on the sorting allocation fa; e (a) ; t (a)g to be incentive compatible, is continuouslydi¤erentiable and strictly increasing.

Theorem 10.1.1. A necessary and su¢ cient condition for sorting ((10.1) and (10.2)) to be incentivecompatible for all t (a) is that

1. v (e; a) is strictly supermodular2. @v

@e < 0 everywhere

Proof. We shall also suppose that v (�; �) is twice continuously di¤erentiable. The utility function of theagent is given by

t (a) + v (e (a) ; a) ,

where a is the true type of agent a and e (a) is the signal the informed agent sends to make the uninformedagent believe he is of type a.(Su¢ ciency) For e (a) to be incentive compatible, a must locally solve for the �rst order conditions of theagent at a = a, namely the true type:

t0 (a) +@v (e (a) ; a)

@e

de

da= 0, at a = a (10.5)

But Fix the payments t (a) as a function of the type a to be t (a) and suppose without loss of generalitythat t (a) is continuously di¤erentiable. The equality (10.5) de�nes a di¤erential equation for the separatinge¤ort level

de

da=

t0 (a)@v(e(a);a)

@e

(10.6)

for which a unique solution exists given a initial condition, say t (0) = 0. The essential element is that the�rst order conditions is only an optimal condition for agent with type a and nobody else, but this followsfrom strict supermodularity. As

t0 (a)

and

de

da

are independent of the true type, a, it follows that if

@2v (e; a)

@e@a> 0,

10.3 Signalling versus Disclosure 43

for all e and a, then (10.5) can only identify truthtelling for agent a.(Necessity) For any particular transfer policy t (a), we may not need to impose the supermodularity con-dition everywhere, and it might often by su¢ cient to only impose it locally, where it is however necessaryto guarantee local truthtelling, i.e.

@v (e (a) ; a)

@e@a> 0

at e = e (a). However, as we are required to consider all possible transfers t (a) with arbitrary positive slopes,we can guarantee that for every a and every e there is some transfer problem t (a) such that e (a) = e by(10.6), and hence under the global condition on t (a), supermodularity becomes also a necessary conditionfor sorting.

10.2 Supermodular and Single Crossing

In the discussion above we restricted our attention to quasilinear utility function. We can generalize all thenotions to more general utility functions. We follow the notation in ?. Let

U (x; y; t) : R3 ! R

where x is in general the signal (or allocation), y an additional instrument, such as money transfer and tis the type of the agent.

De�nition 10.2.1 (Spence-Mirrlees). The function U is said to satisfy the (strict) Spence-Mirrleescondition if

1. the ratio

UxjUyj

is (increasing) nondecreasing in t, and Uy 6= 0, and2. the ratio has the same sign for every (x; y; t).

De�nition 10.2.2 (Single-Crossing). The function U is said to satisfy the single crossing property in(x; y; t) if for all (x0; y0) � (x; y)

1. whenever U (x0; y0; t) � U (x; y; t), then U (x0; y0; t0) � U (x; y; t0) for all t0 > t;2. whenever U (x0; y0; t) > U (x; y; t), then U (x0; y0; t0) > U (x; y; t0) for all t0 > t;

? show that the notions of single crossing and the Spence Mirrlees conditions are equivalent.Thus the supermodularity condition is also often referred to as the single-crossing or Spence-Mirrlees

condition, where Mirrlees used the di¤erential form for the �rst time in ?. The notions of supermodu-larity and single-crossing are exactly formulated in ? and for some corrections ? Some applications tosupermodular games are considered by ? and ?. The mathematical theory is, inter alia, due to ? and ?.

10.3 Signalling versus Disclosure

Disclosure as private but certi�able information. Signalling referred to situations where private informationis neither observable nor veri�able. That is to say, we have considered private information about such thingsas individual preferences, tastes, ideas, intentions, quality of projects, e¤ort costs, etc., which cannot reallybe measured objectively by a third party. But there are other forms of private information, such as an

44 10. Spence-Mirrlees Single Crossing Condition

individual�s health, the servicing and accident history of a car, potential and actual liabilities of a �rm,earned income, etc., that can be certi�ed or authenticated once disclosed. For these types of informationthe main problem is to get the party who has the information to disclose it. This is a simpler problemthan the one we have considered so far, since the informed party cannot report false information. It canonly choose not to report some piece of information it has available. The main results and ideas of thisliterature is based on ? and ?.

10.4 Reading

The material of this lecture is covered by Section 4 of ? and for more detail on the Spence-Mirrleesconditions, see ?.

Part IV

Moral Hazard

10.6 Binary Example 47

Today we are discussing the optimal contractual arrangement in a moral hazard setting

10.5 Introduction and Basics

In contractual arrangements in which the principal o¤ers the contract, we distinguish between

hidden information - �adverse selection�

and

hidden action - �moral hazard�

The main trade-o¤ in adverse selection is between e¢ cient allocation and informational rent. In moralhazard settings it is between risk sharing and wok incentives. Today we are going to discuss the basicmoral hazard setting. As the principal tries to infer from the output of the agent about the e¤ort choice,the principal engages in statistical inference problems. Today we are going to develop basic insights intothe optimal contract and use this is an occasion to introduce three important informational notions:

1. monotone likelihood ratio2. garbling in the sense of Blackwell3. su¢ cient statistic

The basic model goes as follows. An agent takes action that a¤ect utility of principal and agent. Principalobserves �outcome�x and possible some �signal� s but not �action�a of agent. Agent action in absenceof contract is no e¤ort. In�uence agents action by o¤ering transfer contingent on outcome.

Example 10.5.1. Employee employer e¤ect (non observable), Property insurance (�re, theft)

FB: agents action is observable (risk sharing)SB: agents action is unobservable (risk sharing-incentive)

10.6 Binary Example

The agent can take action a 2 fal; ahg at cost c 2 fcl; chg. The outcomes x 2 fxl; xhg occur randomly,where the probabilities are governed by the action as follows.

pl = Pr (xh jal ) < ph = Pr (xh jah )

The principal can o¤er a wage, contingent on the outcome w 2 fwl; whg to the agent and the utility of theagent is

u (wi)� ci

and of the principal it is

v (xi � wi)

where we assume that u and v are strictly increasing and weakly concave.

48

10.6.1 First Best

We consider initially the optimal allocation of risk between the agents in the presence of the risk and withobservable actions. If the actions are observable, then the principal can induce the agent to choose thepreferred action a� by

w (xi; a) = �1

if a 6= a� for all xi. As the outcome is random and the agents have risk-averse preference, the optimalallocation will involve some risk sharing. The optimal solution is characterized by

maxfwl;whg

fpiv (xh � wh) + (1� pi) v (xl � wl)g

subject to

piu (wh) + (1� pi)u (wl)� ci � U; (�)

which is the individual rational constraint. Here we de�ne

w (xi; a�) , wi

This is a constrained optimization problem, and the �rst order conditions from the Lagrangian

L (wl; wh; �) = piv (xh � wh) + (1� pi) v (xl � wl) + � (piu (wh) + (1� pi)u (wl)� ci)

are given by

V 0 (xi � wi)U 0 (wi)

= �;

which is Borch�s rule.

10.6.2 Second Best

Consider now the case in which the action is unobservable and therefore

w (xi; a) = w (xi)

for all a. Suppose the principal wants to induce high e¤ort, then the incentive constraint is:

phu (wh) + (1� ph)u (wl)� ch � plu (wh) + (1� pl)u (wl)� cl

or

(ph � pl) (u (wh)� u (wl)) � ch � cl (10.7)

as pl ! ph wh�wl must increase, incentives become more high-powered. The principal also has to respecta participation constraint (or individual rationality constraint)

phu (wh) + (1� ph)u (wl)� ch � �U (10.8)

We �rst show that both constraints will be binding if the principal maximizes

maxfwh;wlg

ph (xh � wh) + (1� ph) (xl � wl)

10.7 General Model with �nite outcomes and actions 49

subject to (10.7) and (10.8). For the participation constraint, principal could lower both payments and bebetter o¤. For the incentive constraint, subtract

(1� ph) "u0 (wh)

from wh and add

ph"

u0 (wl)

to wl. Then the incentive constraint would still hold for " su¢ ciently small. Consider then participationconstraint. We would substract utility

(1� ph) "

from u (wh) and add

ph"

to u (wl), so that the expected utility remains constant. But the wage bill would be reduced for the principalby

"ph (1� ph)�

1

u0 (wh)� 1

u0 (wl)

�> 0;

since wh > wl by concavity. The solution is :

u (wl) = U �chpl � phclph � pl

and

u (wh) = U �chpl � phclph � pl

+ch � clph � pl

10.7 General Model with �nite outcomes and actions

Suppose

xi 2 fx1; :::; xIg

and

aj 2 fa1; :::aJg

and the probability

pij = Pr (xi jaj )

the utility is u (w)� a for agent and x� w for the principal.

50

10.7.1 Optimal Contract

The principals problem is then

maxfwigIi=1;j

(IXi=1

(xi � wi) pij

)

given the wage bill the agent selects aj if and only if

IXi=1

u (wi) pij � aj �IXi=1

u (wi) pik � ak (�k)

and

IXi=1

u (wi) pij � aj � U (�)

Fix aj , then the Lagrangian is

L (wij ; �; �) =(

IXi=1

(xi � wi) pij

)+Xk 6=j

(IXi=1

u (wi) (pij � pik)� (aj � ak))

+�

(IXi=1

u (wi) pij � aj

)Di¤erentiating with respect to wij yields

1

u0 (wi)= �+

Xk 6=j

�k

�1� pik

pij

�; 8i (10.9)

With a risk-avers principal the condition (10.9) would simply by modi�ed to:

v0 (xi � wi)u0 (wi)

= �+Xk 6=j

�k

�1� pik

pij

�; 8i (10.10)

In the absence of an incentive problem, or �k = 0, (10.10) states the Borch rule of optimal risk sharing:

v0 (xi � wi)u0 (wi)

= �; 8i.

which states that the ratio of marginal utilities is equalized across all states i.We now ask for su¢ cient conditions so that higher outcomes are rewarded with higher wages. As wi

increases as the right side increases, we might ask for conditions when the rhs is increasing in i. To dothis we may distinguish between the �downward�binding constraints (�k; k < j) and the �upward�bindingconstraints (�k; k > j). Suppose �rst then that there were only �downward�binding constraints, i.e. eitherj = J , or �k = 0 for k > j. Then a su¢ cient for monotonicity in i would clearly be that

pikpij

is decreasing in i for all k < j.

10.7 General Model with �nite outcomes and actions 51

10.7.2 Monotone Likelihood Ratio

The ratio

pikpij

are called likelihood ratio. We shall assume a monotone likelihood ratio condition, or 8i < i0;8k < j :pikpij

>pi0kpi0j

or reversing the ratio

pijpik

<pi0jpi0k

so that the likelihood ratio increases with the outcome. In words, for a higher outcome xi0 , it becomes morelikely, the action pro�le was high (j) rather than low (k). By modifying the ratio to:

pi0kpik

<pi0jpij

we get yet a di¤erent interpretation. An increase in the action increases the relative probability of a highoutcome versus a low outcome.The appearance of the likelihood ratio can be interpreted as the question of estimating a �parameter�a

from the observation of the �sample�xi as the following two statements are equivalent:

aj is the maximum likelihood estimator of a given xi , 8k; pikpij � 1.

As is explained in Hart-Holmström [1986]:

The agent is punished for outcomes that revise belief about aj down, while he is rewarded for out-comes that revise beliefs up. The agency problem is not an inference problem in a strict statisticalsense; conceptually, the principal is not inferring anything about the agent�s action from x becausehe already knows what action is being implemented. Yet, the optimal sharing rule re�ects preciselythe pricing of inference.

�1� pik

pij

�> 0 if pik < pij�

1� pikpij

�< 0 if pik > pij

The monotonicity in the optimal contract is then established if we can show that the downward bindingconstraints receive more weight than the upward binding constraints. In fact, we next give conditions whichwill guarantee that �k = 0 for all k > j.

10.7.3 Convexity

The cumulative distribution function of the outcome is convex in a: for aj < ak < al and � 2 [0; 1] suchthat for

ak = �aj + (1� �) al

we have

52

Pik � �Pij + (1� �)Pil

The convexity condition then states that the return from action are stochastically decreasing. Next weshow in two steps that monotone likelihood and convexity are su¢ cient to establish monotonicity. Theargument is in two steps and let aj be the optimal action.First, we know that for some j < k, �j > 0. Suppose not, then the optimal choice of ak would be same

if we were to consider A or fak; :::; aJg. But relative to fak; :::; aJg we know that ak is the least cost actionwhich can be implemented with a constant wage schedule. But extending the argument to A, we know thata constant wage schedule can only support the least cost action a1.Second, consider fa1; :::; akg. Then ak is the most costliest action and by MLRP wi is increasing in i.

We show that ak remains the optimal choice for the agent when we extend his action to A. The proof isagain by contradiction. Suppose there is l > k s.th.

IXi=1

u (wi) pik � ak <IXi=1

u (wi) pil � al:

and let j < k be an action where �k > 0 and hence

IXi=1

u (wi) pik � ak =IXi=1

u (wi) pij � aj :

Then there exists � 2 [0; 1] such that

ak = �aj + (1� �) al

and we can apply convexity, by �rst rewriting pik and using the cumulative distribution function

IXi=1

u (wi) pik � ak =IXi=1

(u (wi)� u (wi+1))Pik + u (wI)� ak

By rewriting the expected value also for aj < ak < al, we obtain

IXi=1

(u (wi)� u (wi+1))Pik + u (wI)� ak �

IXi=1

(u (wi)� u (wi+1))Pij + u (wI)� aj

!

+(1� �)

IXi=1

(u (wi)� u (wi+1))Pil + u (wI)� al

!The later is of course equivalent to

IXi=1

u (wi) pij � aj

!+ (1� �)

IXi=1

u (wi) pil � al

!

The inequality follows from the convexity and the fact that we could assume wi to be monotone, so thatu (wi)� u (wi+1) � 0.A �nal comment on the monotonicity of the transfer function : Perhaps a good reason why the function

wi must be monotonic is that the agent may be able to arti�cially reduce performance. In the aboveexample, he would then arti�cially lower it from xj to xi whenever the outcome is xj .

10.8 Information and Contract 53

10.8 Information and Contract

10.8.1 Informativeness

We now ask how does the informativeness of a signal structure a¤ect the payo¤ of the agent. Consider astochastic matrix Q which modi�es the probabilities pij such that

pkj =IXi=1

qkipij ; (10.11)

such that qki � 0 and

KXk=1

qki = 1

for all i. The matrix is called stochastic as all its entries are nonnegative and every column adds up toone. Condition (10.11) is a generalization of the following idea. Consider the information structure givenby pij . Each time the signal xi is observed, it is garbled by a stochastic mechanism that is independent ofthe action aj , which may be interpreted here as a state. It is transformed into a vector of signals bxk. Theterm qki can be interpreted as a conditional probability of bxk given xi. Clearly pkj are also probabilities.Suppose the outcomes undergo a similar transformation so that

KXk=1

xkpkj =IXi=1

xipij ;

for all j. Thus the expected surplus stays the same for every action. It can be achieved (in matrix language)by assuming that x = Q�1x.In statistical terms, inferences drawn on aj after observing bx with probabilities bp will be less precise

than the information drawn from observing x with probability p.Consider now in the (p; x) model a wage schedule bwi which implements aj . Then we �nd a new wage

schedule for the (p; x) problem based on bw such that aj is also implementable in (p; x) problem, yet wewill see that it involves less risk imposed on the agent and hence is better for the principal. Let

u (wi) =KXk=1

qkiu ( bwk) (10.12)

We can then writeIXi=1

piju (wi) =IXi=1

pij

KXk=1

qkiu ( bwk)!

=KXk=1

pkju ( bwk)But the implementation in (p; x) is less costly for the principal than the one in (p; x) which proves theresult. But (10.12) immediately indicates that it is less costly for the principal as the agent has a concaveutility function. This shows in particular that the optimal contract is only a second best solution as the�rst best contains no additional noise.

54

10.8.2 Additional Signals

Suppose besides the outcome, the principal can observe some other signal, say y 2 Y = fy1; ::::; yl; :::; yLg,which could inform him about the performance of the agent. The form of the optimal contract can thenbe veri�ed to be

1

u0�wli� = �+X

k 6=j�k

1� p

lik

plij

!.

Thus the contract should integrate the additional signal y if there exists some xi and yl such that for ajand ak

plikplij

6= pl0

ik

pl0ij

;

but the inequality simply says that

x

is not a su¢ cient statistic for (x; y) as it would be if we could write the conditional probability as follows

f (xi; yl jaj ) = h (xi; yl) g (xi jaj ) :

Thus, Hart-Holmström [1986] write:

The additional signal s will necessarily enter an optimal contract if and only if it a¤ects theposterior assessment of what the agent did; or perhaps more accurately if and only if s in�uencesthe likelihood ratio.

10.9 Linear contracts with normally distributed performance and exponentialutility

This constitutes another �natural� simple case. Performance is assumed to satisfy x = a + ", where " isnormally distributed with zero mean and variance �2. The principal is assumed to be risk neutral, whilethe agent has a utility function :

U(w; a) = �e�r(w�c(a))

where r is the (constant) degree of absolute risk aversion (r = �U 00=U 0), and c(a) = 12ca

2:We restrict attention to linear contracts:

w = �x+ �:

A principal trying to maximize his expected payo¤ will solve :

maxa;�;�

E(x� w)

subject to :

10.9 Linear contracts with normally distributed performance and exponential utility 55

E(�e�r(w�c(a))) � U(w)

and

a 2 argmaxa

E(�e�r(w�c(a)))

where U(w) is the default utility level of the agent, and w is thus its certain monetary equivalent.

10.9.1 Certainty Equivalent

The certainty equivalent w of random variable x is de�ned as follows

u (w) = E [u (x)]

The certainty equivalent of a normally distributed random variable x under CARA preferences, hence wwhich solves

�e�rw = E��e�rx

�has a particularly simple form, namely

w = E [x]� 12r�2 (10.13)

The di¤erence between the mean of random variable and its certain equivalent is referred to as the riskpremium:

1

2r�2 = E [x]� w

10.9.2 Rewriting Incentive and Participation Constraints

Maximizing expected utility with respect to a is equivalent to maximizing the certainty equivalent wealthbw(a) with respect to a, where bw(a) is de�ned by�e�r bw(a) = E(�e�r(w�c(a)))

Hence, the optimization problem of the agent is equivalent to:

a 2 argmax f bw(a)g =2 argmax

��a+ � � 1

2ca2 � r

2�2�2

�which yields

a� =�

c

Inserting a� into the participation constraint

��

c+ � � 1

2c��c

�2� r

2�2�2 = �w

56

yields an expression for �,

� = �w +r

2�2�2 � 1

2

�2

c

This gives us the agent�s e¤ort for any performance incentive �. The principal then solves :

max�

��

c� ( �w + r

2�2�2 +

1

2

�2

c)

�The �rst order conditions are

1

c� (r��2 + �

c) = 0;

which yields :

�� =1

1 + rc�2

E¤ort and the variable compensation component thus go down when c (cost of e¤ort); r (degree of risk aver-sion), and �2 (randomness of performance) go up, which is intuitive. The constant part of the compensationwill eventually decrease as well as r; c or �2 become large, as

� = �w +

1

2

r�2 � 1c

(1 + rc�2)2

!:

10.10 Readings

The following are classic papers in the literature on moral hazard: ?, ?, ?, ?, ?, ?,?,?.

[Lecture 7]

Part V

Mechanism Design

11. Introduction

Mechanism design. In this lecture and for the remainder of this term we will look at a special class ofgames of incomplete information, namely games of mechanism design. Examples of these games include: (i)monopolistic price discrimination, (ii) optimal taxation, (iii) the design of auctions, and (iv) the provisionof public goods.The adverse selection model when extended to many agent will ultimately form the theoretical center-

piece of the lectures. In the later case we refer to it as the

many informed agents:mechanism design(social choice)

�mechanism design�problem which encompasses auctions, bilateral trade, public good provision, taxationand many other general asymmetric information problems. The general mechanism design problem can berepresented in a commuting diagram:

private information social choice(type) of agent i 2 I (allocation, outcome)

f�i 2 �igi2I �! f : � ! X �! x 2 X&

si : �i !Mi g :M ! X

& %fmi 2Migi2Imessages:from agentto principal

The upper part of the diagram represents the social choice problem, where the principal has to decide onthe allocation contingent on the types of the agent. The lower part represents the implementation problem,where the principals attempts to elicit the information by announcing an outcome function which mapsthe message (information) he receives from the privately informed agents into allocation. From the pointof view of the privately informed agents, this then represents a Bayesian game in which they can in�uencethe outcome through their message.Information extraction, truthtelling, and incentives.All these examples have in common there is

a �principal�(social planner, monopolist, etc.) who would like to condition her action on some informationthat is privately known by the other players, called �agents�. The principal could simply ask the agentsfor their information, but they will not report truthfully unless the principal gives them an incentive to doso, either by monetary payments or with some other instruments that she controls. Since providing theseincentives is costly, the principal faces a trade-o¤ that often results in an e¢ cient allocation.Principal�s objective. The distinguishing characteristic of the mechanism design approach is that

the principal is assumed to choose the mechanism that maximizes her expected utility, as opposed to usinga particular mechanism for historical or institutional reasons. (Caveat: Since the objective function of the

60 11. Introduction

principal could just be the social welfare of the agents, the range of problem which can be studies is ratherlarge.)Single agent. Many applications of mechanism design consider games with a single agent (or equivalent

with a continuum of in�nitesimal small agents). For example, in a second degree price discrimination bya monopolist, the monopolist has incomplete information about the consumer�s willingness to pay for thegood. By o¤ering a price quantity schedules he attempts to extract the surplus from the buyer.Many agents. Mechanism design can also be applied to games with several agents. The auctions we

studied are such an example. In a public good problem, the government has to decide whether to supplya public good or not. But as it has incomplete information about how much the good is valued by theconsumers, it has to design a scheme which determines the provision of the public good as well as transferto be paid as a function of the announced willingness to pay for the public good.Three steps. Mechanism design is typically studied as three-step game of incomplete information,

where the agents� type - e.g. their willingness to pay - are private information. In step 1, the principaldesigns a �mechanism�, �contract�, or �incentive scheme�. A mechanism is a game where the agents sendsome costless message to the principal, and as a function of that message the principal selects an allocation.In step 2, the agents simultaneously decide whether to accept or reject the mechanism. An agent who rejectsthe mechanism gets some exogenously speci�ed �reservation utility�. In step 3, the agents who acceptedthe mechanism play the game speci�ed by the mechanism.Maximization subject to constraints. The study of mechanism design problems is therefore formally

a maximization problem for the principal subject to two classes of constraints. The �rst class is calledthe �participation� or �individual rationality� constraint, which insures the participation of the agent.The second class are the constraints related to thruthtelling, what we will call �incentive compatibility�constraints.E¢ ciency and distortions. An important focus in mechanism design will be how these two set

of constraints in their interaction can prevent e¢ cient outcomes to arise: (i) which allocation y can beimplemented, i.e. is incentive compatible?, (ii) what is the optimal choice among incentive compatiblemechanisms?, where optimal could be e¢ cient, revenue maximizing.

12. Adverse selection: Mechanism Design with One Agent

Often also called self-selection, or �screening�. In insurance economics, if a insurance company o¤ers atari¤ tailored to the average population, the tari¤ will only be accepted by those with higher than averagerisk.

12.1 Monopolistic Price Discrimination with Binary Types

A simple model of wine merchant and wine buyer, who could either have a coarse or a sophisticated taste,which is unobservable to the merchant. What qualities should the merchant o¤er and at what price?The model is given by the utility function of the buyer, which is

v (�i; qi; ti) = u (�i; qi)� ti = �iqi � ti; i 2 fl; hg (12.1)

where �i represent the marginal willingness to pay for quality qi and ti is the transfer (price) buyer i hasto pay for the quality qi. The taste parameters �i satis�es

0 < �l < �h <1. (12.2)

The cost of producing quality q � 0 is given by

c (q) � 0; c0 (q) > 0; c00 (q) > 0. (12.3)

The ex-ante (prior) probability that the buyer has a high willingness to pay is given by

p = Pr (�i = �h)

We also observe that the di¤erence in utility for the high and low valuation buyer for any given quality q

u (�h; q)� u (�l; q)

is increasing in q. (This is know as the Spence-Mirrlees sorting condition.). If the taste parameter �i werea continuous variable, the sorting condition could be written in terms of the second cross derivative:

@2u (�; q)

@�@q> 0,

which states that taste � and quality q are complements. The pro�t for the seller from a bundle (q; t) isgiven by

� (t; q) = t� c (q)

62 12. Adverse selection: Mechanism Design with One Agent

12.1.1 First Best

Consider �rst the nature of the socially optimal solution. As di¤erent types have di¤erent preferences, theyshould consume di¤erent qualities. The social surplus for each type can be maximized separately by solving

maxqif�iqi � c (qi)g

and the �rst order conditions yield:

qi = q�i () c0 (q�i ) = �i ) q�l < q

�h.

The e¢ cient solution is the equilibrium outcome if either the monopolist can perfectly discriminatebetween the types (�rst degree price discrimination) or if there is perfect competition. The two outcomesdi¤er only in terms of the distribution of the social surplus. With a perfectly discriminating monopolist,the monopolist sets

ti = �iqi (12.4)

and then solves for each type separately:

maxfti;qig

� (ti; qi)() maxfti;qig

f�iqi � c (qi)g ;

using (12.4). Likewise with perfect competition, the sellers will break even, get zero pro�t and set prices at

ti = c (q�i )

in which case the buyer will get all the surplus.

12.1.2 Second Best: Asymmetric information

Consider next the situation under asymmetric information. It is veri�ed immediately that perfect discrim-ination is now impossible as

�hq�l � t�l = (�h � �l) q�l > 0 = ��hq�h � th (12.5)

but sorting is possible. The problem for the monopolist is now

maxftl;ql;th;qhg

(1� �) tl � c (ql) + � (th � (c (qh))) (12.6)

subject to the individual rationality constraint for every type

�iqi � ti = 0 (IRi) (12.7)

and the incentive compatibility constraint

�iqi � ti = �iqj � tj (ICi) (12.8)

The question is then how to separate. We will show that the binding constraint are IRl and ICh, whereasthe remaining constraints are not binding. We then solve for tl and th, which in turn allows us to solve forqh; and leaves us with an unconstrained problem for ql.Thus we want to show

(i) IRl binding, (ii) ICh binding, (iii) qh � ql (iv) qh = q�h (12.9)

12.1 Monopolistic Price Discrimination with Binary Types 63

Consider (i). We argue by contradiction. As

�hqh � th =ICh

�hql � tl =�h>�l

�lql � tl (12.10)

suppose that �lql � tl > 0, then we could increase tl; th by a equal amount, satisfy all the constraints andincrease the pro�ts of the seller. Contradiction.Consider (ii) Suppose not, then as

�hqh � th > �hql� tl =�h>�l

�lql � tl(IRl)

= 0 (12.11)

and thus th could be increased, again increasing the pro�t of the seller.(iii) Adding up the incentive constraints gives us (ICl) + (ICl)

�h (qh � ql) = �l (qh � ql) (12.12)

and since:

�h > �l ) qh � ql = 0: (12.13)

Next we show that ICl can be neglected as

th � tl = �h (qh � ql) = �l (qh � ql) : (12.14)

This allows to say that the equilibrium transfers are going to be

tl = �lql (12.15)

and

th � tL = �h (qh � qL)) th = �h (qh � qL) + �lql.

Using the transfer, it is immediate that

qh = q�h

and we can solve for the last remaining variable, ql.

maxqlf(1� p) (�lql� (c (ql)) + p (�h (q�h � qL) + �lql � c (q�h)))g

but as q�h is just as constant, the optimal solution is independent of constant terms and we can simplifythe expression to:

maxqlf(1� p) (�lql � c (ql))� p (�h � �l) qlg

Dividing by (1� p) we get

maxql

��lql � c (ql)�

p

1� p (�h � �l) ql�

for which the �rst order conditions are

�l � c0 (ql)�p

1� p (�h � �l) ql = 0

This immediately implies that the solution ql:

64 12. Adverse selection: Mechanism Design with One Agent

() c0 (ql) < �l () ql < q�l

and the quality supply to the low valuation buyer is ine¢ ciently low (with the possibility of completeexclusion).Consider next the information rent for the high valuation buyer, it is

I (ql) = (�h � �l) ql

and therefore the rent is increasing in ql which is the motivation for the seller to depress the quality supplyto the low end of the market.The material is explicated in ?, Chapter 2.

[Lecture 8]

12.2 Continuous type model

In this lecture, we consider the adverse selection model with a continuous type model and a single agent.This will set the state for the multiple agent model.

12.2.1 Information Rent

Consider the situation in the binary type model. The allocation problem for the agent and principal canwell be represented in a simple diagram.The information rent is then

I (�h) = (�h � �l) ql

and it also represents the di¤erence between the equilibrium utility of low and high type agent, or

I (�h) = U (�h)� U (�l)

If we extend the model and think of more than two types, then we can think of the information rent asresulting from tow adjacent types, say �k and �k�1. The information rent is then given by

I (�k) = (�k � �k�1) qk�1

As the information rent is also the di¤erence between the agent�s net utility, we have

I (�k) = U (�k)� U (�k�1) ;

well as we see not quite, but more precisely:

I (�k) = U (�k j�k )� U (�k�1 j�k ) ;

where

U (�k j�l )

denotes in general the utility of an agent of (true) type �l, when he announces and pretends to be of type�k. As we then extend the analysis to a continuous type model, we could ask how the net utility of theagent of type �k evolves as a function of k. In other words as �k�1 ! �k:

U (�k j�k )� U (�k�1 j�k )�k � �k�1

=(�k � �k�1)(�k � �k�1)

qk�1

12.2 Continuous type model 65

and assuming continuity:

qk�1 ! qk;

we get

U 0 (�) = q (�)

which is identical using the speci�c preferences of our model to

U 0 (�) =@u

@�(q (�) ; �) :

Thus we have a description of the equilibrium utility as a function of q (�) alone rather than (q (�) ; t (�)).The pursuit of the adverse selection problem along this line is often referred to as �Mirrlees�trick�.

12.2.2 Utilities

The preferences of the agent are described by

U (x; �; t) = u (x; �)� t

and the principals

V (x; �; t) = v (x; �) + t

and u; v 2 C2. The type space � 2 � � R+.The social surplus is given by

S (x; �) = u (x; �) + v (x; �) :

The uncertainty about the type of the agent is given by:

f (�) ; F (�) .

We shall assume the (strict) Spence-Mirrlees conditions

@u

@�> 0;

@2u

@�@x> 0:

12.2.3 Incentive Compatibility

De�nition 12.2.1. An allocation is a mapping

� ! y (�) = (x (�) ; t (�)) .

De�nition 12.2.2. An allocation is implementable if y = (x; t) satis�es truthtelling:

u (x (�) ; �)� t (�) � u�x�b�� ; ��� t�b�� ; 8�;b� 2 �:

De�nition 12.2.3. The truthtelling net utility is denoted by

U (�) , U (� j� ) = u (x (�) ; �)� t (�) ; (12.16)

and the net utility for agent �, misreporting by telling � is denoted by

U�� j��= u

�x�b�� ; ��� t�b�� .

66 12. Adverse selection: Mechanism Design with One Agent

We are interested in (i) describing which contracts can be implemented and (ii) in describing optimalcontracts.

Theorem 12.2.1. The direct mechanism y (�) = (x (�) ; t (�)) is incentive compatible if and only if:

1. the truthtelling utility is described by:

U (�)� U (0) =�Z0

u� (x (s) ; s)) ds; (12.17)

2. x (s) is nondecreasing.

Remark 12.2.1. The condition (12.17) can be restated in terms of �rst order conditions:

dU

d�=@u

@y

@y

@�+@u

@�= 0: (12.18)

Proof. Necessity. Suppose truthtelling holds, or

U (�) � U�� j��;8�;b�;

which is:

U (�) = U�b� j�� M

= U�b��+ u�x�b�� ; ��� u�x�b�� ;b�� ;

and thus

U (�)� U�b�� � u�x�b�� ; ��� u�x�b�� ;b�� : (12.19)

A symmetric condition gives us

U���� U (�) � u

�x (�) ; �

�� u (x (�) ; �) : (12.20)

Combining (12.19) and (12.20), we get:

u (x (�) ; �)� u�x (�) ; �

�� U (�)� U

�b�� � u�x�b�� ; ��� u�x�b�� ;b�� :Suppose without loss of generality that � > �, then monotonicity of x (�) is immediate from

@2u

@�@x.

Dividing by � � �; and taking the limit as

� ! b�at all points of continuity of x (�) yields

dU (�)

d (�)= u� (x (�) ; �) ;

and thus we have the representation of U . As x (�) is nondecreasing, it can only have a countable number ofdiscontinuities, which have Lebesgue measure zero, and hence the integral representation is valid, indepen-dent of continuity properties of x (�), which is in contrast with the representation of incentive compatibilityby the �rst order condition (12.18).

12.3 Optimal Contracts 67

Su¢ ciency. Suppose not, then there exists � and b� s.th.U�b� j�� = U (�) : (12.21)

Suppose without loss of generality that � � �. The inequality (12.21) implies that the inequality in (12.19)is reversed:

u�x�b�� ; ��� u�x�b�� ;b�� > U (�)� U �b��

integrating and using (12.17) we get

�Z�

us

�x�b�� ; s� ds > �Z

us (x (s) ; s) ds

and rearranging

�Z�

hus

�x�b�� ; s�� us (x (s) ; s)i ds > 0 (12.22)

but since

@2u

@�@x> 0

and the monotonicity condition implied that this is not possible, and hence we have the desired contradic-tion.

[Lecture 9]

12.3 Optimal Contracts

We can now describe the problem for the principal.

maxy(�)

E� [v (x (�) ; �) + t (�)]

subject to

u (x (�) ; �)� t (�) � u�x���; ��� t���; 8�; �

and

u (x (�) ; �)� t (�) � �u.

The central idea is that we restate the incentive conditions for the agents by the truthtelling utility asderived in Theorem 12.2.1.

68 12. Adverse selection: Mechanism Design with One Agent

12.3.1 Optimality Conditions

In this manner, we can omit the transfer payments from the control problem and concentrate on the optimalchoice of x as follows:

maxx(�)

E� [S (x (�) ; �)� U (�)] (12.23)

subject to

dU (�)

d�= u� (x (�) ; �) (12.24)

and

x (�) nondecreasing (12.25)

and

U (�) = �u; (12.26)

which is the individual rationality constraint. Next we simplify the problem by using integration by parts.Recall, that integration by parts, used in the following formZ

dU (1� F ) = U (1� F ) +ZUf

and henceZUf = �U (1� F ) +

ZdU (1� F )

As

E� [U (�)] =Z 1

0

U (�) f (�) d� = � [U (�) (1� F (�))]10 +Z 1

0

dU (�)

d�

1� F (�)f (�)

f (�) d� (12.27)

we can rewrite (12.23) and using (12.24), we get

maxE��S (x (�) ; �)� 1� F (�)

f (�)u� (x (�) ; �)� U (0)

�(12.28)

subject to (12.25) and (12.26).and thus we removed the transfers out of the program.

12.3.2 Pointwise Optimization

Consider the optimization problem for the principal and omitting the monotonicity condition, we get

� (x; �) = S (x; �)� 1� F (�)f (�)

u� (x; �) : (12.29)

Assumption 12.3.1 � (x; �) is quasiconcave and has a unique interior maximum in x 2 R+ for all � 2 �:

Theorem 12.3.2. Suppose that Assumption 12.3.1 holds, then the relaxed problem is solved at y = y (�)if:

12.3 Optimal Contracts 69

1. �x (x (�) ; �) = 0:2. the transfer t (�) satisfy

t (�) = u (x (�) ; �)�

0@U (0) + �Z0

u� (x (s) ; s) ds)

1AThen y = (x; t) solve the relaxed program.

The idea of the proof is simple. By integration by parts we removed the transfers. Then we can choosex to maximize the objective function and later choose t to solve the di¤erential equation. Given that@u=@� > 0; the IR contract can be related as

U (0) = �u.

It remains to show that x is nondecreasing.

Assumption 12.3.3 @2� (x; �) =@x@� � 0.

Theorem 12.3.4. Suppose that assumptions 12.3.1 and12.3.3 are satis�ed. Then the solution to the relaxedprogram satis�es the original solution in the unrelaxed program.

Proof. Di¤erentiating

�x (x (�) ; �) = 0 for all �

and hence we obtain

=) dx (�)

d�= ��x�

�xx

and as

�xx � 0

has to be satis�ed locally, we know that x (�) has to be increasing, which concludes the proof.

Next we interpret the �rst order conditions and obtain �rst results:

Sx (x (�) ; �) =1� F (�)f (�)

ux� (x(�) ; �) = 0

implies e¢ cient provision for � = 1 and underprovision of the contracted activity for all but the highesttype � = 1: To see that x (�) < x� (�), we argue by contradiction. Suppose not, or x (�) � x� (�). Observe�rst that for all x and � < 1

@� (x; �)

@x<@S (x; �)

@x(12.30)

by:

@2u (x; �)

@x@�> 0. (12.31)

Thus at x� (�), by de�nition

@S (x� (�) ; �)

@x= 0

70 12. Adverse selection: Mechanism Design with One Agent

and using (12.30)

@� (x� (�) ; �)

@x<@S (x� (�) ; �)

@x= 0

But as � (x; �) is quasiconcave the derivative with respect to x satis�es the (downward single crossingcondition), which implies that for

@� (x (�) ; �)

@x= 0

to hold x (�) < x� (�).The �rst order condition can be written as to represent the trade-o¤:

f (�)Sx (x (�) ; �)increase in joint surplus

= [1� F (�)] ux� (x (�) ; �)increase in agent rents

.

Consider then the monopoly interpretation, where:

maxp(1� F (p)) (p� c)

and the associated �rst order condition is given by:

f (p) (p� c)marginal

= 1� F (p)infra marginal

The material in this lectures is bases on ?.

[Lecture 10]

13. Mechanism Design Problem with Many Agents

13.1 Introduction

Mechanism design explores the means of implementing a given allocation of available resources when theinformation is dispensed in the economy. In general two questions arise:�! can y (�) be implemented incentive compatible?�! what is the optimal choice among incentive-compatible mechanisms?Today we shall introduce several general concepts:

� revelation principle� e¢ ciency� quasi-linear utility

We emphasize that we now look at it as a static game of the agents and hence the designer is acting non-strategically. The optimal design is a two stage problem. Mechanism design takes a di¤erent perspectiveon the game and the design of games.

13.2 Model

Consider a setting with I agents, I = f1; :::; Ig. The principal, mechanism designer, center, is endowed withsubscript 0. They must make a collective choice among a set of possible allocations Y .Each agent observes privately a signal �i 2 �i, his type, that determines his preferences over Y ,

described by a utility function ui (y; �) for all i 2 I. The prior distribution over types P (�) is assumedto be common knowledge. In general the type could contain any information agent i possesses about hispreferences, but it could also contain information about his neighbor, competitors, and alike.

De�nition 13.2.1. A social choice function is a mapping:

f : �1 � ::::��I ! Y:

The problem is that � = (�1; ::; �I) is not publicly observable when the allocation y is to be decided.However each agent can send a message to the principal mi 2 Mi. The message space can be arbitrary,with

M =I�i=1Mi.

When the agents transmit a message m, the center will choose a social allocation y 2 Y as a function ofthe message received.

De�nition 13.2.2. An outcome function is a mapping

g :M1 � :::�MI ! Y:

72 13. Mechanism Design Problem with Many Agents

Each agent transmits the message independent and simultaneous with all other agents. Thus a mecha-nism de�nes a game with incomplete information for which must choose an equilibrium concept, denotedby c. A strategy for agent i is a function

mi : �i !Mi.

De�nition 13.2.3. A mechanism � = (M1; ::;MI ; g (�)) is a collection of strategy sets (M1; :::;MI) andan outcome function

g :M1 � :::�MI ! Y:

With a slight abuse of notation, we use the same notation, Mi, for the set of strategies and their rangefor a particular agent i. Graphically, we have the following commuting diagram:

�1 � :::��If (�)! Y

&mg;c (�)

%g (�)

M1 � :::�MI

wheremg;c is the mapping that associates to every I-tuple of true characteristics �, the equilibrium messagesof this game for the equilibrium concept c. The strategy space of each individual agent is often called themessage space.

13.3 Mechanism as a Game

The central question we would like to ask whether a particular objective function, or social choice functionf (�) can be realized by the game which is the mechanism.

De�nition 13.3.1. A mechanism � = (M1; ::;MI ; g (�)) implements the social choice function f (�) if thereis an equilibrium pro�le

(m�1 (�1) ; :::;m

�I (�I))

of the game induced by � such that

g (m�1 (�1) ; :::;m

�I (�I)) = f (�1; :::; �I) .

The identi�cation of implementable social choice function is at �rst glance a complex problem because wehave to consider all possible mechanism g (�) on all possible domains of strategies M . However a celebratedresult (valid for all of the implementation versions above), the revelation principle, simpli�es the task.

De�nition 13.3.2. A mechanism is direct if Mi = �i and g (�) = f (�) for all i.

De�nition 13.3.3. A direct mechanism is said to be revealing if mi (�i) = �i for each � 2 �.

De�nition 13.3.4. The social choice function f (�) is truthfully implementable (or incentive compatible)if the direct revelation mechanism

� = (�; f (�))

has an equilibrium (m�1 (�1) ; :::;m

�I (�I)) in which m

�i (�i) = �i for all �i 2 �i, for all i.

13.3 Mechanism as a Game 73

We shall test these notions with our example of adverse selection with a single buyer. Suppose themessage space is

M

and the outcome function is

g :M ! X � T

The best response strategy of the agent is a mapping

m� : � !M

such that

m� (�) 2 argmaxm2M

fu (g (m) ; �)g

and he obtains the allocation

g (m� (�)) (13.1)

Therefore we say that � = (g (�) ;M) implement f , where f satis�es

g (m� (�)) = f (�) (13.2)

for every �. Thus based on the indirect mechanism � = (g (�) ;M) we can de�ne a direct mechanism assuggested by (13.2):

�d = (f (�) ; �) .

Theorem 13.3.1 (Revelation Principle).If the social choice function f (�) can be implemented through some mechanism, then it can be implementedthrough a direct revelation (truthful) mechanism: �d = (f (�) ; �).

Proof. Let (g;M) be a mechanism that implements f (�) through m� (�) be the equilibrium message:

f (�) = g (m� (�))

Consider the direct mechanism

�d = (f (�) ; �) .

If it were not truthful, then an agent would prefer to announce �0 rather than �, and he would get

u (f (�) ; �) < u�f(�0

�; �)

but by de�nition of implementation, and more precisely by (13.2), these would imply that

u (g (m� (�)) ; �) < u�g�m� ��0� ; ���

which is a contradiction to the hypothesis that m� (�) would be an equilibrium of the game generated bythe mechanism

� (g (�) ;M) :

74 13. Mechanism Design Problem with Many Agents

The revelation principle states that it su¢ cient to restrict attention to mechanisms which o¤er a menuof contracts: agent announces �, and will get

y (�) = (x (�) ; t (�)) (13.3)

but many mechanisms in the real world are indirect in the sense that goods are provided in di¤erentqualities and/or quantities and the transfer is then a function of the choice variable x.We can use the taxation principle which says that every direct mechanism is equivalent to a nonlinear

tari¤. How? O¤ering a menu of allocations fx (�)g�2� and let

t (x) , t (x (�)) = t (�)

be the corresponding nonlinear tari¤, so that the menu is

fx; t (x)gx2X

The proof that the nonlinear tari¤ is implementing the same allocation as the direct mechanism is easy.Suppose there are �; �0 such that x (�) = x

��0�. If t (�) 6= t

��0�, then the agent would have an interest

to misrepresent his state of the world, and hence the direct mechanism couldn�t be truthtelling. It followsthat t (�) = t

��0�. In consequence, the transfer function can be uniquely de�ned as

if x = x (�)) t (x) = t (�) ;

which leads to the nonlinear tari¤. The notion of a direct mechanism may then represent a normativeapproach, whereas the indirect mechanism represents a positive approach.

13.4 Second Price Sealed Bid Auction

Consider next the by now familiar second price auction and let us try to translate it into the language ofmechanism design. Consider the seller with two bidders and �i 2 [0; 1]. The set of allocations is Y = X�T .where yi = (xi; ti), where xi 2 f0; 1g denotes the assignment of the object to bidder i: no if 0, yes if 1. Theutility function of agent i is given by:

ui (yi; �i) = �ixi � ti

The second price sealed bid auction implements the following social choice function

f (�) = (x0 (�) ; x1 (�) ; x2 (�) ; t0 (�) ; t1 (�) ; t2 (�))

with

x1 (�) = 1; if �1 � �2; = 0 if �1 < �2:

x2 (�) = 1; if �1 < �2; = 0 if �1 � �2:x0 (�) = 0; for all �;

and

t1 (�) = �2x1 (�)

t2 (�) = �1x2 (�)

t0 (�) = t1 (�)x1 (�) + t2 (�)x2 (�)

The message mi is the bid for the object:

13.4 Second Price Sealed Bid Auction 75

mi : �i !Mi = R+

The outcome function is

g :M ! Y

with

g =

8>><>>:x1 = 1; t1 = m2

x2 = 0; t2 = 0; if m1 � m2

x1 = 0; t1 = 0x2 = 1; t2 = m1

; if m1 < m2

A variety of other examples can be given:

Example 13.4.1 (Income tax). x is the agent�s income and t is the amount of tax paid by the agent; � isthe agent�s ability to earn money.

Example 13.4.2 (Public Good). x is the amount of public good supplied, and ti is the consumer i�s monetarycontribution to �nance it; �i indexes consumer surplus from the public good.

The example of the second price sealed bid auction illustrates that as a general matter we need notonly to consider to directly implementing the social choice function by asking agents to reveal their type,but also indirect implementation through the design of institutions by which agents interact. The formalrepresentation of such institutions is known as a mechanism.

[Lecture 12]

14. Dominant Strategy Equilibrium

Example 14.0.3 (Voting Game). Condorcet paradox

Next we consider implementation in dominant strategies. We show that in the absence of prior restric-tions on the characteristics, implementation in dominant equilibria is essentially impossible.

General Environments. Next, we prove the revelation principle for the dominant equilibrium concept.Equivalent results can be proven for Nash and Bayesian Nash equilibria.

De�nition 14.0.1. The strategy pro�le m� = (m�1 (�1) ; :::;m

�I (�I)) is a dominant strategy equilibrium of

mechanism � = (M; g (�)) if for all i and all �i 2 �i :

ui (g (m�i (�i) ;m�i) ; �i) � ui (g (m0

i;m�i) ; �i)

for all m0i 6= mi, 8m�i 2M�i.

Theorem 14.0.1 (Revelation Principle). Let � = fM; g (�)g be a mechanism that implements thesocial choice function f (�) for the dominant equilibrium concept, then there exists a direct mechanism� 0 = f�; f (�)g that implements f (�) by revelation.

Proof. (Due to ?.). Let m� (�) = (:::;m�i (�i) ; :::) be an I�tuple of dominant messages for (M; g (�)). De�ne

g� to be the composition of g and m�.

g� , g �m�

or

g� (�) , g (m� (�)) . (14.1)

By de�nition

g� (�) = f (�) .

In fact

� � = (�; g� (�))

is a direct mechanism. Next we want to show that the mechanism implements f (�) as an equilibriumin dominant strategies. The proof is by contradiction. Suppose not, that is the transmission of his true

characteristics is not a dominant message for agent i. Then there exists��i; ��i

�such that

ui

�g���i; ��i

�; �i

�> ui (g

� (�) ; �i) :

But by the de�nition (14.1) above, the inequality can be rewritten as

ui

�g�m���i; ��i

��; �i

�> ui (g (m

� (�)) ; �i) :

But this contradicts our initial hypothesis that � = (M; g (�)) implements f (�) in a dominant strategyequilibrium, as m� is obviously not an equilibrium strategy for agent i, which completes the proof.

78 14. Dominant Strategy Equilibrium

The notion of implementation can then be re�ned in De�nition ?? in a suitable way. This is imple-mentation in a very robust way, in terms of strategies and in informational requirements as the designerdoesn�t need to know P (�) for the successful implementation. But can we always implement in dominantstrategies:

De�nition 14.0.2. The social choice function f (�) is dictatorial if there is an agent i such that for all� 2 �,

f (�) 2 fx 2 X : ui (x; �i) � ui (y; �i) for all y 2 Xg .

Next, we can state the celebrated result from ? and ?.

Theorem 14.0.2 (Gibbard-Satterthwaite). Suppose that X contains at least three elements, and thatRi = P for all i, and that f (�) = X. Then the social choice function is truthfully implementable if andonly if it is dictatorial.

Quasi-Linear Environments. The quasilinear environment is described by

vi (x; t; �i) = ui (x; �i)� ti.

Denote the e¢ cient allocation by x� (�). Then the generalization of the Vickrey auctions states: ?, ?, ?.

Theorem 14.0.3 (Vickrey-Clark-Groves). The social choice function f (�) = (x�; t1 (�) ; :::; tI (�)) istruthfully implementable in dominant strategies if for all i:

ti (�i) =

24Xj 6=i

uj (x� (�) ; �j)

35�24Xj 6=i

uj�x��i (��i) ; �j

�35 : (14.2)

Proof. If truth is not a dominant strategy for some agent i, then there exist �i;b�i; and ��isuch thatui

�x��b�i; ��i� ; �i�+ ti �b�i; ��i

�> vi (x

� (�i; ��i) ; �i) + ti (�i; ��i) (14.3)

Substituting from (14.2) for ti�b�i; ��i

�and ti (�i; ��i), this implies that

IXj=1

uj

�x��b�i; ��i� ; �j� > IX

j=1

uj (x� (�) ; �j) ; (14.4)

which contradicts x� (�) being an optimal policy: Thus, f (�) must be truthfully implementable in dominantstrategies.

This is often referred to as the pivotal mechanism. It is ex-post e¢ cient but it may not satisfy budgetbalance:X

i2Iti (�i) = 0.

We can now weaken the implementation requirement in terms of the equilibrium notion.

15. Bayesian Equilibrium

In many cases, implementation in dominant equilibria is too demanding. We therefore concentrate next onthe concept of Bayesian implementation. Note however that with a single agent, these two concepts areequivalent.

De�nition 15.0.3. The strategy pro�le s� = (s�1 (v1) ; :::; s�I (vI)) is a Bayesian Nash equilibrium of mech-

anism � = (S1; ::; SI ; g (�)) if for all i and all v 2 � :

Ev�i�ui�g�s�i (vi) ; s

��i (v�i)

�; vi�jvi�� Ev�i

�ui�g�si; s

��i (v�i)

�; vi�jvi�

for all si 6= s�i (v), 8s��i (v).

15.1 First Price Auctions

We replace the type notation v by the vector of private valuations

v = (v1; :::; vI)

and denote the distribution over types by symmetric

f (vi) ; F (vi) .

For agent i to win with value v, he should have a higher value than all the other agents, which happenswith probability

G (v) , (F (v))I�1

and the associated density is

g (v) = (I � 1) f (v) (F (v))I�2 :

This is often referred to as the �rst order statistic of the sample of size I � 1.

Theorem 15.1.1. The unique symmetric Bayes-Nash equilibrium in the �rst price auction is given by

b� (v) =

Z v

0

y

�g (y)

G (v)

�dy

Remark 15.1.1. The expression in the integral is thus the expected value of highest valuation among theremaining I � 1 agents, provided that the highest valuation is is below v. (Thus the bidder acts if he wereto match the expected value of his competitor.)

80 15. Bayesian Equilibrium

Proof. We start the proof by making some �inspired guesses,� and then go back to rigorously prove theguesses. This technique is how the �rst economist to study auctions might have originally derived theequilibrium. That economist might have made the following good guesses about the equilibrium b� (�)being sought:(i) All types of bidder submit bids: b (v) 6= ; for all v � 0:(ii) Bidders with greater values bid higher: b (v) > b (z) for all v > z:(iii) The function b (�) is di¤erentiable.Obviously, many functions other than b� (�) satisfy Guesses (i) � (iii). We now show that the only

possible equilibrium satisfying them is b� (�). From (ii) and (iii), b (�), has an inverse function, which wedenote as � (�) ; that satis�es b

���b��= b for all numbers b in the range of b (�) : Value v = �

�b�is the

value a bidder must have in order to bid b when using strategy b (�). Because b (�) strictly increases, theprobability that bidder i wins if he bids b is

Q�b�= F

���b��n�1

, G (� (b)) : (15.1)

Why? Well, observe that since b (�) is strictly increasing and the other bidders�values are continuouslydistributed, we can ignore ties: another bidder j bids the same b only if his value is precisely vj = �

�b�;

which occurs with probability zero. So the probability of bidder i winning with bid b is equal to the prob-ability of the other bidders bidding no more than b: Because b (�) strictly increases, this is the probabilitythat each other bidder�s value is not more than �

�b�: This probability is G (� (b)) = F

���b��n�1

; the

probability that the maximum of the bidders�value is no more than � (b) : Thus, the expected pro�t of atype v bidder who bids b when the others use b (�) is

��v; b�=�v � b

�G���b��: (15.2)

Because of (iii), the inverse function � (�) is di¤erentiable. Letting �0�b�be its derivative at b, the partial

derivative of ��v; b�with respect to b is

�b (v; b) = �G (� (b)) + (v � b) g (� (b))�0 (b) : (15.3)

Since b (�) is an equilibrium, b (v) is an optimal bid for a type v bidder, i.e. b = b (v) maximizes � (v; b) :The �rst order condition �b (v; b (v)) = 0 holds:

�G (� (b (v))) + (v � b (v)) g (� (b(v))�0 (b (v)) 0: (15.4)

Use � (b (v)) = v and �0 (b (v)) = 1=b0 (v) to write this as

�G (v)) + (v � b (v)) g (v)b0 (v)

= 0: (15.5)

This is a di¤erential equation that almost completely characterizes the exact nature of b (�) : It is easy tosolve. Rewrite it as

G (v) b0 (v) + g (v) = vg (v) : (15.6)

Since g (v) = G0 (v), the left side is the derivative of G (v) b (v) : So we can integrate both sides from anyv0 to any v (using �y�to denote the dummy integration variable):

G (v) b (v)�G (v0) b (v0) =Z v

0

yg (y) dy (15.7)

15.2 Optimal Auctions 81

>From Guess A, all types bid, so we can take v0 ! 0:We know b (0) � 0; as r = 0: Hence, G (v0) b (v0)! 0as v0 ! 0: Take this limit in (6:5) and divide by G (v) to obtain

b (v) =

Z v

0

y

�g (y)

G (v)

�dy = b� (v) : (15.8)

This is the desired result.

Finally, we can rewrite (??) using integration by parts:Zu0v = uv �

Zuv0

as

b� (v) = v �Z v

0

�F (y)

F (v)

�I�1dy

15.2 Optimal Auctions

15.2.1 Revenue Equivalence

We are now describing the optimal auction and the revenue equivalence result on the basis of the result weobtained earlier. Recall the following theorem which describes the incentive compatible equilibrium utilityof agent i:

Theorem 15.2.1. The direct mechanism y (v) = (q (v) ; t (v)) is incentive compatible if and only if:

1. the truthtelling utility is described by:

Ui (vi)� Ui (0) =viZ0

uvi (qi (si) ; si)) dsi; (15.9)

2. qi (si) is nondecreasing.

The utility function of agent i in the single unit auction is

u (vi; qi) = viqi

where

qi = Pr (xi = 1)

Thus we can rewrite the theorem in our context as follows

Theorem 15.2.2. The direct mechanism to sell a single object y (v) = (q (v) ; t (v)) is incentive compatibleif and only if:

1. the truthtelling utility is described by:

Ui (vi)� Ui (0) =viZ0

qi (si) dsi; (15.10)

82 15. Bayesian Equilibrium

2. qi (si) is nondecreasing.

We can now state the celebrated revenue equivalence theorem:

Theorem 15.2.3 (Revenue Equivalence).Given any two auctions mechanism (direct or not) which agree

1. on the assignment probabilities qi (v) for every v2. on the equilibrium utility Ui (0) for all

lead to identical revenues.

In fact, we have the remarkable revenue equivalence theorem. The argument relies on the important(and more general) revelation principle. To see which social choice rules are implementable by some game,it is enough to focus attention on direct revelation games where players truthfully report their typesand an allocation is chosen (perhaps randomly) as a function of their reported types.

vi � bi =viZ0

qi (si) dsi

15.2.2 Optimal Auction

We described the optimization problem by the principal with the single agent as follows:

maxEv�S (q (v) ; v)� 1� F (v)

f (v)uv (q (v) ; v)� U (0)

�(15.11)

and the extension to many agents is trivial:

maxq(v)

Ev

"S (q (v) ; v)�

IXi=1

�1� Fi (vi)fi (vi)

uiv (qi (v) ; vi)� Ui (0)�#

(15.12)

and making use of the auction environment we get

maxq(v)

1Zv1=0

::

1ZvI=0

IXi=1

qi (v)

�vi �

�1� Fi (vi)fi (vi)

��!�"IYi=1

fi (vi)

#dv1 � � � dvI

where

q = (q1; :::; qI) (15.13)

and

qi � 0;IXi=1

qi � 1.

But the optimization problem can be solved pointwise for every v = (v1; :::; vI):

maxq

IXi=1

qi (v)

�vi �

�1� Fi (vi)fi (vi)

��

15.2 Optimal Auctions 83

Let us denote, following Myerson, by �vi the virtual utility of agent i

~vi , vi �1� Fi (vi)fi (vi)

It is then optimal to set the reserve price for every agent such that ri = vi satis�es:

vi �1� Fi (vi)fi (vi)

= 0

84 15. Bayesian Equilibrium

Theorem 15.2.4 (Optimal Auction). The optimal policy q� = (q�1 ; :::; q�I ) is then given by:

1. q�i = 0 if all ~vi < 02.P

i q�i = 1 if 9~vi � 0

3. q�i > 0) ~vi = max f~v1; :::; ~vIg.

15.3 Additional Examples

15.3.1 Procurement Bidding

a single buyer and many sellers competing to sell the service or product. The private information is nowregarding the cost of providing the service.the virtual utility is now

ci +Fi (ci)

fi (ci)

information rent with privately informed buyers is due to the fact that buyers wish to understatetheir valuation.information rent with privately informed sellers is due to the fact that sellers wish to overstate their

cost.

15.3.2 Bilateral Trade

Consider a seller (with valuation v1) and a buyer (with valuation v2) who wishes to engage in a trade:virtual utility now becomes identical to:

Z Z �v2 �

1� F2 (v2)f2 (v2)

���v1 +

F1 (v1)

f1 (v1)

�point wise optimization occurs when�

v2 �1� F2 (v2)f2 (v2)

���v1 +

F1 (v1)

f1 (v1)

�> 0

but for

v2 = v1

the di¤erence of the virtual utilities is negative

�v2 �

1� F2 (v2)f2 (v2)

���v1 +

F1 (v1)

f1 (v1)

�< 0

and thus in general ine¢ cient trade with two sided private information

15.3.3 Bilateral Trade: Myerson and Sattherthwaite

15.3.4 Readings

The surveys by ? and ? are highly recommended. The notion of incentive compatible and the basic structureof a collective decision rule preserving some kind of informational decentralization is due to ?. The variousnotions of e¢ ciency are discussed in ?.

]

16. E¢ ciency

A game theorist or a mediator who analyzes the Pareto e¢ ciency of behavior in a game with incompleteinformation must use the perspective of an outsider, so he cannot base his analysis on the players�privateinformation. An outsider may be able to say how the outcome will depend on the players�types. Thatis, he can know the mechanisms but not its outcome. Thus, Holmstrom and Myerson (1983) argued thatthe concept of e¢ ciency for games with incomplete information should be applied to the mechanisms,rather than to outcomes, and the criteria for determining whether a particular mechanism � is e¢ cientshould depend only on the commonly known structure of the game, not on the privately own types of theindividual players.

16.1 First Best

A de�nition of Pareto e¢ ciency in a Bayesian collective-choice problem is: A mechanism is e¢ cient ifand only if no other feasible mechanism can be found that might make some other individuals better o¤and would certainly not make other individuals worse o¤. However, this de�nition is ambiguous in severalways. In particular, we must specify what information is to be considered when determining whether anindividual is �better o¤�or �worse o¤.�A mechanism can now be thought of as contingent allocation plan

: � ! � (Y )

and consequently

(y j� )

expresses the conditional probability that y is realized given a type pro�le �.One possibility is to say that an individual is made worse o¤ by a change that decreases his expected

utility payo¤ as would be computed before his own type or any other individuals�type are speci�ed. Thisstandard is called the ex ante welfare criterion. Thus, we say that a mechanism is ex ante Pareto superiorto another mechanism � if and only if

X�2�

Xy2Y

p (�) (y j� )ui (y; �) �X�2�

Xy2Y

p (�)� (y j� )ui (y; �) ;8i 2 I; (16.1)

and this inequality is strict for at least one player in I. Next de�ne

Ui (� j�i ) =X

��i2��i

Xy2Y

pi (��i j�i )� (y j� )ui (y; �)

and notice that

86 16. E¢ ciencyX�2�

Xy2Y

p (�)� (y j� )ui (y; �) =X�i2�i

pi (�i)Ui (� j�i ) : (16.2)

Another possibility is to say that an individual is made worse o¤ by a change that decreases hisconditionally expected utility, given his own type (but to given the type of any other individuals). Anoutside observer, who does not know any individual�s type, would then say that a player i �would certainlynot be made worse o¤ (by some change of mechanism)�in this sense if this conditionally expected utilitywill not be decreased (by the change) for any possible type of player i: This standard is called the interimwelfare criterion, because it evaluates each player�s welfare after he learns his own type but before helearns any other player�s type. Thus, we say that a mechanism is interim Pareto superior to anothermechanism � if and only if

Ui ( j�i ) � Ui (� j�i ) ;8i 2 I;8�i 2 �i; (16.3)

and this inequality is strict for at least one type of one player in I.Yet another possibility is to say that an individual is made worse o¤ by a change that decreases his

conditionally expected utility, given the types of all individuals. An outsider observer would then say thata player �would certainly not be made worse o¤�in this sense if his conditionally expected utility wouldnot be decreased for any possible combination of types for all the players. This standard is called ex postwelfare criterion, because it uses the information that would be available after all individuals have revealedtheir types. Thus, we say that a mechanism is ex post pareto superior to another mechanism � if andonly if

Xy2Y

(y j� )ui (y; �) �Xy2Y

� (y j� )ui (y; �) ;8i 2 I;8� 2 �; (16.4)

and this inequality is strict for at least one player in N and at least one possible combination of types � in� such that p (�) > 0:Given any concept of feasibility, the three welfare criteria (ex ante, interim, ex post) give rise to three

di¤erent concepts of e¢ ciency For any set of mechanisms � (to be interpreted as the set of �feasible�mechanisms in some sense), we say that a mechanism � is ex ante e¢ cient in the set � if and only if �is in � and there exists no other mechanism v that is in � and is ex ante Pareto superior to �: Similarly� is interim e¢ cient in � if and only if � is in � and there exists no other mechanism v that is in � andis interim Pareto superior to �; and � is ex post e¢ cient in � if and only if � is in � and there exists noother mechanism v that is in � and is ex post Pareto superior to �.

16.2 Second Best

Similar e¢ ciency notions can be de�ned in a second best environment. The notions will be unaltered, thechange is in the set of mechanisms which we are considered to be feasible. The set may be restricted, e.g.,to the set of incentive feasible mechanism, that is all mechanisms which satisfy the incentive compatibilitycondition.

17. Social Choice

Can individual preferences be aggregated into social preferences, or more directly into social decisions.Denote by X the set of alternatives. Each individual i 2 I has a rational preference relation �i. Denotethe set of preferences by R and P.

17.1 Social Welfare Functional

De�nition 17.1.1. A social welfare functional is a rule

F : RI ! R:The strict preference relation derived from F is denoted by Fp.

De�nition 17.1.2. F is paretian if for any pair (x; y) 2 X and for all �I2 A,8i; x �i y ) xFpy:

Example 17.1.1. Borda count: x �i y �i z ) ci (x) = 1; ci (y) = 2; ci (z) = 3.

De�nition 17.1.3. F satis�es independence of irrelevant alternatives if for �I and �0Iwith the propertythat 8i; 8 (x; y) :

�i jfx; yg = �0i jfx; yg ) F jfx; yg = F 0 jfx; yg :Example 17.1.2. Borda count doesn�t satisfy IIA for as we change from x �i z �i y; and y �j x �j z, tox �i y �i z; and y �j z �j x, the relative position remains unchanged yet the Borda count changes.Example 17.1.3. Condorcet paradox with majority voting: x �1 y �1 z; z �2 x �2 y; y �3 z �3 x yieldscyclic and non-transitive preferences.

De�nition 17.1.4. F is dictatorial if 9i such that 8 (x; y) 2 X;8 �I2 A,x �i y ) xFpy:

Theorem 17.1.1. Suppose jXj � 3 and A = R. Then every F which is paretian and satis�es independenceof irrelevant alternatives is dictatorial.

Proof. See ?.

The result demonstrates the relevance of procedures and rules for social aggregation. To escape theimpossibility results, one may either (i) restrict the domain of the set of preferences or (ii) weaken therequirement of rational preferences.

F : A(i)! R

(ii)

17.2 Social Choice Function

De�nition 17.2.1. A social choice function f : A !X assigns f (�I) 2 X for every �I2 A.

88 17. Social Choice