EC2066 Vle
EC2066 Vle
EC2066 Vle
A. Daripa
EC2066
2016
Undergraduate study in
Economics, Management,
Finance and the Social Sciences
This subject guide is for a 200 course offered as part of the University of London
International Programmes in Economics, Management, Finance and the Social Sciences.
This is equivalent to Level 5 within the Framework for Higher Education Qualifications in
England, Wales and Northern Ireland (FHEQ).
For more information about the University of London International Programmes
undergraduate study in Economics, Management, Finance and the Social Sciences, see:
www.londoninternational.ac.uk
This guide was prepared for the University of London International Programmes by:
Dr Arup Daripa, Lecturer in Financial Economics, Department of Economics, Mathematics and
Statistics, Birkbeck, University of London.
This is one of a series of subject guides published by the University. We regret that due to
pressure of work the author is unable to enter into any correspondence relating to, or arising
from, the guide. If you have any comments on this subject guide, favourable or unfavourable,
please use the form at the back of this guide.
Contents
1 Introduction 1
1.1 Routemap to the subject guide . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Introduction to the subject area and prior knowledge . . . . . . . . . . . 2
1.3 Syllabus . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.4 Aims of the course . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.5 Learning outcomes for the course . . . . . . . . . . . . . . . . . . . . . . 3
1.6 Overview of learning resources . . . . . . . . . . . . . . . . . . . . . . . . 4
1.6.1 The subject guide . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.6.2 Essential reading . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.6.3 Further reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.6.4 Online study resources . . . . . . . . . . . . . . . . . . . . . . . . 6
1.6.5 The VLE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.6.6 Making use of the Online Library . . . . . . . . . . . . . . . . . . 7
1.7 Examination advice . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.7.1 Format of the examination . . . . . . . . . . . . . . . . . . . . . . 7
1.7.2 Types of questions . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.7.3 Specific advice on approaching the questions . . . . . . . . . . . . 8
2 Consumer theory 11
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.1.1 Aims of the chapter . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.1.2 Learning outcomes . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.1.3 Essential reading . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.2 Overview of the chapter . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.3 Preferences, utility and choice . . . . . . . . . . . . . . . . . . . . . . . . 12
2.3.1 Preferences and utility . . . . . . . . . . . . . . . . . . . . . . . . 12
2.3.2 Indifference curves . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.3.3 Budget constraint . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.3.4 Utility maximisation . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.4 Demand curves . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
i
Contents
4 Game theory 49
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
4.1.1 Aims of the chapter . . . . . . . . . . . . . . . . . . . . . . . . . . 49
4.1.2 Learning outcomes . . . . . . . . . . . . . . . . . . . . . . . . . . 49
4.1.3 Essential reading . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
ii
Contents
iii
Contents
iv
Contents
8 Monopoly 129
8.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
8.1.1 Aims of the chapter . . . . . . . . . . . . . . . . . . . . . . . . . . 129
8.1.2 Learning outcomes . . . . . . . . . . . . . . . . . . . . . . . . . . 129
8.1.3 Essential reading . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
8.2 Overview of the chapter . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
8.3 Properties of marginal revenue . . . . . . . . . . . . . . . . . . . . . . . . 130
8.4 Profit maximisation and deadweight loss . . . . . . . . . . . . . . . . . . 130
8.5 Price discrimination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
8.6 Natural monopoly . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
8.7 A reminder of your learning outcomes . . . . . . . . . . . . . . . . . . . . 135
8.8 Test your knowledge and understanding . . . . . . . . . . . . . . . . . . . 135
8.8.1 Sample examination questions . . . . . . . . . . . . . . . . . . . . 135
v
Contents
9 Oligopoly 137
9.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137
9.1.1 Aims of the chapter . . . . . . . . . . . . . . . . . . . . . . . . . . 137
9.1.2 Learning outcomes . . . . . . . . . . . . . . . . . . . . . . . . . . 137
9.1.3 Essential reading . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
9.2 Overview of the chapter . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
9.3 Cournot competition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
9.3.1 Collusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
9.3.2 Cournot with n > 2 firms . . . . . . . . . . . . . . . . . . . . . . 141
9.3.3 Stackelberg leadership . . . . . . . . . . . . . . . . . . . . . . . . 142
9.4 Bertrand competition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
9.4.1 Collusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
9.5 Bertrand competition with product differentiation . . . . . . . . . . . . . 143
9.5.1 Sequential pricing . . . . . . . . . . . . . . . . . . . . . . . . . . . 144
9.6 A reminder of your learning outcomes . . . . . . . . . . . . . . . . . . . . 146
9.7 Test your knowledge and understanding . . . . . . . . . . . . . . . . . . . 146
9.7.1 Sample examination questions . . . . . . . . . . . . . . . . . . . . 146
vi
Contents
vii
Contents
viii
Chapter 1
Introduction
1
1. Introduction
Nevertheless, a thorough reading of, and careful note-taking from, the recommended
textbook and the subject guide is a prerequisite for successful problem solving. The
subject guide aims to indicate as clearly as possible the key elements in microeconomic
analysis that you need to learn. The subject guide also presents detailed algebraic
derivations for a variety of topics. For each topic, you should consult both the subject
guide and the suggested parts of the textbook to understand fully the economic
principles involved.
The appropriate analysis will depend on the specific facts of a problem. However, you
are not expected to know the detailed facts about specific economic issues and policies
mentioned in textbook examples. Rather, you should use these examples (and the
end-of-chapter Sample examination questions) to aid your understanding of how
economic principles can be applied creatively to the analysis of economic problems.
If you are taking this course as part of a BSc degree you will also have passed ST104a
Statistics 1 and MT105a Mathematics 1 or MT1174 Calculus before beginning
this course. Every part of the syllabus can be mastered with the aid of diagrams and
relatively simple algebra. The subject guide indicates the minimum level of
mathematical knowledge that is required. Knowledge (and use in the examination) of
sophisticated mathematical techniques is not required. However, if you are
mathematically competent you are encouraged to use mathematical techniques when
these are appropriate, as long as you recognise that some verbal explanation is always
necessary.
1.3 Syllabus
The course examines how economic decisions are made by households and firms, and
how they interact to determine the quantities and prices of goods and factors of
production and the allocation of resources. Further, it examines the nature of strategic
2
1.4. Aims of the course
Producer theory: production and cost functions, firm and industry supply.
Game theory: normal-form and extensive form games, Nash equilibrium and
subgame perfect equilibrium, repeated games and cooperative equilibria.
Market failures arising from monopoly, externalities and public goods. The role of
policy.
3
1. Introduction
Each chapter includes a list of the learning outcomes that are specific to it. However,
you also need to go beyond the learning outcomes of each single chapter by developing
the ability of linking the concepts introduced in different chapters, in order to approach
the examination well.
The Essential reading lists the relevant textbook chapters and sections of chapters,
even though a more detailed indication of the required reading is listed throughout
the chapter.
The sections that follow specify in detail what you are expected to know about
each topic. The relevant sections of the recommended textbooks are referred to.
Wherever necessary, the sections integrate the textbook with additional material
and explanations. Finally, they draw attention to any problems that occur in
textbook expositions and explain how these can be overcome.
The boxes that appear in some of the sections give you exercises based on the
material discussed.
The learning outcomes show you what you should be able to do by the end of the
chapter.
A final section gives you questions to test your knowledge and understanding.
4
1.6. Overview of learning resources
This is available as an e-book at a discounted price via the VLE. Please visit
the course page for details.
Students may use the previous edition instead:
Note that the title of the twelfth edition differs slightly from that of the eleventh. If
using this edition, students should refer to the reading supplement on the VLE for
customised references.
N&S is more adequate for some parts of the syllabus while less so for others; as a
consequence we integrate some topics more with the extra material provided in the
subject guide. The textbook employs verbal reasoning as the main method of
presentation, supplemented by diagrammatic analyses. The textbooks use of algebra is
not uniformly satisfactory. The subject guide supplements the textbook in many cases
in this regard. There are also some references to the following textbook:
Perloff, J.M. Microeconomics with Calculus. (Pearson Education, 2014) 3rd edition
[ISBN 9780273789987].
Detailed reading references in this subject guide refer to the editions of the textbooks
listed above. New editions of these textbooks may have been published by the time you
study this course. You can use a more recent edition of any of the textbooks; use the
detailed chapter and section headings and the index to identify relevant readings. Also,
check the virtual learning environment (VLE) regularly for updated guidance on
readings.
Unless otherwise stated, all websites in this subject guide were accessed in February
2016. We cannot guarantee, however, that they will stay current and you may need to
perform an internet search to find the relevant pages.
Besanko, D. and R. Braeutigam Microeconomics. (John Wiley & Sons, 2014) 5th
edition, international student version [ISBN 9781118716380].
Varian, H.R. Intermediate Microeconomics, a Modern Approach. (W.W. Norton,
2014) 9th edition [ISBN 9780393920772].
Pindyck, R.S. and D.L. Rubinfeld Microeconomics. (Pearson, 2014) 8th edition
[ISBN 9781292081977].
5
1. Introduction
Electronic study materials: All of the printed materials which you receive from
the University of London are available to download, to give you flexibility in how
and where you study.
Discussion forums: An open space for you to discuss interests and seek support
from your peers, working collaboratively to solve problems and discuss subject
material. Some forums are moderated by an LSE academic.
Videos: Recorded academic introductions to many subjects; interviews and
debates with academics who have designed the courses and teach similar ones at
LSE.
Recorded lectures: For a few subjects, where appropriate, various teaching
sessions of the course have been recorded and made available online via the VLE.
Audio-visual tutorials and solutions: For some of the first year and larger later
courses such as Introduction to Economics, Statistics, Mathematics and Principles
of Banking and Accounting, audio-visual tutorials are available to help you work
through key concepts and to show the standard expected in examinations.
Self-testing activities: Allowing you to test your own understanding of subject
material.
Study skills: Expert advice on getting started with your studies, preparing for
examinations and developing your digital literacy skills.
6
1.7. Examination advice
Note: Students registered for Laws courses also receive access to the dedicated Laws
VLE.
Some of these resources are available for certain courses only, but we are expanding our
provision all the time and you should check the VLE regularly for updates.
This subject guide is reproduced in colour on the VLE and you may find it easier to
understand if you access the online PDF.
7
1. Introduction
Numerical questions will sometimes require the use of a calculator. A calculator may be
used when answering questions on the examination paper for this course and it must
comply in all respects with the specification given in the Regulations.
Questions will not require knowledge of empirical studies or institutional material.
However, you will be awarded some marks for supplementary empirical or institutional
material which is directly relevant to the question.
Occasionally, you may be unsure exactly what a question is asking. If there is some
element of doubt about the interpretation of a question, state at the beginning of
your answer how you interpret the question. If you are very uncertain of what is
required, and the question is in Section B, do another question.
Explain briefly what you are doing: an answer that is simply a list of equations or
numbers will not be credited with full marks even if it gets to the correct solution.
Moreover, by explaining what you are doing, you will be awarded some marks for
correct reasoning even if there are mistakes in some part of the procedure.
It is essential to attempt eight questions in Section A. Even if you think you do not
know the answer, at least define any terms or concepts which you think may be
relevant (including those in the question!) and, if possible, present the question in
diagrammatic or algebraic form. The same applies to a specific part of a multi-part
question in Section B. The examiners can give no marks for an unattempted
question, but they can award marks for relevant points. A single mark may make
the difference between passing and failing the examination.
Although you should attempt all the questions and parts of questions that are
required, to avoid wasting time you should make sure that you do no more than is
required. For example, if only two parts of a three-part question need to be
answered, only answer two parts.
Note the importance of key words. In some of the True or false? type questions,
the words must, always, never or necessarily usually invite you to explain why
the statement is false. Notice that this is simply a way in which you can start
approaching the problem, but there is no way to know in advance the correct
answer without analysing every specific question.
It is worth noting that in this type of question, simply writing true or false will
not earn you any marks, even if it happens to be the right answer. The examiners
are looking for reasoning, not blind guesses.
8
1.7. Examination advice
For many questions, good answers will require diagrammatic and/or algebraic
analysis to complement verbal reasoning. Good diagrams can often save much
explanation but free-standing diagrams, however well-drawn and labelled, do not
portray sufficient information to the examiners. Diagrams need to be explained in
the text of the answer. Similarly, symbols in algebraic expressions should be defined
and the final line of an algebraic presentation should be explained in words.
The examiners are primarily looking for analytical explanations, not descriptions.
On reading a question, your first thought should be: what is the appropriate
hypothesis, theory, concept or model to use?
where available, past examination papers and Examiners commentaries for the
course which give advice on how each question might best be answered.
9
1. Introduction
10
Chapter 2
Consumer theory
2.1 Introduction
How do people choose what bundle of goods to consume? We cannot observe this
process directly, but can we come up with a model to capture the decision-making
process so that the predictions from the model match the way people behave? If we can
build a sensible model, we should be able to use the model to understand how choice
varies when the economic environment changes. This should also help us design
appropriate policy. This is the task in this chapter and as we go through the various
steps, you should keep this overarching goal in your mind and try to see how each piece
of analysis fits in the overall scheme.
Once we build such a model, we use it to analyse how optimal consumption choice
responds to price and income variations. We also extend the analysis to cover
labour-leisure choice as well as intertemporal consumption choice.
draw indifference curve diagrams starting from the utility function of a consumer
solve the consumers utility maximisation problem and derive the demand for a
consumer with a given utility function and budget constraint
11
2. Consumer theory
derive the present discounted value of payment streams and explain bond pricing.
12
2.3. Preferences, utility and choice
model of choice that has some predictive power, we do need to put some restrictions on
preferences to rule out irrational behaviour. Just a few relatively sensible restrictions
allow us to build a model of choice that has great analytical power. Indeed, this idea of
creating an analytical structure that can be manipulated to understand how changes in
the economic environment affect economic outcomes underlies the success of economics
as a tool for investigating the world around us.
Section 2.2 of N&S sets out three restrictions on preferences: completeness, transitivity
and non-satiation (more is better). These restrictions allow us to do something very
useful. Once we add some further technical requirements (a full specification must await
Masters level courses but the main extra condition we need is that preferences have
certain continuity properties), these restrictions allow us to represent preferences by a
continuous function. This is known as a utility function.
Note that a utility function is an artificial concept no-one actually has a utility
function (you knew that of course, since you surely do not have one). But because we
can represent preferences using such a function, it is as if agents have a utility function.
All subsequent analysis using utility functions and indifference curves has this as if
property.
Preferences generally give us rankings among bundles rather than some absolute
measure of satisfaction derived from bundles. You might prefer apple juice to orange
juice but would have difficulty saying exactly how much more satisfaction you derive
from the former compared to the latter. Preferences, therefore, typically give us an
ordinal ranking among bundles of goods. Since utility is simply a representation of
preferences, it is also an ordinal measure. This means that if your preferences can be
represented by a utility function, then a positive transformation of this function which
preserves the ordering among bundles is another function that is also a valid utility
function. In other words, there are many possible utility functions that can represent a
given set of preferences equally well.
However, there are some instances where we use cardinal utility and make absolute
comparisons among bundles. Money, for example, is a cardinal measure you know that
20 pounds is twice as good as 10 pounds. In general, though, you should understand
utility as an ordinal concept.
Once we can represent preferences using a continuous utility function, we can draw
indifference curves. An indifference curve is the locus of different bundles of goods that
yield the same level of utility. In other words, an indifference curve for utility function
u(x, y) is given by u(x, y) = k, where k is some constant. As we vary k, we can trace out
the indifference map. Note that an indifference curve is simply a level curve of a utility
function. Just as you draw contours on a map to represent, say, a mountain, so
indifference curves drawn for two goods are contour plots of a utility function over these
two goods. You can see Figure 3.3 in Perloff for a pictorial representation. You should
read carefully the discussion in N&S (Sections 2.3 to 2.5) on indifference curves. You
13
2. Consumer theory
should know how different types of preferences generate different types of indifference
curves.
Activity 2.1 For each of the following utility functions, write down the equation for
an indifference curve and then draw some indifference curves.
(b) u(x, y) = x + y.
5. Every bundle of goods lies on some indifference curve (follows from completeness).
Typically, preferences have the following property. Consider a point where a lot of y and
very little x is being consumed. Starting from any such point, a consumer is willing to
give up a lot of y in exchange for another unit of x while retaining the same level of
utility as before. As we keep adding units of x and reducing y while keeping utility
constant (i.e. we are moving down an indifference curve), consumers are willing to give
up less and less of y in return for a further unit of x. One way to interpret this is that
14
2.3. Preferences, utility and choice
people typically have a taste for variety and want to avoid extremes (i.e. avoid
situations where a lot of one good and very little of the other good is being consumed).
This implies that MRS falls along an indifference curve. This property is referred to as
indifference curves being convex to the origin in some textbooks. This works as a
visual description, but you should be aware that in terms of mathematics, this is not a
meaningful description there is no mathematical concept where something is convex
relative to something else. The correct idea of convex indifference curves is as follows.
Consider a subset S of Rn . S is a convex set if the following property holds: if points s1
and s2 are in S, then a convex combination s1 + (1 )s2 is also in the set S for any
0 < < 1.
Now consider any indifference curve yielding utility level u. Consider the set of all
points that yield utility u or more. This is the set of all points on an indifference curve
plus all points above. Call this set B. Diminishing MRS implies that B is a convex set.
Figure 2.1 below shows a convex indifference curve. Note that the set B (part of which
is shaded) is a convex set. Try taking any two points in B and then making a convex
combination. You will find that the combinations are always inside B.
Next, Figure 2.2 shows an example of non-convex indifference curves. Note that the set
B of points on or above the indifference curve is not convex. If you combine points such
as a1 and a2 in the diagram, for some values of , the convex combinations fall outside
the set B.
Note that when two goods are perfect substitutes, you get a straight line indifference
curve. At the other extreme, the two goods are perfect complements (no
substitutability) and the indifference curve is L-shaped. Indifference curves with
diminishing MRS lie in between these two extremes.
15
2. Consumer theory
Here is an activity to get you computing the MRS for different utility functions.
Activity 2.2 Compute the MRS for the following utility functions.
(a) u(x, y) = xy.
(b) u(x, y) = ln x + ln y.
Once we have specified our model of preferences, we need to know the set of goods that
a consumer can afford to buy. This is captured by the budget constraint. Since
consumers are generally taken to be price-takers (i.e. what an individual consumer
purchases does not affect the market price for any good), the budget line is a straight
line. See Section 2.7 of N&S for the construction of budget sets. You should be aware
that budget lines would no longer be a straight line if a consumer buys different units at
different prices. This could happen if a consumer is a large buyer in a market or if the
consumer gets quantity discounts. See Application 2.6 in N&S for an example.
The consumer chooses the most preferred point in the budget set. If preferences are
such that indifference curves have the usual convex shape, the best point is where an
indifference curve is tangent to the budget line. This is shown as point A in Figure 2.3.
At A the slope of the indifference curve coincides with the slope of the budget
16
2.3. Preferences, utility and choice
constraint. So we have:
MUx Px
= .
MUy Py
Multiplying both sides by 1 we can write this as the familiar condition:
Px
MRSxy = .
Py
Let us derive this condition formally using a Lagrange multiplier approach. This is the
approach you are expected to use when faced with optimisation problems of this sort.
Note that the more is better assumption ensures that a consumer spends all income (if
not, then the consumer could increase utility by buying more of either good). Therefore,
the budget constraint is satisfied with equality. It follows that the consumer maximises
u(x, y) subject to the budget constraint Px x + Py y = M . Set up the Lagrangian:
L = u(x, y) + (M Px x + Py y).
Px u/x
= = MRSxy .
Py u/y
17
2. Consumer theory
Second-order condition
Figure 2.4: Violation of the second-order condition under non-convex preferences. Note
that MRS is not always diminishing. Point A satisfies the first-order condition MRS equal
to price ratio, but is not optimal. (Source: Schmalensee, R. and R.N. Stavins (2013) The
SO2 allowance trading system: the ironic history of a grand policy experiment, Journal
of Economic Perspectives, Vol. 27, pp.10321. Reproduced by kind permission of the
American Economic Association.)
Read Sections 2.7 to 2.9 of N&S carefully and work through all the examples therein.
Note that if two goods are perfect substitutes or complements, the tangency condition
does not apply. For perfect substitutes, there is either a corner solution or the budget
line coincides with the indifference curve. In the latter case, any point on the budget
line is optimal. For the case of perfect complements, the optimum occurs where the kink
in the indifference curve just touches the budget line. Note that this is not a tangency
point the slope of the indifference curve is undefined at the kink. N&S clarifies these
cases with appropriate diagrams.
18
2.3. Preferences, utility and choice
Demand functions
The maximisation exercise above gives us the demand for goods x and y at given prices
and income. As we vary the price of good x, we can trace out the demand curve for
good x. See Section 3.6 of N&S for a discussion. The activities below compute demand
functions in specific examples.
u(x, y) = x y
max x y subject to x + py = M.
x, y
This implies:
x1 y 1
= .
x y 1 p
Simplifying:
y 1
= .
x p
Using this in the budget constraint and solving, we get the demand functions:
M
x(p, M ) =
+
M
y(p, M ) = .
p( + )
u(x, y) = min{x, y}
19
2. Consumer theory
Since we are at the kink, it must be that x = y. Using this in the budget
constraint, we get the demand functions:
M
x(p, M ) =
+ p
M
y(p, M ) = .
+ p
Figure 2.5: The optimum occurs at point E. Note that this is not a tangency point. The
slope of the indifference curve at the kink is undefined.
Income changes
Section 3.2 of N&S explains the classification of goods according to the response of
demand to income changes.
20
2.4. Demand curves
Note that it is not possible for all goods to be inferior. This would violate the more is
better assumption. The full argument is left as an exercise.
Activity 2.3 It is not possible for all goods to be inferior. Provide a careful
explanation of this statement.
The income-consumption curve of a consumer traces out the path of optimal bundles as
income varies (keeping all prices constant). Using this exercise, we also plot the
relationship between quantity demanded and income directly. The curve that shows this
relationship is called the Engel curve. See Perloff Section 4.2 for an exposition of the
income-consumption curve and the Engel curve.
The slope of the income-consumption curve indicates the sign of the income elasticity of
demand (explained below).
Price changes
See Sections 3.3 to 3.8 of N&S. It is very important to understand fully the
decomposition of the total price effect into income and substitution effects. This
decomposition is, of course, a purely artificial thought experiment. But this thought
experiment is extremely useful in understanding how the demand for different goods
responds to a change in price at different levels of income and given different
opportunities to substitute out of a good. You should understand how these effects
(and, therefore, the total price effect) differ across normal and inferior goods, and
understand how the effect known as Giffens paradox can arise.
The idea of income and substitution effects can help us understand the design of an
optimal tax scheme. See Section 3.4 of N&S for a discussion of this issue.
Finally, you should also study the impact on the demand for a good by changes in the
price of some other good, and how this effect differs depending on whether the other
good is a substitute or a complement.
Example 2.3 Suppose u(x, y) = x1/2 y 1/2 . Income is M = 72. The price of y is 1
and the price of x changes from 9 to 4. Calculate the income effect (IE) and the
substitution effect (SE).
Let (px , py ) denote the original prices and let p0x denote the lower price of x.
Under the original prices, the Marshallian demand functions (you should be able to
calculate these) are:
M
x =
2px
and:
M
y = .
2py
21
2. Consumer theory
where the subscript of u is a reminder that this is the original utility level (before
the price change).
The total price effect (PE) from a price fall is:
M M
PE = 0
.
2px 2px
which implies:
M M M
p = .
2 p0x py 2 px py
Using the values supplied, M = 24 so that M M = 48.
Under a reduced income of 48, and given the new price p0x = 4, the demand for x is
6. The original demand for x was 4. Under the compensated price change, the
demand is 6. Therefore, the SE is 2. It follows that the rest of the change must be
the IE. Since the total price effect is 5, the IE is 3.
In terms of algebra:
M M M
SE = 0
.
2px 2px
The IE is the remainder of the price effect, so that:
M M M M M
IE = .
2p0x 2px 2p0x 2px
Simplifying:
M
IE = .
2p0x
Looking ahead, the M we calculated here is known as the compensating
variation. We will study this concept later in this chapter.
22
2.4. Demand curves
Figure 2.6: As the price of x falls, the change from A to C shows the total price effect.
The movement from A to B (under a compensated price change) shows the substitution
effect, while the movement from B to C shows the income effect.
From individual demand curves, we can construct the market demand curve by
aggregating across individuals. See N&S Section 3.10 for a discussion.
dQ/Q P dQ
= = .
dP/P Q dP
23
2. Consumer theory
Price elasticity of demand is the most common measure of elasticity and often referred
to as just elasticity of demand.
Other than price elasticity, we can define income elasticity and cross-price elasticity.
Income elasticity
Cross-price elasticity
Let us consider the elasticity of demand for good i with respect to the price of good j.
The cross-price elasticity of demand for good i is given by:
Pj dQi
ij = .
Qi dPj
This is negative for complements and positive for substitutes.
You should take a long look at the elasticity estimates presented in Section 3.16 of
N&S. Practical knowledge of elasticities forms an important part of designing and
understanding a variety of tax and subsidy policies in different markets.
24
2.4. Demand curves
For a normal good, the compensated demand curve is less elastic compared to the
uncompensated demand curve.
For an inferior good, the compensated demand curve is more elastic compared to
the uncompensated demand curve.
You should understand that all three properties result from the fact that only the
substitution effect matters for the change in compensated demand when the price
changes.
The next example asks you to calculate the compensated demand curve for a
CobbDouglas utility function.
Example 2.4 Suppose u(x, y) = x1/2 y 1/2 . Income is M . Calculate the compensated
demand curves for x and y.
To do this, we must first calculate the Marshallian demand curves. These are given
by (you should do the detailed calculations to show this):
M M
x= and y= .
2px 2py
This is the value of income which is compensated to keep utility constant at the level
given by the original choice of x and y. It follows that the compensated demand
functions are:
M
r
py
xc = =V
2px px
and:
M px
r
yc = =V .
2py py
Note that the Marshallian demand for x does not depend on py , but the Hicksian, or
compensated, demand does. This is because changes in py require income
adjustments, which generate an income effect on the demand for x.
25
2. Consumer theory
When drawing demand curves, we typically draw the inverse demand curve (price on
the vertical axis, quantity on the horizontal axis). In such a diagram, the consumer
surplus (CS) is the area under the (inverse) demand curve and above the market price
up to the quantity purchased at the market price. This is the most widely-used measure
of welfare. We can measure the welfare effect of a price rise by calculating the change in
CS (denoted by CS).
Much of our discussion of policy will be based on this measure. Any part of CS that
does not get translated into revenue or profits is a deadweight loss. The extent of
deadweight loss generated by any policy is a measure of inefficiency associated with that
policy.
However, CS is not an exact measure because of the presence of an income effect.
Ideally, we would use the compensated demand curve to calculate the welfare change.
CV and EV give us two such measures. You should use these measures to understand
the design of ideal policies, but when measuring welfare change in practice, use CS.
CV is the amount of money that must be given to a consumer to offset the harm
from a price increase, i.e. to keep the consumer on the original indifference curve
before the price increase.
EV is the amount of money that must be taken away from a consumer to cause as
much harm as the price increase. In this case, we keep the price at its original level
(before the rise) but take away income to keep the consumer on the indifference curve
reached after the price rise.
Consider welfare changes from a price rise. For a normal good, we have
CV > CS > EV, and for an inferior good we have CV < CS < EV. The measures
would coincide for preferences that exhibit no income effect.
The example that follows shows an application of these concepts.
Example 2.5 The government decides to give a pensioner a heating fuel subsidy of
s per unit. This results in an increase in utility from u0 before the subsidy to u1
after the subsidy. Could the government follow an alternative policy that would
result in the same increase in utility for the pensioner, but cost the government less?
Let us show that an equivalent income boost would be less costly. Essentially, the
EV of a price fall is lower than the expenditure on heating after the price fall. The
intuition is that a per-unit subsidy distorts choice in favour of consuming more
heating, raising the total cost of the subsidy. To put the same idea differently, an
equivalent income boost would raise the demand for fuel through the income effect.
26
2.4. Demand curves
But a price fall (the fuel subsidy results in a lower effective price) causes an
additional substitution effect boosting the demand for heating.
To see this, consider Figure 2.7. The initial choice is point A and after the subsidy
the pensioner moves to point B. How much money is the government spending on
the subsidy? Note that after the subsidy, H1 units of heating fuel are being
consumed. At pre-subsidy prices, buying H1 would mean the pensioner would have
E 0 of other goods. Since the price of the composite good is 1, M is the same as total
income. It follows that the amount of income that would be spent on heating to buy
H1 units of heating at pre-subsidy prices is given by M E 0 . Similarly, at the
subsidised price, the amount of income being spent on heating fuel is M B 0 . The
difference B 0 E 0 is then the amount of the subsidy. This is the same length as
segment BE.
Once we understand how to show the amount of spending on the subsidy in the
diagram, we are ready to compare this spending with an equivalent variation of
income. This is added in Figure 2.8 below.
The pensioners consumption is initially at A, and moves to B after the subsidy.
Since the composite good has a price of 1, the vertical distance between the budget
lines (segment BE) shows the extent of the expenditure on the subsidy (as explained
above). An equivalent variation in income, on the other hand, would move
consumption to C. It follows that DE is the equivalent variation in income, which is
smaller than the expenditure on the subsidy. Therefore, a direct income transfer
policy would be less costly for the government.
27
2. Consumer theory
28
2.5. Labour supply
max u(N, M )
N, M
subject to:
Z = 24 N
M = wZ + M .
max u(N, M )
N, M
M + wN = 24w + M .
At the optimum we have the slope of the indifference curve (MRS) equal to the slope
of the budget line (w). Therefore:
MUN
= w.
MUM
How does the optimal choice of labour respond to a change in w? We can analyse this
using income and substitution effects. In this case, a change in w also changes income
directly (as you can see from the budget constraint), so the exercise is a little different
compared to that under standard goods.
Suppose w rises.
Income effect
The rise in w raises income at current levels of labour and leisure. Assuming leisure
is a normal good (this should be your default assumption), this raises the demand
for leisure.
29
2. Consumer theory
Substitution effect
The rise in w makes leisure relatively more expensive, causing the agent to substitute
away from leisure. This reduces demand for leisure.
The two effects go in opposite directions, therefore the direction of the total effect is
unclear. In most cases, the substitution effect dominates, giving us an upward-sloping
labour supply function. However, it is possible, especially at high levels of income (i.e.
when wage levels are high), that the income effect might dominate. In that case we
would get a backward-bending labour supply curve which initially slopes upward but
then turns back and has a negative slope.
If leisure is, on the other hand, an inferior good, the two effects would go in the same
direction and labour supply would necessarily slope upwards.
Figure 2.9 (a) below shows a backward-bending labour supply curve while Figure 2.9
(b) shows an increasing labour supply curve.
30
2.6. Saving and borrowing: intertemporal choice
max u(C0 , C1 )
C0 , C 1
C1 Y1
C0 + = Y0 + .
1+r 1+r
This is similar to a standard optimisation problem with two goods, C0 and C1 , where
the price of the former is 1 and the price of the latter is 1/(1 + r). Unsurprisingly, the
optimum satisfies the property that:
MUC0
= 1 + r.
MUC1
Income effect
For a borrower, Y0 < C0 , so that a rise in the interest rate lowers income tomorrow.
Given consumption is normal in both periods, the agent should consume less in
period 0 (borrow less).
Substitution effect
A rise in the interest rate makes immediate consumption more costly. Therefore,
the substitution effect suggests that the individual should choose to lower C0 and,
therefore, borrow less.
Since the two effects go in the same direction, the direction of change is unambiguous: a
rise in the rate of interest lowers borrowing.
31
2. Consumer theory
Income effect
For a saver, Y0 > C0 , so that a rise in the interest rate raises income tomorrow. Given
consumption is normal in both periods, the agent should consume more in period 0
(save less).
Substitution effect
A rise in the interest rate makes immediate consumption more costly. Therefore,
the substitution effect suggests that the individual should choose to lower C0 (save
more).
Since the two effects go in opposite directions, the total effect on saving is uncertain.
Usually, the substitution effect dominates so that agents save less when the interest rate
rises, but it could go the other way.
Figure 2.10 shows the intertemporal budget constraint. The endowment point is
(Y0 , Y1 ). As the rate of interest increases, the budget constraint pivots around the
endowment point as shown.
Note that consumers reaching an optimum in the part of the budget constraint above
the endowment point are savers, and those reaching an optimum somewhere in the
lower part are borrowers.
32
2.7. Present value calculation with many periods
Activity 2.5 Using Figure 2.10 above, explain that a saver cannot become a
borrower if the rate of interest rises.
PV = y0 + y1 + 2 y2 + + n yn .
PV = y(1 + + 2 + + n ).
S = 1 + + 2 + + n.
Then:
S = + 2 + + n+1 .
We have:
S S = 1 n+1 .
Therefore:
1 n+1
S= .
1
So:
1 n+1
PV = y .
1
If the payoff stream is infinite, the present value is very simple:
2 1
PV = y(1 + + + ) = y .
1
33
2. Consumer theory
2.7.1 Bonds
A bond typically pays a fixed coupon amount x each period (next period onwards) until
a maturity date T , at which point the face value F is paid. The price of the bond, P , is
simply the present value given by:
P = x + 2 x + + T F.
Note that the price declines if falls, which happens if r rises. Therefore, the price of a
bond has an inverse relationship with the rate of interest.
A special type of bond is a consol or a perpetuity that never matures. The price of a
consol has a particularly simple expression:
P = x + 2 x + = x .
1
Now = 1/(1 + r). Therefore:
1/(1 + r) 1
= = .
1 r/(1 + r) r
It follows that:
x
P = .
r
This makes the inverse relationship between P and r clear.
draw indifference curve diagrams starting from the utility function of a consumer
solve the consumers utility maximisation problem and derive the demand for a
consumer with a given utility function and budget constraint
34
2.9. Test your knowledge and understanding
derive the present discounted value of payment streams and explain bond pricing.
2. The Hicksian demand curve for a good must be more elastic than the Marshallian
demand curve for a good. Is this true or false? Explain.
3. Savers gain more when the rate of interest rises. Is this true or false? Explain.
35
2. Consumer theory
36
Chapter 3
Choice under uncertainty
3.1 Introduction
In the previous chapter, we studied consumer choice in environments that had no
element of uncertainty. However, many important economic decisions are made in
situations involving some degree of risk. In this chapter, we cover a model of
decision-making under uncertainty called the expected utility model. The model
introduces the von NeumannMorgenstern (vNM) utility function. This is unlike the
ordinal utility functions we saw in the previous chapter and has special properties. In
particular, the curvature of the vNM utility function can indicate a consumers
attitude towards risk. Once we introduce the model, we use it to derive the demand for
insurance and also introduce a measure of the degree of risk aversion.
explain the nature of the vNM utility function and calculate the expected utility
from a gamble
explain the different risk attitudes and what they imply for the vNM utility
function
analyse the demand for insurance and show the relationship between insurance and
premium
calculate the ArrowPratt measure of risk aversion for different specifications of the
vNM utility function.
37
3. Choice under uncertainty
3.3 Preliminaries
You should already be familiar with concepts such as probability and expected value.
Do familiarise yourself with these concepts if this is not the case. Section 4.1 of N&S
discusses these.
Random variable
A variable that represents the outcomes from a random event. A random variable
has many possible values, and each value occurs with a specified probability.
38
3.5. Risk aversion
choice under uncertainty that might be deemed reasonable. They showed that under
their axioms, there exists a function u such that the gamble can be evaluated using the
following expected utility formulation:
E(U (G)) = p1 u(x1 ) + p2 u(x2 ) + + pn u(xn ).
The function u is known as the von NeumannMorgenstern (vNM) utility function.
The vNM utility function is somewhat special. It is not entirely an ordinal function
like the utility functions you saw in the last chapter. Starting from a vNM u function,
we can make transformations of the kind a + bu, with b > 0 (these are called positive
affine transformations), without changing the expected utility property but not any
other kinds of transformations (for example, u2 is not allowed). The reason is that, as
we discuss below, the curvature of the vNM utility function captures attitude towards
risk. Transformations other than positive affine ones change the curvature of this
function, and therefore the transformed u function would not represent the same
risk-preferences as the original. Thus vNM utility functions are partly cardinal.
Note that the expected utility representation is very convenient. Once we know the
vNM utility function u, we can evaluate any gamble easily by simply taking the
expectation over the vNM utility values.
39
3. Choice under uncertainty
Figure 3.1: The vNM utility function for a risk-averse individual. Note that the function
is concave and u(E(G)) > E(U (G)) so that the agent does not accept a fair gamble.
Kims utility depends on wealth W . Kims vNM utility function is given by:
u(W ) = W.
Kims wealth is uncertain. With probability 0.5 wealth is 100, and with probability 0.5
a loss occurs so that wealth becomes 64. In what follows, we will assume that Kim can
only buy full insurance. In other words, the insurance company offers to pay Kim 36
whenever the loss occurs and in exchange Kim pays them a premium of R in every state
(i.e. whether the loss occurs or not). In what follows, we will calculate the maximum
and minimum value of R.
40
3.6. Risk aversion and demand for insurance
How do we know Kim would be prepared to pay to buy full insurance? You can draw a
diagram as above to show that Kim would be prepared to pay a positive premium if
wealth is fully insured. Alternatively, you could point out that the expected utility of
the uncertain wealth (which is 9) is lower than the utility of expected wealth since:
u(E(W )) = 0.5 100 + 0.5 64 = 82 = 9.055.
This implies that the premium that Kim is willing to pay is positive.
You could also point out that W is a concave function (check that the second
derivative is negative), implying that Kim is risk-averse. Then draw the CE point as
above and point out that since expected wealth exceeds the certainty equivalent, the
premium is positive.
Let us now calculate the maximum premium that Kim would be willing to pay to
buy full insurance.
First, calculate the certainty equivalent of the gamble Kim is facing. This is given by:
u(CE) = E(U ).
Therefore:
CE = 9
implying that CE = 81. Therefore, the maximum premium Kim is willing to pay is
100 81 = 19.
There is another way of finding this, which considers the maximum premium in terms of
expected wealth (i.e. how much expected wealth would Kim give up in order to fully
insure?). You need to understand this, since some textbooks use this way of identifying
the premium. For example, this is the approach adopted by Perloff. Unfortunately,
textbooks (including Perloff) never make clear exactly what they are doing, which can
be very confusing for students. Reading the exposition here should clarify the matter
once and for all.
The maximum premium in terms of expected wealth is calculated as follows. Note that
under full insurance the gross expected wealth Kim would receive is E(W ) = 82. We
also know that CE = 81. Therefore, the maximum amount of expected wealth Kim
would give up is 82 81 = 1. (Note that this should explain why in the diagram on risk
aversion in Perloff Section 16.2, and subsequent solved problems, the risk premium is
identified as the difference between expected wealth and the CE.)
To connect this approach to the one above, consider the actual premium and coverage.
Kim loses 36 with probability 0.5. So full insurance means a coverage of 36, which is
paid when the loss occurs. In return, Kim pays an actual premium of 19 in each state.
Therefore, the change in expected wealth for Kim is:
41
3. Choice under uncertainty
Once again, the purpose of writing this out in detail is to make you aware that
textbooks vary in their treatment of this. Some talk about premium in terms of
expected wealth, while others calculate the actual premium, but they do not make it
clear what it is that they are doing. In answering questions of this sort in an
examination, it is easiest (and clearest) to calculate the actual premium. You can follow
the other route and define the premium in terms of expected wealth, but in that case
you should make that clear in your answer.
Next, we calculate the minimum premium.
Assuming the insurance company is risk-neutral, it must break even. So the minimum
premium (or fair premium) is Rmin such that it equals the expected payout, which is
0.5 (100 64) = 18. (Note that this is simply 100 E(W ), where E(W ) is expected
wealth, which is 82 in this case.)
As above, the other way of answering the question is to say that in terms of the expected
wealth that Kim needs to give up, the minimum is zero. Think of this as follows. Kim
simply hands over her actual wealth to the insurance company, and in return receives
the expected wealth in all states. The insurance company is risk-neutral, and in
expected wealth terms it is giving and receiving the same amount, and breaks even.
42
3.7. Risk-neutral and risk-loving preferences
Note that the second-order condition for a maximum is satisfied since the agent is
risk-averse implying that u00 < 0 (u is concave). Therefore, we have:
u0 (WL ) (1 p)r
0
= .
u (WN ) (1 r)p
If insurance is fair, that implies the insurance company breaks even, i.e. the expected
payout pX equals the expected receipt rX. Since pX = rX, we have:
p = r.
It follows that:
u0 (WL ) = u0 (WN ).
Since u0 is a decreasing function (because u00 < 0), it is not possible to have this equality
if WL 6= WN . (Note that if u0 was a non-monotonic function and was going up and
down, it would be possible to have WL different from WN but have the same value of u0
at these two different points.)
It follows that WL = WN , i.e. we have:
W L + X rX = W rX
implying that X = L. Therefore, the agent would optimally fully insure (cover the
entire loss) at the fair premium.
Note also what would happen if r > p. Then (1 p)r > (1 r)p. Therefore:
u0 (WL ) (1 p)r
= > 1.
u0 (WN ) (1 r)p
This implies that:
u0 (WL ) > u0 (WN ).
Again, because u0 is a decreasing function (u00 < 0), this implies that WL < WN , which
in turn implies that X < L. Thus if the premium exceeds the fair premium, less than
full insurance would be purchased.
43
3. Choice under uncertainty
Figure 3.2: The vNM utility function for a risk-neutral individual. Note that the function
is linear and u(E(G)) = E(U (G)) so that the agent is indifferent between a fair gamble
and the safe alternative of 15.
Figure 3.2 shows the vNM utility function for a risk-neutral agent. The figure refers to
the gamble introduced in the section on risk aversion (either keep 15 or invest in a fair
gamble yielding 20 with probability 0.5, and 10 with probability 0.5).
A risk-loving agent, on the other hand, prefers a risky bet to a safe alternative when
they have the same expected outcome. In other words, a risk-loving agent would prefer
to accept a fair gamble. As Figure 3.3 below shows, for a risk-loving agent, the CE of a
gamble is higher than the expected value of the gamble (you would have to pay a
risk-loving agent to give up a risky gamble in favour of the safe alternative of getting
the expected value of the gamble).
Figure 3.3 shows the vNM utility function for a risk-loving agent.
44
3.8. The ArrowPratt measure of risk aversion
Figure 3.3: The vNM utility function for a risk-loving individual. Note that the function
is convex and u(E(G)) < E(U (G)) so that the agent prefers a fair gamble to the safe
wealth of 15.
positive risk measure under risk aversion. These help to interpret the ArrowPratt
measure of risk aversion, which is given by:
u00
= 0.
u
That is, the ArrowPratt measure of risk aversion is 1 times the ratio of the second
derivative and the first derivative of the vNM utility function. This is the most
common measure of risk aversion. There are other measures, which you will encounter
in Masters level courses.
It can be shown that the larger the ArrowPratt measure of risk aversion, the more
small gambles an individual will take. A derivation of this result must await a Masters
level course as well.
As noted above, for a risk-averse individual, u00 < 0, so the minus sign in front makes
the measure a positive number. For a risk-neutral agent, u00 = 0 so that = 0, and for a
risk-loving agent u00 > 0 so that the measure is negative.
Example 3.1 Let us calculate the ArrowPratt measure for different specifications
of the vNM utility function.
45
3. Choice under uncertainty
( 1)W 2 1
= 1
= .
W W
Note that as increases, the degree of risk aversion decreases. As goes to 1,
the risk aversion measure goes to 0, which is right since at = 1 the agent is
risk-neutral.
46
3.11. Test your knowledge and understanding
2. Suppose u(W ) = 1/W . What is the risk attitude of this person? Calculate the
ArrowPratt measure of risk aversion for this preference.
3. Suppose an agent has vNM utility function u(W ). Under what conditionp would
a + bu(W ) also be a valid vNM utility function for this agent? Would u(W ) be
a valid vNM utility function for this agent?
4. Suppose u(W ) = ln W for an agent. The agent faces the following gamble: with
probability 0.5 wealth is 100, and with probability 0.5 a loss occurs so that wealth
becomes 64. The agent can buy any amount of insurance: a coverage of X can be
purchased by paying premium rX.
(a) Work out the insurance coverage X that the agent would optimally purchase
as a function of r.
(b) Plot the optimal X as a function of r.
(c) Calculate the value of r for which full insurance is purchased.
(d) Calculate the value of r for which no insurance is purchased.
47
3. Choice under uncertainty
48
Chapter 4
Game theory
4.1 Introduction
In many economic situations agents must act strategically by taking into account the
behaviour of others. Game theory provides a set of tools that enables you to analyse
such situations in a logically coherent manner. For each concept introduced, you should
try to understand why it makes sense as a tool of analysis and (this is much harder) try
to see what its shortcomings might be. This is not the right place to ask questions such
as what is the policy-relevance of this concept?. The concepts you come across here are
just tools, and that is the spirit in which you should learn them. As you will see, some
of these concepts are used later in this course (as well as in a variety of other economics
courses that you might encounter later) to analyse certain types of economic
interactions.
explain why Nash equilibrium is the central solution concept and explain the
importance of proving existence
explain the idea of refining Nash equilibria in extensive-form games using backward
induction and subgame perfection
49
4. Game theory
explain the multiplicity of equilibria in repeated games and state the folk theorem
for the Prisoners Dilemma game.
50
4.3. Simultaneous-move or normal-form games
Suppose Si denotes the set of strategies of player i. In the example above, S1 = {U, D}
and S2 = {L, R}. Let si denote a strategy of i (that is, si is in the set Si ). If there are n
players, a strategy profile is (s1 , s2 , . . . , sn ). For any such strategy profile, there is a
payoff for each player. Let ui (s1 , s2 , . . . , sn ) denote the payoff of player i.
Continuing the example above, suppose u1 ({U, L}) = u1 ({D, R}) = 1 and
u1 ({D, L}) = u1 ({U, R}) = 2. Furthermore, suppose u2 ({U, L}) = u2 ({D, R}) = 2,
u2 ({D, L}) = 3 and u2 ({U, R}) = 1.
We write this in a convenient matrix form (known as the normal form) as follows. The
first number in each cell is the payoff of player 1 and the second number is the payoff of
player 2. Note that player 1 chooses rows and player 2 chooses columns.
Player 2
L R
Player 1 U 1, 2 2, 1
D 2, 3 1, 2
(the ith element is missing). With this notation, we can write the payoff of player i as
ui (si , si ).
Player 2
C D
Player 1 C 2, 2 0, 3
D 3, 0 1, 1
Here each player has a dominant strategy, D. This is the well-known Prisoners
Dilemma game. Rational players, playing in rational self-interest, get locked into a
51
4. Game theory
Dominant strategy
ui (si , si ) > ui (si , si ) for all si different from si and for all si .
That is, si performs better than any other strategy of player i no matter what others
are playing.
Even if a player does not have a dominant strategy, he might have one or more
dominated strategies. A dominated strategy for i is a strategy of i (say si ) that yields a
lower payoff compared to another strategy (say s0i ) irrespective of what others are
playing. In other words, the payoff of i from playing si is always (i.e. under all possible
choices of other players) lower than the payoff from playing s0i . Since si is a dominated
strategy, i would never play this strategy. Thus we can eliminate dominated strategies.
Indeed, we can eliminate such strategies not just once, but in an iterative fashion.
If in some game, all strategies except one for each player can be eliminated by
iteratively eliminating dominated strategies, the game is said to be dominance solvable.
Consider the game in Section 4.3. Note that no strategy is dominant for player 1, but
for player 2 L dominates R. So we can eliminate the possibility of player 2 playing R.
Once we do this, in the remaining game, for player 1 D dominates U (2 is greater than
1). So we can eliminate U . We are then left with (D, L), which is the equilibrium by
iteratively eliminating dominated strategies. The game is dominance solvable.
Here is another example of a dominance solvable game. We find the equilibrium of this
game by iteratively eliminating dominated strategies.
Player 2
Left Middle Right
Top 4, 3 2, 7 0, 4
Player 1 Middle 5, 5 5, 1 4, 2
Bottom 3, 5 1, 5 1, 6
52
4.3. Simultaneous-move or normal-form games
3. In the remaining game, for player 1, Top is dominated by Middle. Eliminate Top.
Player 2
A2 B2 C2
A1 3, 1 1, 3 4, 2
Player 1 B1 1, 0 3, 1 3, 0
C1 2, 3 2, 0 3, 2
As noted above, the problem with dominance criteria is that they apply only to some
games. For games that do not have dominant or dominated strategies, the idea of
deriving an equilibrium using dominance arguments does not work. If we cannot derive
an equilibrium by using dominant strategies or by (iteratively) eliminating dominated
strategies, how do we proceed?
If we want to derive an equilibrium that does not rely on specific features such as
dominance, we need a concept of equilibrium that applies generally to all games. As we
show below, a Nash equilibrium (named after the mathematician John Nash) is
indeed such a solution concept.
Before proceeding further, we need to clarify something about the nature of strategies.
In the discussion above, we identified strategies with single actions. For example, in the
Prisoners Dilemma game, we said each player has the strategies C and D. However,
this is not a full description of strategies. A player could also do the following: play C
with probability p and play D with probability (1 p), where p is some number
between 0 and 1. Such a strategy is called a mixed strategy; while a strategy that just
chooses one action (C or D) is called a pure strategy.
We start by analysing Nash equilibrium under pure strategies. Later we introduce
mixed strategies. We then note that one can prove an existence result: all games have at
least one Nash equilibrium (in either pure or mixed strategies). This is why Nash
equilibrium is the central solution concept in game theory.
53
4. Game theory
To find out the Nash equilibrium of the game above, we must look for the mutual best
responses. Let us check the best response of each player. Player 1s best response is as
follows:
Note from these that the only mutual best response is (B1 , B2 ). This is the only Nash
equilibrium in this game.
You could also check as follows:
From the three steps above, we can conclude that (B1 , B2 ) is the only Nash equilibrium.
54
4.3. Simultaneous-move or normal-form games
Let us also do a slightly different exercise. Suppose you want to check if a particular
strategy profile is a Nash equilibrium. Suppose you want to check if (A1 , C2 ) is a Nash
equilibrium. You should check as follows.
If player 2 plays C2 , player 1 cannot do any better by changing strategy from A1 (4 is
better than 3 from B1 or 3 from C1 ). However, player 2 would not want to stay at
(A1 , C2 ) since player 2 can do better by switching to B2 (3 from B2 is better than 2
from C2 ). We can therefore conclude that (A1 , C2 ) is not a Nash equilibrium.
You can similarly check that all boxes other than (B1 , B2 ) have the property that some
player has an incentive to switch to another box. However, if the players are playing
(B1 , B2 ) no player has an incentive to switch away. Neither player can do better by
switching given what the other player is playing. Since player 2 is playing B2 , player 1
gets 3 from B1 which is better than 1 from A1 or 2 from C1 . Since player 1 is playing
B1 , player 2 gets 1 from B2 which is better than 0 from A2 or C2 . Therefore (B1 , B2 ) is
a Nash equilibrium.
Nash equilibrium is not necessarily unique. Consider the following game.
Player 2
A B
A 2, 1 0, 0
Player 1 B 0, 0 1, 2
Note there are multiple pure strategy Nash equilibria. Both (A, A) and (B, B) are Nash
equilibria.
Note that while a dominant strategy equilibrium is also a Nash equilibrium, Nash
equilibrium does not require dominance. However, the greater scope of Nash equilibrium
comes at a cost: it places greater rationality requirements on players. To play (B1 , B2 ),
player 1 must correctly anticipate that player 2 is going to play B2 . Such a requirement
is even more problematic when there are multiple Nash equilibria. On the other hand, if
players have dominant strategies, they do not need to think at all about what others are
doing. A player would simply play the dominant strategy since it is a best response no
matter what others do. This is why a dominant strategy equilibrium (or one achieved
through iterative elimination of dominated strategies) is more convincing than a Nash
equilibrium. However, as noted before, many games do not have dominant or dominated
strategies, and are therefore not dominance solvable. We need a solution concept that
applies generally to all games, and Nash equilibrium is such a concept.
Player 2
A2 B2
A1 3, 1 2, 3
Player 1 B1 2, 1 3, 0
55
4. Game theory
Notice that this game has no pure strategy Nash equilibrium. However, as you will see
below, the game does have a Nash equilibrium in mixed strategies.
Let us first define a mixed strategy.
Mixed strategy
In the game above, A1 and B1 are the pure strategies of player 1. A mixed strategy of
player 1 could be A1 with probability 1/3, and B1 with probability 2/3. Notice that a
pure strategy is only a special case of a mixed strategy.
A Nash equilibrium can then be defined in the usual way: a profile of mixed strategies
that constitute a mutual best response.
A mutual best response in mixed strategies has an essential property that makes it easy
to find mixed strategy Nash equilibria. Let us consider games with two players to
understand this property.
Suppose we have an equilibrium in which both players play strictly mixed strategies:
player 1 plays A1 with probability p and B1 with probability (1 p) where 0 < p < 1,
and player 2 plays A2 with probability q and B2 with probability (1 q) where
0 < q < 1.
In this case, whenever player 1 plays A1 , she gets an expected payoff of:
What must be true of these expected payoffs that player 1 gets from playing A1 and
B1 ? Suppose 1 (A1 ) > 1 (B1 ). Then clearly player 1 should simply play A1 , rather than
any strictly mixed strategy, to maximise her payoff. In other words, player 1s best
response in this case would be to choose p = 1, rather than a strictly mixed strategy
p < 1. But then we do not have a mixed strategy Nash equilibrium.
Similarly, if 1 (A1 ) < 1 (B1 ), player 1s best response would be to choose p = 0 (i.e. just
to play B1 ), and again we cannot have a mixed strategy Nash equilibrium.
So if player 1 is going to play a mixed strategy in equilibrium it must be that she is
indifferent between the two strategies. How does such indifference come about? This is
down to player 2s strategy choice. Player 2s choice of q must be such that player 1 is
indifferent between playing A1 or B1 . In other words, in equilibrium q must be such
that 1 (A1 ) = 1 (B1 ), i.e. we have:
3q + 2(1 q) = 2q + 3(1 q)
56
4.3. Simultaneous-move or normal-form games
1 = 3p
Figure 4.1 below shows these best response functions and shows the equilibrium point
(where the two best response functions intersect).
Figure 4.1: The best-response functions. They cross only at E, which is the mixed strategy
Nash equilibrium. There are no pure strategy Nash equilibria in this case.
57
4. Game theory
Activity 4.1 Find the mixed strategy Nash equilibrium in the following game. Also
show all Nash equilibria of the game in a diagram by drawing the best response
functions.
Player 2
A B
A 2, 1 0, 0
Player 1 B 0, 0 1, 2
Existence theorem
Every game with a finite number of players and finite strategy sets has at least one
Nash equilibrium.
Nash proved the existence theorem for his equilibrium concept using a mathematical
result called a fixed-point theorem. Take any strategy profile and compute the best
response to it for every player. So the best response is another strategy profile. Suppose
we do this for every strategy profile. So long as certain conditions are satisfied, a
fixed-point theorem says that there is going to be at least one strategy profile which is a
best response to itself. This is, of course, a Nash equilibrium. Thus upon setting up the
strategy sets and the best response functions properly, a fixed-point theorem can be
used to prove existence. You will see a formal proof along these lines if you study game
theory at more advanced levels. Here, let us point out the importance of this result.
The existence theorem is indeed very important. It tells us that no matter what game
we look at, we will always be able to derive at least one Nash equilibrium. If existence
did not hold for some equilibrium concept (for example, games do not necessarily have a
dominant strategy equilibrium), we could derive wonderful properties of that concept,
but we could not be sure such derivations would be of any use. The particular game
58
4.4. Sequential-move or extensive-form games
that we confront might not have an equilibrium at all. But being able to prove existence
for Nash equilibrium removes such problems. Indeed, as noted before as well, this is
precisely what makes Nash equilibrium the main solution concept for
simultaneous-move games.
59
4. Game theory
For the sake of brevity of notation, we write these as follows. Just as we read words
from left to right, we read strategies from left to right. So we write the strategy if
player 1 plays a1 , play b2 and if player 1 plays a2 , play b1 as b2 b1 . Reading from left to
right, this implies that the plan is to play b2 at the left node and play b1 at the right
node. This is precisely what the longer specification says.
So the strategy set of player 2 is {(b1 b1 ), (b1 b2 ), (b2 b1 ), (b2 b2 )}.
Suppose instead of 2 actions, player 2 could choose between b1 , b2 and b3 at each node.
In that case, player 2 would have 3 3 = 9 strategies.
60
4.4. Sequential-move or extensive-form games
Suppose player 1 had 3 strategies a1 , a2 and a3 , and after each of these, player 2 could
choose between b1 and b2 . Then player 2 would have 2 2 2 = 8 strategies.
Player 2
b1 b1 b1 b2 b2 b1 b2 b2
Player 1 a1 3, 1 3, 1 1, 0 1, 0
a2 4, 1 0, 1 4, 1 0, 1
Note that if we pair, say, a1 with b2 b1 , only the first component of player 2s strategy is
relevant for the payoff. In other words, since player 1 plays a1 , the payoff is generated by
player 2s response to a1 , which in this case is b2 . Similarly, if we pair a2 with b2 b1 , the
second component of player 2s strategy is relevant for the payoff. Since player 1 plays
a2 , we need player 2s response to a2 , which in this case is b1 , to determine the payoff.
Once we write down the normal form, it is easy to find Nash equilibria. Here let us only
consider pure strategy Nash equilibria. There are three pure strategy Nash equilibria in
this game. Finding these is left as an exercise.
Example 4.1 Find the pure strategy Nash equilibria of the extensive-form game
above.
It is easiest to check each box. If we start at the top left-hand box, player 1 would
switch to a2 . So this is not a Nash equilibrium. From (a2 , b1 b1 ) no player can do
better by deviating. Therefore, (a2 , b1 b1 ) is a Nash equilibrium.
Next try (a1 , b1 b2 ). This is indeed a Nash equilibrium. Note that you must write
down the full strategy of player 2. It is not enough to write (a1 , b1 ). Unless we know
what player 2 would have played in the node that was not reached (in this case the
node after a2 was not reached), we cannot determine whether a strategy is part of a
Nash equilibrium. So while (a1 , b1 b2 ) is indeed a Nash equilibrium, (a1 , b1 b1 ) is not.
Finally, (a2 , b2 b1 ) is also a Nash equilibrium. These are the three pure strategy Nash
equilibria of the game.
61
4. Game theory
player does not observe some past moves, what he believes took place becomes
important. We do not study these problems here the analysis below simply shows you
how to represent situations of imperfect information in some simple cases.
Consider the extensive-form game introduced at the start of this section. Suppose at the
time of making a decision, player 2 does not know what strategy player 1 has chosen.
Since player 2 does not know what player 1 has chosen, at the time of taking an action
player 2 does not know whether he is at the left node or at the right node. To capture
this situation of imperfect information in our game-tree, we say that the two decision
nodes of player 2 are in an information set. We represent this information set as in
Figure 4.3: by connecting the two nodes by a dotted line (another standard way to do
this is to draw an elliptical shape around the two nodes you will see this in N&S).
Note that player 2 knows he is at the information set (after player 1 moves), but does
not know where he is in the information set (i.e. he does not know what player 1 has
chosen). Since player 2 cannot distinguish between the two nodes inside his information
set, he cannot take different actions at the two nodes. Therefore, the strategy set of 2 is
simply {b1 , b2 }. In other words, now, for both players the strategy set coincides with the
action set. This is not surprising: since player 2 takes an action without knowing what
player 1 has done, the game is the same as a simultaneous-move game. Indeed, the
normal form of the game above coincides with a game in which the two players choose
strategies simultaneously. This is shown below. Now the Nash equilibrium is simply
(a2 , b1 ).
Player 2
b1 b2
a1 3, 1 1, 0
Player 1 a2 4, 1 0, 1
62
4.5. Incredible threats in Nash equilibria and subgame perfection
Firm I
A F
In 2, 1 1, 2
Firm E Out 0, 2 0, 2
Note that there are two pure strategy Nash equilibria: (In, A) and (Out, F). The latter
equilibrium involves an incredible threat. Clearly, if the entrant does decide to enter the
63
4. Game theory
market, the incumbent has no incentive to choose F. Hence the threat of F is incredible.
Yet, Nash equilibrium cannot preclude this possibility. Out is the best response to F,
and once Firm E decides to stay out, anything (and in particular F) is a best response
for Firm I.
The game in Figure 4.5 presents another example.
When you write the normal form and work out the Nash equilibria, you should see
that (R, rr) is a Nash equilibrium. Look at the game above and see that this
involves an incredible threat. Player 2s strategy involves playing r after L. Player
1s strategy is taking a best response given this, and so player 1 is playing R. Given
that player 1 is playing R, the threat is indeed a best response for player 2 (indeed,
given that player 1 plays R, anything that player 2 can choose after L is trivially a
best response since it does not change the payoff, but, of course, not every choice
would lead to an equilibrium in which player 1s best response is R).
64
4.5. Incredible threats in Nash equilibria and subgame perfection
Subgame
A subgame is a part of a game that starts from a node which is a singleton (i.e. a
subgame does not start at an information set), and includes all successors of that node.
If one node in an information set belongs to a subgame, so do all other nodes in that
information set. In other words, you cannot cut an information set so that only part of
it belongs to a subgame. That would clearly alter the information structure of the
game, which is not allowed.
We can now define a subgame perfect equilibrium.
65
4. Game theory
What player 2 plays at the unreached node is crucial. If player 2 played ` after L, R
would not be the optimal choice. Therefore, you must specify player 2s full strategy
and identify (R, `r) as the SPNE.
In these games, the SPNE is unique, but it need not be. The next activity presents an
example.
Activity 4.3 Derive the pure strategy subgame perfect Nash equilibria of the game
in Figure 4.2.
66
4.5. Incredible threats in Nash equilibria and subgame perfection
decides between A and B, and player 2 decides between C and D. When player 2 makes
his decision, he knows that player 1 has decided to come in (if not then player 2 would
not have been asked to play), but without knowing what player 1 has chosen between A
and B. In other words, the situation is just as if once player 1 comes in, player 1 and
player 2 play a simultaneous-move game (as the game structure shows, they do not
move simultaneously player 1 moves before player 2, but since player 2 has no
knowledge of player 1s move, it is similar to the decision problem faced in a
simultaneous move game).
1. Pure strategy Nash equilibria. Let us first identify the pure strategy Nash
equilibria. Note that player 1 has 4 strategies: Out A, Out B, In A and In B, while
player 2 has 2 strategies: C and D. You might think strategies like Out A do not make
sense, but in game theory we try to model the thought process of players, and even if
player 1 stays out, she would do so only after thinking about what she would have done
had she entered. Strategies reflect such thinking (it is as if player 1 is saying I have
decided to finally stay out, but had I come in I would have chosen A).
Let us now write down the normal form of the game.
Player 2
C D
Out A 1, 3 1, 3
Player 1 Out B 1, 3 1, 3
In A 2, 2 2, 0
In B 0, 2 5, 5
You should be able to see from this that the pure strategy Nash equilibria are
(Out A, C), (Out B, C) and (In A, D).
67
4. Game theory
2. Pure strategy subgame perfect Nash equilibria. Note that backward induction
does not work here: we cannot fold back given the information set of player 2. However,
subgame perfection still works. Let us see how applying subgame perfection can refine
the set of Nash equilibria.
Note that apart from the whole game, there is just one strict subgame, which starts at
the node after player 1 chooses In. Below we write down the normal form of the
subgame.
Player 2
C D
Player 1 A 2, 2 2, 0
B 0, 2 5, 5
As you can see, the subgame has two pure strategy Nash equilibria: (B, C) and (A, D).
If (B, C) is played in the subgame, player 1 compares 1 (from Out) with 0 (from In
followed by (B, C)) and decides to stay out. Therefore, a SPNE of the whole game is
(Out B, C).
If, on the other hand, (A, D) is played in the subgame, player 1 compares 1 with 2 and
decides to come in. Therefore, another SPNE of the whole game is (In A, D).
It follows that the pure strategy SPNE of the whole game are (Out B, C) and (In A, D).
Another way to derive these is as follows. Since the set of SPNE is a subset of the set of
Nash equilibria, and since (B, C) and (A, D) are the Nash equilibria in the subgame, it
must be that any Nash equilibria of the whole game that involves playing either (B, C)
and (A, D) in the subgame are subgame perfect. Considering the set of Nash equilibria
derived above, we can immediately infer that (Out A, C) is Nash but not subgame
perfect, while the other two are subgame perfect.
Player 2
C D
Player 1 C 2, 2 0, 3
D 3, 0 1, 1
In a one-shot game, rational players simply play their dominant strategies. So (D, D) is
the only possible equilibrium. Suppose the game is repeated. Can we say something
about the behaviour of players in such supergames that differs from the behaviour in
the one-shot game?
First, consider the case of a finite number of repetitions. Say the game is played twice.
Would anything change? The answer is no. In the second round, players simply face a
one-shot game and they would definitely play their dominant strategies. Given that
(D, D) will be played in the next period, playing anything other than D today makes no
sense. Therefore, in each period players would play (D, D). But this logic extends to any
68
4.6. Repeated Prisoners Dilemma
finite number of repetitions. If the game is played a 100 times, in the last period (D, D)
will be played. This implies that (D, D) will be played in the 99th period, and so on.
While the logic is inescapable, actual behaviour in laboratory settings differs from this.
Faced with a large finite number of repetitions, players do cooperate for a while at least.
Therefore, it is our modelling that is at fault. To escape from the logic of backward
induction, we can assume that when a game is repeated many times, players play them
as if the games are infinitely repeated. In that case, we must apply forward-looking logic
as there is no last period from which to fold backwards.
(A quick note: you should be aware that there are other games with multiple Nash
equilibria where some cooperation can be sustained even under a finite number of
repetitions. You will encounter these in more advanced courses. Here we only consider
the repeated Prisoners Dilemma.)
Let us now analyse an infinitely repeated Prisoners Dilemma game.
First, we need to have an appropriate notion of payoffs in the infinitely repeated game.
Each player plays an action (in this case either C or D) in each period. So in each
period, the players end up playing one of the four possible action profiles (C, C), (C, D),
(D, C) or (D, D). Let at denote the action profile played in period t. Then in period t,
player i receives the payoff ui (at ). The payoff of player i in the repeated game is simply
the discounted present value of the stream of payoffs.
Let denote the common discount factor across players, where 0 < < 1. If todays
date is 0, and a player receives x in period t, the present value of that payoff is t x. The
discount factor can reflect players time preference. This can also arise from a simple
rate of interest calculation, in which case can be interpreted as 1/(1 + r), where r is the
rate of interest. Note that higher values of indicate that players are more patient (i.e.
value future payoffs more). If is very low, the situation is almost like a one-shot game,
since players only value todays payoff, and place very little value on any future payoff.
Given such discounting, the payoff of player i in the repeated game is:
ui (a0 ) + ui (a1 ) + 2 ui (a2 ) + .
More concisely, the payoff is:
X
t ui (at ).
t=0
If the payoff is the same every period (x, say), this becomes:
x
x(1 + + 2 + ) = .
1
69
4. Game theory
Trigger strategy
Start by playing C (that is, cooperate at the very first period, when there is no
history yet).
In period t > 1:
if (C, C) was played last period, play C
if anything else was played last period, play D.
Suppose each player follows this strategy. Note that cooperation (playing (C, C)) would
work only until someone deviates to D. After the very first deviation, each player
switches to D. Since anything other than (C, C) implies playing (D, D) next period,
once a switch to (D, D) has been made, there is no way back: the players must play
(D, D) forever afterwards. This is why this is a trigger strategy.
Another way of stating the trigger strategy is to write in terms of strategy profiles.
In period t:
if (C, C) is played in t 1, play (C, C)
otherwise play (D, D).
Let us see if this will sustain cooperation. Suppose a player deviates in period t. We
only need to consider what happens from t onwards. The payoff starting at period t is
given by:
3 + + 2 + = 3 + .
1
If the player did not deviate in period t, the payoff from t onwards would be:
2
2 + 2 + 2 2 + = .
1
For deviation to be suboptimal, we need:
2
>3+
1 1
which implies:
1
> .
2
Thus if the players are patient enough, cooperation can be sustained in equilibrium. In
other words, playing (C, C) always can be the outcome of an equilibrium if the discount
factor is at least 1/2.
70
4.6. Repeated Prisoners Dilemma
For example, note that always playing (D, D) is an equilibrium no matter what the
value of is. Each player simply adopts the strategy play D initially, and at any period
t > 1, play D irrespective of history. Note that both players adopting this strategy is a
mutual best response. Therefore, we can sustain (1, 1) in equilibrium. In fact, by using
suitable strategies, we can sustain many more in fact infinitely more payoffs as
equilibrium outcomes. For the Prisoners Dilemma game, we will describe the set of
sustainable payoffs below.
The result about the large set of payoffs that can be sustained as equilibrium outcomes
is known as the folk theorem. These types of results were known to many game
theorists from an early stage of development of non-cooperative game theory. While
formal proofs were written down later, we cannot really trace the source of the idea,
which explains the name.
Here we state a folk theorem adapted to the repeated Prisoners Dilemma game. To
state this, we will need to compare payoffs in the repeated game to payoffs in the
one-shot game that is being repeated. The easiest way to do that is to normalise the
repeated game payoff by multiplying by (1 ). Then if a player gets 2 every period,
the repeated game payoff is (1 ) 2/(1 ) = 2. As you can see, this normalisation
implies that the set of normalised repeated game payoffs now coincide with the set of
payoffs in the underlying one-shot game. So now we can just look at the set of payoffs of
the one-shot game and ask which of these are sustainable as the normalised payoff in
some equilibrium of the repeated game.
For the rest of this section, whenever we mention a repeated game payoff, we always
refer to normalised payoffs. Note that in this game, a player can always get at least 1 by
simply playing D. It follows that a player must get at least 1 as a (normalised) repeated
game payoff.
To see the whole set of payoffs that can be sustained, let us plot the payoffs from the
four different pure strategy profiles. These are shown in Figure 4.8 below. Now join the
payoffs and form a convex set as shown in the figure. We now have a set of payoffs that
can arise from pure or mixed strategies.
The Folk theorem is the following claim. Consider any pair of payoffs (1 , 2 ) such
that i > 1 for i = 1, 2. Any such payoff can be supported as an equilibrium payoff for
high enough .
As noted above, for the Prisoners Dilemma game we can also sustain the payoff (1, 1)
as an equilibrium outcome irrespective of the value of .
The set of payoffs that can be supported as equilibrium payoffs in our example is shown
as the shaded part in Figure 4.8.
(a) Find conditions on the discount factor under which cooperation can be
sustained in the repeated game in which the above game is repeated infinitely.
71
4. Game theory
Figure 4.8: The set of payoffs that can be supported as equilibrium payoffs under an
infinitely-repeated game.
(a) First, note that player 2 has no incentive to deviate from (C, C). To ensure that
player 1 does not deviate, consider the following strategy profile. Play (C, C)
initially. At any period t > 0, if (C, C) has been played in the last period, play
(C, C). Otherwise, switch to (D, D).
Under this strategy profile, player 1 will not deviate if:
3
7+2 6
1 1
which implies > 4/5.
(b) We want to support alternating between (C, C) and (D, D), starting with
(C, C) in period 0, as an equilibrium.
Note first that player 2 has no incentive to deviate in odd or even periods.
Player 1 cannot gain by a one-shot deviation in odd periods, when (D, D) is
supposed to be played. So the only possible deviation is by player 1 in even
periods, when (C, C) is supposed to be played.
To prevent such a deviation, consider the following strategy profile.
Start by playing (C, C) in period 0.
In any odd period t (where t = 1, 3, 5, . . .) play (D, D) (irrespective of
history).
72
4.6. Repeated Prisoners Dilemma
Note that this is a version of the trigger strategy. After any deviation from
cooperation in even periods, (D, D) is triggered forever.
If player 1 does not deviate in any even period t, player 1s payoff (t onwards) is:
Vt = 3 + 2 + 3 2 + 2 3 +
= 3(1 + 2 + 4 + ) + 2(1 + 2 + 4 + )
3 + 2
= .
1 2
If player 1 does deviate in even period t, player 1s payoff (t onwards) is:
Vtdev = 7 + 2 + 2 2 + = 7 + 2 .
1
Therefore, player 1 prefers not to deviate in either period 0 or in any even
period if:
3 + 2
> 7 + 2
1 2 1
which simplifies to 5 2 4 > 0, implying:
r
4
> 0.89.
5
Note also the repeated game payoff generated in this equilibrium. To see this, first
normalise the payoff by multiplying by (1 ). The normalised payoff of player 1
starting any even period is:
3 + 2 3 + 2
(1 ) 2
= .
1 1+
Similarly, the normalised payoff of player 1 starting any odd period is
(2 + 3)/(1 + ). Note that either payoff goes to 5/2 as 1.
For player 2, the normalised payoff from the equilibrium is (2 + )/(1 + ) starting
any even period, and (1 + 2)/(1 + ) starting any odd period. Note that either
payoff goes to 3/2 as 1.
In other words, this exercise shows you an example of an equilibrium that sustains a
payoff in the interior of the set of payoffs which is sustainable according to the folk
theorem (which, in this case, is anything strictly above (2, 1), or the point (2, 1)
itself).
73
4. Game theory
explain why Nash equilibrium is the central solution concept and explain the
importance of proving existence
explain the idea of refining Nash equilibria in extensive-form games using backward
induction and subgame perfection
explain the multiplicity of equilibria in repeated games and state the folk theorem
for the Prisoners Dilemma game.
Player 2
A2 B2 C2
A1 3, 3 1, 4 0, 5
Player 1 B1 2, 1 3, 2 1, 0
C1 1, 0 0, 1 1, 0
2. Identify actions and strategies of each player in each of the following games (Figure
4.9 to Figure 4.11).
74
4.8. Test your knowledge and understanding
75
4. Game theory
(a) Write down the actions and strategies for each player.
(b) Identify the pure and mixed strategy Nash equilibria.
(c) Identify the pure and mixed strategy subgame perfect Nash equilibria.
(a) Write down the actions and strategies for each player.
(b) Identify the pure strategy Nash equilibria.
(c) Identify the pure strategy subgame perfect Nash equilibria.
76
4.8. Test your knowledge and understanding
(a) Write down the actions and strategies for each player.
(b) Identify the pure strategy Nash equilibria.
(c) Identify the pure strategy subgame perfect Nash equilibria.
6. Find the pure strategy subgame perfect Nash equilibria for the game in Figure 4.9
in Question 2.
7. Suppose the following game is repeated infinitely. The players have a common
discount factor , where 0 < < 1.
Player 2
C D
Player 1 C 3, 2 0, 3
D 5, 0 2, 1
(a) Find conditions on the discount factor under which cooperation (which implies
playing (C, C) in each period) can be sustained as a subgame perfect Nash
equilibrium of the infinitely-repeated game. Your answer must specify the
strategies of players clearly.
(b) In a diagram, show the set of payoffs that can be supported in a repeated
game equilibrium according to the folk theorem.
77
4. Game theory
78
Chapter 5
Production, costs and profit
maximisation
5.1 Introduction
Over the next few chapters we develop models to analyse different market structures. In
particular, we analyse two extremes: a competitive market and a monopoly. Later on we
also analyse oligopoly. To develop these models, we need to make use of a variety of
concepts on the nature of technology, costs and optimisation by firms. This chapter
introduces various concepts to this end, many of which you should already be familiar
with from previous courses.
explain the concepts of average and marginal product and their relationship
derive the optimal input purchasing decision of a firm in the short run
derive the optimal input mix decision in the long run
explain the concepts of isoquants and marginal rate of technical substitution
explain how the shape of isoquants depends on input substitutability
explain the concepts of marginal cost, average cost, fixed cost, sunk cost and
variable cost
explain the various cost curves faced by a firm in the short run and in the long run
derive the profit-maximising output of a firm
explain the shut-down decision of a firm.
79
5. Production, costs and profit maximisation
Q = f (K, L).
f (K, L)
MPL = .
L
Similarly, the marginal product of capital is given by:
f (K, L)
MPK = .
K
We will assume that beyond some level of an input marginal product is diminishing. In
other words, beyond some level of labour (which could be zero), MPL is decreasing in
80
5.5. The short run: one variable factor
labour, and beyond some level of capital MPK is decreasing in capital. In the region
with diminishing marginal product, we have:
2 f (K, L)
<0
L2
and:
2 f (K, L)
< 0.
K 2
The average product of labour is given by:
Q
APL = .
L
You can define the average product of capital similarly.
What is the relationship between average and marginal product? You should
understand the general relationship between average and marginal. Suppose you play a
computer game and your average score is 60. The next time you play, you score 100.
What will happen to your average score? It will go up. Here the marginal score
exceeded the average, so the average goes up. Similarly, if the marginal were below the
average, the average would fall. The same principle applies to average and marginal
product (and to other comparisons such as between average and marginal cost that we
would make later). The average product rises if the average is below the marginal and
falls if the average is above the marginal. At low levels of output, the marginal product
is higher than the average product, and then eventually it decreases. Once it crosses
below the average product, the average product falls. It follows that the crossing point
is the highest point on the average curve.
Example 5.1 Plot the average and marginal product of labour for the following
production function:
Q = 10L + 30L2 L3 .
These are shown in Figure 5.1 below.
81
5. Production, costs and profit maximisation
Figure 5.1: Average and marginal product of labour for Activity 5.1.
On the other hand, one extra unit of labour costs w (i.e. w is the marginal cost of
labour). Thus the profit-maximising level is given by MRPL = w.
Let us derive this formally. Let P (Q) be the (inverse) demand function facing the firm.
This is simply P if the firm is a price-taker (a competitive firm), but under imperfect
competition, market price depends on firm output.
The firm maximises profit given by:
P (Q) f (K, L) wL rK
where K is the fixed level of K in the short run, and where Q = f (K, L). Differentiating
with respect to L, we get:
Q
P 0 (Q) f + P (Q)fL w = 0.
L
Since Q/L = fL , we have:
fL P 0 (Q) f + P (Q) = w.
fL P 0 (Q) Q + P (Q) = w.
Now, total revenue is P (Q) Q. So marginal revenue is P 0 (Q) Q + P (Q). Therefore, the
equation above is simply:
MPL MRL = w.
Note that for a competitive firm, price is given and does not vary with firm output.
Therefore, P 0 (Q) = 0, and the optimal labour hire equation becomes:
P fL = w
82
5.6. The long run: both factors are variable
How does demand for labour change as the wage rate varies in the short run? Note that
the marginal product of labour decreases with labour. Marginal revenue decreases as
output increases, and output increases as labour increases. Therefore, as labour
increases, MR falls. For a competitive firm MR is constant at price P . Thus either both
MR and MP fall as labour increases (imperfect competition) or MR is constant (at price
P ) while MP falls for competitive firms. Therefore, MRP falls as labour increases. It
follows that if the wage rate falls, more labour is hired.
Figure 5.2 below shows the effect of a fall in the market wage rate.
5.6.1 Isoquants
With two variable factors (typically this would be the case in the long run), we can
represent the production function by its level curves. An isoquant traces out the
different combinations of K and L that produce a given level of output. Analogously to
an indifference curve, the slope of an isoquant is given by the marginal rate of
technical substitution (MRTS). The marginal rate of technical substitution of L for
K is the amount by which capital can be reduced if the firm hires one extra unit of
labour. Analogously to an indifference curve:
MPL
MRTSLK = .
MPK
83
5. Production, costs and profit maximisation
84
5.7. Cost curves
TC(Q) = F + VC(Q).
It follows that:
F VC(Q)
AC(Q) = + .
Q Q
The first component is the average fixed cost, and the second is average variable cost.
Example 5.2 Plot the AC, AVC and MC for the cost curve:
Example 5.3 Let C(Q) = F + VC(Q) be the cost function of a firm where F is the
fixed cost and VC(Q) is the variable cost. Show algebraically that MC passes
through the lowest points of AC and AVC. Assume C 00 > 0 and V 00 > 0.
Note that dC/dQ = dVC(Q)/dQ = MC.
Let us find the minimum point of AC, which is C(Q)/Q. We have:
dC(Q)/Q 1
= 2 [Q C 0 (Q) C(Q)].
dQ Q
At the minimum point:
dC(Q)/Q
= 0.
dQ
85
5. Production, costs and profit maximisation
Figure 5.3: Average cost, average variable cost and marginal cost curves for Activity 5.2.
In the long run, all factors are variable, and the optimal input mix is chosen according
to the long-run optimisation rule derived previously. Since capital can be varied in the
long run, short-run costs are at least as high as the long-run costs, and if in the short
run the firm chooses anything other than the capital size associated with the lowest
long-run average costs, short-run costs are strictly higher.
See N&S Section 7.4 for an explanation using isoquants.
The long-run average cost curve (LRAC) is an envelope of the short-run average cost
curves (SRACs), as shown in Figure 5.4. Note that the long-run average cost is lower
than the minimum short-run average cost over the region where the firm experiences
increasing returns to scale. This is the region where the LRAC is falling. Similarly, over
the region where the LRAC is increasing, the firm experiences decreasing returns to
scale and, again, the LRAC is lower than the minimum of the SRAC. For the right
long-run plant size, the minimum point of the SRAC is also the minimum point of the
LRAC. Note also that the LRMC is flatter, i.e. more elastic, compared to the SRMC.
86
5.7. Cost curves
Figure 5.4: The long-run average cost curve is an envelope of the short-run average cost
curves.
As shown in Figure 5.4, if in the long run a firm experiences increasing returns to scale
at low levels of output, but decreasing returns at higher levels of output, the long-run
average cost is U-shaped, and the long-run MC cuts this from below at the lowest point.
However, if the firm experiences constant returns to scale, the long-run average cost
curve is a horizontal line at the level of the fixed long-run average cost. In this case
LRAC and LRMC are the same.
Q = L0.5 K 0.5 .
In the short run, K is fixed at K = 100. The wage rate is 25 and the rental rate of
capital is 8. Derive the short-run and long-run cost functions.
To produce Q in the short run, we need L so that L0.5 K 0.5 = Q, i.e. we have:
Q2
L(Q) = .
K
Therefore, the short-run cost function is:
Q2
CSR (Q) = 8K + 25 .
K
You can easily work out the short-run AC, AVC and MC. These are shown in Figure
5.5 below for K = 100.
To derive the long-run cost, use the Lagrangian method to minimise cost for a given
level of output. This gives us the tangency condition price ratio equal to MRTS.
Using this:
w 0.5L0.5 K 0.5 K
= 0.5 0.5
= .
r 0.5L K L
87
5. Production, costs and profit maximisation
Now:
5
Q = (25L/8)0.5 L0.5 = L
2 2
which implies:
2 2
L= Q.
5
Therefore:
CLR (Q) = 20 2 Q.
In this case LRAC = LRMC = 20 2.
Figure 5.6 below shows the short run cost and MC curves (the MC curves are the
dashed lines) for fixed short-run K levels of K = 100 and K = 200.
The minimum of SRAC with capital fixed at K occurs at 2 2K/5 (you should
derive this explicitly). The LRAC is the envelope of SRAC as K varies.
In this case we have a constant-returns-to-scale technology, so the LRAC is
horizontal and coincides with the LRMC.
Example 5.5 Consider a firm with production function Q = min{K, L}. The
wage rate is w, and the rental rate of capital is r. The firm wants to produce Q0
units of output. In the short run, K is fixed at K.
(a) Calculate the short-run optimal cost as a function of , , input prices w, r and
Q0 .
88
5.7. Cost curves
Figure 5.6: Short-run average cost curves for K = 100 (SRAC1) and K = 200 (SRAC2),
the associated marginal cost curves (shown as dashed lines), and the long-run average
(as well as marginal) cost curve.
(b) Calculate the long-run optimal cost as a function of , , input prices w, r and
Q0 .
(a) We consider the short-run optimum. Note that it is impossible to produce more
than K. Thus if Q0 exceeds this level, it cannot be produced in the short run.
Let us assume Q0 6 K. Then we need L so that L = Q0 , or:
Q0
L= .
The optimal short-run cost for Q0 6 K is, therefore, given by:
Q0
CSR = rK + w .
(b) We now consider the long-run optimum. To minimise cost, note that we must
have K = L. To produce Q0 , L and K must be such that:
K = L = Q0 .
Therefore:
Q0
K=
and:
Q0
L= .
It follows that the minimised value of cost is:
r w
C(Q0 ) = Q0 + .
89
5. Production, costs and profit maximisation
d dC(Q)
(P (Q) Q) = 0.
dQ dQ
For a price-taking firm, the price P does not depend on output, so d(P Q)/dQ = P , and
profit is maximised at the output where P = MC.
To decide the optimal level of output, a firm must consider two things.
Profit-maximising output
The firm must set MR equal to MC to determine the profit-maximising output level
Q .
Shut-down decision
In the short run, a firm continues to operate even if revenues do not cover
all costs. If fixed costs on equipment such as plants and machinery have been
incurred, this cannot be altered in the short run. So the short-run shut-down
decision pays no attention to fixed costs and only considers if the firm is covering
its variable costs. The firm should continue to operate so long as it covers its
variable costs. In other words, the firm shuts down production in the short run
if P Q < VC, or:
P < AVC.
In the long run all factors are variable. Therefore, the firm shuts down if it
cannot cover all its costs even at the minimum level of LRAC. In other words,
a firm shuts down in the long run if:
P < LRACmin .
You should note that if some part of the fixed costs is sunk to start with, this plays no
role in the calculation to shut down even in the long run. This part of the cost is not
recoverable by shutting down, so the shut down decision does not take it into account.
90
5.9. A reminder of your learning outcomes
explain the concepts of average and marginal product and their relationship
derive the optimal input purchasing decision of a firm in the short run
explain the concepts of marginal cost, average cost, fixed cost, sunk cost and
variable cost
explain the various cost curves faced by a firm in the short run and in the long run
2. Consider a firm with production function Q = K L . The wage rate is w, and the
rental rate of capital is r. Calculate the short-run optimal cost as a function of , ,
input prices w, r and Q.
3. Consider a firm with production function Q = min{K, L}. The wage rate is w,
and the rental rate of capital is r. Derive the long-run cost function of the firm.
4. Consider a firm with production function Q = K 1/2 L1/2 . The wage rate is w, the
rental rate of capital r. Derive the long-run cost function of the firm.
5. A firms short-run cost curve is C(Q) = 100 + 50Q bQ2 + Q3 /2. For what values
of b are cost, average cost and average variable cost positive at all levels of Q?
6. Consider a firms input choice in the long run. Suppose the production function of
a firm is given by Q = L K . The wage rate is w, and the capital rental rate is r.
(a) Calculate the firms optimal input mix for producing output Q0 .
91
5. Production, costs and profit maximisation
(b) Suppose = = 0.5, w = 12, r = 3 and Q0 = 40. Calculate the optimal input
mix.
(c) What is the cost at the optimal input mix derived in (b)?
(d) Calculate the marginal products of labour and capital at the optimal choice.
92
Chapter 6
Perfect competition in a single market
6.1 Introduction
The analysis of consumer behaviour in Chapter 2 showed how we can derive the
demand curve in a market. In the first part of this chapter, we analyse the behaviour of
competitive firms, and show how we can derive the supply curve in a market. In the
second half of this chapter, we study the partial equilibrium competitive model that
puts demand and supply together to derive the equilibrium in individual markets. Note
that this is different from a general equilibrium analysis that considers all markets in
an economy together, which we do in the next chapter. The partial equilibrium
approach allows us to understand in a simple setting how small price-taking agents
(both buyers and sellers) come together in a competitive market. Furthermore, we can
use this supply-demand model to carry out a partial equilibrium analysis of the impact
of policy on specific markets.
explain the supply decision of a price-taking firm in the short run and in the long
run
derive the industry supply curve in the short run and in the long run
calculate the market equilibrium price and quantity given supply and demand
curves
analyse the impact of entry and exit decisions of firms and their impact on industry
supply
relate the nature of input costs to the long-run industry supply curve
93
6. Perfect competition in a single market
explain the concept of producer surplus and use this in conjunction with consumer
surplus as a measure of welfare
use the measure of welfare to evaluate the impact of a variety of tax and subsidy
policies on the market
use the measure of welfare to evaluate the impact of policies: examples include
price ceilings (such as rent control), price floors (such as minimum price guarantees
for agricultural products), permits and quotas, price support policies as well as
trade policies involving tariffs and quotas.
94
6.4. Supply decision by a price-taking firm
In discussing markets, we often talk about market supply and market demand and how
the intersection of the two determines market equilibrium. But does any kind of market
structure produce a market supply curve? In fact, this is not true. A market supply
curve is an appropriate concept only when the firms in an industry are price-takers. In
this case, we can associate a specific firm and industry output to each price and,
therefore, trace out firm and industry supply curves, respectively. But if a firm has
control over price, that is, if the market price depends on firm output, the firm would
consider market demand and see at which point on market demand it should produce.
In this case, there is no unique relationship between price and output. As demand
conditions vary, so would the price-quantity pair of a firm vary. Therefore, we cannot
trace out a supply curve. So keep in mind that whenever you say supply curve you are
in fact talking about price-taking firms and, by implication, a competitive market
structure.
A price-taking firm faces a notional perfectly-elastic demand curve. Essentially, the firm
can sell any amount at the given price, P . The idea is that a firm in a competitive
market is a very small part of the overall market so that whatever amount such a firm
produces is a very small part of the overall supply in the market. Therefore, variations
in the output of a single firm have a vanishingly small impact on the market price,
hence the notional firm demand curve is horizontal. Figure 6.1 below shows profit
maximisation by a firm in the short run.
Figure 6.1: Short-run profit maximisation by a competitive firm. The demand curve facing
the firm is perfectly elastic at price P (i.e. the firm is a price-taker). Profit is maximised
at P = MC and the maximised value of profit is (P AC(Q )) Q .
95
6. Perfect competition in a single market
Example 6.1 Derive the short-run firm supply curve given the following cost
functions:
AVC = 3 + 5q
and:
MC = 3 + 10q.
In this case, MC is always above AVC, so the (inverse) short-run supply curve
of a price-taking firm is given by:
P = 3 + 10q.
You can invert this to get the usual form of supply curve:
P 3
q= .
10
AVC = 10 6q + q 2 .
6 + 2q = 0
or q = 3. At q = 3, AVC = 1. Therefore:
AVCmin = 1.
The (inverse) supply curve is given by P = MC for P > 1. For lower values of P
the supply is zero. Therefore, the supply curve is given by:
3q 2 12q + (10 P ) = 0.
96
6.5. Market supply and market equilibrium
97
6. Perfect competition in a single market
Example 6.2 Consider the firm supply functions derived in Activity 6.1. Suppose
there are n identical firms in the industry. Derive the market supply function in each
case.
(a) We have q = (P 3)/10. Since there are n firms, the market supply is Q = nq.
Therefore, the market supply curve is given by:
P 3
Q=n .
10
(b) Here, we have q = 2 + P + 2/ 3 for P > 1. Therefore, the industry supply is:
r
P +2
Q = 2n + n
3
for P > 1. As before, for P < 1, supply is zero.
The intersection of market demand and market supply determines the market
equilibrium price and quantity.
and the long-run market supply is simply the firms supplies added together.
However, once we allow entry and exit, this is a little different. The process of entry and
exit implies that firms always end up earning a zero profit and producing at the
minimum long-run average cost. How this minimum behaves as the market output
expands determines the shape of the long-run market supply curve.
If input costs do not rise as new firms enter, the long-run supply curve is horizontal at
the minimum LRAC. In this case the long-run market supply equation is:
P = LRACmin .
If input costs rise, we have an increasing-cost industry and you should be able to show
that in this case we get a long-run market supply curve that slopes upwards. Similarly,
if we have a decreasing-cost industry, the long-run market supply curve slopes
downwards. However, whether there are any realistic cases of a downward-sloping
supply function is unclear see N&S for a discussion.
Just as we can calculate demand elasticities, we can calculate supply elasticities in the
short and long runs. The elasticity of supply is simply:
P dQS
S = .
QS dP
98
6.6. Producer surplus
This is positive in all cases other than for the long-run supply under a decreasing-cost
industry (if this is at all possible).
Furthermore, the long-run supply curves are more elastic than their short-run
counterparts (why?).
99
6. Perfect competition in a single market
the elasticities of supply and demand. You need to understand these issues clearly. N&S
develops this in Section 9.10. See also Perloff, Section 9.5.
The general principle is that the more inelastic end of the market bears more of any
(per-unit) tax, and benefits more from any (per-unit) subsidy.
100
6.7. Applications of the supply-demand model: partial equilibrium analysis
Example 6.4 Suppose the demand curve for wheat is Q = 100 10P , and the
supply curve is Q = 10P . What are the effects of a subsidy of s = 1 per unit on the
equilibrium, government subsidy costs, consumer surplus, producer surplus and total
welfare?
In the following discussion, again we denote consumer surplus by CS, producer
surplus by PS and deadweight loss by DWL.
In Figure 6.4 below, the initial equilibrium is E, where P = 5 and Q = 50. The CS is
(10 5) 50/2 = 125 and the PS is 5 50/2 = 125. The total surplus is 250 and
there is no DWL.
After the subsidy, the supply curve becomes P = Q/10 s = Q/10 1. The new
equilibrium is P = 4.5, Q = 55. The buyers pay 4.5 per unit, while sellers receive
4.5 + 1 = 5.5 per unit.
Note that the incidence is split equally between the two sides of the market. This
results from the fact that the supply and demand elasticities have the same absolute
value.
Now consider the impact of the subsidy.
The new CS is the triangle ABC which has an area of (10 4.5) 55/2 = 151.25.
The new PS is the triangle FGH which has an area of 5.5 55/2 = 151.25. The
government spending on subsidy is 1 55 = 55, which is the rectangle GBCH. Also,
there is now a DWL given by the shaded triangle EHC which has an area of
1 5/2 = 2.5.
Another way to calculate the DWL is to compare the total surplus (TS) before and
after the policy. Before the policy the TS is 250. After the policy, the TS is
151.25 + 151.25 55 = 247.5. Therefore, the DWL is 250 247.5 = 2.5.
Note that the DWL arises because after the subsidy production extends to levels
101
6. Perfect competition in a single market
where the marginal benefit (given by the demand curve) is lower than the marginal
cost (given by the supply curve). Note also the comparison with the DWL from a
tax. A per-unit tax gives rise to a DWL because it reduces quantity so some of the
surplus remains unexploited. For a subsidy, on the other hand, the DWL arises
because production extends beyond the efficient quantity.
102
6.7. Applications of the supply-demand model: partial equilibrium analysis
Figure 6.5: Deadweight loss under rent control with different supply elasticities.
Figure 6.6 shows the outcome under the most pessimistic case. The deadweight loss is
larger in this case.
103
6. Perfect competition in a single market
Figure 6.6: Outcome under a price ceiling, the most pessimistic case. In this case
consumers between QD QC and QD obtain the good. The surplus area A is now lost,
but area B is gained, so the deadweight loss is given by area A minus area B.
production level Q2 would be happy to supply the product at the price floor, therefore a
misallocation is possible, which would raise the deadweight loss further.
Figure 6.7: Price floor at PF . The deadweight loss is given by area A plus area B.
104
6.7. Applications of the supply-demand model: partial equilibrium analysis
6.7.4 Quota
Figure 6.8 shows the impact of a quota at Q1 . Note that, as in the case of a price floor
above, some of the surplus is lost, and out of the rest, some of the consumer surplus is
shifted to suppliers. Note also that there is an implicit assumption that the sellers with
the lowest cost (those at the lowest end of the supply curve) are the ones who obtain
the quota. Again, a misallocation would raise the deadweight loss. However, irrespective
of the initial allocation of quotas, if they are tradeable, it is likely that they would end
up in the hands of the most efficient sellers. In this respect, the implicit assumption is
perhaps not too restrictive in this case. When you study the commons problem in
Chapter 12 of the subject guide, you will come across an instance of transferable quotas:
an arrangement known as ITQs, used in managing fishing levels in several countries.
Figure 6.8: Quota at Q1 . Once the quota is imposed, the market price rises to P1 . The
deadweight loss is given by area A plus area B.
Under a price support policy for agricultural products, the government maintains a
minimum market price in order to ensure a minimum income level for farmers. To ensure
that the price does not fall below the announced support price, the government buys up
any excess supply at the support price. Figure 6.9 shows the effect of such a policy.
105
6. Perfect competition in a single market
Let us calculate the deadweight loss from the policy in Figure 6.9 above, and then try
to see whether there is a different policy which attains the same objective of raising the
income level for farmers, but incurs a smaller deadweight loss.
In this case, the change in consumer surplus is (A + B), the change in producer
surplus is +(A + B + C). Next, the cost to the government is PS (Q2 Q1 ). Adding
these up, the net loss from the policy is PS (Q2 Q1 ) minus the triangle C (area
B + D + E + F + G). This is shown in Figure 6.10 below.
More generally, this shows why income generation for farmers through price support
policies (the EU common agricultural policy is an important example) is inefficient. The
policy gets farmers to produce more (even though demand is less at a higher price) and
then spends government revenue to buy up the unsold amount. This is, of course, a
poor use of resources.
A better solution is to simply give the amount of money A + B + C to farmers directly,
and not interfere with the market. This avoids the unnecessary excess production, and
incurs no deadweight loss.
Suppose that the government wants to raise the price of a commodity on the domestic
market by reducing imports. Assume that the importing country is a price-taker in
world markets. Below, we compare the welfare effects of an import tariff with an import
quota.
The best way to proceed is to compare the two policies using a diagram. As shown in
Figure 6.11, a tariff can reduce imports and raise the domestic price as follows. Without
intervention, the domestic price is equal to the world price, PW . At this price, domestic
consumption is QD , of which QS is produced domestically, and QD QS is imported. A
106
6.7. Applications of the supply-demand model: partial equilibrium analysis
Figure 6.10: Price support and deadweight loss. The deadweight loss is the pattern-filled
area which is PS (Q2 Q1 ) minus the triangle C.
Figure 6.11: The welfare effects of an import tariff and an import quota.
tariff of:
T = P PW
raises the price to P , increases domestic production to Q0S , and reduces exports to
Q0D Q0S .
The welfare loss from this policy is as follows. Consumer surplus changes by
107
6. Perfect competition in a single market
W = (A + B + C + D) + A + D = B C.
explain the supply decision of a price-taking firm in the short run and in the long
run
derive the industry supply curve in the short run and in the long run
calculate the market equilibrium price and quantity given supply and demand
curves
analyse the impact of entry and exit decisions of firms and their impact on industry
supply
relate the nature of input costs to the long-run industry supply curve
explain the concept of producer surplus and use this in conjunction with consumer
surplus as a measure of welfare
use the measure of welfare to evaluate the impact of a variety of tax and subsidy
policies on the market
use the measure of welfare to evaluate the impact of policies: examples include
price ceilings (such as rent control), price floors (such as minimum price guarantees
for agricultural products), permits and quotas, price support policies as well as
trade policies involving tariffs and quotas.
108
6.9. Test your knowledge and understanding
2. In the short run, the demand for cigarettes is totally inelastic. In the long run,
suppose that it is perfectly elastic. What is the impact of a cigarette tax on the
price that consumers pay in the short run and in the long run?
3. Suppose the market demand curve is infinitely elastic. In a diagram, show the
deadweight loss from a subsidy. What is the incidence of any subsidy?
4. Suppose the government sets a price ceiling below the unregulated competitive
equilibrium price. Show the effect on the market equilibrium, consumer surplus,
producer surplus and total welfare.
5. Consider an agricultural price support policy where the government announces the
support price PS . Farmers then choose how much to produce, and sell everything to
consumers at the price at which everything can be sold. Let this price be P . The
government then pays farmers a deficiency payment equal to (PS P ) per unit on
the entire quantity produced.
Suppose the demand curve for wheat is Q = 100 10P and the supply curve is
Q = 10P . Suppose the government announces a support price of PS = 6 using a
deficiency payment programme.
(a) Calculate the quantity supplied, the price at which output is sold and the
deficiency payment.
(b) Calculate the effect on consumer surplus, producer surplus and total welfare.
109
6. Perfect competition in a single market
110
Chapter 7
General equilibrium and welfare
7.1 Introduction
Previously, we have studied competitive equilibrium in specific markets. This chapter
now introduces the idea of an economy-wide general equilibrium. The analysis of
competitive markets allowed us to evaluate welfare in a specific market. A general
equilibrium allows an evaluation of welfare in the entire economy. To simplify the
exposition and make the main ideas clear, we focus extensively on an exchange
economy. We consider the conditions for a general equilibrium and discuss the two
welfare theorems in detail. Later on, we assume that there is one agent (or a
representative agent) in an economy to introduce production in a simple setting.
construct an Edgeworth box and depict the contract curve, the set of possible
equilibrium exchanges given an endowment point, and depict the equilibrium price
ratio and allocation
derive algebraically the contract curve, the equilibrium price ratio and allocation
state and provide an informal proof outline for the first welfare theorem in the
context of an exchange economy with two agents and two goods as well as in the
context of an economy with production
discuss the policy implications of the first welfare theorem and understand the
situations in which the theorem might not apply
111
7. General equilibrium and welfare
state and provide an informal proof outline for the second welfare theorem in the
context of an exchange economy with two agents and two goods
discuss the policy implications of the second welfare theorem and understand the
situations in which the theorem might not apply
derive the equilibrium conditions in an economy with production.
General equilibrium
A general equilibrium for a pure exchange economy consists of prices (p1 , . . . , pn ) and
quantities ((x11 , . . . , x1n ), . . . , (xI1 , . . . , xIn )) such that:
at the given prices, for each i, the bundle (xi1 , . . . , xin ) maximises utility of agent
i, subject to the agents budget constraint
all markets clear, i.e. for each market j, total demand equals total supply:
I
X I
X
xij = eij .
i=1 i=1
112
7.3. General equilibrium in an exchange economy
The final consumption bundles of agents are together known as the equilibrium
allocation.
Here we study a simple case with two agents (A and B) and two goods (X and Y ).
General equilibrium in such a two-person exchange economy can be usefully studied
using an Edgeworth box diagram. You should be able to construct such a diagram
and depict the set of Pareto efficient outcomes (the contract curve), be able to identify
the possible set of Pareto-improving trades given initial endowments, be able to identify
the possible set of equilibrium exchanges, and be able to show the equilibrium price
ratio and equilibrium allocation.
Let XA and YA denote the endowments of goods X and Y , respectively, for agent A.
Similarly, let XB and YB denote the endowments of goods X and Y , respectively, for
agent B. Let:
X = XA + XB and Y = YA + YB .
In an Edgeworth box, the contract curve shows the Pareto efficient allocations. This is
shown in Figure 7.1 below.
Figure 7.1: The contract curve traces out the Pareto efficient allocations in the Edgeworth
box. These are points of tangency between the indifference curves of A and B. Once we
reach such a point of tangency, it is impossible to make one agent better off without
making the other agent worse off. Therefore, if through trading a point on the contract
curve is reached, no further trade is possible. Conversely, if trading has reached any point
not on the contract curve, further Pareto-improving trades are possible.
max uA (XA , YA )
XA , YA , XB , YB
113
7. General equilibrium and welfare
uA = X Y 1 and uB = X Y 1 .
(a) The contract curve can be derived by equating the MRS of the two agents:
YA YB
= .
1 XA 1 XB
YA Y YA
= .
1 XA 1 X XA
114
7.3. General equilibrium in an exchange economy
(1 ) Y XA
YA = .
XA ( ) + (1 )X
Y
YA = XA
X
which is obviously a straight line (a ray from the origin).
Figure 7.2 plots the contract curve for = 0.4 and various values of assuming
X = 50 and Y = 100.
Figure 7.3 shows the set of allocations that are possible to achieve by trading starting
from the initial allocation, E. These allocations give at least as much utility as the
endowment to either agent. In other words, each point in this set of allocations
(excluding E itself and the other crossing point between the indifference curves) Pareto
dominates E.
We can now introduce prices. The slope of the price line is PY /PX . Trading can take
place along any such price line. However, as Figure 7.4 shows, any arbitrary price line
cannot guarantee an equilibrium.
Figure 7.5 shows the property that the price line must satisfy in order to produce an
equilibrium. It must be tangent to the indifference curves of both agents. This can of
course happen only on the contract curve, since this also requires the indifference curves
themselves to be tangent to each other.
115
7. General equilibrium and welfare
Figure 7.3: E is the endowment point. The shaded area enclosed by the indifference curves
passing through E shows the allocations that are Pareto-improving compared to E, i.e.
this is the area of mutually-beneficial trades. The segment FG is the part of the contract
curve on which equilibrium trades lie.
Figure 7.4: The price line shown here is not consistent with equilibrium. Note that there
is excess demand for Y (the total demand for Y exceeds the total endowment of Y ) and
excess supply of X. So neither market is in equilibrium. Starting from this situation, the
price of X would fall and the price of Y would rise, so that the price line should become
steeper and steeper until the points L (where As utility is maximised given the prices)
and M (Bs utility is maximised given the prices) coincide and we get an equilibrium.
116
7.3. General equilibrium in an exchange economy
Figure 7.5: The equilibrium price line (the slope of the price line is the ratio of prices) is
tangent to the indifference curves of both A and B at the same point. Since the indifference
curves are therefore also tangent to each other, this point must be on the contract curve.
Here H is the equilibrium allocation.
Relative prices
Note that the price line and, therefore, the equilibrium allocation would remain
unchanged if all prices are multiplied by the same positive number. Essentially, in a
general equilibrium we can only determine relative prices. If you study economics at
Masters level, you will see that this is a consequence of Walras law which states that if
out of n markets, n 1 are in equilibrium, then so must be the remaining one. This
implies that there are only n 1 independent prices. In our context with two goods, if
the market for good X is in equilibrium, so must be the market for good Y . So there is
only one independent price. This is the relative price of good X, say. In other words, we
can only determine the ratio PX /PY and not the levels of PX and PY separately. So we
could declare PY to be equal to 1 and determine PX . (It is as if the price of good X is in
units of good Y .)
Example 7.2 Consider an economy with two goods, x and y, and two agents, A
and B. Suppose:
uA = x y 1 and uB = x y 1 .
Suppose the endowments are (Ax , Ay ) for A and (Bx , By ) for B. Derive the
equilibrium price ratio and the equilibrium allocation.
The budget constraint of A is:
px xA + py yA = px Ax + py Ay .
The budget constraint of B is:
px xB + py yB = px Bx + py By .
117
7. General equilibrium and welfare
Let MA = px Ax + py Ay and MB = px Bx + py By .
You should derive explicitly the demand functions by maximising utility for each
agent subject to the budget constraint of that agent. The demand functions are:
MA
xA =
px
(1 )MA
yA =
py
MB
xB =
px
(1 )MB
yB = .
py
xA + xB = Ax + Bx .
Solving:
p(Ay + By ) = (1 )Ax + (1 )Bx
which gives us:
(1 )Ax + (1 )Bx
p= .
Ay + By
We have derived the equilibrium price ratio py /px (i.e. the relative price of y).
Substituting this in the demand functions, we can obtain the equilibrium allocation.
For example:
xA = (Ax + pAy )
where p has been derived above. You can derive the rest of the allocation in this
manner.
118
7.4. Existence of equilibrium
7.5 Uniqueness
General equilibrium in an economy need not be unique. Figure 7.6 shows three
equilibrium allocations, F, G and H, starting from endowment E, supported by different
equilibrium price lines.
119
7. General equilibrium and welfare
The terms Pareto optimal, Pareto efficient, and efficient are used interchangeably.
Note that the notion of efficiency in economics says nothing about equity. Contract
curves include points where one consumer gets a very high utility while another gets a
very low utility. Later, when we study monopoly, you will see that if a monopolist can
extract the entire surplus, the market outcome is efficient, while if the monopolist
cannot do so the outcome is typically inefficient even though consumers do receive some
surplus. To take equity into account, we would need stronger notions involving social
justice criteria. However, except for a small set of topics where such notions are called
for, we do not worry about equity, and simply consider efficiency as the welfare measure.
The first theorem of welfare economics says that the competitive equilibrium is Pareto
efficient. In other words, the process of decentralised decision-making by agents and the
equilibrating force of market prices together ensure that there are no gains from trade
left unexploited. Starting from any equilibrium allocation, if we want to make any one
agent better off, we can do that only by making someone else worse off.
The first theorem of welfare economics applies in a very general setting, and is an
absolute blockbuster. This remarkable result is fundamental to our understanding of
markets.
The first thing it tells us is about the remarkable ability of the price mechanism to
eliminate inefficiencies. Even though markets are a conglomeration of a large number of
agents trying to maximise their own utility without paying attention to what others
might be doing, the price mechanism creates the right incentives for agents to exploit
every trading opportunity. Wherever any opportunity for efficiency-enhancing trade is
left, prices will be such that agents will have the incentive to exploit these
opportunities. This is the invisible hand at work the market price mechanism ensures
that each agent acting in pure self-interest results in a Pareto optimal outcome.
You should be able to show the result easily in a two-agent two-good exchange
economy. Looking at Edgeworth box diagrams, you should be able to see that it is
impossible to make anyone better off without making the other agent worse off. Let us
try to outline an informal proof. Suppose we have an equilibrium that is not Pareto
efficient. Then there must be some other allocation inside the Edgeworth box that
makes both agents better off. Since each agent is choosing the best point on the price
line (which is the budget line for both agents), to make agent A better off we must
choose an allocation that lies to the right of the price line, and to make B better off we
need to choose an allocation to the left of the price line. But there is no point common
to the set of points to the right and left of the price line. Therefore, it is impossible to
120
7.6. Welfare theorems
make both agents better off (or make one agent better off and the other left indifferent).
It follows that the equilibrium must be Pareto efficient.
From this it follows that a set of competitive markets functions very well using this
mechanism, and there is no case for efficiency-enhancing government intervention. The
invisible hand works well.
7.6.2 Implications
While this theorem is fundamental to our understanding of economics, you should also
be aware that the argument can be taken too far. The doctrine of laissez faire
advocates that the state should not intervene in markets. The doctrine has influenced
the economic stance of many conservative political parties across the world over the last
few centuries, who typically argue for a smaller state.
To see the flaw in this doctrine, you need to understand the concept of a market
failure. A market failure occurs when the equilibrium in a market involves some
deadweight loss, i.e. the equilibrium admits some degree of inefficiency. The theory of
general equilibrium that we described above assumes away such problems, but these are
relevant in many situations, as we describe below. Indeed, from this point onwards, the
rest of the course considers topics where such failures do arise. In such cases, we can
make a case for the visible hand of the government, intervening through policies to try
to reduce or eliminate the market inefficiency. Of course, if such failures arise only in
rare cases or only in markets that have a small role to play in the overall economy, we
would still conclude that there is little role for policy and the laissez faire type
arguments are largely correct.
In fact, market failures are a central feature of large and important parts of the
economy. Let us try to understand why. For market failures to arise, there must be
some factor impeding the functioning of competitive markets or the markets are not
competitive to start with.
Consider, first, the problem of public goods. These are goods that we consume together
and consumption is neither rival nor excludable. As you should know, in this case
competitive markets work poorly. If we try to use private provision through competitive
markets, it results in under-provision. However, public goods such as national defence,
roads, basic public health, basic education, defence against natural catastrophes etc. are
important for any economy. Provision of such goods using receipts from taxes is a
crucial role for the government in the economy.
Second, in some situations one agents actions affect other agents directly. But the agent
does not take this externality into account when acting. The result is that the market
outcome is inefficient. The individually-optimal levels of negative externality-generating
activities are higher than the socially optimal level, and that for positive
externality-generating activities are lower than the socially optimal level. The invisible
hand does not work well in such cases individual optimisation does not lead to social
optimality. Are these problems important? Indeed, some of our biggest problems are
caused by negative externalities. These include global warming, pollution, congestion,
urban noise, and so on. In all these cases, government policy has a crucial role to play
to bring individual incentives in line with the social optimum.
Third, the market itself might not be made up of a large number of price-taking firms.
121
7. General equilibrium and welfare
If only a few firms (or just one firm) are present, they act strategically, taking into
account their own impact on the market price as well as the reactions of other firms. In
this case the market outcome is typically inefficient. Of course, before you can think of
such situations as not conforming to the first welfare theorem, you must investigate the
source of market power enjoyed by firms. If such power is intrinsic to the nature of the
products or technology, then indeed the first welfare theorem does not apply. In these
cases, regulation plays an important role in reducing the deadweight loss. If, on the
other hand, market power results from entry barriers created exogenously by policy
itself, then it is not a fundamental violation of the theorem. In such cases, you should
argue that the policies that create a distortion in the first place should be removed so
that competition and with it efficiency is restored.
Fourth, situations involving asymmetric information often result in market failures.
Such information problems are endemic to markets for information-sensitive financial
assets, insurance markets, as well as credit markets. In these cases, the role of policy is
less clear since policymakers themselves are likely to face the same information
problems. However, some of the problems can be reduced by building the right
institutions. Examples of such institutions include rules for stock trading that seek to
minimise information problems, adding deductibles or forcing people to belong to a
large insurance pool to reduce the information problems faced by an insurer, building a
long-term relationship with a creditor to reduce information problems in lending. The
primary role of government policy in such cases is to try to facilitate such institutions.
To summarise, the first welfare theorem is one of the most (if not the most) important
results in economics. It provides a benchmark against which to judge the social welfare
from any market outcome. There are various instances where market outcomes do
admit inefficiency, in which case there is a role for policy.
Finally, note that policy itself can be subject to problems. Governments themselves
might sometimes serve special interests, and in such cases interventions need not be
efficiency-enhancing. Along with market failures, you should also keep in mind the
possibility of such government failures.
The first welfare theorem says that any general equilibrium allocation is efficient. The
second theorem makes a sort of reverse statement. It says that if all agents have convex
preferences, any efficient allocation can be obtained as an equilibrium if we can suitably
change the initial endowment point.
This is easy to see in a two-agent two-good economy. As you know, efficient allocations
are points on the contract curve. Let us pick an arbitrary point H on the contract curve.
This is a point of tangency between indifference curves of the two agents, i.e. the
indifference curves have the same slope at H. Consider a price line that also has this
same slope passing through H. This is shown in Figure 7.7 below. If agents are faced
with this price line, which point would they choose to consume at? So long as
preferences are convex (diminishing MRS), both agents would choose H as the
utility-maximising point. Therefore, if we can change the endowment from the point E
(as shown in Figure 7.7) to H, the fact that H is a point of exchange equilibrium implies
that the market process would not move away from H. To summarise, starting from any
122
7.6. Welfare theorems
point H on the contract curve, we have shown that there is a price line which supports
H as a competitive equilibrium if we change the initial allocation suitably.
7.6.4 Implications
What does this mean? To see this, suppose we want a particular distribution of welfare
among agents. Can we achieve this distribution without wasting resources? That is, can
we reallocate resources in an efficient manner? The theorem shows that the answer to
this question is yes.
We noted that a key assumption needed for the theorem is that preferences are convex.
Figure 7.8 below shows why this is the case. As the figure shows, given non-convex
preferences, it might not be possible to find a price line to support a Pareto efficient
allocation as a competitive equilibrium. H is a Pareto efficient allocation, but on the
price line that has the same slope as the common slope of the indifference curves at H,
the optimal choice of agent A is not H, but G. Therefore, H is a Pareto efficient
allocation but it cannot be supported as an equilibrium. In other words, if the
government wanted to achieve a redistribution to the point H and redistributed
endowments suitably, the economy would not stay at point H as the market process
would move it away.
Next, consider the question of the redistribution of resources. How do we do this? In an
ideal world we would use lump-sum taxes and transfers. Indeed, if we could use
lump-sum taxes quite easily in practice, this theorem would spell out the right type of
policy to achieve different welfare distributions.
However, lump-sum taxes are difficult to impose in practice. A lump-sum tax is a tax
123
7. General equilibrium and welfare
imposed on something that an agent cannot change. A tax on labour endowment or any
type of earned income is distortionary any such tax would lead agents to change how
much labour they supply or how much effort they invest in earning. Furthermore, even
if you could find a lump-sum tax (such as a poll tax), people might not fully trust the
government to make use of the proceeds in the right manner. This might make such a
tax politically unpalatable. These problems with lump-sum taxes limit the force of the
second welfare theorem as a tool for the real world.
7.7 Production
The discussion so far has been in the context of an exchange economy. N&S Sections
10.3 and 10.4 introduce a simple model of general equilibrium with production. You
should read this carefully, and understand the informal proof of the first welfare
theorem (in Section 10.4) in such an economy.
Consider an economy with two goods (x and y) and one agent in society. This sort of an
economy is often referred to as a Robinson Crusoe economy. In an economy with many
identical agents you can simply solve the problem of a representative agent. This is
why analysing single-agent economies makes sense. Single-agent economies are used
often by macroeconomists in building a micro-foundation for macroeconomic models.
The utility function of Robinson is given by u(x, y).
The production possibility frontier shows the combinations of x and y that can be
produced using the resources available in the economy. Figure 7.9 shows such a frontier.
Note that the slope of the frontier at any point shows how much y we need to give up to
124
7.7. Production
obtain one extra unit of x. This is called the marginal rate of transformation (MRT).
Typically, the MRT is increasing as we raise x (and lower y). One way to interpret this
is as follows. Some resources in the economy are better at producing x while other
resources are better at producing y (as an example, think of the goods as guns and
butter and the inputs as iron and milk). At low values of x, extra x can be added by
giving up very little y because we are releasing resources that are very good at
producing x but not so good at producing y. As we keep increasing the production of x,
the amount of y we need to give up grows larger and larger as we release resources that
are less and less useful in producing x.
Figure 7.9: A production possibility frontier. This shows the combinations of x and y that
can be produced using the resources available in an economy. All points on the frontier
are technically efficient in the sense that all resources are employed in production.
T /x
.
T /y
T /x
MRT = .
T /y
We are now ready to consider the optimisation problem faced by Robinson. Robinsons
problem is to maximise utility subject to the production constraint:
125
7. General equilibrium and welfare
Example 7.3 There are two goods, X and Y , in an economy. The production
function is given by:
X 2 + Y 2 = 200.
Each individual has the utility function:
Calculate the general equilibrium price ratio and the amounts of X and Y produced
in equilibrium.
The MRT is given by X/Y . The profit-maximising condition is:
X pX
= .
Y pY
Y pX
= .
X pY
Since MRS = MRT, we have:
X Y
=
Y X
or X = Y . Using this in the production function, X = Y = 10. Furthermore, the
equilibrium price ratio is pX /pY = 1.
126
7.9. Test your knowledge and understanding
construct an Edgeworth box and depict the contract curve, the set of possible
equilibrium exchanges given an endowment point, and depict the equilibrium price
ratio and allocation
derive algebraically the contract curve, the equilibrium price ratio and allocation
state and provide an informal proof outline for the first welfare theorem in the
context of an exchange economy with two agents and two goods as well as in the
context of an economy with production
discuss the policy implications of the first welfare theorem and understand the
situations in which the theorem might not apply
state and provide an informal proof outline for the second welfare theorem in the
context of an exchange economy with two agents and two goods
discuss the policy implications of the second welfare theorem and understand the
situations in which the theorem might not apply
2. Consider a pure exchange economy with two agents, Ann and Bob, who consume
only two goods, x and y. In the economy there are 4 units of x and 2 units of y.
Ann is endowed with 1 unit of x and 2 units of y, and Bob is endowed with 3 units
of x and 0 units of y. Their utility functions are:
uA (xA , yA ) = xA yA and uB (xB , yB ) = xB + 2yB .
(a) Draw the Edgeworth box, showing the endowment point, and the indifference
curves.
(b) Derive the contract curve.
(c) Derive the demand functions of the two agents.
(d) Calculate the exchange equilibrium price ratio.
(e) Calculate the exchange equilibrium allocation.
127
7. General equilibrium and welfare
128
Chapter 8
Monopoly
8.1 Introduction
Previously, we analysed competitive markets. In these markets firms are price-takers,
and the market price is determined by the interaction of industry supply and market
demand, a process over which no single firm has any control. In this chapter we develop
the model of the opposite market structure one where the supply side consists of a
single firm. The other dimension in which the analysis differs is that a monopolist is
relatively unconstrained in choosing its pricing strategy. Instead of choosing a single
price, it can choose different prices across different identifiable groups or markets, and
also charge a fee to give a consumer access to purchasing. We analyse the consequences
of these pricing schemes. Finally, certain markets naturally lend themselves to a
monopoly structure. Such natural monopolies also present unique regulatory challenges.
We analyse the problem of natural monopolies and explain the regulatory problems.
derive the marginal revenue curve associated with any demand curve
explain the relationship between marginal revenue and price elasticity of demand as
well as Lerners index
analyse the structure of natural monopoly and the regulatory problem in this case.
129
8. Monopoly
where < 0. To make this fact explicit, it might be useful to rewrite this using the
absolute value of price elasticity:
1
MR = P 1 .
||
Figure 8.1 below shows the relationship between elasticity and MR for the linear
demand curve P = 10 Q.
130
8.4. Profit maximisation and deadweight loss
Figure 8.1: Elasticity and marginal revenue for a linear demand curve.
131
8. Monopoly
Consider the linear demand curve shown in Figure 8.1. Note that at MR = 0 we have
|| = 1. Therefore, for any positive MC, the optimal output-price combination occurs at
the left of the unit elasticity point. At all such points demand is elastic. It follows that
for a linear demand curve, the monopolist always produces at the elastic part of the
curve.
Regulation
Since a monopoly gives rise to an efficiency loss, the question of regulation becomes
relevant. You should be able to understand and analyse regulatory strategies to reduce
monopoly deadweight loss. One way to regulate a monopoly is to specify the price that
a monopolist can charge. By restricting price to the level P = MC(QC ), the regulator
can eliminate deadweight loss. Another policy is to encourage new entry in the industry
so that there is increased competition, which reduces the price towards marginal cost
and reduces deadweight loss.
1
MC = P 1+
which implies:
P MC 1 1
= = .
P ||
The left-hand side is known as the Lerner index, which is a measure of the monopoly
power of a firm. For a competitive firm, P = MC so that the Lerner index is 0. If MC is
0 or P is infinity, the Lerner index is 1, which is its maximum value. For a monopolist,
it lies between 0 and 1 and rises as demand becomes less elastic.
A monopoly firm need not charge a single price to all customers. There are a variety of
ways in which it can practise price discrimination. N&S Section 11.4 discusses perfect
price discrimination, market separation, two-part tariffs, as well as a variety of
non-linear pricing schemes. A durable good monopolist faces competition with its own
future self, which affects pricing. See N&S for details.
In the activity below we develop an example of market separation as well as a two-part
tariff.
132
8.5. Price discrimination
Example 8.1 A monopolist has two customers with the following demand
functions:
Q1 = 10 P1 (Demand of customer 1)
Q2 = 12 P2 (Demand of customer 2).
Here Pi is the price charged to customer i, i {1, 2}. The monopolist has a constant
marginal cost of 1, and no fixed costs.
(a) Suppose the monopolist can differentiate between customers, and customers
cannot trade between themselves, allowing the monopolist to engage in
third-degree price discrimination. What is the price charged to each consumer?
(b) Now suppose the monopolist cannot differentiate between customers and must
charge them the same price. Calculate the monopolists optimal single price P
as well as the quantity sold to each customer.
(c) Suppose the monopolist cannot differentiate between customers. However, in
addition to a per-unit price P , the monopolist can also charge a fixed fee F . A
customer must pay this fee irrespective of the quantity purchased when a
positive amount is purchased. Derive the monopolists optimal price and fee.
(a) With third-degree price discrimination, the monopolist sets MR = MC for each
customer. You should be able to show that Q1 = 4, P1 = 6, Q2 = 6 and P2 = 8.
The monopolists profit is 52 and the total surplus is 78.
(b) In this case the monopolist sets a single price to maximise profit given total
demand, which is Q = 24 2P . You should be able to show that in this case
Q = 10 and P = 7. The profit is 50 and the total surplus is 79.
(c) At any price P < 10, if the fixed fee is set equal to the surplus of consumer 1
(given by (10 P )2 /2), the total demand is Q1 + Q2 = 24 2P , and the
revenue is:
(10 P )2
2 + (P 2)(24 2P ).
2
Maximising, the first-order condition is:
2(10 P ) + 24 2P 2(P 2) = 0
which implies P = 4. The fixed fee is then 18, and the revenue is 36 + 32 = 68.
If the fixed fee is set equal to the surplus of consumer 2, then consumer 1 does
not buy. The monopolist is just dealing with consumer 2. In this case, since the
monopolist can set a fixed fee to extract all consumer surplus, it is best to
maximise the surplus by setting price equal to marginal cost, and then
extracting the full surplus.
Therefore, the optimal price is P = 2 and the fixed fee is:
(14 2)2
F = = 72.
2
The profit of the monopolist is 72, which is higher than the previous cases. The
total surplus in this case is also 72, which is lower than the total surplus in (a)
and (b).
133
8. Monopoly
Figure 8.3: A natural monopoly with low fixed MC and falling AC. Note that the
monopolist would incur a loss if under marginal cost pricing. If regulated price is set
at AC, the shaded area is the deadweight loss. This is, of course, much lower than the
deadweight loss without regulation.
A standard monopolist can be regulated by setting the regulated price equal to marginal
cost at the efficient quantity. However, such an approach runs into difficulty here. Since
the AC is falling, the MC is below the AC, implying that if the regulated price is set
equal to the MC, the monopolist would incur a loss and quit rather than produce. As
you can see from Figure 8.3, pricing at the MC implies a loss for the monopolist. The
government then must give a lump-sum subsidy to the monopolist to cover the losses.
However, governments typically do not want to be in a position to subsidise a
monopolist forever. In that case, the government can regulate the monopolist by setting
the price equal to average cost. This implies that the deadweight loss is not fully
eliminated, as shown in Figure 8.3. Alternatively, the monopolist might be allowed to
charge a higher price to some consumers and a lower price to others. A multi-product
monopolist might be allowed to make a profit on some products while making a loss on
134
8.7. A reminder of your learning outcomes
others through marginal cost pricing, while making a zero profit overall.
derive the marginal revenue curve associated with any demand curve
explain the relationship between marginal revenue and price elasticity of demand as
well as Lerners index
analyse the structure of natural monopoly and the regulatory problem in this case.
2. Suppose the cost function of a monopolist is C(Q) = 10Q 2Q2 + 3Q3 . Derive the
supply function of the monopolist.
3. Suppose a firm has cost function ln Q + 1, where Q > 1. Suppose the government
wants to regulate the firm and set the regulated price equal to the MC. Propose a
policy that achieves this goal.
P1 = 24 Q1 and P2 = 12 0.5Q2 .
135
8. Monopoly
(d) Now suppose regulation forces the monopolist to charge a single price.
Calculate the profit and deadweight loss in this case. Would both markets be
served? Does the regulation reduce the deadweight loss?
5. Suppose a monopolist can sell to two different customers with demand curves:
Q1 = 24 P1
Q2 = 24 2P2 .
The monopolist has a constant marginal cost of 6. Calculate the optimal two-part
tariff.
136
Chapter 9
Oligopoly
9.1 Introduction
When an industry comprises of a few firms, the nature of competition is very different
from that under perfect competition. In trying to maximise profit, a firm must not only
take into account the impact of its own supply on the market price, but must also
account for the strategic responses of other firms. We have already analysed this type of
interaction when we studied game theory, and identified Nash equilibrium as the central
solution concept. Recall that a Nash equilibrium is a mutual best response: given what
others are doing, each firm must play a best response. We now use this solution concept
to analyse the outcome under quantity and price competition.
137
9. Oligopoly
This is the best response function of firm 1 (also called the reaction function of firm
1).
After optimisation, we impose symmetry: q1 = q2 = q. This implies:
a c bq
q= .
2b
138
9.3. Cournot competition
Solving:
ac
q = .
3b
The equilibrium price is:
ac 3a 2a 2c 1 2
p = a 2b = + = a + c.
3b 3 3 3 3 3
Let D denote the profit of either firm under duopoly. This is given by (p c)q .
Therefore:
(a c)2
D 1 2 ac (a c) a c
= a+ cc = = .
3 3 3b 3 3b 9b
The best response functions and the CournotNash equilibrium are shown in Figure 9.1
below.
Here we imposed symmetry after optimisation. However, this method does not work if
the firms are not symmetric. Suppose the marginal costs are c1 and c2 . In this case, you
must derive the two best response functions, and then solve two simultaneous equations
for q1 and q2 .
The best response functions are:
a c1 q 2
q1 =
2b 2
and:
a c2 q 1
q2 = .
2b 2
Using the value of q2 in the best response of firm 1, we have:
a c1 a c2 q 1
q1 = + .
2b 4b 4
139
9. Oligopoly
Simplifying:
3q1 a 2c1 + c2
= .
4 4b
This gives us:
a 2c1 + c2
q1 = .
3b
Using this in the best response of firm 2, we have:
a c2 a 2c1 + c2 a 2c2 + c1
q2 = = .
2b 6b 3b
Note that an increase in c1 reduces the equilibrium output of firm 1, but raises that of
firm 2. Graphically, the best response function of firm 1 moves inwards, so that the
Nash equilibrium point slides up along the best response function of firm 2.
9.3.1 Collusion
We now return to the symmetric firms model. Suppose the firms could collude by some
means. Let us calculate the output, price and profit in this case.
Let QM = q1 + q2 be the collusive output. The joint profit is:
M (QM ) = (a bQM c)QM .
The first-order condition for a maximum is M /QM = 0 which implies:
ac
QM = .
2b
Note that QM is lower than total output under Cournot. Figure 9.1 in the previous
section shows the collusive output line, which joins the monopoly outputs of the two
firms.
The market price is:
ac a+c
pM = a b = .
2b 2
The total profit is:
(a c)2
M a+c ac
= c = > 2D .
2 2b 4b
The monopoly outcome cannot be sustained in a one-shot game the best response of
firm 1 to any q2 is to produce more than the collusive output. This is clear from Figure
9.1 in the previous section.
However, in an infinitely-repeated game, the collusive output can be sustained as the
outcome of an equilibrium if the players are sufficiently patient. You already know this
result from the theory of repeated games. Let us do some calculations for this specific
case to reinforce the point.
Suppose two symmetric firms engage in an infinitely-repeated game where in each stage
they play a Cournot game. Suppose each firm produces QM /2 under collusion. Each
firm has a discount factor , where 0 < < 1.
Now, suppose each firm adopts the following trigger strategy. Start by producing
QM /2 at t = 0. In any subsequent period t > 1, do as follows:
140
9.3. Cournot competition
p = a b(q1 + + qn ) = a b(qi + Qi ).
Note that the second mover produces half the output of the first mover. Compared to
the CournotNash outputs, the first mover produces more and, given this, the second
movers best response is to produce a lower output. The total output is:
3 (a c)
Q =
4 b
which is greater than the total output under Cournot competition. Consequently, the
market price:
1 3
p = a + c
4 4
is lower than that under Cournot competition. The profit of the Stackelberg leader (firm
1 in this case) is:
ac (a c)2
SL = (p c) =
2b 8b
which is higher than the profit under Cournot competition. This is, of course,
unsurprising. The Stackelberg leader could choose any point on the followers best
response function. If the leader chooses the CournotNash quantity, the followers best
response would also be to choose the CournotNash quantity. However, the leader can
do even better given the freedom to choose any point on the followers best response
function. Therefore, the Stackelberg leader must earn a greater profit compared to that
under Cournot competition. The follower, on the other hand, earns a lower profit given
by:
S (a c)2
F = .
16b
9.4.1 Collusion
If firms play an infinitely-repeated game, and are patient enough, we know from game
theory that they can sustain a high collusive price.
q 1 = P 1 + P2
143
9. Oligopoly
and:
q 2 = P2 + P1 .
Why is this a model of differentiated products? Note that the demand for the product
of firm 1, say, does not vanish if P1 > P2 . In this case, the demand for firm 2s product
will be higher, but firm 2 will not take over the entire market. The underlying idea here
is that firms have different products that are substitutable, but not completely so. This
could be because the products serve the same basic purpose but have slightly different
features so that some people like firm 1s product better while others prefer firm 2s
product. As firm 1s price increases, some of the customers (who only consider firm 1s
product to be marginally better than firm 2s) will switch to buying firm 2s product
lowering demand for firm 1 and raising that for firm 2. Only at very high levels of P1 (in
the above example, this would require P1 > P2 + ) would q1 go to zero.
Let us now calculate the Nash equilibrium. The profit of firm 1 is given by:
1 = (P1 c)( P1 + P2 ).
( P1 + P2 ) (P1 c) = 0.
P = + c.
Each firm charging a price of + c is the Nash equilibrium. The profit of each firm is 2 .
Figure 9.2 shows the shape of the reaction functions. Note that the reaction functions
slope upwards in this case because an increase in P1 , say, allows firm 2 to raise its price
optimally.
144
9.5. Bertrand competition with product differentiation
Figure 9.2: The figure plots the best response functions assuming the values = 3 and
c = 1. In this case the Nash equilibrium is P1 = P2 = 4.
Maximising, we get:
3
P1 = c + .
2
5
P2 = c + .
4
92
3 3 5
1 = (P1 c)( P1 + P2 ) = + =
2 2 4 8
while:
252
2 = .
16
145
9. Oligopoly
2. Consider the Cournot duopoly in which two firms, 1 and 2, produce a homogeneous
good, and have total costs C1 (q) = 3q and C2 (q) = 2q, respectively. The market
demand is Q = 63 P . Find the equilibrium quantities, price and profits. Which
firm produces more? Which firm earns more?
3. The (inverse) market demand for a good is P = a bQ, and all firms have a
constant marginal cost c.
(a) Derive the perfect competition market quantity, price, industry profit,
consumer surplus and total surplus.
(b) Derive the monopoly market quantity, price, industry profit, consumer surplus
and total surplus.
(c) Assuming that there are two firms that compete by setting quantities
simultaneously, derive the Cournot duopoly market quantity, price, industry
profit, consumer surplus and total surplus.
(d) What can you say about the different market structures if you compare them?
146
9.7. Test your knowledge and understanding
4. Consider two firms that produce a homogeneous good and compete by setting
quantities. Suppose that firm 1 has a cost function C1 (q) = 100 + 5q, and that firm
2 has a cost function C2 (q) = 300 + 5q. The market demand is Q = 105 P .
(a) Assuming that they choose their quantities simultaneously, determine their
equilibrium quantities, price and profits.
(b) Assuming that firm 1 chooses its quantity first, determine the equilibrium
quantities, price and profits.
(c) Assuming that firm 2 chooses its quantity first, determine the equilibrium
quantities, price and profits.
5. Suppose two firms (firm 1 and firm 2) sell differentiated products and compete by
setting prices. The demand functions are:
P2
q 1 = 2 P1 +
2
and:
P1
q 2 = 2 P2 + .
2
Firms have a zero cost of production.
(a) Find the Nash equilibrium in the simultaneous-move game. Also find the
quantities sold by each firm and the profit of each firm.
(b) Find the subgame perfect Nash equilibrium if firm 1 moves before firm 2. Also
find the quantities sold by each firm and the profit of each firm.
147
9. Oligopoly
148
Chapter 10
Asymmetric information: adverse
selection
10.1 Introduction
For a wide range of products (for example, things you buy from a grocery store) the
markets are large and decentralised, and exchange is anonymous. There are, however,
important exceptions. Consider the market for used cars. It is difficult to tell how good
or bad a used car is unless you use it for some time. The seller, on the other hand,
knows how good the car is.
In such situations, you would care about the identity of the seller. In fact, in such cases
anonymous exchange is often impossible, as that would lead to a breakdown of the
market. The customers are willing to pay a price based on the average quality. However,
this price is not acceptable to the better-than-average quality car owner. This leads the
owners of better-than-average cars to leave the market. Knowing this, customers would
revise average quality downwards and consequently be willing to pay a lower price. But
the same phenomenon then applies again, leading to further falls in average quality and
willingness to pay, and so on.
The problem is called the lemons problem or adverse selection.
The same problem arises in selling insurance. Consider the problem of selling health
insurance to older people who know much more about their state of health than an
insurance company. If the insurance company sets a price based on the average medical
expenditure, this would typically lead to a market price (in this case, the insurance
premium) that is too high for the healthiest individuals and only the people who feel
that they are very likely to claim the insurance would buy insurance. This would lead,
as in the case of used cars above, the market to fail.
Banks (or other lenders) face the same problem in credit markets. If a high rate is set in
expectation of high average risk, safer borrowers will exit the market and only the
riskiest borrowers will want to borrow. Again, as in the market for lemons, this causes a
market failure.
A different sort of problem arises when you, as the manager of a company, are trying to
hire some salespeople. As the job requires a door-to-door sales campaign, you cannot
supervise them directly. And if the workers choose not to work very hard, they can
always blame it on the mood of the customers. If you pay them a fixed wage
independent of the sales they achieve, they are unlikely to work hard.
This problem is known as moral hazard.
Note the difference between adverse selection and moral hazard. In the first case, the
asymmetry in information exists before you enter into the exchange (buy a used car, or
149
10. Asymmetric information: adverse selection
sell insurance). In the latter, however, the asymmetry in information arises after the
wage contract is signed.
This is why another name for adverse selection is hidden information and another
name for moral hazard is hidden action.
In this chapter and the next, we will consider certain adverse selection and moral
hazard problems and remedies in some detail.
150
10.2. Overview of the chapter
151
10. Asymmetric information: adverse selection
N&S briefly discusses this problem in the last part of Section 15.4. Perloff Section 18.3
contains a more detailed informal discussion of the model and policy implications. Here
we present an example with two qualities.
This is based on the famous article that gave rise to the field of asymmetric information:
George A. Akerlof (1970) The market for lemons: quality uncertainty and the market
mechanism, Quarterly Journal of Economics, volume 84, pages 488500.
Suppose there are four kinds of cars there are new cars and old cars, and in each of
these two categories there may be good cars and bad cars. Buyers of new cars purchase
them without knowing whether they are good or bad. After using a car for some time,
the owners can find out whether the car is good. This leads to an asymmetry: sellers of
used cars have more information about the quality of the specific car they offer for sale.
Consider a market for used cars. There are some low quality cars and some high quality
cars. Potential sellers have a car each, and there are many more buyers than possible
sellers in the market. A high quality car rarely breaks down. A low quality car provides
a poorer ride quality over longer journeys and also breaks down with higher probability
compared to high quality cars.
A seller values a high quality car at 1,000 and a low quality car at 300. A buyer values a
high quality car at 1,300 and a low quality car at 400. All agents are risk-neutral.
Assume that the sellers get the entire surplus from trade.
Market failure
Suppose quality is observable to sellers but not to buyers. Buyers only know that a
fraction of 1/2 of the cars in the market are high quality and the rest are low quality.
Let us analyse the market outcome in this case.
The average value of buyers is 850. If all cars are in the market, this is the most buyers
would pay. But at this price high quality cars withdraw. So only low quality cars would
sell. Knowing this, buyers would be willing to pay at most 400. Assuming sellers get the
entire surplus, the market price is 400 and only low quality cars exchange hands. The
market outcome is not efficient since the gains from trading high quality cars are not
exploited.
Suppose low quality cars break down with probability 0.8, and high quality cars break
down with probability 0.1. Suppose the sellers have an option of promising a refund of
R if the car breaks down.
Let us see if such a policy can lead to a separating equilibrium. In a separating
equilibrium, it must be the case that sellers of high quality cars sell at 1,300 and
promise a refund of R in the event of a breakdown, while sellers of low quality cars sell
152
10.5. A model of price discrimination
which implies R 6 3000. This constraint is called the participation constraint (PC) of
high quality sellers.
Also, the low quality sellers must prefer to sell at 400 without a guarantee rather than
imitate the high quality sellers (offer a refund of R and sell at 1,300). This implies:
which implies R > 1125. This is called the incentive compatibility constraint (IC) of low
quality sellers.
The value of R must be such that the PC of high quality sellers and the IC of low
quality sellers are both satisfied. This implies that the range of values of R for which
the separating equilibrium holds is 1125 6 R 6 3000.
Forced refunds
Suppose the government decides to force each seller to offer a full refund if the car sold
by the seller breaks down. How does this change the market outcome? Is the market
outcome efficient?
Note that the maximum possible price in the market is 1,300. But even at a price of
1,300, 0.2 1300 = 260 is below 300. It follows that low quality sellers would not
participate in the market, and only high quality sellers would remain and sell at 1,300.
But this is not efficient as the gain from trade of low quality cars is not realised.
In other words, forcing sellers to issue a refund destroys the role of a refund policy in
separating the two qualities. Since all sellers must issue a refund if the car breaks down,
low quality sellers prefer to withdraw from the market, making it impossible to exploit
gains from trading low quality cars. Policy choices that are not sensitive to the role of
an instrument in providing information to uninformed agents do not necessarily
promote efficiency.
153
10. Asymmetric information: adverse selection
(t, q) = t q 2
u(t, q, ) = q t.
Here is a parameter that reflects how much the buyer cares about quality. is usually
referred to as the buyers type. This is the buyers private information.
In this model we suppose can take two values, 1 and 2 , where 2 > 1 .
A contract is a quality and price pair (q, t) offered by the seller. We assume, for
simplicity, that the seller has all the bargaining power.
The buyer gets 0 utility if they do not buy. Therefore, any contract must give the buyer
at least 0. This is called the participation constraint (PC) of the buyer.
The indifference curve of a buyer of type i , i {1, 2} is given by:
i q t = constant.
t q 2 = constant.
The slope of the iso-profit curve at quality q is 2q. Note that the slope does not depend
on t, only q. So if we fix any q and change t, the slope does not change. In other words,
all iso-profit curves have the same slope at points along any vertical line (at all values of
t). Figure 10.1 shows the indifference map of the buyer of type 2 and that of the seller.
The arrows show the direction of improvement.
Figure 10.1: The indifference map of a buyer of type 2 and that of the seller.
154
10.5. A model of price discrimination
However, since the seller has all the bargaining power, there is no reason to give the
buyer any more than 0, and thus the constraint holds with equality. Thus the seller
solves:
max ti qi2 subject to i qi ti = 0
qi , ti
Thus qi = i /2, and ti = i qi = i2 /2. Figure 10.2 below shows the optimal contracts
under full information. The optimal contract for type i is obtained at the point of
tangency between the iso-profit curve and the indifference curve of type i at the
reservation utility level (here the reservation utility is 0).
155
10. Asymmetric information: adverse selection
However, type 2 does want to take the contract meant for type 1 . From Figure 10.2,
you should be able to see that the indifference curve of type 2 passing through (q1 , t1 )
is associated with a higher utility compared to the original indifference curve of 2 .
You can also see this easily using algebra. Type 2 gets a 0 payoff from (q2 , t2 ). But the
utility from (q1 , t1 ) is:
1 1
2 q1 t1 = 2 1 12 = (2 1 ) > 0.
2 2
Alternatively, note that 2 q1 t1 > 1 q1 t1 = 0. Therefore, the full information
contracts are not incentive compatible for type 2 .
Indeed, given any choice of (q1 , t1 ), incentive compatibility for type 2 requires that
(q2 , t2 ) must be on the indifference curve of type 2 passing through (q1 , t1 ). This is
shown in Figure 10.3 below. Note that for any (q2 , t2 ) on the solid indifference curve for
type 2 , that type must be indifferent between accepting (q2 , t2 ) and (q1 , t1 ). If (q2 , t2 ) is
anywhere above this line, incentive compatibility does not hold, and type 2 would
rather choose (q1 , t1 ). If (q2 , t2 ) is anywhere below this line, type 2 would strictly prefer
it, but then the seller is not choosing optimally they could raise t2 (say) and still
satisfy incentive compatibility.
Which point should the seller choose on the solid indifference curve for type 2 ? Where
the sellers iso-profit curve is tangent to this line. The line has a slope of 2 , while the
slope of the iso-profit curve is 2q. Therefore, the optimal q2 equals 2 /2, which is the
same as q2 under full information.
So the only thing that changes is the price t2 . As Figure 10.4 below shows, given any
choice of (q1 , t1 ) on the participation line of type 1 , and given q2 = q2 = 2 /2, t2 must
be lowered to satisfy incentive compatibility for type 2 .
156
10.5. A model of price discrimination
Under full information, the contracts for the two types can be chosen independently,
but this is no longer possible under asymmetric information. Now, for any choice of
(q1 , t1 ), t2 has to be adjusted down to prevent type 2 from taking the contract meant
for type 1 . In the process, type 2 gets a surplus over its participation level (which is
0). This surplus is called the information rent of type 2 , shown in Figure 10.4.
In terms of algebra, the seller must offer contracts (q1 , t1 ) and (q2 , t2 ) such that:
1 q1 t1 = 0
q2 = q2 = 2 /2.
Since q2 is known, given any choice of q1 , the two constraints (PC1 ) and (IC2 ) above
determine t1 and t2 , respectively. How does the seller choose q1 ? We do not solve this
problem here, but the seller must consider the proportion of the two types in the
population and determine an optimal value. The exact q1 can be determined as a
function of the proportion and 1 and 2 , but this is not important for us here.
The lesson that you should take away from this is the difference that asymmetric
information makes. Under full information, each contract can be determined separately
157
10. Asymmetric information: adverse selection
Separating equilibrium
B > c` y .
Furthermore, low types should not want to send the same signal, so we must have:
B 6 ch y .
(We assume that when indifferent, types behave benignly, i.e. high types send and low
types do not send the signal. We could, of course, not assume this and require instead
the inequalities to be strict.) Thus, for a separating equilibrium, we must be able to
choose a y such that B/ch 6 y 6 B/c` .
158
10.7. A reminder of your learning outcomes
Pooling equilibrium
A second type of equilibrium is a pooling equilibrium. This has some subtleties which
you will appreciate more when you study game theory at a more advanced level. But let
us give a preliminary idea here.
Consider a pooling equilibrium in which both types of workers acquire an education.
Suppose the firm requires yb years of education and suppose all types acquire this much
education. Then the firm only knows that all types are in the market and pays an
average wage w, reflecting the average productivity.
Acquiring more education does not help, so there is no such deviation possible. But is it
possible for a type to deviate to lower than yb years of education? Clearly, if anyone
would want to deviate, it is the low type for whom the cost of education is higher. But
in equilibrium the firm expects all types of workers to get yb years of education. If some
worker deviates to no education, what would the firm believe about this workers type?
This is a rather complex question: what is a reasonable belief when you see something
that you are not expecting to see in equilibrium? Game theorists have devoted a lot of
time to answering this question, and have proposed many ways to try to deal with this
subtle issue, but there is no definitive answer. Anyway, here let us suppose that any
deviation to fewer than yb years of education is met with the worst belief about the
deviator: the firm believes this worker is a low type and would pay w` . In this case,
deviation is not beneficial if:
wn ch yb > w` n.
b = (w w` )n. The above condition can be written as:
Let B
B
b > ch yb.
If this holds, we can have a pooling equilibrium where all types acquire yb years of
education.
159
10. Asymmetric information: adverse selection
2. Consider a second-hand car market with two kinds of cars, type A which are
completely reliable, and type B which break down with probability 1/2. There are
20 car owners, 10 with each kind of car, and 20 potential buyers. The car owners
and buyers value a car at 1,000 and 1,500, respectively, if it works and both owners
and buyers value a non-working car at zero. Both buyers and owners are assumed
to be price-takers and risk-neutral. Finally, each seller is aware of the type that her
car belongs to, but this information is not available to buyers.
(a) What values do buyers and sellers place on type A and type B cars?
(b) How many cars, and of what type(s) are supplied at each price?
(c) What is the demand for cars at each price?
(d) How many equilibria are there in the market?
160
Chapter 11
Asymmetric information: moral
hazard
11.1 Introduction
Moral hazard arises when the behaviour of an individual changes as a result of entering
into a contract. The owner of a bike that is fully insured against theft might be more
likely to forget to lock it. The true level of care (or effort) is hidden from the insurer.
The asymmetry in information here is described as hidden action, as distinct from that
of hidden information or adverse selection.
The issue can be studied more generally as a principal-agent problem. Many economic
transactions have the feature that unobservable actions of one individual have direct
consequences for another, and the affected party may seek to influence behaviour
through a contract with the right incentives. In the above example, the insurance
company (the principal) is affected by the unobservable carelessness of the insured (the
agent): it then chooses a contract with only partial insurance to preserve the right
incentives. Other economic relationships of this type include shareholders and
managers, manager and salespersons, landlord and tenants etc.
In the following sections, we specify a model to capture an effort choice problem and
suggest contractual solutions.
explain the problem of effort choice and analyse effort choice in a principal-agent
model
161
11. Asymmetric information: moral hazard
1. Let e denote the effort exerted by the agent on the project. A can either work hard
(choose e = eH ) or take it easy (choose e = eL ). An informational asymmetry arises
when the choice of e cannot be monitored by P .
2. Let the random variable denote the (observable) profit of the project. Profit is
affected by the effort level chosen, but is not fully determined by it. (If it was fully
determined, the principal can infer the true effort choices by observing the
realisation of : we would no longer be in a situation of hidden actions.) Higher
effort leads to higher profit in a probabilistic sense. Assume that can be either
high (H ) or low (L ). Let P ( | e) be the probability of under effort e.
162
11.3. Effort choice and incentive contracts: a formal model
Under high effort, the high profit is more likely. Therefore, we have:
P (H | eH ) > P (H | eL ).
Profit
10 0
Agents effort eH 2/3 1/3
eL 1/3 2/3
Moral hazard arises from the fact that the realised profit is not fully informative
about effort. If, on the other hand, the probabilities were (1, 0) under eH and (0, 1)
under eL , there would be no moral hazard, as profit would reveal effort completely.
3. The principal is risk-neutral: she maximises expected profit net of any wage
payments to the agent.
4. The agent is risk-averse in wage income, and dislikes effort. His utility function
takes the form:
v(w, e) = u(w) g(e)
where w is the wage received from the principal, g(e) is the disutility of effort and
u is the vNM utility function, where u0 > 0, u00 0 (implying risk aversion).
Finally, high effort is more costly so g(eH ) > g(eL ).
Note the central conflict of interest here. The principal would like the agent to choose
higher effort, since this leads to higher expected profits. But, other things being equal,
the agent prefers low effort. However, (to anticipate our conclusion), if the
compensation package is carefully designed, their interests could be more closely
aligned. For instance, if the agents compensation increases with profitability (by means
of a profit-related bonus, say), the agent would be tempted to work hard (because high
effort will increase the relative likelihood of more profitable outcomes). However, this
will also make the agents compensation variable (risky), and that is not so efficient
from a risk-sharing perspective. The inefficiency in risk-sharing will be an unavoidable
consequence of asymmetric information.
5. The principal pays the agent a profit-contingent wage schedule {wH , wL }. Here wH
is the wage paid if the profit is H , and wL is the wage paid if the profit is L . In
general, wH > wL . Note that if wH = wL = w , the principal is paying a fixed wage
w that does not depend on the profit level. A fixed wage gives full insurance to the
agent (income does not vary no matter what the profit is). However, if wH > wL ,
the wage schedule carries risk.
6. Let u0 be the reservation utility of the agent. This is the expected utility the agent
gets from alternative occupations. Any wage contract that the principal offers the
163
11. Asymmetric information: moral hazard
agent must give the agent a net payoff of at least u0 , otherwise the agent would not
accept the contract. This is known as the agents participation constraint (PC).
Let us now show how to determine the optimal wage schedule under full information
and under asymmetric information.
Which wage schedule should the principal choose to implement eH ? If the agent chooses
effort eH , the principal gets the surplus:
P (H | eH ) (H wH ) + P (L | eH ) (L wL ).
The principal wants to maximise this surplus subject to the agents PC.
We will do this by Lagrangian optimisation. But even before doing that, the optimal
arrangement should be intuitively clear. Choose a wage schedule (wH , wL ) which just
satisfies the agents PC, i.e. PC holds with equality. Suppose the wage schedule is risky,
i.e. wH > wL . Starting from this, can the principal do better? Recall that the principal
is risk-neutral while the agent is risk-averse. If the principal gives insurance to the agent
by replacing the risky wage schedule with a fixed wage equal to the expected wage, the
agent would be strictly better off while the principal would remain indifferent. But now
the agents utility is strictly above u0 , so the principal can reduce the fixed wage down
to a level that will again make the agent indifferent between the utility from
participating and u0 . In the process, the principal is strictly better off.
In other words, when the principal gives the agent a risky wage, there is inefficient
risk-sharing. Facing a risky wage, the agent would pay to buy income insurance. But
the principal is risk-neutral, so she can give this insurance to the agent and extract that
payment (i.e. reduce the expected wage). In yet other words, starting from a fixed wage
that just satisfies PC, if the principal replaces the fixed wage with a risky wage, the
agent will not participate as his payoff will fall below u0 . The principal has to also
164
11.4. Full information: observable effort
compensate the agent for the addition of risk in order to satisfy the PC. So putting risk
on the agent is costly for the principal. Since effort is directly observable, there are no
incentives to worry about and risk serves no purpose. Therefore, when effort is
observable, it is optimal for the principal to pay a fixed wage that does not depend on
the profit level.
Let us now show this by doing a Lagrangian optimisation. However, you should really
not rely on algebra for this the intuition should be clear to you.
The principals problem is to maximise surplus (i.e. net-of-wages profit) subject to the
PC of the agent. Note that there is no reason for the principal to leave any extra
surplus for the agent, so the optimal wage schedule would be such that PC binds (holds
with equality). With this in mind, the principals problem is:
max P (H | eH ) (H wH ) + P (L | eH ) (L wL )
wH , wL
subject to:
(H | eH ) u(wH ) + (L | eH ) u(wL ) g(eH ) = u0 .
Set up the Lagrangian with as the Lagrange multiplier, and derive the first-order
conditions:
L
= P (H | eH ) + P (H | eH ) u0 (wH ) = 0
wH
L
= P (L | eH ) + P (L | eH ) u0 (wL ) = 0.
wL
We ignore the derivative with respect to which recovers the participation constraint.
Simplifying, we get:
1
= u0 (wL ).
u0 (wH ) =
00 0
Since u < 0, u is a decreasing function. Therefore, from the equalities above, we get
wH = wL .
Therefore, to implement eH , the principal optimally offers a fixed wage w such that:
u(w ) g(eH ) = u0 .
u(w ) g(eL ) = u0 .
P (H | eH ) H + P (L | eH ) L w .
165
11. Asymmetric information: moral hazard
The expected surplus from eL , which can be implemented by paying a fixed wage w , is
given by:
P (H | eL ) H + P (L | eL ) L w .
The principal should compare the two and implement the effort that yields a higher
surplus.
u(w ) g(eL ) = u0 .
The wage schedule must satisfy the agents PC, i.e. it must be such that the agents
payoff under eH is at least as large as u0 :
The compensation package must also be such that the agent prefers to exert eH
rather than eL . In other words, the wage schedule must create the right incentives
for choosing high effort over low effort. This is called the incentive compatibility
constraint IC. This is given by:
166
11.5. Asymmetric information: unobservable effort
Note that a fixed wage cannot satisfy the IC. If the wage is fixed, we have
wH = wL = w. Then the left-hand side is u(w) g(eH ), while the right-hand side is
u(w) g(eL ). Since g(eH ) > g(eL ), the IC is violated. Putting some risk on the agent is
unavoidable in providing an incentive for high effort.
To determine the optimal values of wH and wL , we need to maximise the surplus of the
principal subject to the PC and IC. However, in our simple model there are just two
wage levels and two constraints. Therefore, once we put numerical values for the various
probabilities and effort costs, the constraints determine the values of wH and wL fully
(so no maximisation is necessary). The following section contains an example.
However, in more general models that you would encounter if you pursued economics at
a Masters level, there could be (say) more than two profit levels, which would indicate
that the compensation package would consist of more than two wage levels. In such
cases, some further maximisation needs to be done.
The following activity goes through a detailed example of the theory we described
above.
(a) If effort is observable, the principal optimally offers a fixed wage to implement
any effort level. Provide a brief intuition for this result.
Since the agent is risk-averse and the principal is risk-neutral, an optimal
risk-sharing result is for the principal to take all the risk. If the agent were to
carry unwanted risk, he would have to be given a higher expected wage than in
the optimal case. Since this is not necessary for incentive reasons because effort
is observable, the first best solution is for the principal to take all the risk and
completely insure the agent.
You can check this formally using the derivation presented in Section 11.4.1 of
the subject guide.
(b) If effort is observable, what wage w should the principal offer if she wants to
implement e1 ? What wage w implements e2 ? Which induced effort level
167
11. Asymmetric information: moral hazard
u(w ) g(e1 ) = u0 = 0.
Since (e1 ) > (e2 ), the principal will implement e1 and pay a fixed wage of
25/9.
(c) Suppose, next, that the agents choice of effort level is not observable. In this
circumstance, a contract consists of an output-contingent wage schedule
{wH , wL }. What wage schedule will implement e1 in this case? What expected
net return does the principal get in this case? How does this compare with the
value in (b), where effort was observable?
State-dependent contracts are {wH , wL }. Implementing e1 at minimum expected
wage cost requires minimising 2wH /3 + wL /3 subject to the participation
constraint (PC1 ) as well as the incentive constraint (IC) of the agent. Solving
the two constraints completely determines wH and wL , so that there is no
further minimisation to be done. We have:
2 1 5
(P C1 ) wH + wL 0
3 3 3
and:
2 1 5 1 2 4
(IC) wH + wL wH + wL .
3 3 3 3 3 3
Since there are two variables to be determined, and two inequalities, it is
possible to find wH and wL so that both bind. (IC) implies wH = 1 + wL .
We can use this in (P C1 ) and obtain:
2(1 + wL ) + wL = 5 (wL = 1, wH = 4)
168
11.6. Risk-neutral agent
Note that the profit was 35/9 under full information, and here it is 33/9. Why
does the principals profit fall by 2/9 under asymmetric information? This is
because the principal is now putting risk on the agent and, therefore, has to pay
a higher expected wage to compensate. Hence the profit is lower.
(d) If effort is not observable, what contract is best for the principal? Should she
implement e1 or e2 ?
Implement e2 . Any flat wage that satisfies the agents participation constraint
implements e2 . The cheapest way for the principal to do this is to offer the first
best flat wage w1 = 16/9.
As before, (e2 ) = 14/9. Since (e1 ) > (e2 ), the principals optimal choice
under asymmetric information is e1 .
v(w, ei ) = w g(ei ).
Under full information, to implement the high effort e1 , the principal pays w = 5/3 and
to implement e2 , pays w = 4/3. The principals profit from e1 is 5 and from e2 the
principals profit is 2.
Now consider the scenario under asymmetric information. We already know that the
principal can implement e2 by the same fixed wage as above, and earn 2 as before
asymmetric information has no impact on the profit from implementing the low effort.
However, what is the profit from implementing e1 ? Recall that some loss of profit was
inevitable when the agent was risk-averse. Here, suppose the principal simply asks for a
fixed fee of F = 5. The agent is the residual claimant and bears all the risk. Will this
work?
If the agent adopts e1 , he gets 20/3 5/3 F = 5 5 = 0, while the payoff from
adopting e2 is 10/3 4/3 F = 2 5 < 0. Note that the payoff from e1 satisfies both
169
11. Asymmetric information: moral hazard
the participation constraint and the incentive compatibility constraint. Therefore, the
agent would adopt e1 , and the principal gets 5, which is the same as under full
information. So, for the principal, asymmetry of information does not impose any loss
compared to full information because putting risk on the agent is not costly.
explain the problem of effort choice and analyse effort choice in a principal-agent
model
1. Car insurance and home insurance typically include a deductible. Briefly explain
the reason behind this feature.
170
11.8. Test your knowledge and understanding
(a) Suppose the effort level chosen by the agent is observable by the principal. A
wage contract then specifies an effort level (e1 or e2 ), and an output-contingent
wage schedule {wH , wL }. Here wH is the wage paid if the project yields 8, and
wL is the wage paid if the project yields 0. If effort is observable, it is optimal
for the principal to choose a fixed wage contract (that is, set wH = wL ) for
each effort level. Informally, explain the intuition for this result.
(b) If effort is observable, which effort level should the principal implement? What
is the best wage contract that implements this effort?
(c) If effort is not observable, which effort level should the principal implement?
What is the best wage contract that implements this effort?
171
11. Asymmetric information: moral hazard
172
Chapter 12
Externalities and public goods
12.1 Introduction
As noted in Chapter 7, the presence of externalities, as well as public goods, gives rise
to both qualitatively and quantitatively important departures from the first welfare
theorem. Much of economic policy is focused on dealing with these departures.
In some situations, the utility of an agent is affected by the level of activity of another
agent directly. However, the acting agent does not take this externality into account.
The result is that the market outcome is inefficient. The individually-optimal level of an
activity that generates a negative externality is higher than the socially-optimal level,
and that for an activity that generates a positive externality is lower than the
socially-optimal level. The invisible hand does not work well in such cases individual
optimisation does not lead to social optimality.
Some of the biggest problems facing the world are caused by negative externalities.
These include global warming, pollution, congestion, urban noise, and so on. In all these
cases, government policy has a crucial role to play to bring individual incentives in line
with the social optimum.
Competitive markets also work poorly in the presence of public goods. However, public
goods such as national defence, roads, basic public health, basic education and defence
against natural catastrophes are important for any economy. Provision of such goods
using receipts from taxes is a crucial role for the government in the economy.
When you study these topics, you should try to understand the nature of different
policies: their merits, demerits and limits. You should also try to understand how much
of a difference good economics can make to the world. Unfortunately, good economics is
not always appreciated or endorsed by politicians, industry leaders or the wider
population. The cap-and-trade markets in the United States are a case in point.
Economists have been urging governments to have such markets since at least the
1960s. It took another three decades for the first such system to be set up, amid great
scepticism expressed by politicians as well as industry leaders. This was the United
States acid rain programme, part of the Clean Air Act (1990). The caps on NOx and
SO2 took effect in 1995. Within a few years, it was clear that the system was a huge
success. Yet, even now, scepticism is regularly expressed about emissions trading by the
opponents of the policy in the United States Congress, whose entrenched positions
effectively eliminated any chance of the Obama administration establishing a
nationwide CO2 emissions trading system. While we highlighted the case of the United
States here, disputes about economic policy arise across the world, and, unfortunately,
bad economics often drives good economics away. As students of economics, you should
try and understand these problems well, so that through individuals like yourself, good
173
12. Externalities and public goods
explain the concept of externalities and show how externalities lead to market
failures
explain the content of the Coase theorem and analyse the Coasian property rights
approach to solving the problem of externalities
explain the concept of public goods and analyse how private provision leads to an
inefficiently low provision of such goods
explain the concept of the tragedy of the commons and demonstrate how strategic
behaviour leads to over-exploitation of open-access resources in simple models
174
12.2. Overview of the chapter
12.3 Externalities
See N&S Sections 16.1 to 16.4 for basic concepts. Here we consider a simple model of a
society with two agents, 1 and 2. Agent 1 decides on the level x of an action. Assume
that x is costless. To model externalities, we assume that not just agent 1 but also
agent 2 is affected by agent 1s choice.
Agents have quasilinear utility. The utility function of agent i is i (x) + mi , where
i = 1, 2. Here you can think of good m as money. Let m be the total amount of good m.
Furthermore, i (x) denotes the utility of i from x. We assume that 01 > 0 and 001 < 0.
Also, 002 < 0. What about the sign of 02 ? If this is positive, we have a case of a positive
externality, while 02 < 0 implies a case of a negative externality. Below we consider the
case of a positive externality briefly, but most of our analysis is concerned with the case
of a negative externality.
Note that the utility function is linear in mi , but not in x. Hence the name quasilinear.
Individual optimum
Next, since x is costless, the individual optimum for agent 1 is reached at x such that:
01 (x ) = 0.
Pareto optimum
The Pareto optimal outcome can be found simply by maximising the sum
1 (x) + 2 (x). This is a consequence of the quasilinear utility function: we can simply
compare the utilities from x with money amounts. Suppose we have a negative
externality. Suppose the marginal utility of agent 1 from x is higher than the marginal
disutility of agent 2 from x. Suppose we increase x and keep agent 2s utility the same
by transferring money to agent 2 equal to the extra disutility. Since the extra utility
gained by agent 1 from the increase in x is greater than the money transferred to agent
2, agent 1 has a net gain. So whenever agent 2s marginal disutility is lower than agent
1s marginal utility, we can make a Pareto improvement. This leads to the idea that at
the optimum the sum of marginal utilities must be zero. A similar intuition applies to
175
12. Externalities and public goods
the case of a positive externality. We will show this more formally when discussing
public goods. This gives us the following.
The Pareto optimal level of x is denoted by xo . This maximises the total 1 (x) + 2 (x).
Therefore, xo is such that:
01 (xo ) + 02 (xo ) = 0.
Note that the second-order condition holds since 1 and 2 are both concave by
assumption.
In what follows we use the terms Pareto optimal, efficient and socially optimal
interchangeably.
Positive externality
Consider first the case of a positive externality. In this case the social benefit of x
obviously exceeds the individual benefit of agent 1, and, as shown in Figure 12.1 below,
the level x undertaken by agent 1 is lower than the efficient level xo .
Negative externality
From here onwards we consider the case of a negative externality. The policy solutions
analysed later can be appropriately modified (for example, a tax solution becomes a
subsidy solution) for the case of a positive externality.
In this case 02 > 0. Figure 12.2 below shows the functions 01 and 02 . At the
socially-optimal outcome these are equal. Note that now the individual optimum
exceeds the socially-optimal level. While 01 is the marginal social benefit, 02 is the
marginal social cost. In deciding the level of x, agent 1 does not take this social cost
into account and, therefore, the level of x chosen by agent 1 is too high relative to the
socially-optimal level.
176
12.3. Externalities
Other than tax, a quota at xo would also, of course, restore social optimality. However,
the tax and quota solutions depend on the government knowing a lot about the
preferences of the agents. Absent such information, the correct level of tax or quota
177
12. Externalities and public goods
The Coase theorem is not really a theorem in the usual formal sense. It is an informal
statement that expresses an idea that Coase expounded in his famous article on the
social cost of externalities (Coase, 1960).
The idea is as follows. Consider a situation involving two parties where an externality
arises from the action of one party. The Coase theorem says that under conditions of
perfect competition (that is, under the standard assumption in the economic analysis of
competitive markets that there are no transaction costs of trading), if property rights
are well-defined, bargaining between parties would lead to the efficient outcome. This is
true irrespective of who the property rights are allocated to (to the party generating the
externality or to the party affected).
This approach says that it is unnecessary to provide incentives through taxes, subsidies
or quotas. Indeed, once property rights are well-defined, it allows agents to trade in the
externality-generating activity, and as in any other market under perfect competition,
all beneficial trading opportunities are exploited. Furthermore, taxes can therefore only
cause harm by distorting incentives. From a legal point of view, the implication is that
if we do have a situation of no transaction costs, legal institutions do not matter. If one
party generates, say, a negative externality for another, the law can hold either party
liable for the damages and it would still result in an efficient outcome.
Coases argument shows that it might be logically inconsistent to treat externalities as a
departure from otherwise perfect markets. If markets are indeed perfect, and some legal
liability regime is in place, rational agents should be able to use the same tools that
other markets use, to solve the problem of social cost from externalities. Under perfectly
competitive markets and a well-defined property rights regime, social costs therefore
coincide with private costs, and tax-based solutions are unnecessary and even
potentially harmful.
This is the more formal message in terms of economic theory. Coases intended message
was not that the problem of social cost does not exist or that legal institutions do not
matter. He intended to argue that transaction costs do matter and ignoring them
gives us predictions that are unrealistic.
However, both messages are important. As economists looking at policy, you should
carefully consider the case for a property rights based approach among other
alternatives. Indeed, such ideas have influenced the creation of various pollution rights
markets that have allowed reducing pollution in an efficient manner. Furthermore, you
should also be aware that different types of transaction costs do arise in applications,
and one must take that into account to judge the efficacy of a rights-based solution.
Let us now explore the idea of property rights formally in our model.
Suppose agent 2 has the right to an externality-free environment. If there is no
negotiation, agent 2 can then prevent agent 1 from taking any positive level of activity.
If the two agents negotiate, agent 2 can ask for a payment of T and allow agent 1 to
take some positive level of x. To determine T and x, consider agent 2s optimisation
problem. For any T and x, agent 2s payoff is 2 (x) + T . Agent 2 must maximise this
178
12.3. Externalities
subject to satisfying agent 1s participation constraint, which is 1 (x) T > 1 (0). Note
that agent 2 does not need to leave a surplus for agent 1, so she can choose T such that
agent 1s participation constraint binds. Therefore, agent 2s problem is to:
Since the last term is just a constant, the first-order condition is 02 (x) + 01 (x) = 0.
Therefore, the optimised value of x coincides with the socially optimal xo .
Next, suppose agent 1 has the right to choose any level of x. If there is no negotiation,
agent 1 would choose the previously-derived individually-optimal level x . If the two
agents negotiate, agent 1 can ask for a payment of T and reduce x to some level below
x . To determine T and x, consider agent 1s optimisation problem:
Since the last term is just a constant, the first-order condition is 01 (x) + 02 (x) = 0.
Therefore, the optimised value of x again coincides with the socially optimal xo .
The above analysis demonstrates that irrespective of who has the property rights, so
long as it is well-defined, costless negotiation leads to the socially-optimal outcome.
However, a variety of different transaction costs can arise in reality, limiting the force of
the solution in many cases. Even if there are two (or very few) parties, the parties might
not have all the relevant information about each other. Preferences might be private
information, a situation known formally as incomplete information. Under incomplete
information, bargaining is typically inefficient (i.e. not all gains from trade can be
exploited). Furthermore, even with just two parties, one or both could comprise of
many individuals (an upstream chemicals firm polluting a river, generating a negative
externality for the downstream community of fishermen). Negotiation then depends on
those individuals being able to speak with one voice through a leader. There might also
be free-rider problems if committing to a negotiation is costly in terms of time and
effort. Yet further, there might be many parties involved, making negotiation difficult or
even a complete non-starter.
Coases own work, coupled with his editorship of the Journal of Law and Economics
spanning a period of 17 years, helped create the field of law and economics. Indeed,
game theory and information theory allow questions about the design of legal
institutions, especially those concerned with tort law (contract law), to be very much
part of the field of enquiry of microeconomics.
As noted above, Coases ideas have influenced the creation of pollution rights markets
that play an important role in controlling harmful emissions. Below, we consider the
approach in more detail, and especially consider the United States market for permits
for sulphur dioxide emissions.
179
12. Externalities and public goods
Note also that N&S mentions Obamas cap-and-trade policy proposals of 2009 in
Application 16.4. This is misleading in light of later events. The partisan policy stance
that has characterised American politics in recent years meant that the bill did not
stand any chance of passing and was withdrawn. Later, individual states were given the
responsibility of developing their own systems. We mention this in our discussion of
cap-and-trade below.
180
12.3. Externalities
in the market is a good signal to judge the policy. A very high price implies that the
cap is very low, and the cost this imposes on firms might be very high. On the other
hand, a market price close to zero implies that there is little demand in the market
relative to supply, so that the system does not do much to reduce the costs of
abatement. In other words, in the latter case the cap is ineffective.
In reality, cap-and-trade has not been an unqualified success. In the EU ETS, there
were so many permits issued in the first phase (200507) that the permit price came
down to 0 in 2007. The later phases of the system have recovered from the collapse,
but there have been persistent criticisms alleging oversupply and consequently
low-price permits which, in turn, imply low incentives for reducing emissions.
Schmalensee and Stavins (2013) describe how the SO2 emissions trading system in
the US, which was very successful for the first decade of its operation starting in the
mid-1990s, collapsed later because of the wider regulatory environment.
As the authors note, beginning in 1995 and over the subsequent decade, the SO2
allowance trading programme performed exceptionally well along all relevant
dimensions. SO2 emissions from electric power plants decreased 36 per cent from
15.9 million to 10.2 million tonnes between 1990 and 2004, even though electricity
generation from coal-fired power plants increased 25 per cent over the same period.
By 2010, SO2 emissions had declined further, to 5.1 million tonnes.
The costs of achieving these environmental objectives with cap-and-trade were
significantly less than they would have been with a command-and-control regulatory
approach. Cost savings were at least 15 per cent, and perhaps as much as 90 per
cent, compared with counterfactual policies that specified the means of regulation in
various ways and for various portions of the programmes regulatory period. In
addition to efficiency achieved through trading, there is evidence that the
programme brought down abatement costs over time by providing incentives for
innovation and diffusion that were generally much stronger than those provided by
traditional command-and-control regulation.
As the authors point out, it was widely recognised by the late 1990s that the SO2
cap needed to be lowered significantly to attain environmental goals. The Bush
administration promulgated its Clean Air Interstate Rule in 2005, with the purpose
of lowering the cap on SO2 emissions significantly, and allowance prices rose in
anticipation of this lower cap. After peaking in 2005 at more than $1,200 per tonne,
SO2 allowance prices dropped just as fast as they had risen, aided by an
announcement from the US Environmental Protection Agency that it would
re-examine the Clean Air Interstate Rule and speculation about impending legal
challenges. The act was later vacated by the courts. The Obama administration
similarly tried and failed to reduce the cap through legislation. The end result is
that state-level and source-specific regulation has been imposed. As the authors
point out, the allowance market remains nominally in place, but the imposition of
prescriptive state-level and source-specific regulation has virtually eliminated the
demand for federal SO2 allowances. By the time of the Environmental Protection
Agencys 2012 auction, market-clearing prices had fallen to virtually zero.
181
12. Externalities and public goods
Figure 12.4: SO2 allowance prices in the US, 19942012. Source: Schmalensee and Stavins
(2013).
X = xi + xi .
182
12.4. Public goods
We present the details of the argument below. However, once you have understood the
argument, you can put it at the back of your mind and simply use the result.
Under the quasilinear P
utility specified above, the Pareto optimal outcome can be found
simply by maximising i ui (X) cX.
Proof: For simplicity, let us show this assuming n = 2. The same method can be used
for any value of n. Note that having paid the cost cX, the total amount of the m-good
that is available for division between the agents is m cX.
With this in mind, and with two agents, the Pareto optimal outcome can be found by:
Using the last constraint and substituting the value of m2 , this can be rewritten as:
183
12. Externalities and public goods
The efficient provision of the public good is X o which satisfies the first-order
condition: X
u0i (X o ) = c.
i
But
P since u00i
< 0, u0i (X)
is a decreasing function of X for each i. This implies that
0
i ui (X) decreases as X rises. It follows that:
Xo > X
which implies that private provision of a public good leads to an inefficiently low
provision.
Consider
P 0 o the simple case in which individuals are identical so that ui = u. Then
0 o
i ui (X ) = nu (X ). Then the socially-optimal provision is given by:
nu0 (X o ) = c
while the private provision level is given by:
u0 (X ) = c.
Figure 12.5 shows X o and X . Clearly, X o is larger than X .
184
12.5. The commons problem
Figure 12.5: Private provision of a public good. The public good is provided at level X
under private provision while the socially-optimal level is X 0 .
As noted in the last section, public goods are neither excludable nor rival. Private
goods, on the other hand, are both rival and excludable. However, many resources are
rival but not excludable. This gives rise to an over-exploitation issue known as the
commons problem or the tragedy of the commons.
Section 5.7 of N&S presents a simple game where the Nash equilibrium shows the
tragedy of the commons. In this section we develop a further simple model of resource
extraction.
12.5.1 Evidence
Stavins (2011) notes that the consequences of open access predicted by theory have
been validated repeatedly with empirical data. A study of the Pacific halibut fishery in
the Bering Sea estimated that the efficient number of ships was nine, while the actual
number was 140. An examination of the New England lobster fishery found that in 1966
the efficient number of traps set would have been about 450,000, while the actual
number was nearly one million. Likewise, an analysis of the North Atlantic stock of
minke whale found that the efficient stock size was about 67,000 adult males, whereas
the open-access stock had been depleted to 25,000. In terms of social costs, an analysis
of two lobster fisheries in eastern Canada found that losses due to unrestricted entry
amounted to about 25 per cent of market value of harvests, due mainly to excess
deployment of resources for harvest, with fishery effort exceeding the efficient level by
some 350 per cent.
185
12. Externalities and public goods
Socially-optimal level Lo
Let us first work out the socially-optimal level of L. This is given by L0 such that:
1
=c
2 L0
implying L0 = 1/4c2 .
Privatisation solution
As we know from above, the efficient solution is obtained when n = 1. If a single agent
is assigned the right to extract, efficient extraction results. However, for certain types of
commons resources such as fishing assigning property rights to a single agent may
not be feasible.
186
12.6. A reminder of your learning outcomes
Just as in pollution control (described above), the government could set an overall
quota and issue individual transferable quotas (ITQs) that can be traded. Trading ITQs
ensures that these go to those with the lowest cost of exploitation.
Stavins (2011) writes: ITQ systems have been used successfully in some 150 major
fisheries of 170 species in seventeen countries some with very significant fishing
industries, such as Australia, Canada, Iceland, and New Zealand. . . . In fact, New
Zealand regulates virtually its entire commercial fishery this way. Since 1986, the system
has been effective, largely eliminating overfishing, restoring stocks to sustainable levels,
and increasing fishermens profits. Several ITQ systems are in operation in the United
States, including ones for Alaskas Pacific halibut and Virginias striped-bass fisheries.
Tax
Consider a tax of t per unit of extraction. This raises the marginal cost of extraction to
c + t so that, after the tax, the equilibrium level of L is:
2
1 1
Lt = 1 .
(c + t)2 2n
Set t such that Lt = L0 , i.e. we have:
2
1 1 1
1 = 2
(c + t)2 2n 4c
which gives us:
n1
t= c.
n
This tax would restore efficiency. Note, however, that setting a tax optimally requires a
lot of information, which the policymaker might not have.
Informal cooperation
Note that the Nash equilibrium in the game above is akin to the Prisoners Dilemma
equilibrium. If all agents could agree to hold back their extraction, everyone would be
better off. However, more exploitation than this level is best for each individual, so that
no-one holds back in equilibrium.
Just as in the Prisoners Dilemma, there is scope for cooperation in equilibrium if the
game is repeated. This type of approach can explain how several rural societies,
especially in third-world countries, have been able to overcome a variety of problems
involving local commons resources.
187
12. Externalities and public goods
explain the concept of externalities and show how externalities lead to market
failures
analyse remedial policies based on taxes or subsidies
explain the content of the Coase theorem and analyse the Coasian property rights
approach to solving the problem of externalities
explain how an overall quota can be implemented efficiently using cap-and-trade
systems, how such systems have reduced pollution from several sources and how
such systems can sometimes run into difficulties
explain the concept of public goods and analyse how private provision leads to an
inefficiently low provision of such goods
explain the concept of the tragedy of the commons and demonstrate how strategic
behaviour leads to over-exploitation of open-access resources in simple models
explain the policy responses to commons problems.
2. Your neighbour has an electrical device that is costly to run, but kills insects very
effectively. The longer they keep the device on, the fewer insects appear in your
garden.
(a) Explain the externality in this scenario.
(b) Adapt the simple model developed in this chapter of the subject guide to
reflect the underlying externality in this situation and show formally how a tax
policy would work.
(c) Show formally how a Coasian property rights solution would work.
3. Explain how an emissions trading system operates and discuss its merits.
5. Consider an environment where two firms operate with the following profit
functions:
1 = pq1 3q12 q2
2 = pq2 2q22 .
188
12.7. Test your knowledge and understanding
(a) Compute the level of production that each firm chooses when maximising its
profit individually.
(b) Compute the efficient level of production for each firm.
(c) Compute the per-unit tax that induces the efficient level of production. Do you
have to impose a tax on both firms?
6. This is a more difficult question. Consider the following two-player public goods
game. Players decide simultaneously whether to contribute to the public good. Each
player derives a benefit of 1 if at least one of them contributes and zero otherwise.
The cost of contribution for each player i = 1, 2 is given by ci and is known only to
that player. Player 1 believes that c2 is distributed uniformly on the interval [0, 2].
Similarly, player 2 believes that c1 is distributed uniformly on the interval [0, 2].
Consider the following type of strategy. Each player contributes if their cost is
below a cut-off c .
Compute the Nash equilibrium of this game. What is the likelihood of the public
good being provided in this equilibrium?
189
Notes
Notes
Notes
Comment form
We welcome any comments you may have on the materials which are sent to you as part of your
study pack. Such feedback from students helps us in our effort to improve the materials produced
for the International Programmes.
If you have any comments about this guide, either general or specific (including corrections,
non-availability of Essential readings, etc.), please take the time to complete and return this form.
Name
Address
Email
Student number
For which qualification are you studying?
Comments
Date: