LectureNotes2023 Nov21
LectureNotes2023 Nov21
Markus Brunnermeier
with edits by Caio Almeida, Hao Jin, and Moritz Lenel
Copyright © 2023 Markus Brunnermeier
These lecture notes are currently updated and maintained by Hao Jin ([Link][at][Link]).
Please contact Hao for comments and corrections.
This version: Nov 21, 2023. When this sentence is gone, this will be the complete lecture
notes.
Contents
2.5 Exercises 21
3 State Prices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
3.1 Pricing in One Period Model 23
3.1.1 State Prices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
3.1.2 The Fundamental Theorem of Finance . . . . . . . . . . . . . . . . . . . . . . . . . . 25
3.1.3 State Prices and Incomplete Markets . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
3.2 Asset Pricing Formulas 26
3.2.1 State Price Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
3.2.2 Stochastic Discount Factor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
3.2.3 Equivalent Martingale Measure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
3.2.4 State-Price Beta Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
3.2.5 Four Asset Pricing Formulas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
3.3 Recovering State Prices from Option Prices 29
3.4 Appendix 32
3.5 Exercises 33
We begin our analysis of asset pricing by considering a simple setting in which there are two dates:
today (t = 0) and some future date (t = 1). We know everything about today’s state of the world
s = 0, but we don’t know which one of s = 1, . . . , S states will materialize in the future.
t =1
s=1
s=2
t =0
...
s=S
c1
′
Figure 1.2: Two states (S = 2) and a single Arrow-Debreu security e1 = (1, 0) .
Note that if this is the only security available in the market, we can replicate any security in the
horizontal axis by buying a quantity α ∈ R of e1 , but we cannot replicate any of the securities
outside of the horizontal axis (e.g. the red security). When the available securities are not enough
to replicate all securities in RS we say that markets are incomplete.
If we introduce another asset e2 = (0, 1)′ , all securities in R2 become replicable. In this case, we
say that markets are complete. Note that adding another asset x3 ∈ R2 does not benefit us in any
way, as we were already able to span the whole set R2 with the previous two. The security structure
generated by e1 and e2 is
" #
1 0
X= .
0 1
Note that 1) all payoffs (securities) are linearly independent vectors in RS , 2) markets are complete
by construction (since also J = S).
1.1 States and Securities 9
c2
c1
′
Figure 1.3: Two states (S = 2) and one risk-free security b = (1, 1) .
Under this security structure, all riskless securities on the 45 degree line are replicable but none
of the securities outside of it are replicable (e.g. the red security). Adding a risky security, say
c = (2, 1)′ , allows us to span the whole R2 set. For instance, suppose we wish to replicate the
security d = (1, 2)′ . We can buy 3 securities b (in blue below) and sell short one unit of security c
(in orange), thus obtaining security d (in green):
c2
c1
Figure 1.4: Replicating d = (1, 2)′ with 3 units of bond and −1 (short) unit of the risky asset.
b and c are linearly independent vectors in R2 , and therefore the payoffs that can be obtained with
their linear combinations coincide with those that can be obtained with linear combinations of the
10 Chapter 1. One Period Model
First, the state space itself is an important modeling choice. The relevance of this can hardly
be overstated, because market completeness depends directly on how we specify the state space.
Suppose for example that the security structure looks like:
1 0 9 4 2
2 1 5 2 2
X =
3 1 8 4 9
4 2 1 2 2
4 2 1 2 2
Clearly markets are incomplete, but if the state s = 5 is a state of the world equivalent to s = 4 in all
aspects with the only exception that “at noon my cat is on the third floor”, then state s = 5 probably
isn’t a relevant one and for modeling purposes should be disregarded. And after eliminating state
s = 5 markets are complete (in the relevant states s = 1, 2, 3, 4). Another example where the state
space choice is important is in derivatives pricing, where we assume that the state space corresponds
to the space of possible values of the underlying asset.
Second, although we said we would not deal with frictions until the third part of the course, market
incompleteness is a market friction: an incomplete market is equivalent to a market in which there
are infinite transaction costs to trade assets whose payoffs are linearly independent from the existing
assets, for which instead transaction costs are zero. From this point of view, we are dealing with a
very “black and white” setup: either we have no frictions at all (with market completeness) or an
extreme friction (with market incompleteness). Later in this course we will explore the “grey” area
in between.
Finally, trade limitations are another kind of market friction we will encounter in this part of the
course. Suppose there are only two states of the world and the security structure includes only a
risk-less bond, which we can neither buy in large amounts nor short-sell (assume there is a law that
prohibits these practices):
Therefore, we can only achieve payoffs on the 45 degree line between the two purple dots.
• A Portfolio as a vector h ∈ RJ , whose entries represent a quantity for each asset in the
security structure.
J
• The Portfolio Payoff is given by ∑ h j x j = Xh
j=1
• The Asset Span is ⟨X⟩ = z ∈ RS : ∃h ∈ RJ such that z = Xh
1.3 LOOP & No Arbitrage 11
c2
size constraint
short-sale constraint
c1
Note that it is always the case that ⟨X⟩ ⊆ RS , and we have market completeness if and only if
⟨X⟩ = RS , that is, if and only if rank(X) = S. In other words, market completeness refers to a
market structure in which there are at least S linearly independent assets. We say that security j
is redundant if there exists h ∈ RJ such that x j = Xh and h j = 0: when markets are complete and
J > S there are J − S redundant assets. When rank(X) < S markets are incomplete, and if there
are J < S linearly independent assets then S − J linearly independent assets are needed to achieve
market completeness.
Let p ∈ RJ be the vector of prices for each asset. Then the cost of portfolio h is given by
J
p·h ≡ ∑ p jh j
j=1
xj
and if p j ̸= 0, the (gross) return vector for asset j is given by R j = pj .
Lemma 1.3.1 Law of One Price (LOOP) implies that every portfolio with zero payoff has zero
price.
Proof. Trivial. ■
Proof. We prove the contrapositive “if LOOP does not hold, then NSA does not hold either”.
If the LOOP does not hold, then Xh = Xk and p · h ̸= p · k. Assume p · h < p · k, then we have
X(h − k) = 0 ≥ 0 and p · (h − k) < 0, which is a violation of NSA for the portfolio w ≡ h − k.
Similarly, if p · h > p · k then X(k − h) = 0 ≥ 0 and p · (k − h) < 0, again a violation of NSA for
the portfolio q ≡ k − h. ■
1.4 Exercises
1. Determine whether the following statements are true or false. Provide a proof or a counter-
example.
a) Law of one price and complete markets imply no strong arbitrage.
b) Law of one price and complete markets imply no arbitrage.
c) No strong arbitrage and complete markets imply no arbitrage.
2. Suppose there exist three states of the world s = 1, 2, 3 and two assets x1 , x2 .
a) Suppose x1 = (2, 1, 0) and x2 = (0, 1, 0). Describe the asset span. Are markets com-
plete?
b) Suppose p1 = 4 and p2 = 3. What type of no-arbitrage requirements does this market
satisfy?
c) What are the restrictions on p1 and p2 such that this market satisfies LOOP, NSA and
NA? Write each restriction separately.
d) Repeat a), b) and c) for x1 = (1, 1, 0), x2 = (0, 2, 0) and x3 = (0, 1, 1). Hint: it is a
little bit confusing as the price for the third asset is not given. So here is what you are
expected to do for d): 1) Given the payoff vectors for three assets, derive the conditions
for LOOP, NA, NSA to hold respectively. The conditions might be in terms of the three
prices (p1 , p2 , p3 ). 2) After that, plug in (p1 = 4, p2 = 3) in b) to derive the condition
for p3 .
2. Forwards and Options
2.1 Derivatives
While securities represent property rights (hence contracts), derivatives derive their value from
an underlying security. The most popular derivatives are Swaps, Futures and Options. A natural
question to ask is: since a derivative’s value is a function of the underlying security, are derivatives
always redundant assets? We will see that this is not the case in general, and the answer depends on
the fact that functional dependency does not imply linear dependency.
600
Long forward
Short forward
400
200
Payoff
−200
−400
−600
2,600 2,800 3,000 3,200 3,400
Level of S&P 500
Figure 2.1: S&P 500 6-months forward contract, agreed price $3060.
The strike level is normally set so that the value of the contract at initiation is zero. Compared to
an outright long position in the underlying asset, it requires no cash outlay at initiation and gives
the same synthetic exposure to the underlying (that is, if the underlying ends up higher by 10 at
expiration, the forward will pay 10 as well). Assuming that a bond pays $1 at maturity, the relation
between a unit of the underlying and the corresponding forward contract is given by
To see this, suppose the forward contract on a stock S is agreed at time 0 to buy an asset at a future
date T at a specified price K, so that the payoff is (ST − K). Now we ask: what is the fair price of
14 Chapter 2. Forwards and Options
this contract today? If we can replicate this payoff with a combination of stocks and bonds whose
price is known today, we know that the price of the forward is going to be equal to the price of that
combination of stocks and bonds.
Suppose today we buy the stock at price S0 , and sell K bonds maturing at T at a price of e−rT each.
What do we get at time T ? The stock will be worth ST and each bond will be worth 1, so K bonds
will be worth K. Since we sold them, we will have to pay K. In total, we have ST − K. Therefore we
replicated the payoff of the forward contract, and hence the price of the forward contract today will
be equal to the value of the portfolio of long 1 stock and short K bonds today, that is, S0 − Ke−rT ,
that is, the value of the underlying asset minus the strike price times the bond price.
The forward contract can settle in one of two ways: 1) in cash, i.e. parties exchange the difference
between the underlying at maturity and the strike price in dollar value, which is less costly and
more practical than 2) by physical delivery, i.e. parties exchange the underlying at the agreed price.
Credit risk can be an issue for over-the-counter forwards, for which credit checks and bank letters
of credit may be required other than collateral postings, while for exchange-based transaction
counter-party risk is reduced since the clearing house guarantees the transactions.
2.1.2 Options
A call option is a contract that gives the right (not the obligation) to buy an asset at a specified price
on a future date. From the point of view of the buyer, it preserves the upside potential from owning
an asset without the downside risk. The seller has an obligation to sell if the buyer chooses to buy.
The strike price is the price at which the parties agree to exchange the underlying. When the buyer
chooses to buy the asset, we say he exercises the option. The expiration date is the date by which
the buyer has to decide whether to exercise his right. Classified by exercise style, there are three
main classes of options: 1) European options can only be exercised at expiration date 2) American
options can be exercised anytime between inception and expiry and 3) Bermudan options can be
exercised during some specified dates before or on expiration date.
For European options, because the buyer only exercises when the spot price at maturity is higher
than the strike price (otherwise he would rationally buy the security in the market for less). Therefore
the payoff at expiration is given by
max {S − K, 0} ≡ (S − K)+
and the profit is given by the payoff minus the future value of the option price, called option
premium.
■ Example 2.1 Suppose you buy a 6-month call option on S&P 500, with strike price $3,000 and
premium $93.81. Assuming a 2% 6-month risk-free rate, if the index value in 6 month is $3,100
then the profit is given by $3, 100 − $3, 000 − $93.81 × (1 + 2%) = $4.32, but if it turns out to be
$2,900 then the loss is just $93.81 × (1 + 2%) = $95.68. ■
2.1 Derivatives 15
100
$Payoff
0
−100
−200
2,800 2,900 3,000 3,100 3,200
Level of S&P 500
Purchased call
100 Long forward
0
$Profit
−100
−200
−300
2,800 2,900 3,000 3,100 3,200
Level of S&P 500
Similarly, a put option gives the buyer the right but not the obligation to sell an asset at a determined
price on a future date. The seller of a put is obligated to buy if called upon to do so by the buyer.
Similarly to a call option, for a buyer the payoff is max (K − S, 0) ≡ (K − S)+ and the profit is
given by the payoff minus the future value of the option price. It is worthwhile to remark that while
a call option increases in value when the underlying rises in value, the opposite is true for a put
option.
1. Homeowner insurance: because insurance pays only in the case of a damage to the house,
it can be thought of as a put option on the value of the house.
2. Equity-Linked Certificates of Deposit: this contract pays the invested amount plus 70% of
the gain in the S&P 500 index. For instance, suppose we invest $10,000 when the S&P 500
16 Chapter 2. Forwards and Options
n o
S&P500Final
is at 1700, then the payoff would be 10, 000 × 1 + 0.7 max 1700 − 1, 0 .
Below is a short recap of the profit functions of the derivatives discussed so far:
c1 = (0, 1, 2, . . . , S − 2, S − 1)′
c2 = (0, 0, 1, . . . , S − 3, S − 2)′
..
.
Which together with the stock give rise to the security structure
1 ··· 0
0
2 1 · · · 0
X = .
.. . . ..
.. . . .
S S −1 ··· 1
2.3 Forwards Revisited 17
This is an upper triangular S × S matrix whose determinant (the product of diagonal entries) is one.
Therefore X is full rank and markets are complete.
Another way of getting to the same result is pricing by arbitrage. Suppose that at t = 0 we observe
p
that F0,T > S0 , then we could buy the stock today at S0 , sell the forward at F0p and pocket the
p
difference F0,T − S0 > 0. In t = T our stock is worth ST , and we owe ST from the forward contract
we sold in t = 0: as a result, in t = T we receive 0. This would constitute an arbitrage. A similar
p
argument can be used for the case when F0,T < S0 . Therefore, absence of arbitrage requires that
p
F0,T = S0 .
When there are dividends the two pricing arguments do not hold any longer: the holder of the stock,
unlike that of the forward, will now receive dividends in the period [0, T ]. As a consequence, it
p p
must be that F0,T < S0 since F0,T = S0 − PV (Dividend Payments in [0, T ]). In particular:
• With discrete dividends Dti for t1 , ... ,tn ∈ [0, T ] and assuming reinvestment at the risk-free
n n
p
rate, we have F0,T = S0 − ∑ PV0 (Dti ) = S0 − ∑ Dti e−rti
i=1 i=1
p
• With continuous dividends with annualized dividend yield δ , we have F0,T = S0 e−δ T
Note that this only applied to deterministic dividends: when dividends are stochastic the securities
structure turns from a matrix into a cube and we can no longer use our one period model.
2.3.2 Forwards
Obviously the forward price is just the future value of the prepaid forward:
18 Chapter 2. Forwards and Options
Indexes are an example of assets with continuous dividends. We call forward premium the quantity
F0,T
S0 , which can be used to infer the current stock price from the forward price. The annualized
1 F0,T
forward premium is π = T ln S0 .
We can also use a no-arbitrage argument to price a forward: assuming continuous dividends with
rate δ , we can buy e−δ T units of the stock worth S0 for total price of S0 e−δ T by borrowing the full
amount. At t = 0 there is no cash outlay; however, at t = T the portfolio is worth ST − S0 e(r−δ )T .
Since the long forward payoff is ST − F0,T , this implies that to exclude arbitrage it must be the case
that F0,T = S0 e(r−δ )T .
It follows that
Long forward
3000$ bond
3,000
Forward + bond
2,000
Payoff
1,000
An interesting application of this fact is the so-called cash and carry arbitrage: a market maker can
make a risk-less profit by (for instance) selling short a forward contract and going long a synthetic
forward: the payoff from this strategy is F0,T − S0 e(r−δ )T .
A natural question at this point is: is the forward price a market prediction of the future price? The
answer is no: the formula F0,T = S0 e(r−δ )T shows clearly that the forward price provides no more
information than r, δ and S0 .
2.4 Options Revisited 19
Consider two European options, one call and one put, with the same strike K and time to expiry T .
The put-call parity relation requires that
Note that if F0,T = K the long call short put portfolio above is equivalent to a synthetic forward,
and in fact will have zero price.1 With a dividend stream {Dti }ni=1 we can rewrite the above relation
as
Because an American option can be exercised at any time, while a European option can only be
exercised at maturity, it must be the case that
CA (K, T ) ≥ CE (K, T )
PA (K, T ) ≥ PE (K, T )
In general, the American call option price cannot exceed the stock price (otherwise you would
never buy the option but just buy the stock). Moreover, the European call option cannot be lower
than 1) the price implied by put-call parity by setting to zero the put price, or 2) zero, whichever is
highest. That is,2
Similarly, the American put option price cannot exceed the strike price (otherwise you would never
buy the option but just buy the bond). Moreover, the European put option cannot be lower than 1)
the price implied by put-call parity by setting to zero the call price, or 2) zero, whichever is highest.
That is,
1An alternative definition of “at-the-money” option is to say that an option is at the money when the forward price
equals the strike price. Under this definition, a “long call short put” portfolio replicates a forward when the two options
are at the money.
2 Since max e−rT F
−rT F
+
0,T − S0 , 0 = e 0,T − S0 and CE (K, T ) > 0.
20 Chapter 2. Forwards and Options
Rationally, we never exercise an American call option on a stock with no dividends: this is
because3
That is, for a holder of an American option it is always best to sell the option rather than exercise it
early. By LOOP, the price of an American call option on a stock with no dividends is the same as
that of an European option. Note that this is not true for a dividend-paying stock, as well as for an
American put on a non-dividend-paying stock.
An American option (both put and call) with more time to expiration is at least as valuable as
an American option with less time to expiration. This is because the longer option can easily be
converted into the shorter option by exercising it early. European call options on dividend-paying
stock and European puts may be less valuable than an otherwise identical option with less time to
expiration. 4 A European call option on a non-dividend paying stock will be more valuable than an
otherwise identical option with less time to expiration.
Let K1 < K2 , then we know that C(K1 ) ≥ C(K2 ) and P(K1 ) ≤ P(K2 ) (since their payoff is more
likely to be positive at t = T ). A less obvious fact is that C(K1 ) −C(K2 ) ≤ K2 − K1 , because the
maximum payoff for a collar (long option with low strike, short option with high strike) with strikes
K1 < K2 is K2 − K1 , and the price of the collar cannot exceed its maximum payoff. In the same way,
for put options we have P(K2 ) − P(K1 ) ≤ K2 − K1 . Finally, the option price is convex with respect
to its strike: for K1 < K2 < K3 ,
To conclude, it is worthwhile to note that many of these results on option price bounds can be
derived within our (very simple) one period model: they are incredibly robust. Once we adopt a
more specific setting (like the famous Black-Scholes model) we can use more sophisticated tools
such as needs dynamic replication, and the results become deeper but at the same time more hinging
on the specific model used.
2.5 Exercises
1. Similar to slides 12-13 in Lecture slides Part 2 Forwards and Options, please use put options
instead of call options to complete the market given the stock’s payoff as in slide 13. Please
specify the number of put options used and the strike prices. Please write down the payoff
matrix for the assets (with rows as states and columns as assets) and explain why this payoff
matrix indicates a complete market.
2. a) Suppose there are 3 call options traded on a stock with strike prices equal to 40, 50, and
60 and with prices C(40) = 8, C(50) = 6, and C(60) = 2. Show that prices allow for
arbitrage and provide an arbitrage portfolio with initial price equal to zero. What kind
of strategy is this?
b) We have 3 call options on a single stock with strike prices K1 = 100, K2 = 110, and
K3 = 120. We know that C(K = 100) = 10 and C(K = 110) = 8.
i. Can you infer the price of the call with strike 120? Explain your answer.
ii. If C(K = 120) = 6, provide a portfolio strategy with total price <= 0 and positive
payoff at maturity.
iii. For standardization, assume we could ONLY long/short EXACTLY ONE call
with strike K = 120. Please specify the position, amount and strike price for all
instruments used in your strategy.
3. a) For European options with the same strike price K and time to expiration T , let C(K, T )
be the price for such call and P(K, T ) the price for such put at time 0. Construct an
arbitrage strategy when we have C(K, T ) > P(K, T ) + e−rT (F0,T − K). Here F0,T is
the forward contract price on the same underlying stock that expires at T , and r is the
risk-free interest rate if we borrow/lend money. For simplicity and standardization,
assume that:
• For the European call, we can ONLY long/short ONE unit.
• Let the arbitrage strategy’s payoff at time T be ZERO under all circumstances.
Please specify the amount and position for all instruments (European call, etc.) in your
strategy.
b) Suppose a stock index is currently trading at $300, the dividend yield on the index is 3%
per annum, and the risk-free interest rate is 8% per annum. What is the lower bound for
the price of a 6-month European call option on the index when the strike price is $290?
c) Now assume a stock currently sells for $32. A 6-month call option with a strike of
$30 has a premium of $4.29, and a 6 month put with the same strike has a premium
of $2.64. Assume a 4% continuously compounded risk-free rate. What is the present
value of dividends payable over the next 6 months?
22 Chapter 2. Forwards and Options
† 5. Suppose there exist only a risk-free asset x1 = (1, 1, . . . , 1)′ and a risky asset x2 ̸= x1 and S
states of the world. Let p1 and p2 be the prices of these two assets. A forward contract on
the stock is an agreement to pay an amount F at a future date t = T in exchange for the
payment xsj when the state s ∈ {1, 2, . . . , S} realizes, with no cash flow exchange at time t = 0.
Assuming arbitrage opportunities are ruled out, find the fair value of F.
3. State Prices
v(z) ≡ {p · h : z = Xh}
If LOOP holds then v is a linear functional, that is, a function mapping ⟨X⟩ onto R such that:
3. v(0) = 0, since Xh = 0 always has the solution h = 0 and because v is single-valued (by
LOOP) it must be that p · h = p · 0 = 0 is the only price of the payoff z = 0.
The converse is also true: if there exists a linear functional v defined in ⟨X⟩, then LOOP holds.
Recall that es ∈ RS is the standard basis: a vector with the sth entry equal to 1 and all other entries
1 That is, for all z1 , z2 ∈ ⟨X⟩ and α, β ∈ R such that αz1 + β z2 ∈ ⟨X⟩ it holds that v(αz1 + β z2 ) = αv(z1 ) + β v(z2 ).
24 Chapter 3. State Prices
equal to zero. The next proposition addresses the relationship between v, V and q.
Proposition 3.1.1 if LOOP holds and q is a vector of state prices, then V (z) = q · z for all z ∈ ⟨X⟩.
To see this we only need to show that also for z ∈ ⟨X⟩ we have V (z) = q · z (= v(z)). Suppose that q
is a vector of state prices and LOOP holds, then for z ∈ ⟨X⟩ we have
!
J J J S
xj ·q hj = ∑ ∑ xsj · qs
v (z) = p · h = ∑ p jh j = ∑ hj =
j=1 j=1 j=1 s=1
!
S J S
= ∑ ∑ xsj · h j qs = ∑ zs qs = q · z
s=1 j=1 s=1
Moreover the converse is also true, and therefore the valuation function V (z) = q · z is a linear
functional for all z ∈ RS if and only if q is a vector of state prices and LOOP holds.
■ Example 3.1 Below is a graphical example of state prices. Given the securities structure
!
1 2
X=
1 1
3p1 − p2 = q1 + 2q2
by LOOP. ■
c2
c1
Figure 3.1: Replicating x3 with 3 units of bond and −1 (short) unit of the risky asset.
3.1 Pricing in One Period Model 25
Proposition 3.1.2 Security prices exclude arbitrage if and only if there exists a valuation functional
with q ≫ 0.
It is hard to overstate the importance of this theorem: the absence of arbitrage is equivalent to the
existence of a vector of positive state prices.
z2
⟨X⟩
(1, 1)
p(1, 1)
q̂ q∗
q z1
p(1, 1)
Figure 3.2: Pricing kernel is the unique projection of any state price vector q on ⟨X⟩.
In the picture above, the orange line is the set of q’s consistent with prices and the security structure.
Of all q ∈ R2 consistent with no arbitrage however, there is a very special one that also belongs to
⟨X⟩: it is the unique projection of any state price vector q on ⟨X⟩. In our example, this would be
′
achieved for q1 = p21 since then q = p21 , p21 = p21 × (1, 1)′ ∈ ⟨X⟩. We call this state price pricing
kernel and denote it as q∗ .
Now we are ready to state the important relation between state prices and market complete-
ness:
Proposition 3.1.4 Under no arbitrage, markets are complete if and only if there exists a unique
valuation functional.
26 Chapter 3. State Prices
Proof. If markets are complete and there is no arbitrage, then for a given Xand p the system
X ′ q = p has a unique solution q ≫ 0 (positivity follows from the assumption of no arbitrage).
We prove the if direction by contrapositive statement. If markets are not complete, then there exists
a vector v ∈ RS such that v ̸= 0 and X ′ v = 0. Under no arbitrage, there must exist some positive
q ≫ 0 and some α ∈ R such that q + αv ≫ 0 and X ′ (q + αv) = p (provided α is small enough).
Since X ′ (q + αv) = p this is also a valid state price vector. Hence, there are an infinite number of
strictly positive state prices. ■
Note that p j = E m · x j = E [m] E x j + Cov m, x j , and since for a risk-free bond xsb = 1 for
all s, we have pb = E [m] = R1f where R f is the gross risk-free return. Therefore, for any asset j,
E[x j ]
p j = R f +Cov m, x j . Typically, Cov m, x j < 0.
j 1
Defining R j ≡ xp j , we get E m · R j = 1. Since for a risk-free bond R f = E[m]
, we can write
j f
E m· R −R = 0, or
E m · R j − R f = E [m] E R j − R f +Cov m, R j = 0
That is,
Cov m, R j
E Rj −Rf = −
E [m]
which implies that the excess return for a generic asset j is determined solely by the covariance
with the stochastic discount factor. This also means that an investor is only compensated (with a
higher return) for holding systematic risk, not idiosyncratic risk.
q∗
1
∗
π.1 ∗
Consider the stochastic discount factor obtained from the pricing kernel m ≡ .
. . Note that m
q∗S
πS
3.2 Asset Pricing Formulas 27
which means that there exists a vector h∗ ∈ RJ such that m∗ = Xh∗ . Therefore for any asset j we
can write
p j = E m∗ · x j
p = E X ′ m∗ = E X ′ Xh∗ = E X ′ X h∗
y = Xb + ε
We find the linear combination of X that is “closest” to y by minimizing the variance of the residual
ε. We do this by forcing the residual to be “orthogonal” to X, E [Xε] = 0. The projection of y
onto X is defined as the fitted value Xb = X (E [X ′ X])−1 E [X ′ y]. This idea is often illustrated by a
residual vector ε that is perpendicular to a plane defined by the variable X. Thus, when the inner
product is defined by a second moment, the operation “project y onto X” is a regression.
Finally, note that we can represent our previous discussion for state prices by shrinking the axes by
a factor π:
z2
⟨X⟩
p(1, 1)π2
(1, 1)
m∗
m
z1
p(1, 1)π1
∗
π.1
Consider the stochastic discount factor obtained from the pricing kernel m ≡ .
. , and define its
q∗S
πS
m∗
return as R∗ = pm∗ ≡ αm∗ for α > 0 3 . Then we can write
Cov R∗ , R j
E Rj −Rf = −
E [R∗ ]
Cov(R∗ ,R j )
Defining β j ≡ Var(R∗ ) we can write for the asset j:
Var (R∗ )
E R j − R f = −β j
E [R∗ ]
Var (R∗ )
E [R∗ ] − R f = −
E [R∗ ]
E R j − R f = β j E [R∗ ] − R f
which, if we assume a linear model for R j and R∗ can be specified and tested empirically as
Rkj − R f = β j R∗k − R f + εk
1
EQ x j
3. Equivalent Martingale Measure: p j = 1+r f
As a last remark, note that whenever markets are incomplete, the multiplicity of state price vectors q
translates directly into the multiplicity of stochastic discount factors m and of equivalent martingale
measures π̂.
Fix a strike K > 0, and a (small) number ε > 0. Consider the following portfolio:
State Payoff
ST ∈ [0, K − ε) 0
ST ∈ [K − ε, K) x (ST − (K − ε)) = x (ST − K) + xε
ST ∈ [K, K + ε) x (ST − (K − ε)) − 2x (ST − K) = −x (ST − K) + xε
ST ∈ [K + ε, +∞) x (ST − (K − ε)) − 2x (ST − K) + x (ST − (K + ε)) = 0
The payoff of this portfolio (symmetric butterfly) looks like the following:
Portfolio Payoff
xε
K −ε K K +ε ST
Figure 3.4
This portfolio corresponds the payoff of an Arrow-Debreu security for state ST = K if two conditions
are satisfied:
30 Chapter 3. State Prices
1. The area under the payoff diagram (the sum of all possible payoffs) must be equal to one.
2. The option portfolio must yield a payoff of zero for all values of the underlying asset different
from K.
(xε)×(2ε) 1
The area under the payoff diagram is equal to 2 = xε 2 : to achieve (1), we set x = ε2
; to
achieve (2), we take the limit for ε → 0. Now that we have replicated the payoff of an Arrow-Debreu
security, we can use the law of one price to determine the state prices: for a fixed ε > 0 the price of
the portfolio is
C (K + ε, T ) − 2C (K, T ) +C (K − ε, T )
V (K, ε) =
ε2
and therefore
C (K + ε, T ) − 2C (K, T ) +C (K − ε, T ) ∂ 2C (K, T )
q (K) = limV (K, ε) = lim =
ε→0 ε→0 ε2 ∂ K2
∂ 2C(K,T )
We established an important fact: because q(K) > 0 if and only if ∂ K2
> 0, by the fundamental
∂ 2C(K,T )
theorem of finance there is no arbitrage if and only if for all K > 0 we have ∂ K2
> 0, that is, if
option prices are convex.
∂ 2C(K,T )
There is another important relationship related to state prices and ∂ K2
. We can write the option
price using the equivalent martingale measure representation as follows:
∂C (K, T )
= −e−rT EQ [I {ST − K ≥ 0}] = −e−rT PQ {ST ≥ K} = −e−rT 1 − FSQT (K)
∂K
where FSQT is the cumulative distribution function for the random variable ST under the equivalent
martingale measure Q.4 Differentiating again with respect to K we get
∂ 2C (K, T )
= e−rT fSQT (K)
∂ K2
∂ 2C (K, T )
2
= q(K) = e−rT fSQT (K)
∂K
From the findamental theorem of finance, it follows that there exists a (positive) equivalent martin-
gale measure density function fSQT if and only if no arbitrage holds.
4We define I {ST − K ≥ 0} as a function that is equal to one if ST − K ≥ 0 and zero otherwise. Note that we can take
the partial derivative of C(K, T ) with respect to K since EQ (ST − K)+ = R fSQT (x) (x − K)+ dx = K∞ fSQT (x) (x − K) dx
R R
is smooth in K.
3.3 Recovering State Prices from Option Prices 31
Note that this implies that we can evaluate any future payoff h(ST ) in 3 ways:
∂ 2C (K, T )
Z Z Z
−rT −rT
p=e Q
E [h (ST )] = e fSQT (x) h (x) dx = (x) h (x) dx = q(x)h (x) dx
R R ∂ K2 R
• Absence of arbitrage
• Existence of a positive equivalent martingale measure density function
• Convexity of options
Note that in the market we only directly observe the option prices for some discrete and finite set
of strikes: but if all the option prices C(K, T ) for the strikes K ∈ {0, ∆, 2∆, . . . , N∆} are observable
∂ 2C(K,T )
(for N large and ∆ > 0 small enough) we can approximate ∂ K2
in the following way
∂ 2C (K, T ) C (K + ∆, T ) − 2C (K, T ) +C (K − ∆, T )
≈
∂ K2 ∆2
and therefore we can also calculate the empirical market-implied probability distribution of ST ,
fSQT , and the empirical state price density q.
Going back to our finite-state one-period model, if all the option prices C(s, T ) for the strikes
s ∈ {0, 1, 2, . . . , S} are observable, we have
π̂s
△2C (s, T ) ≡ C (s − 1, T ) − 2C (s, T ) +C (s + 1, T ) = qs =
(1 + r f )
32 Chapter 3. State Prices
3.4 Appendix
The following is an alternative derivation from option prices of the state price density function. We
can construct the following portfolio: for some ε > 0 and δ > 0 and a fixed ŜT
Payoff
ε
ŜT − δ2 ŜT + δ2
ŜT − δ2 − ε ŜT + δ2 + ε ST
Figure 3.5
h i
1
If we buy ε units of this portfolio, when ST ∈ ŜT − δ2 , ŜT + δ2 the total payoff is equal to 1. The
total value of this portfolio is
1 δ δ δ δ
C K = ŜT − − ε, T −C K = ŜT − , T −C K = ŜT + , T +C K = ŜT + + ε, T
ε 2 2 2 2
Finally, dividing by δ and letting δ → 0, we obtain the continuum-states version of a vector of state
∂ 2C(K=ŜT ,T )
prices, the state price density function ∂ K2
.
We now have the technology to price this. Its value will be equal to the integral of the state price
density over the interval [1700, 1750], that is, for T = 1,
Z 1750 2
∂ C(K, T ) ∂C (K = 1750, 1) ∂C (K = 1700, 1)
dK = −
1700 ∂ K2 ∂K ∂K
Finally, note that this is equivalent to a portfolio comprising a long position in a binary 1700 call
and a short position in a 1750 binary call.
3.5 Exercises 33
3.5 Exercises
2. Suppose there are S possible states of the world in t = T and each has a (physical) probability
of occurrence ηs > 0 with for ∑Ss=1 ηs = 1. Consider the vector µ ∈ RS with s = 1, ..., S,
µs = ηqss where q is the known state-price vector. Write E(y) = ∑Ss=1 ηs ys for any y ∈ RS .
a) Consider an asset with payoff x = (x1 , ..., xS ). Show that the price of this asset must be
E(z), where zs = µs xs . Interpret this result.
xs
b) Let the rate of return for an asset with price p > 0 in state s be rs = p and let ws = rs µs .
Show that E(w) = 1. Is there some function of the excess return of the asset r − r f such
1
that E[ f (r − r f )] = 0? (Here we define r f such that, rf
= ∑Ss=1 qs ).
3. Consider a market where there are 3 assets with payoff x1 = (1, 0, 1, 2)′ , x2 = (0, 1, 1, 3)′ , x3 =
(0, 1, 4, 0)′ . Assume p1 = 0.8, p2 = 1.1 and p3 = 1.4, find the state price vector q =
(q1 , q2 , q3 , q4 )’.
a) Is q unique? If not, please express the q with free variables starting from q4 . (i.e., if
there is one free variable, express q1 , q2 , q3 with q4 ; if there are two, express q1 , q2 with
q3 , q4 , ...).
b) Now assume q4 = 0.25 in this sub-question only. What is q now? Does LOOP hold? Is
there an arbitrage opportunity? Please explain.
c) Now assume the risk-free interest rate is 25% for this sub-question only. (And the
risk-free bond is available in the market as well.) What is q now? Is there an arbitrage
opportunity? Is the market complete? Please explain. And please use one-period
discrete discounting here (instead of involving continuous exponential terms).
d) What are the ranges of the free variables from your answer in part 1, if there is no
arbitrage? What is the price range for asset X = (1, 0, 0, 1)′ under NA?
4. Suppose there exist 3 states of the world s = 1, 2, 3 and 2 assets x1 = (2, 1, 0) and x2 = (0, 1, 0).
a) Suppose p1 = 1 and p2 = 0.3. What state prices are consistent with these prices?
b) Solve for the unique pricing kernel q∗ .
c) Use the pricing kernel to value a third asset x3 = (0, 1, 1).
d) Now suppose p3 = 0.6. Solve for the state price vector. Does this market permit
arbitrage?
e) Solve for the stochastic discount factor assuming the physical probability is such that
1
π1 = π2 = π3 = 3
f) Solve for the distribution (π̂) under the equivalent martingale measure.
34 Chapter 3. State Prices
5. Suppose there are three assets and three states. The payoff matrix is X = [X1 , X2 , X3 ] ∈ R3×3 ,
where the columns represent assets and rows represent states. Let the assets have linear
independent payoffs. Furthermore, the price vector for the three assets is p ∈ R3 .
a) Please express the state price vector q∗ with X and p, so that it belongs to ⟨X⟩.
Now for the rest of the sub-questions, let X1 = (1, 0, 1)′ , X2 = (0, 1, 1)′ , X3 = (0, 0, 1)′ ,
and p = (2, 1.5, 1)′ .
6. Suppose that the function that determines the price of a call option with its strike is of the
form C(K) = aK 2 + bK + c.
a) Without any computation find three constraints on the parameters for no arbitrage.
(Hint: zero/first/second-order properties for the call price.)
b) Please first write down the state price density function, then price the following options
in terms of a, b and c:
i. A wedding cake option, where the option which gives the payoff 1 if the ST is
between 2200 and 2600, and 0 otherwise.
ii. Price the double barrier option that gives 0 if ST is below 2300 or above 2500 and
gives ST2 if ST is between 2300 and 2500.
iii. Price the put option on the S&P500 with strike K = 2300 and maturity 6 months.
Explain why this price could be off given our assumptions.
(Comment: This question is for illustration only so that you could get familiar with
using state price density to price certain instruments. The assumptions for C(K) in this
question are actually not quite appropriate)
† 7. Suppose a stock index is currently trading at $25, and there are 5 possible states of the world
in t = T such that ST ∈ {15, 20, 25, 30, 35}.
a) Given a zero risk-free interest rate, describe a valid equivalent martingale measure.
b) Under this measure, price call options at K = 15, 20, 25, 30, 35.
c) Use this information to recover state prices.
† 8. Show in detail how to retrieve state prices using put options both in a continuous and discrete
states setup.
4. Empirical Multi-Factor Models
This chapter has not been typed. See the lecture slides for the coverage of this material.
4.1 Exercise
On Canvas you can find two files, 25_Portfolios_5x5.csv and F_F_Research_Data_Factors.csv.
These files are downloaded from Kenneth French’s website that provides a large set data series
useful for empirical, multi-factor asset pricing. As always feel free to collaborate on this question
but try to write your own code.
1. Open the two files in a text editor to read the descriptions and get an idea of the file format.
Have a look at Kenneth French’s website,
[Link]
and locate the two files that we are working with.
2. We will work with the value weighted monthly portfolio returns and the monthly factor series
from July 1926 to July 2022. Use your favorite language/software to read in the data.
3. We start from the hypothesis that mt∗ = a + b(RtMkt − Rtf ). Estimate the regression
for each portfolio i. Is mt∗ in the payoff space? What does β̂i measure statistically? Based on
our hypothesis, what value should α̂i take approximately?
4. Estimate the average excess return of each portfolio i, R̄i − R̄ f . Regress those average returns
on the previous regression coefficients β̂i , so run the following regression:
R̄i − R̄ f = αi + λ β̂i + εi .
Is your estimate of λ significantly different from zero? What is the R2 of this regression?
Should we reject our initial hypothesis that mt∗ = a + b(RtMkt − Rtf )?
5. Split the sample into three different episodes, 1926 − 1964, 1965 − 2004, and 1983 − 2022.
Does the model work better in any of these sub-samples?
6. Repeat the two-stage procedure for the three factor model, based on the new hypothesis that
Are the three λ coefficients significant? How does the model fare across sub-samples? Do
the λ ’s remain stable across samples?
5. The Multi-Period Model
In the first four chapters, we assume that there is only one period (between t = 0 and t = 1). A
natural question to ask is then, how can we deal with a multi-period setup in which several outcomes
are possible after each outcome realizes? One way to tackle this problem is to bring down the
problem to a set of many “smaller” one-period model problems:
s2,1
A1,1
s2,2
A0
s2,3
A1,2
s2,4
t =0 t =1 t =2
Figure 5.1: The structure of security payoffs—two periods—two states at each node
With this approach however, we need a more efficient way to define and work with the information
available at each time step. A natural way to approach this is using the concepts of σ -algebra and
filtration.
T
Definition 5.1.4 — Filtration. A filtration is a collection of algebras FT = {Fs }s=t such that if
u ≤ v then Fu ⊆ Fv .
T
Definition 5.1.5 — Adapted Process. A stochastic process {Yt }t=0 is adapted to the filtration
FT = {Fs }Ts=t if Ys is measurable with respect to Fs for s = t, . . . , T .
Adaptation is the requirement that people cannot see in the future. Moving from a static to a
dynamic setting, we need to adapt our notation:
We will use trees to illustrate dynamics. In Figure 5.3, the tree has 7 events. Every event has a
positive probability attached to it, and we impose the consistency requirement that the probability
of any event equals the sum of the events that follow from its node: for example, denoting πt,s the
probability of At,s , we have π1,1 = π2,1 + π2,2 . Clearly, for each t we have ∑πt,s = 1.
s
A2,1 = s2,1
A1,1
s2,2
A0
s2,3
A1,2
s2,4
t =0 t =1 t =2
Figure 5.3: The structure of security payoffs—two periods—two states at each node
There are two ways to reduce this seemingly more complicated model to the one-period model.
The first is to boil down every event to a final state, and apply what we know from the one-period
model. Note that if we simplify the tree in this way it follows that to achieve completeness we will
need six independent assets, one for each final state.
s2,1 A1,1
A1,1 A1,2
s2,2 s2,1
A0 A0
s2,3 s2,2
A1,2 s2,3
s2,4 s2,4
t =0 t =1 t =2 t =0 t =2
Figure 5.4
The second way is to reduce each branch of the tree to a one-period model. In this case, assets can
40 Chapter 5. The Multi-Period Model
be re-traded conditionally on the the occurrence of event A1,1 or A1,2 . Completion can be achieved
with two categories of assets: short-lived assets pay off only in the period following the one in
which they are traded, while long-lived assets pay off over many periods.
s2,1 s2,1
A1,1 A1,1
s2,2 s2,2
A0 A0
s2,3 s2,3
A1,2 A1,2
s2,4 s2,4
t =0 t =1 t =2 t =0 t =1 t =2
Figure 5.5
A short-lived asset is an asset that pays out only in the period immediately after the asset is issued.
Without uncertainty and with T one-period assets (that is, T assets that can be bought in period
t − 1 at price pt and pays off in period t for t = 1, . . . , T ), completeness requires that we are able
to transfer wealth between any two periods. For example, if zero coupon bonds are traded in
each period and T = 2, at time t = 0 we can buy a quantity p2 of bonds that pay off at time t = 1
at a total cost of p1 p2 , and then in period t = 1 we can invest the proceeds from the bond we
bought to buy one unit of the bond that pays at time t = 2: in this way we replicated a bond that is
bought at t = 0 and pays $1 at time t = 2, and therefore markets are complete. In general, for T
periods the cost of rolling over this kind of short-lived bond strategy is just pt · pt+1 · . . . · pT −1 , and
completeness requires that we are able to transfer wealth between any two periods t and t ′ (not just
consecutive).
With uncertainty the reasoning is very similar. Suppose the event tree is the one in Figure 5.6 and
we want to replicate the payoff of an asset that is bought at t = 0 and pays $1 if event A2,2 = s2,2
realizes. Let q2,2 denote the state price corresponding to A2,2 traded in t = 1. Then, working
backwards, in event A1,1 we can buy a one-period asset that pays $1 in event A2,2 for a price q2,2 ; to
be able to dispose of an amount q2,2 if event A1,1 realizes, in event A0 we need to buy a quantity q2,2
5.2 Dynamic Trading and Market Completeness 41
of the asset that pays $1 if event A1,1 realizes, which costs q1,1 per unit. Thus the price of the payoff
of an asset that is bought at t = 0 and pays $1 if event A2,2 realizes is q1,1 · q2,2 , and we can repeat
the argument for all other events to see that this market achieves dynamic completeness.
s2,1
A1,1
s2,2
A0
s2,3
A1,2
s2,4
t =0 t =1 t =2
Figure 5.6
With long-lived assets, consider a T -period model without uncertainty. There is a single asset
paying $1 in t = T , which is tradable in each period for a price pt : this means that sequentially,
there are T prices for this asset. The payoff can be transferred from period s to period t > s by
purchasing the asset at time s and selling it in period t. For example, suppose there are two assets
that pay off like in Figure 5.7.
Asset 1 Asset 2
s2,1 1 0
p11,1
A1,1
p21,1
s2,2 0 1
A0
s2,3 1 0
p11,2
A1,2
p21,2
s2,4 0 1
t =0 t =1 t =2
Figure 5.7
42 Chapter 5. The Multi-Period Model
j
The assets are long-lived, since they can be re-traded at t = 1 for the price pt,s . Note that we have
two assets, but each assets can be traded in three events; moreover, the price of each asset can be
endogenously determined at time t = 0. Although one may be tempted to think that because we
only have 2 assets markets cannot be complete, in fact dynamic trading provides a way to transfer
wealth in all events:
Asset 1 Asset 2
s2,1 1 0
p11,1
A1,1
p21,1
s2,2 0 1
p10
A0
p20
s2,3 1 0
p11,2
A1,2
p21,2
s2,4 0 1
t =0 t =1 t =2
Figure 5.8
Call trading strategy [ j, At,i ] the cash flow of asset j that is purchased in event At,i and sold one
period later. There are six such trading strategies: [1, A0 ], [1, A1,1 ], [1, A1,2 ], [2, A0 ], [2, A1,1 ], [2, A1,2 ].
For example, trading strategy [1, A1,1 ] costs p11,1 and pays out $1 in the first final state and zero in
all other events; and trading strategy [1, A0 ] costs p10 and pays out p11,1 in event 1, p11,2 in event 2,
and zero in all the final states.
Strategy [1, A0 ] [2, A0 ] [1, A1,1 ] [2, A1,1 ] [1, A1,2 ] [2, A1,2 ]
Event A0 −p10 −p20 0 0 0 0
Event A1,1 p11,1 p21,1 −p11,1 −p21,1 0 0
Event A1,2 p11,2 p21,2 0 0 −p11,2 −p21,2
State s2,1 0 0 1 0 0 0
State s2,2 0 0 0 1 0 0
State s2,3 0 0 0 0 1 0
State s2,4 0 0 0 0 0 1
The payoff matrix is nonsingular (and hence markets are complete) if and only if the subma-
trix
5.3 The Multi-Period Stochastic Discount Factor 43
p11,1 p21,1
p11,2 p21,2
is nonsingular (that is, it has rank equal to 2). The components of this submatrix are the prices
of the two assets conditional on period-1 events. There are cases in which these are colinear in
equilibrium: for instance when the per capita endowment is the same in event A1,1 and A1,2 , in state
s2,1 and s2,3 , and in state s2,2 and s2,4 , and if the probability of reaching state s1,1 after event A1,1 is
the same as the probability of reaching state s2,3 after event A1,2 then the submatrix is singular (but,
one might argue, then events A1,1 and A1,2 are effectively equivalent and could be collapsed into a
single event). Moreover, a random square matrix is regular: this means that generically, the market
is dynamically complete (that is, it is not complete only for information trees and asset structures
which have zero probability of showing up in any application).
We call branching number the maximum number of branches that fan out from any event in the
uncertainty tree: it turns out that this is also the number of assets necessary to achieve dynamic
completeness. This generalizes to the continuum of events and continuous time case, in which a
small number of assets is sufficient to achieve completeness because the need for many assets to
achieve completeness that arises from a large number of possible events is offset by the ability to
trade continuously, thus generating a large number of trading strategies. The classic example is the
Black-Scholes formula: Cox, Ross and Rubinstein showed that is is possible to build a binomial
tree model of Black-Scholes, with constant interest rates and in which the stock can only go up or
down, in which the market is dynamically complete with just two assets: the stock and the risk-free
bond. Thus an option on the stock can be replicated with dynamic delta-hedging.
Recall that no arbitrage requires that there exists a positive stochastic discount factor mt+1 for each
subperiod t such that
Mt+1 = m1 · m2 · . . . · mt+1
get
∞
Assuming for this asset there is a stream of cash flows {xt }t=1 where each xt is a random payoff,
we have for time t = 0
p0 = E0 [M1 (p1 + x1 )]
similarly
M1 p1 = E1 [M2 (p2 + x2 )]
The Law of Iterated Expectations (LIE) states that, because of the property of filtrations that if u ≤ v
then Fu ⊆ Fv , for any random variable X the time-u expectation of the random variable “time-v
expectation of X” is equal to the time-u expectation of X. The intuition is as follows: consider your
current guess of what the weather will be on Sunday, and compare it to the guess that you will have
on Saturday about the weather on Sunday. LIE states that your current guess on your guess for the
Sunday weather is the same as your current guess for the Sunday weather. Therefore we get
p0 = E0 [M2 p2 + M2 x2 + M1 x1 ]
k
p0 = E0 [Mk pk ] + ∑ E0 [Mt xt ]
t=1
∞
p0 = ∑ E0 [Mt xt ]
t=1
∗ = proj (m ∗
In terms of projections, recall that mt+1 t+1 | ⟨Xt+1 ⟩), that is, there exists some ht such
∗ =X
that mt+1 ∗
t+1 ht and
h ′ i h ′ i
∗
pt = Et Xt+1 mt+1 = Et Xt+1 Xt+1 ht∗
So that
h ′ i−1
ht∗ = Et Xt+1 Xt+1 pt
5.4 Martingales 45
And therefore
h ′ i−1
∗
mt+1 = Xt+1 Et Xt+1 Xt+1 pt
Clearly we also have Mt∗ ≡ m∗1 · m∗2 · . . . · mt∗ ∈ ⟨Xt ⟩. Further, we can express mt+1
∗ in terms of the
covariance of returns Σt :
∗
∗ ∗ ′
mt+1 = Et mt+1 + pt − Et mt+1 Et [Xt+1 ] Σt−1 [Xt+1 − Et [Xt+1 ]]
e
And also in terms of covariance of excess returns Ψt ≡ Covt Rt+1 :
∗ 1 1 e ′ −1 e e
mt+1 = − Et Rt+1 Ψt Rt+1 − Et Rt+1
Rtf Rtf
Recall that if the one-period stochastic discount factor mt is not time-varying (that is, the distribution
of mt is i.i.d for each t), then the expectations hypothesis holds and the investment opportunity set
does not vary. The corresponding Rt∗ of a single-factor state-price beta model is then relatively easy
to estimate, since over time we collect more and more realizations of Rt∗ . However, if mt (or the
corresponding Rt∗ ) is time-varying, then we can assume that it depends on some state variable and
need a multi-factor model to account for it. Indeed, suppose that Rt∗ = R∗ (zt ), where zt is some
state variable. For example, suppose that zt can be either 1 or 2 with equal probability. Then we
could take all the periods in which zt = 1 and back out R∗ (1), and find R∗ (2) in a similar way. Is it
possible to do so in reality? The answer is no, because we have hedges across state variables.
5.4 Martingales
Let {Xt } be a stochastic process and {xt } a sequence of its realizations. We say that {Xt } is a
martingale if E [Xt+1 | Xt = xt , Xt−1 = xt−1 , . . . , X1 = x1 ] = xt . Paul Samuelson argued in 1965 that
asset prices have to be martingales in equilibrium:
1
pt = EtQ [pt+1 + dt+1 ]
1 + rtf
This is trivial assuming no dividends, no discounting and risk-neutral agents. Let’s examine these
cases one by one.
With discounting, in order for the price process to follow a martingale we simply need:
Et [δ pt+1 ] = pt
2Where Zt is a Brownian Motion.
46 Chapter 5. The Multi-Period Model
With dividend payments, things are more complicated because pt depends on the dividend of the
asset in period t + 1 but pt+1 does not (it is the ex-dividend price). However, for example, the value
of a fund that keeps reinvesting the dividends follows a martingale (LeRoy 1989): suppose the fund
owns nothing but one unit of asset j. Assuming agents are risk-neutral, the value of the fund at
time 0 is
" #
h i
t j j
f0 = p0 = E ∑ t
δ x = δ E p 1 + x1
t≥1
After receiving a (state contingent) dividend x1j , the fund buys more of asset j at price p1 , thus
j
x1
owning 1 + p1 units of the asset. The expected discounted value of the fund is then
" !#
xj h i
E [ f1 ] = E δ p1 1+ 1 = E δ p1 + x1j = p0 = f0
p1
A similar statement is true of the agent is risk-averse. The difference is that we have to discount with
the risk-free interest rate instead of the agent’s discount factor, and use risk-neutral probabilities
(by now you should have an idea about why it is also called “equivalent martingale”) instead of
objective probabilities. Just like in the one-period model, we define the risk-neutral probability of
event A as
πt,s Mt,s
π Q (At,s ) =
ρt
−1
t
f
where ρt = ∏ Rs is the risk-free discount factor between 0 and t. The initial value of the
s=0
fund is
" # " # " !#
t
f0 = p0 = E ∑Mt xtj =E M1 x1j + ∑ Mt xtj = E m1 x1j + ∑∏ mt xtj =
t≥1 t≥2 t≥2s=2
h i h i
= E m1 x1j + p1 = E M1 x1j + p1
Therefore, the properly discounted (ρ instead of δ ) and properly expected (π Q instead of π) value
of the fund is indeed a martingale.
We can now introduce a fifth asset pricing formula. Let Ps (t, T ) be the time-s price of a (zero-
coupon) bond to be purchased at time t and with maturity T . We know the pricing relation
5.4 Martingales 47
is
Dividing side by side the pricing relation for a generic asset j and rearranging we get
j j " j j
#
ptj Pt+1 (t + 1, T ) mt+1 xt+1 + pt+1 xt+1 + pt+1
F T
= Et = Et
Pt (t, T ) Pt+1 (t + 1, T ) Et [mt+1 Pt+1 (t + 1, T )] Pt+1 (t + 1, T )
Where the expectation is taken with respect to the risk-forward measure FT for which πsFT =
S
πs Pt+1 (t+1,T )mt+1
Et [mt+1 Pt+1 (t+1,T )] (note that πsFT ≥ 0 and ∑ πsFT = 1 guarantee that π FT is a probability distribution).
s=1
This asset pricing formula was speficically devised to price bond options, and boils down to the
risk-neutral measure if t + 1 = T like in the one-period model case.
48 Chapter 5. The Multi-Period Model
5.5 Exercises
1.00 pu1
1.10 p0 3.30
1.00
1.20 5.00
1.00 pd1
1.20 3.50
Note that no asset pays dividends. The price of the one-period risk-free bond is normalized
1
to 1 in each period. Suppose that the risk-neutral probability of an “up” tick is 3 and that of a
2
“down” is 3, for every point in time.
a) Is the market statically complete? Is the market dynamically complete?
b) What is the event price of the event {up, up} and of the event of only one up, i.e.
{down, up} ∪ {up, down}?
Now suppose we change the process in the following manner:
1.10 6.60
1.00 pu1
1.10 p0 3.30
1.00
1.00 5.00
1.00 pd1
1.00 3.50
c) Is the market statically complete? Is the market dynamically complete?
2. Consider a security market model with three dates,t = 0, 1, 2 and five states of the world,
s = 1, ..., 5. Investors have no information at time 0 and full information at time 2. At the
intermediate date, their information partition consists of the two sets A = 1, 2, 3 and B = 4, 5.
There are two assets with the following price and dividend process:
3. Consider a security market with three dates, t = 0, 1, 2, and five states of the world, s = 1, . . . , 5.
The information structure is described by the following sequence of partitions of the state
space:
P0 = {1, ..., 5}
P1 = {A, B,C}
P2 = {{1}, {2}, {3}, {4}, {5}}
where A = {1, 2}, B = {3, 4},C = {5}. There are four securities with terminal payoffs given
by the matrix
2 4 5 1
1 1 2 1
D=
3 2 4 1
4 5 7 1
5 6 10 1
with the j-th column corresponding to the j-th security. The prices of these securities at date
t = 0, 1 are given by
3.6 4 6.6 1
1.5 2.5 3.5 1
3.25 2.75 4.75 1
5 6 10 1
where the j-th column lists the price of the j-th security at date 0, event A, event B, and event
C, respectively. The securities do not pay dividends prior to the terminal date.
a) Verify that this securities market permits no arbitrage.
b) Is the market dynamically complete?
c) Compute the initial no-arbitrage price of the following three securities:
i. A call option on security 1 with an exercise price of 3.5.
ii. A down-and-under call option on security 1 with an exercise price of 3.5. This is a
call option with an extra provision - if the price of security 1 ever drops below the
exercise price, then option becomes worthless.
iii. A convertible security that in each state pays at the terminal date the largest payout
of the securities 1-4 in that state.
6. Multi-Period Model: Options
Consider a European call option with strike K and maturity T . Its payoff is (ST − K)+ , with no
cash flows between t = 0 and t = T . Unfortunately we are unable to statically replicate this payoff
using just the stock a risk-free bond: we need to dynamically hedge: that is, we need to engage in a
strategy where the required stock (and bond) position changes for each period, until maturity. Since
the replication strategy depends on specified random process of stock price, we need to impose a
specific model for the evolution of the stock. In a discrete time setting, the canonical model is the
Cox-Ross-Rubinstein binomial model. We assume that:
uS
S
dS
Suppose we buy (“go long” in trading jargon) ∆ stocks and B bonds. The payoff from portfolio
C0 = ∆S0 + B is
∆uS + B · R f if the stock goes up
f 0
C1 = ∆S1 + B · R =
∆dS + B · R f if the stock goes down
0
Call Cu the option payoff in the “up” state and Cd the option payoff in the “down” state. The
replicating strategy must satisfy
Cu = ∆uS0 + B · R f
Cd = ∆dS0 + B · R f
6.1 One-Period Binomial Tree 51
∆ can be interpreted as the sensitivity of call price to a change in the stock price, or equivalently,
how much of the stock we should hold to hedge the option: for instance, to hedge a long call
position we need to sell ∆ units of the stock. Substituting ∆ and B in C0 we get
Rf −d u−Rf
Cu −Cd uCd − dCu 1
C0 = S0 + f = Cu + Cd
S0 (u − d) R (u − d) R f u−d u−d
1 h Q
Q
i 1
C0 = f
π Cu + 1 − π Cd = f EQ [C1 ]
R R
Therefore the option price is the discounted payoff of the option under the equivalent martingale
(risk-neutral) measure Q. Note that Q is also the probability measure that would justify the current
stock price in a risk-neutral world:
1 h Q
Q
i
S0 = π uS0 + 1 − π dS0
Rf
Note how we never even mentioned the physical probability measure π, since we are working with
relative asset pricing. To price a call or any other derivative we do not need to know the actual
probability P, but Q. We will be taking expectations with respect to Q, so we can put π Q directly
on the branches.
u2 S
uS
S udS
dS
d2S
To price the option at time t = 0 we just apply the same procedure backwards, starting from the
two sub-trees in the final period. To summarize,
The general procedure for a two-period tree is illustrated below (in the figure below, p = π Q ):
Suppose we observe an option price on the market of $36, while the cost of the dynamic replication
strategy is $34.08: then we can take advantage of this arbitrage opportunity by selling the option
and buying the synthetic portfolio. Today we pocket $1.92, and at maturity our cash outflows equal
our inflows.
6.3 The Relation with the Black-Scholes Model 53
and R f = e r f ·T /n where T is the time to expiration and σ is the annualized standard deviation of the
stock log-returns. The general binomial formula for a European call on non-dividend paying stock
n periods from expiration is:
" #
n
1 n! j n− j +
C0 = f n ∑ j! (n − j)! π Q
1 − π Q
u j d n− j S0 − K
(R ) j=0
2
ln
S0
K + r f + σ2 T √
Let d1 = √
σ T
and d2 = d1 − σ T , we can interpret the trading strategy under the Black-
f
Scholes formula as ∆ = N (d1 ) ∈ [0, 1] and B = −Ke−r T N (d2 ). We can also find the price of a put
option: by put-call parity, we have
fT f
P0 = C0 − S0 − Ke−r = Ke−r T N (−d2 ) − S0 N (−d1 )
We call intrinsic value of the option the quantity (St − K)+ and time value of the option the quantity
Ct − (St − K)+ .
6.4 Delta-Hedging
A stock has a delta equal to one, while a portfolio of J assets - where each asset j has notional
J
amount N j - has a delta △ p = ∑ ∆ j N j . We say that the portfolio is delta neutral if ∆ p = 0. Delta-
j=1
hedging a portfolio is only a “perfect” hedge (yielding the same final payoff as the corresponding
option) with continuous trading: computing the delta is a linear approximation to the option value,
but the option price is convex and therefore the second- and higher-order derivatives matter. Delta-
hedging when trading is not continuous is effective only for small price changes. Delta-Gamma
54 Chapter 6. Multi-Period Model: Options
hedging reduces this convexity risk, and involves trading with other options.
One way to get around this problem is to use a different model for the volatility of the stock called
“stochastic volatility”. Other issues in the Black-Scholes include stochastic interest rates, bid-ask
spreads, other transaction costs, etc.
6.7 Exercises
1. Consider a binomial tree, and the stock trades at S0 = 80. At each period the stock can go up
with probability p in which case St+1 = 1.2St or it can go down with probability 1 − p and
St+1 = 0.8St .The risk-free rate for each period is 10% and we are using discrete discounting
of R f = 1 + r f
a) One-Period Model:
i. What is the probability of stock price going up under the risk-neutral measure?
ii. Suppose we long ∆ shares of stock and B dollars in risk-free bond to replicate the
payoff for a European call option with strike K = 80. Please calculate ∆ and B.
iii. What’s the interpretation for ∆? Suppose we long 100 European calls and we
want to hedge the risk that the portfolio/total assets we have changes when the
underlying stock price changes over time. How to use stocks to hedge such risk?
Please indicate the long/short position and the amount of stocks.
iv. For ∆, what’s its corresponding term in Black-Scholes “Greeks?” (Please write in
the form of partial derivatives).
b) Multi-Period Model (3-Period)
i. What is the price of the prepaid forward (at t=0) delivering the stock at t = 3 (you
pay at t = 0).
ii. What is the price at t = 0 of the European call on this stock with K = 100 and
maturity t = 3?
iii. What is the price at t = 0 of an American put on this stock with strike K = 95 and
maturity t = 3? (Hint: early exercise at each node is possible.)
† 2. Consider the binomial model with u = 1.2, d = 0.9, R = 1.1 and initial asset price S = 100.
a) Calculate the price of a call option with exercise price K = 100 and n = 4 periods left
until expiry.
b) Calculate the prices of the European and American put options with the same exercise
price and time to expiry as in (a).
c) Calculate the price of an option which, at the end of the third period, gives its holder
the right to purchase the underlying asset at the minimum price realized over the life of
the option.
† 3. A stock price is currently $30. During each 2-month period for the next 6 month it will
increase by 8% or reduce by 10%. The annual risk-free interest rate is 5% (continuous
compounding). Use a three-step tree to calculate the value of a derivative (called turbo
option) that pays off ST × (30 − ST )+ , where ST is the stock price in 6 months.
7. Risk Preferences and Expected Utility Theory
So far in this course we dealt with relative asset pricing: we derived asset prices through information
available on other, pre-existing assets. In this chapter we will study absolute asset pricing, a task
for which we have to specify the agents’ preferences towards risk. We will see that risk involves
knowledge of the probabilities of uncertain events, and agents’ attitude towards risk determines
how much they would be willing to pay for an asset - that is the core of absolute asset pricing.
A concept related to risk that we will not see in detail is that of uncertainty, which relates to the
agents’ preferences over events whose probability distribution is unknown.
It is important to stress that preferences over risk refer to and are defined over “final payoff gambles”,
that is, after combining any random payoffs, such as the payoff of securities, investments, insurance,
and even non-random endowments.
We will start this chapter trying to answer the following question: is there a criterion that allows us
to compare random payoffs?
t =0 t =1
1
Cost Probabilities π1 = π2 = 2
s=1 s=2
Investment 1 -1000 1050 1200
Investment 2 -1000 500 1600
Investment 3 -1000 1050 1600
It is hard to tell which one is better between investment 1 and 2: the two possible states they have
the same probability and in state 1 investment 1 pays more, while in state 2 investment 2 pays more.
However, investment 3 pays at least as much as investment 1 and 2 in any state of the world! Clearly
from this perspective investment 3 is more desirable, and we say that it state-by-state dominates
investments 1 and 2.
Definition 7.1.1 — State-by-State Dominance. Given two random variables X and Y defined
over the probability space (Ω, F , P), we say that Y state-by-state dominates X if ∀ω ∈ Ω
X (ω) ≤ Y (ω).
We have already developed the notation needed to express state-by-state dominance: recall that for
7.2 Stochastic Dominance 57
x, y ∈ Rn we write
Assuming that x and y are scaled to have the same price, under any of the three conditions above
we would say that y state-by-state dominates x (you will notice that in the list above we see an
“increasing degree” of dominance).
t =0 t =1
1
Cost Probabilities π1 = π2 = 2 Moments
s=1 s=2 E [R] σ (R)
Investment 1 -1000 +5% +20% +12.5% 7.5%
Investment 2 -1000 -50% +60% +5% 55%
Investment 3 -1000 +5% +60% +32.5% 27.5%
Investment 1 has a higher expected return and lower volatility than investment 2. However, it is
not the case that investment 1 state-by-state dominates investment 2! Moreover, while investment
3 state-by-state dominates both investment 1 and 2, it only mean-variance dominates investment
2. Evidently, mean-variance dominance and state-by-state dominance are somehow unrelated
orderings. However, while mean-variance dominance is an incomplete ranking, it does simplify the
problem of ranking investments to two dimensions only (it is as incomplete as R2 ). It looks like,
potentially, a way of getting round the incompleteness problem is to find a criterion that amounts
to some number, for instance, in R, an ordered set. For example, suppose we rank investments
according to which Sharpe Ratio is highest:
Definition 7.1.2 — Sharpe Ratio.
E [R] − r f
Sharpe Ratio =
σ (R)
In the example above, assuming r f = 3%, we get 1.27, 0.04 and 1.07 for investment 1, 2 and 3
respectively, and therefore we would select investment 1 (which, remember, we would eliminate
based on both state-by-state dominance and mean-variance dominance!)
State s1 s2 s3 s4 s5
Probability π1 π2 π3 π4 π5
Payoff X 10 10 20 20 20
Payoff Y 10 20 20 20 30
Payoff 10 20 30
Probability X p10 = π1 + π2 p20 = π3 + π4 + π5 p30 = 0
Probability Y q10 = π1 p20 = π2 + π3 + π4 p30 = π5
This allows us to define preferences px ≻ py over probability distributions over states (rather than
on payoffs x ≻ y ∈ RS ) and will turn out to be useful when thinking of final payoffs. Note that
we can always go from the“payoff per state” to the “probability lotteries” representation, but the
reverse is not true.
Event e1 e2 e3
Payoff 50 100 200
Probability 1 40% 60% 0%
Probability 2 40% 40% 20%
In this case, every investment pays the same amount in each event, but the probabilities of the
events happening vary. To reconcile this with our previous “states of the world” discussion, it is
enough to note that the above example is such that investment 1 pays off 100 in the states s2 and s3
so that e2 = {s2 , s3 }, while investment 2 pays off 100 in state 2 and 200 in state 3 (and obviously
e1 = s1 ).
F(x)
F1
1
0.8
F2
0.6
0.4
0.2
Visually, FSD occurs whenever FA “stays below” FB for the whole domain. State-by-State Domi-
nance implies First Order Stochastic dominance.
1 cdf(x)
0.8
0.6
0.4
0.2
A
0 B
x
−10 −8 −6 −4 −2 0 2 4 6
Payoff 1 4 5 6 8 12
Probability 3 0% 25% 50% 0% 0% 25%
Probability 4 33% 0% 0% 33% 33% 0%
It is easy to verify that FSD does not hold for any investment. However, investment 3 is somewhat
special because its CDF stays below investment 4’s “most of the time”.
F(x)
F4
1
0.8
F3
0.6
0.4
0.2
0 2 4 6 8 10 12x
Proposition 7.2.1 First Order Stochastic Dominance implies Second Order Stochastic Dominance.
Definition 7.2.3 — Mean-Preserving Spread. We say that the random variable XA is a mean-
preserving spread of the random variable XB if XA = XB + ε, where the random variable ε is
independent of XA and XB , and has zero mean and positive variance.
0.4 f (x)
0.3
0.2
0.1
fA
0 fB
x
−6 −4 −2 0 2 4 6
The concept of mean preserving spread and that of second-order stochastic dominance are connected
by the following proposition:
Proposition 7.2.2 Let FA and FB be the CDFs of two random variables XA and XB defined on the
same space with identical means. Then FA SSD FB if and only if XA is a mean-preserving spread of
XB .
which can be considered as infinite-dimensional vectors (in the case of first and second-order
stochastic dominance). In this section we will elaborate another method which aims at simplifying
the problem by assigning to each investment a number u ∈ R, so that every two investments are
comparable. To do this by avoiding arbitrary choices of this function that maps investments onto
numbers, we will assume the existence of regular preferences over investments.
Let’s start with an example. Suppose you can enter the following bet: we flip a fair coin once. If
tails comes up, I’ll give you $1. If heads comes up, we’ll flip again, and if tails comes up I’ll pay
you $2, while if heads comes up we’ll flip again. Every time we flip again, the prize for a tails up
doubles. The question is: what is the fair price for this bet?
1 n
Since for the n-th flip the probability of a tails is 2 and the payoff is 2n−1 , and the number of
flips ranges from 1 to infinity, it is straightforward to see that the expected payoff is
n
∞ ∞
1 1
Expected Payoff = ∑ × 2n−1 = ∑ = ∞
n=1 2 n=1 2
However this is the correct expected value, it does not make sense to pay all the money in the
world for this coin-flipping game (investment), since it pays you less that $64 in 99% of cases.
Clearly this criterion is inappropriate to value uncertain payoffs. This problem is was named the
St. Petersburg paradox and was first proposed by Daniel Bernoulli, who proposed the following
solution: because when you are rich an additional dollar is less valuable to you than when you are
poor, we should weight less the largest payoffs. He proposed to discount large payoffs using the
natural logarithm: introduce the parameter p for the denominator of the probability, in the previous
example equal to 2. Then we can write:
n
∞
1
× ln 2n−1
Bernoulli Value = ∑
n=1 p
∞
= ln 2 ∑ (n − 1)p−n
n=1
∞
= − ln 2 ∑ (1 − n)p−n
n=1
∞
d h (1−n) i
= − ln 2 ∑ p
n=1 d p
" #
∞
d (1−n)
= − ln 2 ∑p
d p n=1
" #
∞
d 1
= − ln 2 ∑ pn
d p n=0
d 1 1
= − ln 2 = ln 2
dp 1− p (1 − p)2
Which for p = 2 is just ln 2. If the value is ln 2 and the function used to discount large payoffs is
the natural logarithm, we deduce that the price we are willing to pay for the coin-flipping game is
62 Chapter 7. Risk Preferences and Expected Utility Theory
just x such that ln x = ln 2, that is, $2. Let’s dig deeper into this idea of specifying a function that
maps uncertain payoffs (investments) to a single number (utility).
Representation of Preferences
Suppose all we care about is consumption in each state. Then, the choice set would consist
of vectors in RS+1 where the first element represents consumption at t = 0, and all other el-
ements represent consumption in the possible S states at t = T . To reconcile this definition
with our one-period model, suppose an investor has an initial wealth W (constant across states
and time) and is considering buying a security j with price p j and payoff x j . The vector
W − p j , (W − p j ) × I + x j = (c0 , c1 , ... cS )T ∈ RS+1 (where I is a S × 1 vector where each el-
ement is equal to one) would be the corresponding consumption profile. Moreover, given two
consumption profiles c, c′ ∈ RS+1 , if an investor chose c over c′ we say that that investor prefers c
to c′ and write c ≻ c′ .
Theorem 7.3.1 — Representation Theorem. Suppose the preference ordering is i) complete ii)
transitive iii) continuous (and satisfies some further regularity conditions that we do not discuss
here). Then the preference ordering can be represented by a utility function, that is, c ≻ c′ if and
only if there exists a utility function U such that U(c) > U(c′ ).
A related but different approach was developed by John von Neumann and Oskar Morgenstern. In
this setup, we define the (finite) set of outcomes X (in the previous setup for instance, this was the
uncertain consumption profile at time t = T ) and we define a lottery as a probability measure over
X, that is a function p : X → [0, 1] such that 1) for x ∈ X : p(x) ≥ 0, 2) ∑ p(x) = 1. The set of all
x∈X
possible lotteries is
( )
P≡ p : X → [0, 1] | ∀x ∈ X : p(x) ≥ 0, ∑ p(x) = 1
x∈X
A preference relation expresses a person’s feelings between pairs of objects in P. We denote the
preference relation by ⪰, which means “at least as good as”. Two properties of a preference relation
are defined below.
Definition 7.3.1
Assume there exists a preference relation ≻ over P and consider the following three axioms:
Definition 7.3.2 — von Neumann - Morgenstern Axioms.
That is, preferences over lotteries correspond to the expected utility of the lotteries. Below is a
graphical representation of an expected utility function over a lottery defined on a two-dimensional
state space:
x2
45◦
U(x1 , x2 ) = k
x1
0 z πx1 + (1 − π)x2 = c
Figure 7.5
The 45° line represents the set of lotteries that pay off the same amount in both states. The
negatively sloped straight line is the set of lotteries in the (x1 , x2 ) plane such that the expectation
of the lottery is equal to a value c > 0, that is, if π is the probability of the state 1 occurring,
π c
c = πx1 + (1 − π) x2 , or x2 = 1−π x1 + 1−π . The convex curve represents the indifference curve for
the agent: to give up a certain consumption amount in state 2, he is willing to take some amount in
state 1 to keep his utility level constant. The curve is the locus of lotteries in the (x1 , x2 ) plane such
that k = πu (x1 ) + (1 − π)u (x2 ), k ∈ R.
Going back to the axioms, consider the Independence axiom: it basically states that if we “dilute”
64 Chapter 7. Risk Preferences and Expected Utility Theory
p and q with some third lottery r in the same way, but we prefer p to q, then our preference ranking
should not change. It can be visualized using trees:
π p π q
p⪰q ⇔ ⪰
r r
Figure 7.6
However, this axiom rarely holds in experiments. For instance, suppose you are given the choice
between lottery p that pays $10 with probability 10% (and zero otherwise) and lottery q that pays
$15 with probability 9% (and zero otherwise). Because the prize is 50% higher in lottery q than
in lottery p and the probability of winning a positive amount is only 10% lower in lottery q than
lottery p, most people in experiments choose lottery q. However, between a lottery u that pays $10
with 100% probability and a lottery v that pays $15 with 90% probability, most people choose the
sure bet implied by lottery u. However, it turns out that lotteries u and v can be mixed with a third
lottery r that pays zero with probability 100%, using a parameter α = 10%, to obtain lotteries p
and q again(see picture below): the preference relation between p and q is now the opposite of
the initial preference, and therefore this constitutes an empirical violation of the independence
axiom.1
10% 10 9% 15
≺
0 0
Figure 7.7
100% 10 90% 15
10% 10%
0 ≻ 0
0 0
Figure 7.8
u(x)
u(E[x])
E[u(x)])
x
x0 CE E[x] x1
And the difference E [x] − u−1 (E [u(x)]) is called risk premium. This fact derives directly from
Jensen’s Inequality:
Theorem 7.4.1 — Jensen’s Inequality. Let g be concave over [a, b] and let x be a random
variable such that P [x ∈ [a, b]] = 1. If the expectations E [x] and E [g (x)] exist, then E [g (x)] ≤
g (E [x]). Furthermore, if g(·) is strictly concave then the inequality is strict.
The following theorems establish a relationship between First and Second Order Stochastic Domi-
nance and Expected Utility:
Theorem 7.4.2 Let FA and FB be two CDFs for the random payoffs x ∈ [a, b]. Then FA FSD FB
if and only if EA [U (x)] ≥ EB [U (x)] for all nondecreasing utility functions U(·).
Proof. (⇐) We prove by contrapositive. Suppose FA does not FSD FB . Then, we can find at least
one x∗ ∈ [a, b] such that FA (x∗ ) > FB (x∗ ). Define nondecreasing utility function u(x) = 1{x > x∗ }.
Clearly,
Z b Z b
EA [u(x)] = fA (x)dx = 1 − FA (x∗ ) < 1 − FB (x∗ ) = fB (x)dx = EB [u(x)]
x∗ x∗
Theorem 7.4.3 Let FA and FB be two CDFs for the random payoffs x ∈ [a, b]. Then FA SSD FB
if and only if EA [U (x)] ≥ EB [U (x)] for all nondecreasing and concave utility functions U(·).
We conclude this subsetion with a parallel between (non-expected) utility representation and
expected utility representation. In the former case, suppose U(c) = U (c0 , c1 , ... cS ) represents a
complete, transitive and continuous preference ordering between consumption profiles. Then also
V (c) = f (U(c)), for f strictly increasing, represents the same preference ordering. In the latter case,
suppose that E [u(c)] represents a preference ordering satisfying the von Neumann-Morgenstern
axioms. Then for a, b ∈ R the affine function v(c) = a + bu(c) represents the same preference
ordering.
Definition 7.5.1
(Y ) ′′
1. Absolute Risk Aversion: RA (Y ) = − uu′ (Y )
(Y ) ′′
2. Relative Risk Aversion: RR (Y ) = −Y · uu′ (Y ) = Y · RA
1
3. Risk Tolerance: RT (Y ) = RA (Y )
There is an interesting link between expected utility and these measures of risk aversion: consider
the following lottery, where Y is the initial wealth of the agent:
π Y +h
1−π Y −h
What is the change in π needed for an agent to be indifferent when h rises? We can Taylor-expand
to the second degree the expected utility in a neighbourhood of Y to obtain:
1 1
π(Y, h) = + hRA (Y ) + o(h)
2 4
π Y (1 + θ )
1−π Y (1 − θ )
Again, the change in π needed for an agent to be indifferent when h rises is obtained by Taylor-
expanding to the second degree the expected utility in a neighbourhood of Y to obtain:
1 1
π(Y, h) = + θ RR (Y ) + o(θ )
2 4
Similarly, suppose we add a zero-mean small risk (lottery) X to the initial wealth Y0 . Since
E [u(Y0 + X)] = u Y CE , we can use a second order Taylor expansion to show that (Hint: let risk
premium Π ≡ Y0 −Y CE , note that u(Y0 − Π) = E[u(Y0 + X)]. Taylor-series expand both sides around
Y0 .)
Var(X)
Y0 −Y CE ≈ RA (Y0 )
2
where Y0 −Y CE is the risk premium, and it is approximately linear in the variance of the additive
risk, with slope equal to half the coefficient of absolute risk aversion.
Similarly, suppose we have a multiplicative risk. Suppose that E [u(Y0 (1 + θ ))] = u (kY0 ), where
1 + θ is a positive random variable with unit mean and k is is the certainty equivalent growth rate.
This time we get
Var(θ )
1 − k ≈ RR (Y0 )
2
Therefore the coefficient of absolute risk aversion is relevant for additive risk, while the relative
risk aversion is relevant for multiplicative risk.
We now introduce two specific functional forms for von Neumann-Morgenstern function U:
Definition 7.5.2
• Constant Absolute Risk Aversion (CARA) utility function: U(x) = −e−ρY , for some
ρ ∈ R+
Y 1−γ i f γ ̸= 1
1−γ
• Constant Relative Risk Aversion (CRRA) utility function: U(x) =
lnY if γ = 1
For example, consider the CRRA utility and a lottery that adds $50,000 or $100,000 to the initial
wealth Y . The certainty equivalent is defined as the number CE ∈ R such that
68 Chapter 7. Risk Preferences and Expected Utility Theory
For Y = 0 we get
While for Y = 100, 000 and γ = 5 the certainty equivalent is already 66,530 (close to γ = 2 for
Y = 0!)
max E U Y0 1 + r f + a r − r f
a∈R
for some initial wealth Y0 . The problem has first order conditions (FOC)
E U ′ Y0 1 + r f + a r − r f r−rf = 0
We can characterize the solution to the problem with the following theorem:
Theorem 7.6.1 Assume U ′ > 0, U ′′ < 0 and let â denote the solution to the problem above.
Then
so W ′ (a) is everywhere decreasing in a. This implies that â > 0 if and only if W ′ (0) > 0, and since
U ′ > 0 it follows that â > 0 if and only if E [r] > r f . The other assertions follow similarly. ■
How do measures of risk aversion depend on the initial wealth Y0 ? This is the subject of Arrow’s
theorems (1971). The first theorem is about the absolute measure of risk aversion
Theorem 7.6.2 — Arrow (1971) Part 1. Let â = â (Y0 ) be the solution to the problem above.
Then
∂ RA ∂ â
(DARA) <0⇒ >0
∂Y0 ∂Y0
∂ RA ∂ â
(CARA) =0⇒ =0
∂Y0 ∂Y0
∂ RA ∂ â
(IARA) >0⇒ <0
∂Y0 ∂Y0
Proof. No proof. ■
∂ RR d â/â
(DRRA) <0⇒ >1
∂Y dY /Y
∂ RR d â/â
(CRRA) =0⇒ =1
∂Y dY /Y
∂ RR d â/â
(IRRA) >0⇒ <1
∂Y dY /Y
Proof. No proof. ■
d â/â
Note that the quantity dY /Y is the wealth wealth elasticity of investment in the risky asset. If
d â/â
dY /Y > 1, then as wealth increases, the optimal proportion invested in the risky assets also increases.
Below we also provide an example for CRRA utility function.
r−rf
E =0
Y0 (1 + r f ) + a (r − r f )
and assuming that r can take on two possible values r1 and r2 with probabilities π and 1 − π,
respectively with r1 < r f < r2 , it is possible to show that
1 + r f E [r] − r f
a
= >0
Y0 (r2 − r f ) (r f − r1 )
70 Chapter 7. Risk Preferences and Expected Utility Theory
That is, a constant fraction of the initial wealth is invested in the risky asset.
How does all this generalize to the J assets case? The answer is provided by the following
theorem:
Theorem 7.6.4 — Cass and Stiglitz. Let the vector
â1 (Y0 )
.
..
âJ (Y0 )
denote the amount optimally invested in the J risky assets if the initial wealth is Y0 . Then
â1 (Y0 ) â1
. .
.. = .. · f (Y0 )
âJ (Y0 ) âJ
In other words, in the case of constant absolute or constant relative risk-aversion preferences, it
is sufficient to offer a mutual fund: the optimal proportion to invest in each asset is given by the
vector [â1 , . . . , âJ ]T , and agents would then invest an amount depending on their initial wealth level
according to the function f (Y0 ).
To conclude this subsection, we generalize the class of utility functions by starting from the absolute
risk aversion (and then backing out the resulting utility function).
We call linear risk tolerance (or hyperbolic risk aversion) any utility function such that
u′′ (c) 1
RA = − =
u′ (c) A + Bc
1 B−1
u(c) = (A + B · c) B
B−1
for which if B → 1 we obtain the log utility u(c) = ln (A + Bc), for B = −1 we obtain the quadratic
B−1
utility − 21 (A − c)2 , and for A = 0 we obtain the CRRA utility function u(c) = 1
B−1 (Bc) B .
7.7 Savings 71
7.7 Savings
Within our one period model, consider an asset structure composed of one risk-free asset with
constant gross return R f . The agent has to make a decision on how much to consume in each period
t = 0 and t = T . If R f increases, we will see two effects: one is the incentive to save more at time
t = 0, since by doing so we can now consume more in t = T - this is called substitution effect - and
the incentive to consume more (save less) at time t = 0, since the agent is effectively richer when
the gross riskless return increases - this is called income effect. In the log utility case, these two
effects cancel each other; in other cases (γ ̸= 1) one effect can prevail over the other.
Suppose now there is a risk-free assets but the endowment in the future period is random. How does
the behavior of the agent change when we increase his exposure to risk? An old idea (going back
to J. M. Keynes) is that people save more when they face greater uncertainty. This phenomenon
is called precautionary savings and it generally arises under two circumstances, one linked to the
shape of the utility function and one to the presence of borrowing constraints.
Let’s jump forward a little and suppose that in a multi-period setting agents maximize the expected
utility
" #
∞
E0 ∑ β t u (ct )
t=0
u′ (ct ) = β (1 + r) Et u′ (ct+1 )
which shows that the marginal rate of intertemporal substitution is higher in the presence of
uncertainty in ct+1 . The difference between the two marginal rates, with and without uncertainty,
is attributed to precautionary savings. to see this, suppose the variance of et+1 increases (in a
mean preserving fashion). Since numerator Et [u′ (ct+1 )] is increasing in the variance of ct+1 , in
order for the above equality to hold ct must decrease, that is, savings increase due to precautionary
savings.
Suppose now there are no risk-free assets and one risky asset with random gross return R.
Theorem 7.7.1 — Rothschild and Stiglitz (1971). Let RA and RB be two return distributions
72 Chapter 7. Risk Preferences and Expected Utility Theory
with identical means such that RB = RA + ε where ε is a white noise, and let sA and sB be the
savings out of Y0 corresponding to the return distributions of RA and RB respectively. Then
u′′′ (w)
PA (w) = −
u′′ (w)
u′′′ (w)
PR (w) = −w ·
u′′ (w)
Note that this depends directly on the curvature of u′ (u′′′ > 0 implies that u′ is convex) and does
not follow directly from risk aversion. It turns out that precautionary savings occur if P(w) > 0.
Moreover, we have the following result:
Theorem 7.7.2 Let RA and RB be two return distributions such that RA SSD RB , and let sA and
sB (respectively) be the savings out Y0 . Then
sA ≥ sB ⇔ cP(c) ≤ 2
sA < sB ⇔ cP(c) > 2
Finally, agents might save precautionarily also because they are concerned that they will face
borrowing constraints in some state in the future.
The first case in which mean variance and vN-M are equivalent is when agents have quadratic
utility:
u(c) = ac − bc2
Therefore, the expected utility is a function of the mean E [c] and variance Var (c) only.
7.8 Mean-Variance Preferences 73
This is also true in another case: suppose all lotteries have normally distributed outcomes (thus
the state space now is infinite). Then, because any linear combination of normally distributed
random variables is also normal, the distribution of any lottery (or linear combination thereof) will
be completely described by its first two moments, mean and variance. Therefore also expected
utility can be expressed as a function of these two numbers only.
The third cases has been covered in section 7.5, where we showed that if the risk (lottery) X is
zero-mean and small then
Var(X)
Y0 −Y CE ≈ RA (Y0 ) .
2
74 Chapter 7. Risk Preferences and Expected Utility Theory
7.9 Exercises
3. Suppose an agent has an income of $10. He has the possibility to buy for $2 a lottery ticket
which pays the winner $19 (and zero otherwise). Suppose the agent has a von-Neumann
Morgenstern utility u(x) = ln(x) and believes that the probability of winning π is 31 .
a) Write the agent’s expected utility.
b) Should the agent buy the ticket?
c) Suppose the agent buys the ticket, but before a winner is selected he decides to sell it.
What is the minimum certain amount that the agent would accept in exchange for the
lottery ticket?
4. To resolve the [Link] Paradox, Bernoulli proposed to give less weight to large payouts
in the computation of the expected value of the game, to account for the fact that people
value money less and less the richer they become (though this certainly does not apply to
MFins). This property is shared by any increasing and concave function: can you guess why
Bernoulli (a pretty smart guy) proposed exactly u(x) = ln(x)? (Hint: consider the increase in
utility u given by a small increase in the payoff x.)
5. Suppose an agent is given the opportunity to enter, for free, a lottery that pays him $50,000
1
with probability 2 and $100,000 with probability 12 . Suppose the agent has CRRA utility:
c1−γ
u(c) =
1−γ
and γ = 7. Find the certainty equivalent of this lottery for the initial wealth Y equal to $0,
$100,000 and $300,000. What happens to the certainty equivalent when the initial wealth
increases? Explain this result.
7.9 Exercises 75
6. Consider a general von Neumann-Morgenstern utility function u(x) and a lottery which adds
an amount h to your personal wealth c in case of victory (which occurs with probability π)
and subtracts the same amount in case of loss:
c+h
π
c
1−π
c−h
1 1
π(c, h) = + hRA (c) + o(h)
2 4
Similarly, consider a lottery which increases your personal wealth c by θ % in case of victory
(which occurs with probability π) and decreases your personal wealth by the same percentage
amount in case of loss:
c(1 + θ )
π
c
1−π
c(1 − θ )
1 1
π(c, θ ) = + θ RR (c) + o(θ )
2 4
−νx
7. Use the CARA function U(x) = − e ν to solve the risk aversion and portfolio allocation
problem in Lecture slides 3, Slide 47 and show that it confirms Arrow’s theorem.
8. As in Lecture slides 3, Slide 51, show that a constant fraction of wealth is invested in risky
asset for an agent using CRRA utility function with r2 > r f > r1 . Show that it works also
with any distribution of r.
9. Consider the following one-period portfolio problem faced by an agent with vN-M utility
U(·):
max E[U(Y0 − s) + δU(sR̃)], s.t. Y0 ≥ s ≥ 0
s
where s is the amount he saves to consume at the final period (1). Observing the variables
appearing in the theorem proposed by Rothschild and Stiglitz (1971) (Lecture 3, Slide 66),
without directly using the theorem, please solve the portfolio problem above for RA and for
RB = RA + ε (i.e. RA s.s.d RB ) to show that:
a) When the agent has mean variance preferences U(x) = x − θ2 x2 , then sA > sB .
76 Chapter 7. Risk Preferences and Expected Utility Theory
10. Assume that an investor with utility u(ct , ct+1 ) = u(ct ) + E[u(ct+1 )] can freely buy or sell
as much of the payoff xt+1 as he wishes, at a price pt . Assume that he has an exogenous
endowment of (et , et+1 ):
a) How much will he buy or sell?
b) What is the relation between the price pt and the payoff xt+1 ?
Notice that et+1 and xt+1 are random variables. For question i), further assume that et+1
and xt+1 follow binomial distributions, with probability π of being in the up state u (et+1 =
eu , xt+1 = xu ) and (1 − π) of being in the down state d (et+1 = ed , xt+1 = xd ). Obtain a
solution when the utility function is u(c) = ln(c). (Obtaining a simplified equation containing
the unknown amount to buy or sell would be fine, as long as you show the equation can be
solved for an explicit solution).
11. Assume CARA preferences U(x) = E[−e−cx ]. An agent values terminal wealth W and begins
with wealth W0 . He can invest in a risk-free asset with net return R f and in two risky assets
with independent net returns Ri ∼ N(ui , vi ), i = 1, 2. Suppose u1 > u2 > R f and 0 < v1 < v2 .
a) Write down the wealth W in terms of W0 , R f , R1 , R2 , and the fraction of wealth invested
in the two risky assets x1 and x2 .
b) Calculate U(W ).
c) Solve for the optimal x1∗ and x2∗ . Show that x1∗ > x2∗ > 0.
d) Explain why the agent invests a positive amount in asset 2 even though it has a lower
mean and higher variance than asset 1.
(A+Bx)1−1/B
12. Consider the following HARA u(x) = B−1 with A > 0 and B ∈ R. Using an idea that
positive affine transformation does not affect the order of preference expressed by this utility,
show that, after the right affine transformation, the limit of u(x) when B goes to zero is a
CARA utility function, i.e. an exponential concave utility function.
8. Mean-Variance Analysis and CAPM
8.1 Preliminaries
In last chapter we introduced the representation of preferences over a consumption profile c =
(c0 , c1 , . . . , cS ) ∈ RS+1
+ . Suppose that, at birth, agent i is given an endowment e = (e0 , e1 , . . . , eS ) ∈
S+1
R+ and has a utility function U i : RS+1 i
+ → R. We say that U is:
∂U i
Concavity implies quasiconcavity. If U i is also differentiable, a standard requirement is that ∂ cs >0
for each s = 0, 1, . . . S. We are now ready to formulate the portfolio consumption problem:
max U i (c0 , c1 , . . . cS )
c,h
0 ≤ c0 ≤ e0 − p · h
J
0 ≤ cs ≤ es + ∑ xsj h j , s = 1, . . . , S
j=1
0 ≤ cT ≤ eT + Xh
where λ ∈ R, µ ∈ RS are the dual variables. Suppose c∗ solves the problem above, then there exist
λ ∈ R and µ ∈ RS such that:
∂U i ∗
(c ) = λ
∂ c0
∂U i ∗
(c ) = µs
∂ cs
λ p = µX
78 Chapter 8. Mean-Variance Analysis and CAPM
S
µs
pj = ∑ λ xsj
s=1
S
∂U i /∂ cs j
pj = ∑ i xs
s=1 ∂U /∂ c0
and also that the utility over uncertain states ũi is a von Neumann-Morgenstern utility:
S
∂ ui /∂ cs
pj = ∑ πs δ ∂ ui /∂ c0 xsj
s=1
pj = xj ·q = E m·xj
∂U i /∂ cs
Finally, note that the vector of marginal rates of substitutions ∂U i /∂ c0
is the state price vector (note
that this is positive by our assumption on U i ).
E R j − R f = β j E [R∗ ] − R f
Cov(R∗ ,R j )
where β j = Var(R∗ ) and R∗ is the return of an asset whose payoff is equal to the discount factor
m∗ derived from the pricing kernel q∗ ∈ ⟨X⟩. This is a very general statement about returns, but it
does not help us identify what assets in reality have a return of R∗ .
8.2 Simple CAPM with quadratic utility functions 79
Suppose all agents have the same quadratic utility u (c0 , c1 ) = v (c0 ) − (c1 − α)2 . Recall that the
S
∂1 u
expected utility is E [U (c)] = ∑ πs u (c0 , c1,s ) and therefore m = ∂0 v where ∂i (·) ≡ ∂ (·)/∂ ci : in the
s=0
quadratic utility case we have
h i
∂1 u = −2 (c1 − α) · · · −2 (cS − α)
Cov m, R j
Cov ∂1 u, R j
j f f
E R −R = − = −R ×
E [m] ∂0 v
j
2Cov c1 , R j
f Cov −2 (c1 − α) , R f
= −R × =R ×
∂0 v ∂0 v
This holds for any asset x j ∈ ⟨X⟩, and can be generalized to any portfolio h:
2Cov c1 , Rh
h i
h f f
E R −R = R ×
∂0 v
Now consider the market portfolio. The above equation must hold for it too, and therefore dividing
side by side we get
E Rh − R f
Cov c1 , Rh
=
E [Rmkt ] − R f Cov (c1 , Rmkt )
where Rmkt is the return of the market portfolio. If moreover agents are homogeneous and live in
an exchange economy, then we know that c1 corresponds to the aggregate endowment and this is
perfectly correlated with Rmkt :
E Rh − R f
Cov Rmkt , Rh
=
E [Rmkt ] − R f Var (Rmkt )
Cov(Rmkt ,Rh )
Define βh = Var(Rmkt )
we can write
h i h i
E Rh = R f + βh E Rmkt − R f
which we call the Security Market Line (SML). Note that in order for the above result to hold, R∗
must be a linear function of Rmkt of the type
a + bRmkt
R∗ =
a + bR f
with b < 0 since we know that the stochastic discount factor m is high (hence high R∗ ) in states
in which the economy is doing poorly, so in which the market portfolio xmkt is low (hence low
Rmkt ).
80 Chapter 8. Mean-Variance Analysis and CAPM
hj pj
where rh = Rh − 1 is the net return1 and w j = , h j is the amount of asset j ∈ 1, 2, . . . , J (as
∑Jj=1 h j p j
usual), so that the price of Rh is still 1. The variance of the portfolio is given by
Note that in the (σh , µh ) space, for a given value of µh and σh we can immediately find the subset
of portfolios which are not dominated by the given µh and σh :
Definition 8.3.2 — Efficient Frontier. For given µ and σ , the Efficient Frontier is the locus of
all non-dominated payoffs in the (σh , µh ) space.
It follows directly from the definition of efficient frontier that no rational investor with mean-
variance preferences would choose to hold a portfolio outside of the efficient frontier.
In the J = 2 case, for portfolio h we have w2 = 1−w1 , so the mean return is µh = w1 µ1 +(1 − w2 ) µ2
and the variance is σh2 = w21 σ12 + (1 − w1 )2 σ22 + 2w1 (1 − w1 ) σ1 σ2 ρ1,2 . Note that in the 2 assets
example, for ρ1,2 = 1 we get
±σh − σ2
w1 =
σ1 − σ2
µ2 − µ1
µh = µ1 + (±σh − σ1 )
σ2 − σ1
±σh + σ2
w1 =
σ1 + σ2
1 Note that we can always write the excess return in gross or net terms: E R j − R f = E r j − r f .
8.3 The Traditional Derivation of CAPM 81
µ2
µh
µ1
σ
σ1 σh σ2
Figure 8.1: The efficient frontier: two perfectly correlated risky assets.
And therefore
σ2 σ1 µ2 − µ1
µh = µ1 + µ2 ± σh
σ1 + σ2 σ1 + σ2 σ1 + σ2
µ2 −µ1
µ2 slope: σ1 +σ2
σ2 σ1
σ1 +σ2 µ1 + σ1 +σ2 µ2
µ1 slope: − σµ21 −µ
+σ2
1
σ
σ1 σ2
Figure 8.2: The efficient frontier: two perfectly negatively correlated risky assets.
µ2
µ1
σ1 σ2 σ
Figure 8.3: The efficient frontier: two risky assets ρ = −0.5, 0, 0.5 (blue, black, orange).
The same concept generalizes to the J assets case: a frontier portfolio has minimum variance among
all feasible portfolios with the same expected portfolio return, so the problem is
1
max w′ Vw
w 2
subject to λ : w′ µ = µh
γ : w′ 1 = 1
where 1 = (1, . . . , 1)′ and µh ∈ R is a given (fixed) expected return. The idea is that we fix
an expected return µh and then look for the minimum variance possible given that expected
return.
∂L
= Vw − λ µ − γ1 = 0
∂w
∂L
= µh − w′ µ = 0
∂λ
∂L
= 1 − w′ 1 = 0
∂γ
where λ and γ are the Lagrange multipliers for the two constraints. Pre-multiplying the first FOC
by µ′ V−1 we get
µ′ w = λ µ′ V−1 µ + γ µ′ V−1 1 ≡ λ B + γA
8.3 The Traditional Derivation of CAPM 83
µh = λ B + γA
1 = 1′ w = λ 1′ V−1 µ + γ 1′ V−1 1 ≡ λ A + γC
Cµh − A
λ=
D
B − Aµh
γ=
D
where D ≡ BC − A2 .
Therefore, pre-multiplying the first FOC by V−1 and plugging in λ and γ we have
Cµh − A −1 B − Aµh −1
w∗ = V µ+ V 1 (8.1)
D D
BV−1 1 − AV−1 µ CV−1 µ − AV−1 1
= + µh = g + hµh (8.2)
| D
{z } | D
{z }
≡g ≡h
Suppose now that for some portfolio with return r we set µh = E [r]. Similarly to the 2 assets
problem, the solution portfolio weights are linear in the (expected) portfolio returns:
w∗ = g + hE [r] (8.3)
Note that for E [r] = 0 and E [r] = 1 we get g and g + h respectively, which are therefore frontier
portfolios as well. We established the following facts:
Proposition 8.3.1 The entire set of frontier portfolios can be generated by g and g + h: any portfolio
in the frontier is a linear combination of these two portfolios.
Proof. Let p be an arbitrary frontier portfolio with expected return equal to E [r p ]. Consider
portfolio weights wg = 1 − E [r p ] and wg+h = E [r p ]. Then
(1 − E [r p ])g + E [r p ](g + h) = g + hE [r p ] = w p
Proposition 8.3.2 Any linear combination of frontier portfolios is also a frontier portfolio: the
portfolio frontier can be described as linear combinations of any two frontier portfolios (not just g
84 Chapter 8. Mean-Variance Analysis and CAPM
and g + h).
Proof. Exercise. ■
A 2 1
2 C
σ (E [r]) = E [r] − + (8.4)
D C C
Definition 8.3.3 The portfolio with the smallest possible variance is called the global minimum
variance portfolio or mvp.
and of a hyperbola in the expected return/standard deviation space. Figure 8.4 and 8.5
visualize the set of frontier portfolios in these spaces.
Given two portfolios A and B in the frontier, we can see that points above B and below A correspond
to portfolios in which we short-sell assets. See Figure 8.6.
Based on the observations above, we have an equivalent definition for efficient frontier.
8.3 The Traditional Derivation of CAPM 85
Definition 8.3.4 — Efficient Frontier. Efficient portfolios are those frontier portfolios for which
the expected return exceeds A/C, the expected return of the minimum variance portfolio.
Proposition 8.3.3 Any convex combination of frontier portfolios is also a frontier portfolio.
86 Chapter 8. Mean-Variance Analysis and CAPM
1
max w′ Vw
w 2
s.t. w′ µ + (1 − w′ 1)r f = µh
where r f is the net risk-free rate. 2 Note that in this problem weights automatically sum to one since
w′ 1 + (1 − w′ 1) = 1, so we don’t have the second constraint in the problem without the risk-free
asset. First order conditions yield
w = λ V−1 (µ − r f 1)
µh − r f
λ=
(µ − r f 1)′ V−1 (µ − r f 1)
V−1 (µ − r f 1) µh − r f
w= ′ −1 (µh − r f ) = V−1 (µ − r f 1)
(µ − r f 1) V (µ − r f 1) H2
q
where the scalar H = B − 2Ar f +Cr2f is the Sharpe ratio. We have two important results at this
point:
Result 1: For any two frontier portfolios p and q, we have E [rq ] − r f = βq,p (E [r p ] − r f ).
E [r p ] − r f (E [rq ] − r f ) (E [r p ] − r f )
Cov (rq , r p ) = w′q Vw p = w′q (µ − r f 1) 2
=
H H2
2
(E[r p ]−r f )
and Var (r p ) = H2
, and dividing side by side yields the result. Importantly, this holds for
any two frontier portfolio p, thus in particular it also holds for the market portfolio.
E [r p ] = r f + Hσ (r p ) . (8.5)
2 That is, r f = R f − 1.
8.3 The Traditional Derivation of CAPM 87
Therefore the efficient frontier with a risk-free asset is linear, and it is called the Capital Market
Line (CML).
E(r)
Mean Variance
Efficient Porfolio
20
L
CM
15
10
5
rf
0
0 10 20 30 40 50 60 σ
E(r)
20
L
CM
15
10
5
rf
0
0 10 20 30 40 50 60 σ
if asset returns are gaussian then also any portfolio is gaussian and therfore preferences can only
depend on the first two moments. Moreover if agents have CARA utility function we know that the
certainty equivalent for a lottery with mean µ and variance σ 2 is going to be µ − ρ2 σ 2 where ρ is
the absolute risk aversion of agents.
E [rmkt ] − r f
E [rh ] = r f + σh (8.6)
σ (rmkt )
Note that the slope is the Sharpe ratio of the market portfolio.
Similarly, since
h E[r ]−r
f
σh H E [rh ] − r f Cov(rh , rmkt )
= = = = βh,mkt
σ (rmkt ) σ (rm kt) E [rmkt ] − r f Var(rmkt )
we can write
which defines the Security Market Line (SML)in the (β , E [r]) space:
E(r) ine
Market Porfolio
r ket L
Ma
urity
Sec
r2
rM
r1
rf
0 0.5 1 1.5 2 β
8.4 Exercises
1. Consider a two-period economy. There is an investor with CRRA utility and no discounting,
so that she maximizes expected utility
1−γ
" #
1−γ
c0 c1
+E ,
1−γ 1−γ
2. Consider an economy in which there are 3 risky assets (A, B and C) and one riskless asset.
Asset A has an expected net return of 15% and the variance of its return is 0.20. Asset B has
an expected net return of 20% and a variance of 0.40. Asset C has an expected net return
of 55% and its variance is 0.60. The covariance between each pair of asset returns is zero.
There are 3000 shares of A available in the economy and the current price of A is 20. There
90 Chapter 8. Mean-Variance Analysis and CAPM
are 1500 shares of B available and its current price is 40. Finally, there are 10,000 shares of
C outstanding and the current price of C shares is 18. Does there exist a riskless rate of return
R f so that given the current prices of the risky asset, the Sharpe-Lintner pricing equation
E R j = R f + β j,m (E Rmkt − R f ) holds? If so, what is that rate? If not, why is it impossible
3. A manufacturing firm has just been founded and needs to build a factory. It has a choice
between 2 types of factory (factory 1 and factory 2) and the cost of building each is K1 = 120
and K2 = 50, respectively. Only one factory can be built at this time. The table below lists
the properties of the payoffs (X1 and X2 ) from the different projects, the gross market rate of
return (Rmkt ) and the gross risk-free rate (R f ). You may assume that all the assumptions of
the CAPM hold and that the management seeks to maximize the market value of the firm at
all times. To simplify calculations, the net risk-free rate of return is set to zero. You may also
assume that any investments the firm makes have a negligible effect on the market portfolio.
E Rmkt = 1.1
E [x1 ] = 300 E [x2 ] = 150
Var[x1 ] = 10, 000 Var[x2 ] = 2, 500 Var[Rmkt ] = 0.01
Corr[X1 , X2 ] = 0.4 Corr[X1 , Rmkt ] = 0.8 Corr[X2 , Rmkt ] = 0.2
a) In an meeting, an executive argues as follows: Since the expected (net) return on factory
2 is higher (200% vs. 150%), that one should be built. Another executive disagrees
and argues that naturally one should always build factory 1, since the expected profit
is larger than if factory 2 were built (180 vs. 100). Which (if any) of the two decision
procedures suggested will maximize the firm value? Why?
b) In the end, the executives ask you to decide which factory to build. What is your
decision? Compute the firm’s total Value V . If 100 shares are outstanding, what is the
stock price P? (Assume the CAPM holds)
c) Suppose factory 2 has been built (do not assume this was the correct answer to part
b) and a year has passed. It is time to expand, and the firm plans to build factory 1.
However, it has no cash to fund the investment and need to raise K1 = 120 in the capital
market. The firm decides to issue shares for a total value of 120. How many shares
should it issue and at what price?
d) In a shareholder meeting, there is a heated argument about whether or not to go ahead
with the building of the factory and the stock issue. An angry shareholder claims that
although expected returns are impressive, factory 1 is an extremely risky investment,
not least because most of its risk is not diversifiable. She argues that exposing the firm
to this risk will reduce the value of her shares. A concerned executive responds by
pointing out that since the values of the 2 factories are negatively correlated, the project
does not really increase the riskiness of the firm’s portfolio of assets that much and is
therefore a good investment. What, in your opinion, is missing from the analysis by the
shareholder and the executive. Are their arguments valid?
8.4 Exercises 91
4. Assume that the market return is in the mean variance frontier. Suppose that there are 3
equally likely states. One security is risk-free and has a gross return R f = 0.9. A second
security has a gross return vector of (0,1,1). The market portfolio has a payoff of (1,2,2).
a) Use the CAPM to price a security that pays off (0,0,1).
b) Find the pricing kernel.
c) Calculate the price of the same security (that pays off (0,0,1)) using the pricing kernel.
Bibliography
[2] Kerry E. Back, Asset Pricing and Portfolio Choice Theory, Oxford University Press,
2010.
[3] Markus K. Brunnermeier, Lecture Notes for the course FIN 501: Asset Pricing, 2013.
[4] John Y. Campbell, Andrew W. Lo, and A. Craig McKinay, The Econometrics of
Financial Markets, Princeton University Press, 1996.
[5] John Y. Campbell and Robert J. Shiller, Valuation Ratios and the Long-Run Stock
Market Outlook, The Journal of Portfolio Management, 1998.
[6] John H. Cochrane, Asset Pricing revised edition, Princeton University Press, 2002.
[8] Stephen Morris, Lecture Notes for the course ECO 317: The Economics of Uncer-
tainty, 2011.