COHERENT MEASURES OF RISK
Philippe Artzner, Université Louis Pasteur, Strasbourg
Freddy Delbaen, Eidgenössische Technische Hochschule, Zürich
Jean-Marc Eber, Société Générale, Paris
David Heath, Carnegie Mellon University, Pittsburgh, Pennsylvania
July 22, 1998
Abstract. In this paper we study both market risks and non-market risks, without
complete markets assumption, and discuss methods of measurement of these risks.
We present and justify a set of four desirable properties for measures of risk, and call
the measures satisfying these properties “coherent”. We examine the measures of risk
provided and the related actions required by SPAN, by the SEC/NASD rules and by
quantile based methods. We demonstrate the universality of scenario-based methods
for providing coherent measures. We offer suggestions concerning the SEC method.
We also suggest a method to repair the failure of subadditivity of quantile-based
methods.
Key words and phrases. aggregation of risks, butterfly, capital requirement, coherent risk measure, concentration of risks, currency risk, decentralization, extremal events risk, insurance risk,
margin requirement, market risk, mean excess function, measure of risk, model risk, net worth,
quantile, risk-based capital, scenario, shortfall, subadditivity, tail value at risk, value at risk.
The authors acknowledge financial support from Société Générale for this work. The views expressed are those of the authors. They thank for useful discussions on a related paper, participants
of the following meetings: Boston University Mathematical Finance Day, March 31, 1996, University of Waterloo Insurance and Finance Conference, May 24, 1996, StudienZentrum Gerzensee
Symposium in Financial Markets, July 15-26, 1996, Latsis Symposium, ETH Zürich, September
24-25, 1996, Universität Karlsruhe Geld, Finanzwirtschaft, Banken, Versicherungen Tagung, December 11-13, 1996, Aarhus University Workshop, February 25-March 1, 1997. They also thank D.
Madan, discussant at the French Finance Association International Meeting, Geneva, June 1996,
F. Diebold, discussant at the Federal Reserve Bank of Atlanta 1997 Financial Markets Conference,
R. Bliss, Ph. Boyle, V. Brousseau, P. Embrechts, A. Hoffman, W. Neuefeind, Ch. Petitmengin,
P. Poncet, J. Renegar, E. Shiu as well as a referee of an earlier version of this paper.
Typeset by AMS-TEX
1
1. Introduction
We provide in this paper a definition of risks (market risks as well as non-market
risks) and present and justify a unified framework for the analysis, construction and
implementation of measures of risk. We do not assume completeness of markets.
These measures of risk can be used as (extra) capital requirements, to regulate the
risk assumed by market participants, traders, insurance underwriters, as well as to
allocate existing capital.
For these purposes, we:
(1) Define “acceptable” future random net worths (see Section 2.1) and provide
a set of axioms about the set of acceptable future net worths (Section 2.2);
(2) Define the measure of risk of an unacceptable position once a reference,
“prudent,” investment instrument has been specified, as the minimum extra capital
(see Section 2.3) which, invested in the reference instrument, makes the future value
of the modified position become acceptable;
(3) State axioms on measures of risk and relate them to the axioms on acceptance
sets. We argue that these axioms should hold for any risk measure which is to be
used to effectively regulate or manage risks. We call risk measures which satisfy
the four axioms coherent;
(4) Present, in Section 3, a (simplified) description of three existing methods
for measuring market risk: the “variance-quantile” method of value-at-risk (VaR),
the margin system SPAN (Standard Portfolio Analysis of Risk) developed by the
Chicago Mercantile Exchange, and the margin rules of the Securities and Exchanges
Commission (SEC), which are used by the National Association of Securities Dealers
(NASD);
(5) Analyze the existing methods in terms of the axioms and show that the last
two methods are essentially the same (i.e., that when slightly modified they are
mathematical duals of each other);
(6) Make a specific recommendation for the improvement of the NASD-SEC
margin system (Section 3.2);
(7) Examine in particular the consequences of using value at risk for risk management (Section 3.3);
(8) Provide a general representation for all coherent risk measures in terms of
“generalized scenarios” (see Section 4.1), by applying a consequence of the separation theorem for convex sets already in the mathematics literature;
(9) Give conditions for extending into a coherent risk measure a measurement
already agreed upon for a restricted class of risks (see Section 4.2);
(10) Use the representation results to suggest a specific coherent measure (see
Section 5.1) called tail conditional expectation, as well as to give an example of
construction of a coherent measure out of measures on separate classes of risks, for
example credit risk and market risk (see Section 5.2).
(11) Our axioms are not restrictive enough to specify a unique risk measure.
They instead characterize a large class of risk measures. The choice of precisely
which measure to use (from this class) should presumably be made on the basis
of additional economic considerations. Tail conditional expectation is, under some
assumptions, the least expensive among these which are coherent and accepted by
regulators since being more conservative than the value at risk measurement.
A non technical presentation of part of this work is given in [ADEH].
2
2. Definition of risk and of coherent risk measures
This section accomplishes the program set in (1), (2) and (3) above, in the
presence of different regulations and different currencies.
2.1 Risk as the random variable: future net worth.
Although several papers (including an earlier version of this one) define risk in
terms of changes in values between two dates, we argue that because risk is related
to the variability of the future value of a position, due to market changes or more
generally to uncertain events, it is better to instead consider future values only.
Notice indeed that there is no need for the initial costs of the components of the
position to be determined from universally defined market prices (think of over-thecounter transactions). The principle of “bygones are bygones” leads to this “future
wealth” approach.
The basic objects of our study shall therefore be the random variables on the set
of states of nature at a future date, interpreted as possible future values of positions
or portfolios currently held. A first, crude but crucial, measurement of the risk of
a position will be whether its future value belongs or does not belong to the subset
of acceptable risks, as decided by a supervisor like:
(a) a regulator who takes into account the unfavorable states when allowing
a risky position which may draw on the resources of the government, for
example as a guarantor of last resort;
(b) an exchange’s clearing firm which has to make good on the promises to all
parties, of transactions being securely completed;
(c) an investment manager who knows that his firm has basically given to its
traders an exit option where the strike “price” consists in being fired in the
event of big trading losses on one’s position.
In each case above, there is a trade-off between severity of the risk measurement,
and level of activities in the supervised domain. The axioms and characterizations
we shall provide do not single out a specific risk measure, and additional economic
considerations have to play a role in the final choice of a measure.
For an unacceptable risk (i.e. a position with an unacceptable future net worth)
one remedy may be to alter the position. Another remedy is to look for some
commonly accepted instruments which, added to the current position, make its
future value become acceptable to the regulator/supervisor. The current cost of
getting enough of this or these instrument(s) is a good candidate for a measure of
risk of the initially unacceptable position.
For simplicity, we consider only one period of uncertainty (0, T ) between two
dates 0 and T . The various currencies are numbered by i , 1 ≤ i ≤ I and, for
each of them, one “reference” instrument is given, which carries one unit of date 0
currency i into ri units of date T currency i . Default free zero coupon bonds with
maturity at date T may be chosen as particularly simple reference instruments in
their own currency. Other possible reference instruments are mentionned in Section
2.3, right before the statement of Axiom T.
The period (0, T ) can be the period between hedging and rehedging, a fixed
interval like two weeks, the period required to liquidate a position, or the length of
coverage provided by an insurance contract.
We take the point of view of an investor subject to regulations and/or supervision
in country 1. He considers a portfolio of securities in various currencies.
3
Date 0 exchange rates are supposed to be one, while ei denotes the random
number of units of currency 1 which one unit of currency i buys at date T .
An investor’s initial portfolio consists of positions Ai , 1 ≤ i ≤ I, (possibly within
some institutional constraints, like the absence of short sales and a “congruence” for
each currency between assets and liabilities). The position Ai provides
Ai (T ) units
P
of currency i at date T. We call risk the investor’s future net worth 1≤i≤I ei ·Ai (T ).
Remark. The assumption of the position being held during the whole period can be
relaxed substantially. In particular, positions may vary due to the agent’s actions or
those of counterparties. In general, we can consider the risk of following a strategy
(which specifies the portfolio held at each date as a function of the market events
and counterparties’ actions) over an arbitrary period of time. Our current results
in the simplified setting represent a first step.
2.2 Axioms on Acceptance sets, i.e. sets of acceptable future net worths.
We suppose that the set of all possible states of the world at the end of the period
is known, but the probabilities of the various states occurring may be unknown or
not subject to common agreement. When we deal with market risk, the state of the
world might be described by a list of the prices of all securities and all exchange
rates, and we assume that the set of all possible such lists is known. Of course,
this assumes that markets at date T are liquid; if they are not, more complicated
models are required, where we can distinguish the risks of a position and of a future
net worth, since, with illiquid markets, the mapping from the former to the latter
may not be linear.
Notation.
(a) We shall call Ω the set of states of nature, and assume it is finite. Considering Ω as the set of outcomes of an experiment, we compute the final net
worth of a position for each element of Ω. It is a random variable denoted
by X. Its negative part, max(−X, 0), is denoted by X − and the supremum
of X − is denoted by kX − k. The random variable identically equal to 1 is
denoted by 1. The indicator function of state ω is denoted by 1{ω} .
(b) Let G be the set of all risks, that is the set of all real valued functions on
Ω. Since Ω is assumed to be finite, G can be identified with Rn , where
n = card(Ω). The cone of non-negative elements in G shall be denoted by
L+ , its negative by L− .
(c) We call Ai,j , j ∈ Ji , a set of final net worths, expressed in currency i, which,
in country i, are accepted by regulator/supervisor
j.
T
(d) We shall denote Ai the intersection j∈Ji Ai,j and use the generic notation
A in the listing of axioms below.
We shall now state axioms for acceptance sets. Some have an immediate interpretation while the interpretation of the third one will be more easy in terms of
risk measure (see Axiom S in Section 2.3.) The rationale for Axioms 2.1 and 2.2 is
that a final net worth which is always nonnegative does not require extra capital,
while a net worth which is always (strictly) negative certainly does.
Axiom 2.1. The acceptance set A contains L+ .
Axiom 2.2. The acceptance set A does not intersect the set L−− where
L−− = {X | for each ω ∈ Ω , X(ω) < 0}.
It will also be interesting to consider a stronger axiom:
4
Axiom 2.2’. The acceptance set A satisfies A ∩ L− = {0}.
The next axiom reflects risk aversion on the part of the regulator, exchange
director or trading room supervisor.
Axiom 2.3. The acceptance set A is convex.
A less natural requirement on the set of acceptable final net worths is:
Axiom 2.4. The acceptance set A is a positively homogeneous cone.
2.3 Correspondence between acceptance sets and measures of risk.
Sets of acceptable future net worths are the primitive objects to be considered
in order to describe acceptance or rejection of a risk. We present here how, given
some “reference instrument”, there is a natural way to define a measure of risk by
describing how close or how far from acceptance a position is.
Definition 2.1. A measure of risk is a mapping from G into IR.
In Section 3 we shall speak of a model-dependent measure of risk when an explicit
probability on Ω is used to construct it (see e.g. Sections 3.1 and 3.3), and of a
model-free measure otherwise (see e.g. Section 3.2). Model-free measures can still
be used in the case where only risks of positions are considered.
When positive, the number ρ(X) assigned by the measure ρ to the risk X will
be interpreted (see Definition 2.2 below) as the minimum extra cash the agent has
to add to the risky position X, and to invest “prudently”, that is in the reference
instrument, to be allowed to proceed with his plans. If it is negative, the cash
amount −ρ(X) can be withdrawn from the position, or received as restitution as
in the case of organized markets for financial futures.
Remark 1. The reader may be surprised that we define a measure of risk on the
whole of G. Why, in particular, should we consider a risk, a final net worth, like
the constant −1 ? No one would or could willingly enter into a deal which for sure
entails a negative of final net worth equal to 1 ! Let us provide three answers:
(a) we want to extend the accounting procedures dealing with future certain
bad events (like loss in inventories, degradation [wear and tear] of physical
plant), into measurement procedures for future uncertain bad events;
(b) actual measurements used in practice seem to be indeed defined only for
risks where both states with positive and states with negative final net worth
exist. Section 4.2 shows that, under well-defined conditions, they can be
extended without ambiguity to measurements for all functions in G ;
(c) multiperiod models may naturally introduce at some intermediate date the
prospect of such final net worths.
Remark 2. It has been pointed out to us that describing risk “by a single number”
involves a great loss of information. However, the actual decison about taking a
risk or allowing one to take it is fundamentally binary, of the “yes or no” type,
and we claimed at the beginning of Section 2.1 that this is the actual origin of risk
measurement.
Remark 3. The expression “cash” deserves some discussion in the case of a publicly
traded company. It refers to an increase in equity. The amount ρ(X) may, for
example, be used to lower the amount of debt in the balance sheet of the company.
We define a correspondence between acceptance sets and measures of risk.
5
Definition 2.2. Risk measure associated to an acceptance set: given the total rate
of return r on a reference instrument, the risk measure associated to the acceptance
set A is the mapping from G to IR denoted by ρA,r and defined by
ρA,r (X) = inf{m | m · r + X ∈ A}.
Remark. Acceptance sets allow us to address a question of importance to an international regulator and to the risk manager of a multinational firm, namely the
invariance of acceptability of a position with respect to a change of currencies. If,
indeed, we have for each currency i, 1 ≤ i ≤ I , ei · Ai = A1 then, for each position providing an acceptable future net worth X in currency i, the same position
provides a future net worth ei /ej · X in currency j, which is also acceptable. The
situation is more complex for unacceptable positions. If a position requires an extra
initial cash of ρAi ,ri (X) units to be invested in the i-th reference instrument, it is
not necessarily true that this amount is equal to the number ρAj ,rj (X) of initial
units deemed sufficient by the regulation(s) in country j, if invested in the j-th
reference instrument, even though we supposed the initial exchange rate to be 1.
Definition 2.3. Acceptance set associated to a risk measure: the acceptance set
associated to a risk measure ρ is the set denoted by Aρ and defined by
Aρ = {X ∈ G | ρ(X) ≤ 0}.
We consider now several possible properties for a risk measure ρ defined on
G. They will be related, in Section 2.4, to the axioms stated above concerning
acceptance sets. For clarity we label the new axioms with letters.
The first requirement ensures that the risk measure is stated in the same units as
the final net worth, except for the use of the reference instrument. This particular
asset is modeled as having the initial price 1 and a strictly positive price r (or
total return) in any state of nature at date T. It is the regulator’s (supervisor’s)
responsibility to accept for r possible random values as well as values smaller than
1.
Axiom T means that adding (resp. subtracting) the sure initial amount α to the
initial position and investing it in the reference instrument, simply decreases (resp.
increases) the risk measure by α.
Axiom T. Translation invariance: for all X ∈ G and all real numbers α, we have
ρ(X + α · r) = ρ(X) − α.
Remark 1. Axiom T ensures that, for each X, ρ(X + ρ(X) · r) = 0. This equality
has a natural interpretation in terms of the acceptance set associated to ρ (see
Definition 2.3 above.)
Remark 2. By insisting on references to cash and to time, Axiom T clearly indicates
that our approach goes much farther than the interpretation given by Wang of an
earlier version of this paper: [Wan], page 3, indeed claims that “the main function
of a risk measure is to properly rank risks.”
Axiom S. Subadditivity: for all X1 and X2 ∈ G, ρ(X1 + X2 ) ≤ ρ(X1 ) + ρ(X2 ).
We contend that this property, which could be stated in the following brisk form
“a merger does not create extra risk,” is a natural requirement:
6
(a) if an exchange’s risk measure were to fail to satisfy this property, then, for
example, an individual wishing to take the risk X1 + X2 may open two
accounts, one for the risk X1 and the other for the risk X2 , incurring the
smaller margin requirement of ρ(X1 ) + ρ(X2 ), a matter of concern for the
exchange;
(b) if a firm were forced to meet a requirement of extra capital which did not
satisfy this property, the firm might be motivated to break up into two
separately incorporated affiliates, a matter of concern for the regulator;
(c) bankruptcy risk inclines society to require less capital from a group without
“firewalls” between various business units than it does require when one
“unit” is protected from liability attached to failure of another “unit”;
(d) suppose that two desks in a firm compute in a decentralized way, the measures ρ(X1 ) and ρ(X2 ) of the risks they have taken. If the function ρ is
subadditive, the supervisor of the two desks can count on the fact that
ρ(X1 ) + ρ(X2 ) is a feasible guarantee relative to the global risk X1 + X2 .
If indeed he has an amount m of cash available for their joint business,
he knows that imposing limits m1 and m2 with m = m1 + m2 , allows him
to decentralise his cash constraint into two cash constraints, one per desk.
Similarly, the firm can allocate its capital among managers.
Axiom PH. Positive homogeneity: for all λ ≥ 0 and all X ∈ G, ρ(λX) = λρ(X).
Remark 1. If position size directly influences risk (for example, if positions are
large enough that the time required to liquidate them depend on their sizes) then
we should consider the consequences of lack of liquidity when computing the future
net worth of a position. With this in mind, Axioms S and PH about mappings from
random variables into the reals, remain reasonable.
Remark 2. Axiom S implies that ρ(nX) ≤ nρ(X) for n = 1, 2, ... . In Axiom PH
we have imposed the reverse inequality (and require equality for all positive λ) to
model what a government or an exchange might impose in a situation where no
netting or diversification occurs, in particular because the government does not
prevent many firms to all take the same position.
Remark 3. Axioms T and PH imply that for each α , ρ(α · (−r)) = α.
Axiom M. Monotonicity: for all X and Y ∈ G with X ≤ Y, we have ρ(Y ) ≤ ρ(X).
Remark. Axiom M rules out the risk measure defined by ρ(X) = −EP [X]+α·σP (X),
where α > 0 and where σP denotes the standard deviation operator, computed
under the probability P. Axiom S rules out the “semi-variance” type risk measure
defined by ρ(X) = −EP [X] + σP ((X − EP [X])− ).
Axiom R. Relevance: for all X ∈ G with X ≤ 0 and X 6= 0, we have ρ(X) > 0.
Remark. This axiom is clearly necessary, but not sufficient, to prevent concentration of risks to remain undetected (see Section 4.3.)
We notice that for λ > 0, Axioms S, PH, M and R remain satisfied by the
measure λ · ρ, if satisfied by the measure ρ. It is not the case for Axiom T.
The following choice of required properties will define coherent risk measures.
Definition 2.4. Coherence: a risk measure satisfying the four axioms of translation invariance, subadditivity, positive homogeneity, and monotonicity, is called
coherent.
7
2.4 Correspondence between the axioms on Acceptance Sets and the
axioms on Measures of risks.
The reader has certainly noticed that we claimed the acceptance set to be the
fundamental object, and discussed the axioms mostly in terms of the associated
risk measure. The following propositions show that this was reasonable.
Proposition 2.1. If the set B satisfies Axioms 2.1, 2.2, 2.3 and 2.4, the risk
measure ρB,r is coherent. Moreover AρB,r = B̄, the closure of B.
Proof of Proposition 2.1. (1) Axioms 2.2 and 2.3 ensure that for each X, ρB,r (X)
is a finite number.
(2) The equality inf{p | X + (α + p) · r ∈ B} = inf{q | X + q · r ∈ B} − α proves
that ρB,r (X + r · α) = ρ(X) − α, and Axiom T is satisfied.
(3) The subadditivity of ρB follows from the fact that if X + m · r and Y + n · r
both belong to B, so does X + Y + (m + n) · r as Axioms 2.3 and 2.4 show.
(4) If m > ρB,r (X) then for each λ > 0 we have λ · X + λ · m · r ∈ B, by Definition
2.3 and Axiom 2.4, and this proves that ρB,r (λ · X) ≤ λ · m. If m < ρB,r (X), then
for each λ > 0 we have λ · X + λ · m · r ∈
/ B, and this proves that ρB,r (λ · X) ≥ λ · m.
We conclude that ρB,r (λ · X) = λ · ρB,r (X).
(5) Monotonicity of ρB,r follows from the fact that if X ≤ Y and X + m · r ∈
B then Y + m · r ∈ B by use of Axioms 2.3 and 2.1, and of Definition 2.3.
(6) For each X ∈ B, ρB,r (X) ≤ 0 hence X ∈ AρB,r . Proposition 2.2 and points
(1) through (5) above ensure that AρB,r is closed, which proves that AρB,r = B̄.
Proposition 2.2. If a risk measure ρ is coherent, then the acceptance set Aρ is
closed and satisfies Axioms 2.1, 2.2, 2.3 and 2.4. Moreover ρ = ρAρ ,r .
Proof of Proposition 2.2. (1) Subadditivity and positive homogeneity ensure that ρ
is a convex function on G, hence continuous, and that the set Aρ = {X | ρ(X) ≤ 0}
is a closed, convex and homogeneous cone.
(2) Positive homogeneity implies that ρ(0) = 0. Together with monotonicity this
ensures that the set Aρ contains the positive orthant L+ .
(3) Let X be in L−− with ρ(X) < 0. Axiom M ensures that ρ(0) < 0, a contradiction. If ρ(X) = 0, then we find α > 0 such that X + α · r ∈ L−− , which provides,
by use of Axiom T, the relation −α ≥ 0, a contradiction. Hence ρ(X) > 0, that is
X∈
/ Aρ , which establishes Axiom 2.2.
(4) For each X, let δ be any number with ρAρ ,r (X) < δ. Then X +δ·r ∈ Aρ , hence
ρ(X + δ · r) ≤ 0, hence ρ(X) ≤ δ, which proves that ρ(X) ≤ ρAρ ,r (X), that is ρ ≤
ρAρ ,r .
(5) For each X, let δ be any number with δ > ρ(X), then ρ(X +δ·r) < 0 and X +
δ · r ∈ Aρ , hence ρAρ ,r (X + δ · r) ≤ 0. This proves that ρAρ ,r (X) ≤ δ and that
ρAρ ,r (X) ≤ ρ(X), hence ρAρ ,r ≤ ρ.
Proposition 2.3. If a set B satisfies Axioms 2.1, 2.2’, 2.3 and 2.4, then the
coherent risk measure ρB,r satisfies the relevance axiom. If a coherent risk measure
ρ satisfies the relevance axiom, then the acceptance set AρB ,r satisfies Axiom 2.2’.
Proof of Proposition 2.3. (1) For an X like in the statement of Axiom R we know
that X ∈ L− and X 6= 0, hence, by Axiom 2.2’, X ∈
/ B, which means ρB,r (X) > 0.
(2) For X ∈ L− and X 6= 0 Axiom R provides ρ(X) > 0 and X ∈
/ B.
8
3. Three currently used methods of measuring market risk
In this section, we give a (simplified) description of three currently used methods
of measuring market risk:
a - SPAN [Sp] developed by the Chicago Mercantile Exchange,
b - the Securities Exchange Commission rules used by the National Association
of Securities Dealers (see [NASD] and [Fed]), similar to rules used by the
Pacific Exchange and the Chicago Board of Options Exchange
c - the quantile-based Value at Risk (or VaR) method [B], [D], [DP], [DPG],
[Risk], [RM].
We examine the relationship of these three methods with the abstract approach
provided in Section 2. We also suggest slightly more general forms for some of the
methods. It will be shown that the distinction made above between model-free and
model-dependent measures of risk actually shows up.
3.1 An organized exchange’s rules: The SPAN computations.
To illustrate the SPAN margin system [Sp] (see also [Ma], pages 7-8), we consider
how the initial margin is calculated for a simple portfolio consisting of units of a
futures contract and of several puts and calls with a common expiration date on
this futures contract. The SPAN margin for such a portfolio is computed as follows:
First, fourteen “scenarios” are considered. Each scenario is specified by an up or
down move of volatility combined with no move, or an up move, or a down move
of the futures price by 1/3, 2/3 or 3/3 of a specified “range.” Next, two additional
scenarios relate to “extreme” up or down moves of the futures price. The measure of
risk is the maximum loss incurred, using the full loss for the first fourteen scenarios
and only 35% of the loss for the last two “extreme” scenarios. A specified model,
typically the Black model, is used to generate the corresponding prices for the
options under each scenario.
The calculation can be viewed as producing the maximum of the expected loss
under each of sixteen probability measures. For the first fourteen scenarios the
probability measures are point masses at each of the fourteen points in the space
Ω of securities prices. The cases of extreme moves correspond to taking the convex
combination (0.35, 0.65) of the losses at the “extreme move” point under study and
at the “no move at all” point (i.e., prices remain the same). We shall call these
probability measures “generalized scenarios”.
The account of the investor holding a portfolio is required to have sufficient
current net worth to support the maximum expected loss. If it does not, then
extra cash is required as margin call, in an amount equal to the “measure of risk”
involved. This is completely in line with our interpretation of Definition 2.3.
The following definition generalizes the SPAN computation and presents it in
our framework:
Definition 3.1. The risk measure defined by a non-empty set P of probability
measures or “generalized scenarios” on the space Ω and the total return r on a
reference instrument, is the function ρP on G defined by
ρP (X) = sup{EP [−X/r] | P ∈ P}.
The scenario-based measures from Definition 3.1 are coherent risk measures:
9
Proposition 3.1. Given the total return r on a reference instrument and the nonempty set P of probability measures, or “generalized scenarios”, on the set Ω of
states of the world, the risk measure ρP of Definition 3.1 is a coherent risk measure. It satisfies the relevance axiom if and only if the union of the supports of the
probabilities P ∈ P is equal to the set Ω.
Proof of Proposition 3.1. Axioms PH and M ensure that a coherent risk measure
satisfies Axiom R if and only if the negative of each indicator function 1{ω} has a
(strictly) positive risk measure. This is equivalent to the fact that any state belongs
to at least one of the supports of the probabilities found in the set P.
Section 4.1 shows that each coherent risk measure is obtained by way of scenarios.
3.2 Some model-free measures of risks: The SEC rules on final net
worth.
The second example of a risk measure used in practice is found in the rules of
the Securities and Exchange Commission and the National Association of Securities
Dealers. Their common approach is to consider portfolios as formal lists of securities
and impose “margin” requirements on them, in contrast to the SPAN approach
which takes the random variables - gains and losses of the portfolios of securities
- as basic objects to measure. In the terminology of [B] we have here something
similar to a “standardized measurement method”.
Certain spread positions like a long call and a short call of higher exercise price,
both calls having same maturity date, are described in [NASD], page 8133, SEC
rule 15c3-1a,(11), as requiring no margin (no “deduction”). No justification is given
for this specification. We shall use the paper [RS] as the basis for explaining, for a
simple example, the computation of margin according to these common rules.
Let A be a portfolio consisting of two long calls with strike 10, two short calls
with strike 20, three short calls with strike 30, four long calls with strike 40 and
one short call with strike 50. For simplicity assume all calls European and exercise
dates equal to the end of the holding period. A simple graph shows that the final
value of this position is never below −10, which should entail a margin deposit of
at most 10.
Under the SEC method, the position A is represented or “decomposed” as a
portfolio of long call spreads. No margin is required for a spread if the strike of the
long side is less than the strike of the short side. A margin of K − H is required for
the spread consisting of a long call with strike K and a short call with strike H, when
H ≤ K. The margin resulting from a representation or “ decomposition” is the
sum of the margins attached to each call spread. The investor is presumably able to
choose the best possible representation. A simple linear programming computation
will show that 30 is the resulting minimum, that is much more than the negative
of the worst possible future value of the position!
Remark 1. This 30 seems to be the result of an attempt to bound the largest payout
which the investor might have to make at the end of the period. In this method,
the current value of his account must be at least as large as the current value of
the calls plus 30.
Remark 2. A careful reading of the SEC rules reveals that one must:
a - first mark the account (reference instruments plus calls) to market,
b - deduct the market value of the calls (long or short),
10
c - then deduct the various “margins” required for the spreads in the chosen
decomposition (we shall call the total as the “margin,”)
d - and then check that this is at least 0.
In the framework of Definition 2.3, this bears some analogy to
a - marking to market both the positions in the “risky” instruments as well as
in the reference one,
b - subtract the market value of the risky part,
c - make sure that the difference is positive.
We now formalize the special role played by the call spreads, which we call
“standard risks,” and the natural margin requirements on them in the SEC rules
approach to risk measurement, following the lines of [RS], Section 4 (see also [CR],
pages 107-109). Given some underlying security, we denote by CK the European
call with exercise price K and exercise date equal to the end of the holding period,
and by SH,K the spread portfolio consisting of “one long CH , one short CK ” , which
we also denote by CH − CK . These spreads shall be “standard risks” for which a
simple rule of margin requirement is given. They are then used to “support” general
portfolios of calls and provide conservative capital requirements.
We describe the extra capital requirement for a portfolio A consisting
Pof aH calls
CH , H ∈ H, H a finite set of strikes. For simplicity we assume that H aH = 0,
i.e., we have no net long or short position. The exchange allows one to compute
the margin for such a portfolio A by solving the linear programming problem:
(3.1)
inf
nH,K
X
nH,K (H − K)+
H,K,H6=K
under the conditions that
for all H, K, H 6= K we have nH,K ≥ 0 and A =
X
nH,K SH,K .
H,K,H6=K
This program provides the holder of portfolio A with the cheapest decomposition
ensuring that each spread showing in it has a non-negative net worth at date T.
Going one step farther than Rudd and Schroeder (pages 1374-1376) we write
explicitly the dual program:
(3.2)
sup
νK
X
νK aK
K
where the sup is taken over all (νK ) satisfying: νH − νK ≤ (H − K)+ .
For the interpretation of this dual problem, we rewrite the preceding program
with the negative πK of the dual variables, getting:
(3.3)
inf
πK
X
πK aK
K
under the conditions that
πH −πK ≥ −(H −K)+ , or πH −πK ≥ 0 if H < K and πH −πK ≥ K −H if H > K
11
the last inequalities being rewritten as
(3.4)
πK − πH ≤ H − K if H > K.
Notice that if we interpret πH as the cash flows associated with the call CH at
expiration date T , the objective function in (3.3) is the cash flow of the portfolio A
at expiration. The duality theorem of linear programming ensures that the worst
payout to the holder of portfolio A, under all scenarios satisfying the constraints
specified in problem (3.3), cannot be larger than the lowest margin accepted by the
exchange. The exchange is therefore sure than the investor commitments will be
fulfilled.
It is remarkable that the primal problem (3.1) did not seem to refer to a model
of distribution for future prices of the call. Yet the duality results in an implicit
set of states of nature consisting of call prices, with a surprise! Our example of
portfolio A in the beginning of this Section has shown indeed that the exchange is,
in some way, too secure, as we now explain.
That the cash flows of the calls must satisfy the constraints (3.4) specified for
problem (3.3) (and indeed many other constraints such as convexity as a function
of strike, see [Me], Theorem 8.4) is well known. For the specific portfolio A studied
in Section 3.2, the set of strikes is H = {10, 20, 30, 40, 50}, and an optimal primal
solution is given by n∗10,20 = 2, n∗40,50 = 1, n∗40,30 = 3 , all others n∗H,K = 0, for a
∗
∗
∗
∗
minimal margin of 30. The cash flows are given by π10
= π20
= π30
= 10 and π40
=
∗
π50 = 0, which provides the value −30 for the minimal cash flow of the portfolio
at expiration. However, this minimal cash flow corresponds to cash flows for the
individual options which cannot arise for any stock price. Indeed, if the stock price
at expiration is S, the cash flow of CH is (S − H)+ , which is obviously convex in H.
∗
∗
∗
Thus since π20
+ π40
< 2π30
, these π ′ s cannot arise as cash flows for any terminal
stock price. Briefly, there are too many scenarios considered, because some of them
are impossible scenarios.
The convexity of the call price as function of the strike can be derived from
the fact that a long “butterfly” portfolio as B20 = C10 − 2C20 + C30 must have a
positive price. Therefore, we submit this butterfly to the decomposition method
and write it as a sum of spreads S10,20 + S30,20 , which requires a margin of 10.
If we instead take the approach of Section 2, looking at random variables, more
precisely at the random net worth at the end of the holding period, we realize that
the butterfly never has negative net worth, or, equivalently, that the net loss it
can suffer is never larger than its initial net worth. The butterfly portfolio should
therefore be margin free, which would imply a margin of only 10 for the original
portfolio A = 2B20 + 2B30 − B40 . In our opinion it is not coherent, in this setting,
to have only the spreads SH,K (for H ≤ K) as margin free portfolios. The method
uses too few standard risks.
In Section 4.2 we present a framework for extensions of risk measurements of
“standard risks” and give conditions under which our construction actually produces coherent measures of risk. The results of Section 4.1 on scenario representation of coherent measures will allow to interpret the extension in terms of scenarios
attached to the original measurement.
3.3 Some model-dependent rules based on quantiles.
The last example of measures of risk used in practice is the “Value at Risk”
(or VaR) measure. It is usually defined in terms of net wins or P/L and therefore
12
ignores the difference between money at one date and money at a different date,
which, for small time periods and a single currency, may be acceptable. It uses
quantiles, which requires us to pay attention to discontinuities and intervals of
quantile numbers.
Definition 3.2. Quantiles: given α ∈]0, 1[ the number q is an α − quantile of the
random variable X under the probability distribution P if one of the three equivalent
properties below is satisfied:
a - P [X ≤ q] ≥ α ≥ P [X < q] ,
b - P [X ≤ q] ≥ α and P [X ≥ q] ≥ 1 − α ,
c - FX (q) ≥ α and FX (q−) ≤ α with FX (q− ) = limx→q,x<q F (x) , where FX is
the cumulative distribution function of X.
Remark. The set of such α-quantiles is a closed interval. Since Ω is finite, there is
a finite left-(resp. right-) end point qα− (resp. qα+ ) which satisfies qα− = inf{x |
P [X ≤ x] ≥ α} [equivalently sup{x | P [X ≤ x] < α}] (resp. qα+ = inf{x |
P [X ≤ x] > α}). With the exception of at most countably many α the equality
− (α) = inf{x | P{X ≤ x} ≥ α}
qα− = qα+ holds. The quantile qα− is the number F ←
defined in [EKM] Definition 3.3.5 (see also [DP]).
We formally define VaR in the following way:
Definition 3.3. Value at risk measurement: given α ∈]0, 1[, and a reference instrument r, the value-at-risk V aRα at level α of the final net worth X with distribution
P, is the negative of the quantile qα+ of X/r, that is
V aRα (X) = − inf{x | P [X ≤ x · r] > α}.
Remark 1. Notice that what we are using for defining V aRα is really the amount
of additional capital that a V aRα type calculation entails.
Remark 2. We have here what is called an “internal” model in [B], and it is not
clear whether the (estimated) physical probability or a “well-chosen” subjective
probability should be used.
We will now show that, while satisfying properties T, PH and M, V aRα fails to
satisfy the subadditivity property.
Consider as an example, the following two digital options on a stock, with the
same exercise date T , the end of the holding period. The first option denoted by A
(initial price u) pays 1000 if the value of the stock at time T is more than a given
U , and nothing otherwise, while the second option denoted by B (initial price l)
pays 1000 if the value of the stock at T is less than L (with L < U ), and nothing
otherwise.
Choosing L and U such that P{ST < L} = P{ST > U } = 0.008 we look for the
1% values at risk of the future net worths of positions taken by two traders writing
respectively 2 options A and 2 options B. They are −2 · u and −2 · l respectively (r
supposed to be one). By contrast, the positive number 1000 − l − u is the 1% value
at risk of the future net worth of the position taken by a trader writing A + B.
This implies that the set of acceptable net worths (in the sense of Definition 2.4
applied to the value at risk measure) is not convex. Notice that this is an even
worse feature than the non-subadditivity of the measurement. We give below one
more example of non-subadditivity.
13
Remark 1. We note that if quantiles are computed under a distribution for which all
prices are jointly normally distributed, then the quantiles do satisfy subadditivity
as long as probabilities of excedence are smaller than 0.5. Indeed, σX+Y ≤ σX + σY
for each pair (X, Y ) of random variables. Since for a normal random variable X we
have
V aRα (X) = −(EP [X] + Φ−1 (α) · σP (X) ),
with Φ the cumulative standard normal distribution and since Φ−1 (0.5) = 0, the
proof of subadditivity follows.
Remark 2. Several works on quantile-based measures (see [D], [Risk], [RM]) consider mainly the computational and statistical problems they raise, without first
considering the implications of this method of measuring risks.
Remark 3. Since the beginning of this century, casualty actuaries have been involved in computation and use of quantiles. The choice of initial capital controls
indeed the probabilities of ruin at date T . Loosely speaking, “ruin” is defined in
(retrospective) terms by the negativity, at date T, of the surplus, defined to be :
Y = capital at date 0 + premium received - claims paid (from date 0 to date T).
Imposing an upper bound 1 − α on the probability of Y being negative determines
the initial capital via a quantile calculation (for precise information, see the survey
article by Hans Bühlmann, Tendencies of Development in Risk Theory, in [Cen],
pages 499-522).
Under some circumstances, related to Remark 1 above, (see [DPP], pages 157,
168), this “capital at risk” is a measure which possesses the subadditivity property.
For some models the surplus represents the net worth of the insurance firm at date
T . In general, the difficulty of assigning a market value to insurance liabilities forces
us to distinguish surplus and net worth.
Remark 4. We do not know of organized exchanges using value at risk as the basis
of risk measurement for margin requirements.
For a second example of non-subadditivity, briefly allow an infinite set Ω and
consider two independent identically distributed random variables X1 and X2 having the same density 0.90 on the interval [0, 1], the same density 0.05 on the interval
[−2, 0]. Assume that each of them represents a future random net worth with positive expected value, that is a possibly interesting risk. Yet, in terms of quantiles,
the 10% values at risk of X1 and X2 being equal to 0, whereas an easy calculation
showing that the 10% value at risk of X1 +X2 is certainly larger than 0, we conclude
that the individual controls of these risks do not allow directly a control of their
sum, if we were to use the 10% value at risk.
Value at risk measurement also fails to recognise concentration of risks. A remarkably simple example concerning credit risk is due to Claudio Albanese (see
[Alba]). Assume that the base rate of interest is zero, and that the spreads on all
corporate bonds is 2%, while these bonds default, independently from company to
company, with a (physical) probability of 1%. If an amount of 1, 000, 000 borrowed
at the base rate is invested in the bonds of a single company, the 5% value at risk
of the resulting position is negative, namely −20, 000 , and there is “no risk”.
If, in order to diversify, the whole amount is invested equally into bonds of
one hundred different companies, the following happens in terms of value at risk.
14
Since the probability of at least two companies defaulting is greater than 0.18
it follows that the portfolio of bonds leads to a negative future net worth with a
probability greater than 0.05 : diversification of the original portfolion has increased
the measure of risk, while the “piling-up” of risky bonds issued by the same company
had remained undetected. We should not rely on such “measure”.
Value at risk also fails to encourage a reasonable allocation of risks among agents
as can be seen from the simple following example. Let Ω consists of three states
ω1 , ω2 , ω3 with respective probabilities 0.94 , 0.03 , 0.03. Let two agents have the
same future net worth X with X(ω1 ) ≥ 0 , X(ω2 ) = X(ω3 ) = −100. If one uses the
5% value at risk measure, one would not find sufficient an extra capital (for each
agent) of 80. But this same capital would be found more than sufficient, for each
agent, if, by a risk exchange, the two agree on the modified respective future net
worths Y and Z, where Y (ω1 ) = Z(ω1 ) = X(ω1 ) , Y (ω2 ) = Z(ω3 ) = −120 , Y (ω3 ) =
Z(ω2 ) = −80. This is not reasonable since the allocation (X + 80, X + 80) Pareto
dominates the allocation (Y + 80, Z + 80) if the agents are risk averse.
In conclusion, the basic reasons to reject the value at risk measure of risks are
the following:
(a) value at risk does not behave nicely with respect to addition of risks, even
independent ones, creating severe aggregation problems.
(b) the use of value at risk does not encourage and, indeed, sometimes prohibits
diversification, because value at risk does not take into account the economic
consequences of the events the probabilities of which it controls.
15
4. Representation Theorems for Coherent Risk Measures
This section provides two representations of coherent risk measures. The first
corresponds exactly to the SPAN example of Section 3.1 and the second is the proper
generalisation of the NASD/SEC examples of Section 3.2. These representation
results are used in Section 5.2 to provide an example of algorithm to measure risks
in trades involving two different sources of randomness, once coherent measures of
risks for trades dealing with only one of these sources have been agreed upon.
4.1 Representation of coherent risk measures by scenarios.
In this section we show that Definition 3.1 provides the most general coherent
risk measure: any coherent risk measure arises as the supremum of the expected
negative of final net worth for some collection of “generalized scenarios” or probability measures on states of the world. We continue to suppose that Ω is a finite
set, otherwise we would also get finitely additive measures as scenarios.
The σ-algebra, 2Ω , is the class of all subsets of Ω. Initially there is no particular
probability measure on Ω.
Proposition 4.1. Given the total return r on a reference investment, a risk measure ρ is coherent if and only if there exists a family P of probability measures on
the set of states of nature, such that
ρ(X) = sup{EP [−X/r] | P ∈ P}.
Remark 1. We note that ρ can also be seen as an insurance premium principle. In
that case, denoting by R the physical measure, we find that the condition R ∈ P (or
in the convex hull of this set), is of great importance. This condition is translated
as follows: for all X ≤ 0 we have ER [−X/r] ≤ ρ(X).
Remark 2. The more scenarios considered, the more conservative (i.e. the larger)
is the risk measure obtained.
Remark 3. We remind the reader about Proposition 3.1. It will prove that Axiom
R is satisfied by ρ if and only if the union of the supports of the probabilities in P
is the whole set Ω of states of nature.
Proof of Proposition 4.1. (1) We thank a referee for pointing out that the mathematical content of Proposition 4.1, which we had proved on our own, is already in
the book [Hu]. We therefore simply identify the terms in Proposition 2, Chapter
10 of [Hu] with these of our terminology of risks and risk measure.
(2) The sets Ω and M of [Hu] are our set Ω and the set of probabilities on Ω.
Given a risk measure ρ we associate to it the functional E ∗ by E ∗ (X) = ρ(−r · X).
Axiom M for ρ is equivalent to Property (2.7) of [Hu] for E ∗ , Axioms PH and T
together are equivalent to Property (2.8) for E ∗ , and Axiom S is Property (2.9).
(3) The “if” part of our Proposition 4.1 is obvious. The “only if” part results
from the “representability” of E ∗ , since Proposition (2.1) of [Hu] states that
ρ(X) = E ∗ (−X/r) = sup{EP [−X/r] | P ∈ Pρ }
where Pρ is defined as the set
{P ∈ M | for all X ∈ G : EP [X] ≤ E ∗ (X) = ρ(−r · X)}
16
= {P ∈ M | for all Y ∈ G : EP [−Y /r] ≤ ρ(Y )} .
Remark 1. Model risk can be taken into account by including into the set P a family
of distributions for the future prices, possibly arising from other models.
Remark 2. Professor Bühlmann kindly provided us with references to works by
Hattendorf, [Hat], Kanner, [Kan], and Wittstein, [Wit], which he had mentionned in
his Göttingen presentation ([Bü]). These authors consider, in the case of insurance
risks, possible losses only, neglecting the case of gains. For example, risk for a
company providing annuities is linked to the random excess number of survivors over
the expected number given by the lifetable. Several of these references, for example
[Hat], §3, page 5, contain an example of a risk measure used in life insurance,
namely the “mittlere Risico” constructed out of one scenario, related to the life
table used by a company. It is defined as the mathematical expectation of the
positive part of the loss, as “die Summe aller möglichen Verluste, jeden multipliciert
in die Wahrscheinlichkeit seines Eintretens”. This procedure defines a risk measure
satisfying Axioms S, PH, M.
Remark 3. It is important to distinguish between a point mass scenario and a simulation trial: the first is chosen by the investor or the supervisor, while the second
is chosen randomly according to a distribution they have prescribed beforehand.
Conclusion. The result in Proposition 4.1 completely explains the occurrence of the
first type of actual risk measurement, the one based on scenarios, as described in
Section 3.1. Any coherent risk measure appears therefore as given by a “worst case
method”, in a framework of generalized scenarios. At this point it should be emphasized that scenarios be announced to all traders within the firm (by the manager)
or to all firms (by the regulator). In the first case, we notice that decentralization
of risk management within the firm is only available after these announcements.
Yet, in quantile-based methods, even after the announcements of individual limits,
there remains a problem preventing decentralized risk management: two operators
ignorant of each other’s actions may well each comply with their individual quantile
limits and yet no automatic procedure provides for an interesting upper bound for
the measure of the joint risk due to their actions. As for the regulation case we
allow ourselves to interpret a sentence from [Stu]: “regulators like Value at Risk,
because they can regulate it” as pointing to the formidable task of building and
announcing a reasonable set of scenarios.
4.2 Construction of coherent risk measures by extension of certain risk
measurements.
We now formalize the attempts described in Section 3.2 to measure risks. Their
basis is to impose margin requirements on certain basic portfolios considered as
“standard risks”, to use combinations of those risks to “support” other risks and
then bound from above required capital, using the margins required for standard
risks.
Definition 4.1. Supports of a risk: given a set Y of functions on Ω, we consider
a family, indexed by Y, of nonnegative numbers µ = (µY )Y ∈Y , all of them but a
finite number being zero, and we say that the couple (µ, γ), where γ is a real number,
17
“supports” X, for X ∈ G, provided
X≥
X
µY Y + γ · r.
Y ∈Y
The set of all such (µ, γ) which support X will be denoted by SY (X).
The idea is now to use these “supports”, made of “standard risks”, to bound
above possible extensions of a function Ψ defined on a subset of G. A consistency
condition is required to avoid supports leading to infinitely negative values.
Condition 4.1. Given a set Y of functions on Ω, and a function Ψ: Y −→ IR,
we say that
PΨ fulfills Condition 4.1 if for each support (µ, γ) of 0, we have the
inequality Y ∈Y µY Ψ(Y ) − γ ≥ 0.
Proposition 4.2. Given a set Y of functions on Ω and a function Ψ: Y −→ IR,
the equality
X
ρΨ (X) =
µY Ψ(Y ) − γ
inf
(µ,γ)∈SY (X)
Y ∈Y
defines a coherent risk measure ρΨ , if and only if Ψ fulfills Condition 4.1. If so,
ρΨ is the largest coherent measure ρ such that ρ ≤ Ψ on Y.
Proof of Proposition 4.2. (1) The necessity of Condition 4.1 is obvious.
(2) Since (0, 0) is a support of the element X = 0 of G and since Condition 4.1
ensures that any support of 0 provides a nonnegative number we find that ρΨ (0) = 0.
Notice that if Condition 4.1 is violated, then we would get ρΨ (0) = −∞.
(3) Axiom S required from a coherent risk measure follows here from the relation
SY (X1 + X2 ) ⊃ SY (X1 ) + SY (X2 ), and Axiom PH is satisfied since, given λ > 0,
(µ, γ) supports X if and only if (λ · µ, λ · γ) supports λ · X. P
(4) For a support (µ, γ) of a risk X let us call the number Y ∈Y µY Ψ(Y ) − γ the
“cost” of the support. By noticing for each risk X and each real α, that the support
(µ, γ) for X+α·r provides the support (µ, γ−α) for X, at a cost lower by the amount
α than the cost of the support of X + α · r we find that ρΨ (X) = ρΨ (X + α · r) + α.
Axiom T is therefore satisfied by ρΨ .
(5) Since for X ≤ Z we have SY (Z) ⊃ SY (X), Axiom M is satisfied by ρΨ .
(6) For any coherent measure ρ P
with ρ ≤ Ψ on Y we must have, for any support
(µ, γ) of X, the inequality ρ(X) ≤ Y ∈Y µY Ψ(Y )−γ and therefore ρ(X) ≤ ρΨ (X).
Remark. As opposed to the case of scenarios based measures of risks, the fewer
initial standard risks are considered, the more conservative is the coherent risk
measure obtained. This is similar to what happens with the SEC rules since Section
3.2 showed us that too many scenarios, and dually, too few standard risks, were
considered.
Condition 4.1 allows one to consider the function ρΨ in particular on the set
Y, the set of prespecified risks. There, it is clearly bounded above by the original
function Ψ. An extra consistency condition will prove helpful to figure out whether
ρΨ is actually equal to Ψ on Y.
Condition 4.2. Given a set Y of functions on Ω and a function Ψ: Y −→ IR, we
say that Condition 4.2 is satisfied
P by Ψ if for each element Z ∈ Y and each support
(µ, γ) of Z we have Ψ(Z) ≤ Y ∈Y µY Ψ(Y ) − γ.
Remark. It is an easy exercise to prove that Condition 4.2 implies Condition 4.1.
18
Proposition 4.3. Given a set Y of functions on Ω and a function Ψ: Y −→ IR+
satisfying Condition 4.2, the coherent risk measure ρΨ is the largest possible extension of the function Ψ to a coherent risk measure.
Proof of Proposition 4.3. (1) Condition 4.2 just ensures that the value at Z ∈ Y of
the original function Ψ is bounded above by the sum obtained with any support of
Z, hence also by their infimum ρΨ (Z), which proves that ρΨ = Ψ on Y.
(2) Let ρ be any coherent risk measure, which is also an extension of Ψ. Since
ρ ≤ Ψ on Y, Proposition 4.3 ensures that ρ ≤ ρΨ .
Propositions 4.2 and 4.3 above applied to (Y, Ψ) = (G, ρ), provide a statement
similar to Proposition 4.1 about representation of coherent risk measures.
Proposition 4.4. A risk measure ρ is coherent if and only if it is of the form ρΨ
for some Ψ fulfilling Condition 4.1.
Remark. It can be shown that for a coherent risk measure ρ built as a ρΨ , the
following set of probabilities
PΨ = {P | for all X ∈ G : EP [−X/r] ≤ Ψ(X)}
is non-empty and verifies the property
ρ(X) = sup{EP [−X/r] | P ∈ PΨ }.
4.3 Relation between scenario probabilities and pricing measures.
The representation result in Proposition 4.1 allows us to approach the problem
of risk concentration for coherent risk measures.
If the position consisting of the short Arrow-Debreu security corresponding to
state of nature ω, has a non-positive measure of risk, that is bankruptcy in the state
ω is “allowed”, the market price of this security should also be non-positive. To
formalize this observation we suppose an arbitrage free market, and denote by Qr
the closed convex set of pricing probability measures on Ω, using the instrument
r as numeraire. Given the coherent risk measure ρB,r associated to r and to an
acceptance set B, simply denoted by ρr (see Proposition 2.2), it will be natural to
assume the following condition:
Condition 4.3. The closed convex set Pρr of probability measures defining the
coherent risk measure ρr has a non empty intersection with the closed convex set
Qr of probability pricing measures.
When Condition 4.3 is satisfied, there is some Q ∈ Qr such that for any future
net worth Y , EQ [−Y /r] ≤ ρr (Y ), hence if Y has a strictly negative price under
Q it cannot be accepted. We interpret this fact in the following manner: if a firm
can, by trading, add a position Y to its portfolio and receive cash at the same time,
without having any extra capital requirement, then there is a bound to the quantity
of Y which the firm can add this way without trigging a request for extra capital.
If Condition 4.3 is not satisfied, then there exists a future net worth Y such that
sup{EQ [Y /r] | Q ∈ Qr } < inf{ES [Y /r] | S ∈ Pρr }.
Hence for each pricing measure Q we have EQ [−Y /r] > ρr (Y ) and therefore the
future net worth Z = Y + ρr (Y ) · r satisifies both conditions ρr (Z) = 0 and
EQ [Z/r] < 0. We have therefore an acceptable position with strictly negative price,
a situation which may well lead to an undetected accumulation of risk.
19
5. Two applications of representations of coherent risk measures
5.1 A proposal: the “worst conditional expectation” measure of risk.
Casualty actuaries have been working for long computing pure premium for
policies with deductible, using the conditional average of claim size, given that the
claim exceeds the deductible, see [HK]. In the same manner, reinsurance treaties
have involved the conditional distribution of a claim for a policy (or of the total
claim for a portfolio of policies), given that it is above the ceding insurer’s retention
level. In order to tackle the question of “how bad is bad”, which is not addressed by
the value at risk measurement, some actuaries (see [Albr], [E]) have first identified
the deductible (or retention level) with the quantile used in the field of financial
risk measurement. We prove below that one of the suggested methods gets us close
to coherent risk measures.
Considering the “lower partial moment” or expectation of the “shortfall”, the
presentation in [Albr] would translate, with our paper’s notations, into measuring
a risk X by the number EP [min (0, −V aRα (X) − X)] .
The presentations in [BEK], [E], use instead the conditional expectation of the
shortfall given that it is positive. The quoted texts (see also [EKM], Definition
3.4.6 as well as the methods indicated there to estimate the whole conditional
distribution) present the terminology “mean excess function”. We suggest the term
tail conditional expectation since we do not consider the excess but the whole of
the variable X :
Definition 5.1. Tail conditional expectation (or “TailVaR”): given a base probability measure P on Ω, a total return r on a reference instrument and a level α, the
tail conditional expectation is the measure of risk defined by
T CEα (X) = −EP [X/r | X/r ≤ −V aRα (X)] .
Definition 5.2. Worst conditional expectation: given a base probability measure P
on Ω, a total return r on a reference instrument and a level α, the worst conditional
expectation is the coherent measure of risk defined by
W CEα (X) = − inf{EP [X/r | A] | P [A] > α}.
Remark. T CEα has been suggested as a possible ingredient of reinsurance treaties
(see [A]).
Proposition 5.1. We have the inequality T CEα ≤ W CEα .
Proof of Proposition 5.1. (1) Let us denote X/r by Y . If FY (qα+ (Y )) > α the set
+
A = {ω | Y (ω) ≤ qα(Y
) } is one used in the definition of W CEα , hence the claim is
true.
(2) If FY (qα+ (Y )) = α it follows from the definition of qα+ and the monotonicity
of FY that for each ε > 0 , FY (ε + qα+ (Y )) > α. Hence, setting Aε = {ω | Y (ω) ≤
ε + qα+ (Y )} we get
W CEα (X) ≥ −EP [Y | Aε ] = −
EP [Y · 1Aε ]
.
P [Aε ]
Since FY is right-continuous, limε→0 P [Aε ] = FY (qα+ (Y )) and Aε ↓ A0 so the right
hand side has the limit −EP [Y | A0 ] = T CEα (X).
20
The paper [Alba] makes numerical studies of portfolios built out of collection of
risky bonds. It looks for a coherent measure which dominates the Value at Risk
measurement and yet gets close to it on a specific bond portfolio.
We interpret and generalize this search as the problem of a firm constrained by
the supervisors along the lines of the quantile risk measurement. Nevertheless, the
firm wishes at the same time to operate on a coherent basis, at the lowest possible
cost. Proposition 5.4 will provide circumstances where the firm’s problem has a
clear-cut solution.
Proposition 5.2. For each risk X one has the equality
V aRα (X) = inf{ρ(X) | ρ coherent and ρ ≥ V aRα }
The proof will use the following
Lemma 5.1. If ρ is the coherent risk measure defined by a set P of probability
measures, then ρ ≥ V aRα if and only if for each B with P [B] > α and each ε > 0
there is a Q ∈ P with Q [B] > 1 − ε.
Proof of Lemma 5.1. (1) Necessity: take X = −r · 1B where P [B] > α. Clearly
V aRα (−r · 1B ) = 1 and hence ρ(−r · 1B ) ≥ 1. This implies that for each ε > 0
there exists Q ∈ P with Q [B] ≥ 1 − ε.
(2) Sufficiency: let −k = V aRα (X) , then P [X ≤ k · r ] ≥ α and for each δ > 0
we have P [X ≤ (k + δ) · r ] > α.
Let Q ∈ P be chosen such that Q [X ≤ (k + δ) · r ] ≥ 1−δ. We obtain EQ [−X/r] ≥
(−k − δ) · (1 − δ) − δ · kX/rk. Since δ > 0 was arbitrary we find that ρ(X) ≥ −k.
Proof of Proposition 5.2. (1) Given any risk X let again −k = V aRα (X). Then
P [X ≤ k · r ] ≥ α and for each δ > 0 , P [X ≤ (k + δ) · r ] > α. We will construct a
coherent risk measure ρ such that ρ ≥ V aRα and ρ(X) ≤ V aRα (X) + δ.
(2) For any set B with P [B] > α , we must have P [B ∩ {X ≥ k · r}] > 0 and
we can define hB as 1B∩{X≥k·r} /P [B ∩ {X ≥ k · r}] and QB = hB · P. Lemma 5.1
shows that the measure ρ built with all the QB dominates V aRα , but for X we
obtain ρ(X) = supQB EQB [−X/r] ≤ −k = V aRα (X).
Definition 3.1 and Proposition 3.1 allow one to address a question by Ch. Petitmengin, Société Générale, about the coherence of the T CEα measure.
Proposition 5.3. Assume that the base probability P on Ω is uniform. If X is a
risk such that no two values of the discounted risk Y = X/r in different states are
ever equal, then T CEα (X) = W CEα (X).
Proof of Proposition 5.3. (1) Given α ∈]0, 1[ let us denote −V aRα (X) by q , the
set {X ≤ q · r} by B and the various values of Y = X/r by y1 < y2 < · · · < yn .
(2) Let k be the integer with 0 ≤ k < n such that α ∈ [ nk , k+1
n ). We will prove
that −V aRα (X) = qα+ (Y ) = q = yk+1 .
(3) For each u > q we have
#{i | yi ≤ u}
>α
n
hence the integer #{i | yi ≤ u} being strictly greater than α · n is at least k + 1.
21
(4) By taking u = yk+1 we actually minimize the integer #{i | yi ≤ u} and
therefore prove the point stated in (2).
(5) The set Y (B) is the set {y1 , . . . , yk+1 } and
T CEα (X) = −E [X/r | X ≤ q · r] = −
y1 + · · · + yk+1
.
k+1
(6) Any set C containing at least k + 1 states of nature and different from B will
provide values for −Y averaging to strictly less than T CEα (X), which therefore
equals W CEα (X).
Proposition 5.4. Assume that the base probability P on Ω is uniform. If a coherent risk measure ρ only depends on the distribution of the discounted risk and is
greater than the risk measure V aRα , then it is greater than the W CEα (coherent)
risk measure.
Proof of Proposition 5.4. (1) Given a risk X, we denote −V aRα (X) simply by q
and X/r by Y . The set A = {ω | Y (ω) ≤ q} has cardinality p > n · α and A is
written after possible renumbering as A = {ω1 , ω2 , . . . , ωp } with Y (ωi ) ≤ Y (ωi+1 )
for 1 ≤ i ≤ p − 1.
(2) Define Ȳ (ωi ) for i ≤ p as y ∗ = (Y (ω1 ) + · · · + Y (ωp ))/p = E [Y | Y ≤ q] and
as Y (ωi ) otherwise.
(3) For a permutation σ of the first p integers, we define Y σ by Y σ (ωi ) = Y (ωσ(i) )
for 1 ≤ i ≤ p , and Y σ (ωj ) = Y (ωj ) for p + 1 ≤ j ≤ n. We then find that Ȳ is also
the average of the p! random variables Y σ .
(4) The assumption that for each risk Z , ρ(Z) only depends on the distribution
of Z/r implies that all the ρ(r·Y σ ) are equal to ρ(X). The convexity of the function
ρ then implies the inequality ρ(X) ≥ ρ(r · Ȳ ).
(5) The last assumption made on ρ implies that ρ(r · Ȳ ) ≥ V aRα (r · Ȳ ).
(6) We have V aRα (r · Ȳ ) = −y ∗ = E [−Y | Y ≤ q] since for i ≤ p , Ȳ (ωi ) ≤
Y (ωp ). Hence ρ(X) ≥ E [−X/r | X ≤ q · r].
(7) For a dense set of random variables X on the finite state space Ω we have,
by Proposition 5.3, the equality E [−X/r | X ≤ q · r] = W CEα (X) hence the inequality ρ(X) ≥ W CEα (X) holds for a dense set of elements X of G.
(8) Both risk measures ρ and W CEα are coherent, hence continuous functions
on G. The inequality ρ ≥ W CEα is therefore true on the whole of G.
5.2 Construction of a measure out of measures on separate classes of
risks.
It is important to realize that Proposition 4.3 can be applied to a set Y of risks
having no structure. It can be the union of a family (Yj )j∈J of sets of risks, where,
for each j a function (Ψj ) is given on Yj , in such a way that Ψj = Ψj ′ on Yj ∩ Yj ′ .
The function Ψ is then defined by its restrictions to each of the Yj .
The different sets Yj may be exchange based risks on the one hand and over the
counter risks on the other hand, or market risks and credit risks in a framework
where a joint internal model would be looked for. Similarly, multi line aggregated
combined risk optimisation tools (see [Sh], 1998) would call for combined measure
of risks. The functions Ψj may come from preliminary rules given by exchanges
and/or by regulators (see [B], 1996). Assuming that Condition 4.2 is being satisfied,
which will depend on inequalities satisfied by the Ψj , Proposition 4.3 allows one to
mechanically compute a coherent risk measure extending the family of the Ψj and
22
dominating any other possible coherent risk measure chosen by exchanges and/or
by regulators to extend the family of the Ψj . It therefore provides a conservative
coherent tool for risk management.
In the special case of Ω = Ω1 × Ω2 with given coherent risk measures ρi , i = 1, 2,
on Gi , we define Yi as the set of all functions on Ω which are of the form fi ◦ pri
where fi is any function on Ωi , and where pri is the projection of Ω on its i-th
factor. We also define Ψi on Yi by the equality Ψi (fi ◦ pri ) = ρi (fi ). Since Y1 ∩ Y2
consists of the constants, the functions Ψ1 and Ψ2 are equal on it and they define
a function Ψ on Y = Y1 ∪ Y2 which satisfies Condition 4.2.
Let Pi be the set of scenarios defining ρi and let P be the set of probabilities on
Ω with marginals in P1 and P2 respectively. We claim that the risk measure ρΨ
on the set G of functions on Ω, that is the largest coherent risk measure extending
both Ψ1 and Ψ2 , is equal to the risk measure ρP , generated, as in Definition 3.1,
by the scenarios in P.
Proposition 5.5. The two coherent risk measures ρP and ρΨ are equal.
Proof of Proposition 5.5. The restriction of ρP to Yi equals Ψi since for each function fi on Ωi we have
ρP (fi ◦ pri ) = sup{EP [−fi ◦ pri /r] | P ◦ pr1−1 ∈ P1 , P ◦ pr2−1 ∈ P2 }
= sup{EP◦pr−1 [−fi /r] | P ◦ pri−1 ∈ Pi }
i
= ρi (fi ) = Ψi (fi ◦ pri ),
which proves that ρP ≤ ρΨ .
To prove the reverse inequality we use point (3) in the proof of Proposition
4.1 and show that if a probability Q on Ω is such that for each function X on Ω,
EQ [−X/r] ≤ ρΨ (X), then Q has its marginals Q1 and Q2 in P1 and P2 respectively.
Choose indeed X = fi ◦pri . We find that EQ [−fi ◦pri /r] = EQi [−fi /r] which proves
that for each fi ∈ Gi one has EQi [−fi /r] ≤ ρΨ (fi ◦ pri ), and therefore Qi ∈ Pi .
References
[A]
[ADEH]
[Alba]
[Albr]
[B]
[BEK]
[Bü]
[Cen]
[CR]
Amsler, M.-H. (1991), Réassurance du risque de ruine, Mitteilungen der Schweiz. Vereinigung der Versicherungsmathematiker, Heft 1, 33-49.
Artzner, Ph., F. Delbaen, J.-M. Eber, and D. Heath (1997), Thinking Coherently, RISK
10, November, 68-71.
Albanese, C. (1997), Credit Exposure, Diversification Risk and Coherent VaR, Working
Paper, Department of Mathematics, University of Toronto, September.
Albrecht, P. (1993), Normal and lognormal shortfall risk, Proceedings 3rd AFIR International Colloquium, Roma, Vol. 2, pp. 417-430.
Basle Committee on Banking Supervision (1996), Amendment to the Capital Accord to
Incorporate Market Risks, Basle, January.
Bassi, F., P. Embrechts, and M. Kafetzaki, Risk management and quantile estimation, Practical Guide to Heavy Tails (Adler, R., Feldman, R. and Taqqu, M., eds.),
Birkhäuser, Boston (to appear).
Bühlmann, H. (1995), Der Satz von Hattendorf und Hattendorf ’s Originalarbeit, Vortrag auf der Jubiläumstagung, Göttingen: “Hundert Jahre Versicherungsmathematik
an Universitäten”.
1989 Centennial Celebration of the Actuarial Profession in North America, Proceedings, Society of Actuaries, 475 N.Martingale Road, Schaumburg, IL.
Cox, J., and M. Rubinstein (1985), Options Markets, Prentice-Hall, Englewood Cliffs,
NJ.
23
[D]
[DP]
[DPG]
[DPP]
[E]
[EKM]
[Fed]
[Hat]
[HK]
[Hu]
[Kan]
[Ma]
[Me]
[NASD]
[Risk]
[RM]
[RS]
[Sh]
[Sp]
[Stu]
[Wan]
[Wit]
Dowd, K. (1998), Beyond Value at Risk, The New Science of Risk Management, Wiley,
Chichester.
Duffie, D., and J. Pan, (1997), An Overview of Value at Risk, Journal of Derivatives 4,
7-49.
Derivatives Policy Group (1995), Framework for Voluntary Oversight, a framework
for voluntary oversight of the OTC derivatives activities of securities firm affiliates to
promote confidence and stability in financial markets, New York.
Daykin, C., T. Pentikainen, and T. Pesonen, (1994), Practical Risk Theory for Actuaries, Chapman & Hall.
Embrechts, P. (1995), A survival Kit to Quantile Estimation, UBS Quant Workshop,
Zürich.
Embrechts, P., C. Klüppelberg, and T. Mikosch, (1997), Modelling Extremal Events,
Springer, New York.
Board of Governors of the Federal Reserve System, Securities Credit Transactions Regulation T, Margin Credit extended by Brokers and Dealers, as amended effective November 25, 1994.
Hattendorff, K. (1868), Über die Berechnung der Reserven des Risico bei der Lebensversicherung, Rundschau der Versicherungen 18, 1-18.
Hogg, R., and S. Klugman (1984), Loss Distributions, Wiley, New York.
Huber, P. (1981), Robust Statistics, Wiley, New York.
Kanner, M. (1867), Bestimmung des mittleren Risico bei Lebensversicherungen, Deutsche Versicherungs-Zeitung.
MATIF (1993), Initial Margin Calculation Method, MATIF, SA, Paris.
Merton, R. (1973), Theory of Rational Option Pricing, Bell Journal of Economics and
Management Science 4, 141-183.
National Association of Securities Dealers (1996), Reprint of the Manual July 1996,
CCR, Chicago.
Risk Special Supplement (1996), Value at Risk, Risk Magazine, June.
RiskMetrics (1995), Technical Document, Morgan Guarantee Trust Company, Global
Research, New York.
Rudd, A., and M. Schroeder (1982), The Calculation of Minimum Margin, Management
Science 28, 1369-1379.
Shimpi, P. (1998), The Convergence of Insurance and Financial Markets, a Practitioner’s View, Presentation, Conference on the Interplay between Finance, Insurance
and Statistics, Cortona, June 8-12.
SPAN (1995), Standard Portfolio Analysis of Risk, Chicago Mercantile Exchange,
Chicago.
Stulz, R. (1996), Rethinking Risk Management, Guest Speaker Conference, French Finance Association Meeting, Geneva, June 24.
Wang, T. (1996), A Characterisation of Dynamic Risk Measures, Working Paper, Faculty of Commerce, U.B.C., First Version, September 17, Second Version, October 10.
Wittstein, T. (1867), Mathematische Statistik und deren Anwendung auf NationalÖkonomie and Versicherungswissenschaft, Hahn’sche Buchhandlung, Hannover.
Institut de Recherche Mathématique Avancée Université Louis Pasteur et CNRS,
et Laboratoire de Recherches en Gestion, F 67084 Strasbourg, France
E-mail address:
[email protected]
Departement für Mathematik, Eidgen össische Technische Hochschule,
ETH-Zentrum, CH 8092 Zürich, Schweiz
E-mail address:
[email protected]
Société G énérale, Direction des marchés de capitaux MARC/SGOP/R&D, Tour
Société G énérale, F 92987 Paris La D éfense, France
E-mail address:
[email protected]
Carnegie Mellon University, Department of Mathematical Sciences, Pittsburgh,
PA 15213-3890, USA
E-mail address:
[email protected]
24