Documente Academic
Documente Profesional
Documente Cultură
Incisive Media
32–34 Broadwick Street
London W1A 2HG
Tel: +44(0) 20 7316 9000
E-mail: books@incisivemedia.com
Sites: www.riskbooks.com
www.incisivemedia.com
Conditions of sale
All rights reserved. No part of this publication may be reproduced in any material form whether
by photocopying or storing in any medium by electronic means whether or not transiently or
incidentally to some other use for this publication without the prior written consent of the copy-
right owner except in accordance with the provisions of the Copyright, Designs and Patents Act
1988 or under the terms of a licence issued by the Copyright Licensing Agency Limited of Saffron
House, 6–10 Kirby Street, London EC1N 8TS, UK.
Warning: the doing of any unauthorised act in relation to this work may result in both civil and
criminal liability.
Every effort has been made to ensure the accuracy of the text at the time of publication, this
includes efforts to contact each author to ensure the accuracy of their details at publication is cor-
rect. However, no responsibility for loss occasioned to any person acting or refraining from acting
as a result of the material contained in this publication will be accepted by the copyright owner,
the editor, the authors or Incisive Media.
Many of the product names contained in this publication are registered trade marks, and Risk
Books has made every effort to print them with the capitalisation and punctuation used by the
trademark owner. For reasons of textual clarity, it is not our house style to use symbols such
as TM, ®, etc. However, the absence of such symbols should not be taken to indicate absence of
trademark protection; anyone wishing to use product names in the public domain should first
clear such use with the product owner.
While best efforts have been intended for the preparation of this book, neither the publisher, the
editor nor any of the potentially implicitly affiliated organisations accept responsibility for any
errors, mistakes and or omissions it may provide or for any losses howsoever arising from or in
reliance upon its information, meanings and interpretations by any parties.
5 Volatility Interpolation 77
Jesper Andreasen and Brian Huge
Danske Bank
6 Random Grids 91
Jesper Andreasen and Brian Huge
Danske Bank
vi
Index 351
vii
ix
xi
xii
modelling, and managed the credit quant team through the credit
crisis. Before joining Barclays in 1999, Christoph worked in theo-
retical and experimental particle physics and was a fellow at CERN
and DESY. He holds a PhD in physics from Hamburg University.
xiii
xiv
xv
authored the book Interest Rate Models: Theory and Practice and has
published extensively in books and international journals, including
13 Cutting Edge articles in Risk magazine. He holds a BSc in applied
mathematics from the University of Padua and a PhD in mathemat-
ical finance from the Erasmus University of Rotterdam.
xvi
Robin Stuart is the head of risk analytics for State Street Corporation
Global Markets, responsible for the modelling of market and coun-
terparty credit risk. Trained in mathematics and theoretical physics,
he held post-doctoral positions at a number of international institu-
tions including CERN and the Max Planck Institute. Robin was
previously a professor of physics at the University of Michigan,
xvii
xviii
xix
The origins of quantitative finance are lost in the mists of time and
are difficult to identify precisely. However, most scholars agree that
they can be traced back certainly as far as the celebrated treatise on
double entry bookkeeping Summa de Arithmetica, Geometria,
Proportioni et Proportionalita (Everything About Arithmetic, Geometry
and Proportion), which was published in 1494 by Luca Pacioli. He, in
turn, credits an even earlier manuscript Delia Mercatura et del
Mercante Perfetto (Of Trading and the Perfect Trader) by his half-
forgotten predecessor Benedetto Cotrugli. Two important treatises
on options trading are Confusion de Confusiones (Confusion of
Confusions), published by Joseph de la Vega in 1688, and Traité de la
Circulation et du Crédit (An Essay on Circulation of Currency and
Credit), published by Isaac de Pinto in 1771. The works by de la
Vega and de Pinto clearly show that trading in options is not a new
phenomenon (as is from time to time wrongly claimed by its detrac-
tors) and has been thriving in Europe at least since the 16th century,
if not earlier. For instance, the Antwerp Exchange, the London
Royal Exchange and the Amsterdam Bourse were opened in 1531,
1571 and 1611, respectively. The reasons for the existence of a
burgeoning trade in options are not difficult to fathom – such
trading is crucial for the smooth functioning of commerce.
Regardless of any disputes about the history of quantitative
finance, there is general consensus that the starting point of modern
quantitative finance was the PhD thesis by Louis Bachelier Théorie
de la Spéculation (The Theory of Speculation), which was published in
1900. In his thesis, Bachelier introduced the stochastic process now
known as Brownian motion, and used it to study the evolution of
stock prices and to develop a theory of option pricing. Bachelier’s
work was forgotten for several decades until it was rediscovered
and published in English by Paul Cootner in 1964.
Although several distinguished scholars contributed to the
progress of quantitative finance in the interim (the name of Paul
xxi
xxii
John Lintner (1965) and Fisher Black and Robert Litterman (1992).
Tobin, and Markowitz and Sharpe were awarded the Nobel Prize in
Economics in 1981 and 1990, respectively. MPT explains the advan-
tages of diversification and shows how to achieve it in the best
possible way, and makes numerous idealised assumptions that it
shares with the BSM framework. In essence, it describes the behav-
iour of rational investors operating in frictionless Gaussian markets
and aiming at maximisation of economic utility.
However, since the 2008 financial crash, practitioners and
academics alike have realised that in markets under duress frictions
become dominant. This means that some large parts of quantitative
finance, including option pricing theory and MPT, have to be rebuilt
in order to account for market frictions in earnest. Some of the fail-
ures of quantitative finance from the pre-crisis build-up and during
the crisis itself were used by ill-informed detractors to claim that
the mathematical modelling of financial markets is futile and there-
fore has no future. This timely book serves as a concise response to
these detractors; it shows very clearly that well thought through
modelling is not only useful but necessary in order to help financial
markets to operate smoothly and perform their social role properly.
The editor, Mauro Cesa, has selected some of the best papers
published in Risk magazine since the beginning of the crisis; he
should be congratulated on his knowledge and taste.
The book consists of three parts and covers several important
topics, including post-BSM derivative pricing, asset allocation and
risk management, and, most importantly, counterparty risk.
Broadly speaking, it addresses the following subjects: (i) choices of
appropriate stochastic processes for modelling primary assets
including cash, bonds (government and corporate), equities,
currencies and commodities; (ii) financial derivatives on primary
assets and their risk-neutral and real-world pricing in different
modelling frameworks; (iii) modern approaches to volatility objects
and model calibration; (iv) asset allocation and related issues in the
presence of market frictions; and (v) credit risk and credit, debt, and
funding value adjustment calculations with and without collateral.
The reader will benefit from the expertise of some of the sharpest
thinkers in the field. Although most of the post-crisis models are
still far from being in the same state of completeness as their pre-
crisis predecessors, after reading the book it becomes clear that in
xxiii
the future these new and more realistic and accurate models will
find wide applications and thus flourish and expand.
Alexander Lipton
Bank of America Merrill Lynch and Imperial College
February 2013
xxiv
Since its inception in 1987, Risk magazine has had the privilege to
publish a collection of articles widely considered to be milestones
of modern quantitative finance, such as Vasicek (2002) on distribu-
tions of portfolio losses and Lipton (2002) on the volatility smile of
exotic options, while Dupire’s (1994), which introduced local vola-
tility, is still considered one of the most influential articles on
derivatives pricing. However, the world of modern quantitative
finance is changing. Where pre-2007 quants dreamed up compli-
cated theorems and designed exotic payouts, the credit crisis has
caused the industry as a whole to question long-held truisms,
including the pricing of something as simple as a plain vanilla
interest rate swap. Quants have also had to refocus their attentions
on capital and funding as a wave of regulatory reform has dramati-
cally reshaped the derivatives industry. As a result of this rapid
change, and adaption to the post-crisis landscape, quants have
generated a new wave of research. The aim of this book is to provide
a comprehensive overview of this new research, the challenges
quants have had to confront during the crisis, and of course their
responses; it will also focus on instruments and methodologies that
emerged or showed resilience during the crisis.
The repercussions of the credit crisis that enveloped global
markets in 2007 were keenly felt and continue to have a widespread
effect in all asset classes, even obliterating some and contributing to
the birth of others. Prior to 2007, a significant portion of financial
research was dedicated to complex credit derivatives. However,
subsequent to the collapse in credit markets in 2007, much of that
research was singled out as blameworthy and a key contributor to
the deterioration of bank’s balance-sheet health and their plum-
meting stock prices. One of the instruments borne out of that
quantitative research, the collateralised debt obligation (CDO), an
instrument that constituted the most toxic component of banks’
xxv
xxvi
xxvii
xxviii
xxix
xxx
xxxi
xxxii
xxxiii
xxxiv
of CCR, CVA and bilateral CVA. It explains how the last of these –
an approach that allows the two parties to agree on a price – is a
function of loss given default (LGD), expected exposure and default
probability, and that it takes into account the joint default proba-
bility and the first-to-default entity. In the general setting of an
asymptotic single risk factor model, Pykhtin shows two applica-
tions for the proposed framework. In the first, a market risk
approach, which is indicated for use by sophisticated banks that
actively and dynamically hedge CCR, it allows banks to calculate
the VaR in the trading book comprising both market risk and CCR
simultaneously. The second approach treats CCR separately from
market risk and is more suitable for banks that do not actively
manage CCR. Finally, Pykhtin discusses the minimum capital
requirements under Basel II and Basel III, and argues that the CVA
capital charge for Basel III, as it is calculated independently from
market risk, could incentivise risk taking. He concludes by
proposing a solution to this issue.
In Chapter 18, “Partial Differential Equation Representations of
Derivatives with Bilateral Counterparty Risk and Funding Costs,”
Christoph Burgard and Mats Kjaer propose a unified framework in
which the creditworthiness of the dealer and its subsequent effects
on funding costs and bilateral counterparty risk are taken into
account. The model is derived as an extension of the Black–Scholes
partial differential equation (PDE) that includes a funding compo-
nent, which may differ for lending and borrowing. The model is
based on a controversial assumption – that there exists the possi-
bility for the bank to buy back its own bonds in order to hedge its
credit risk. Some say this operation cannot be executed, as it is tech-
nically not possible for a bank to have a long position in its own
debt (see, for example, Castagna, 2012). However, assuming DVA is
replicable, the model is presented in two settings. In the first, the
mark-to-market value of a derivative at default includes counter-
party credit risk, while in the second it does not. In the latter
situation, the authors obtain a linear PDE whose Feinman–Kac
representation (a formula that allows for solving certain types of
PDEs) makes it easily tractable. One example shows how large the
impact on CVA can be if funding is taken into account. The work is
considered one of the most influential on the subject.
Damiano Brigo and Massimo Morini, in “Close-out Convention
xxxv
xxxvi
Mauro Cesa
REFERENCES
Basel Committee on Banking Supervision (BCBS), 2010, "Basel III: A Global Regulatory
Framework for more Resilient Banks and Banking Systems," December.
Brace A., 2010, “Multiple Arbitrage Free Interest Rate Curves,” preprint, National
Australia Bank.
Brigo, D. and A. Capponi, 2008, "Bilateral Counterparty Risk with Stochastic Dynamical
Models" (available at SSRN or arXiv.org).
Castagna, A., 2012, “The Impossibility of DVA Replication, Risk, Nove,ber, pp 66–70.
Duffie, D. and M. Huang, 1996, “Swap Rates and Credit Quality,” Journal of Finance,
51(3), pp 921–49.
Giles, M. B. and P. Glasserman, 2006, “Smoking Adjoint: Fast Monte Carlo Greeks,” Risk,
January.
xxxvii
Mercurio, F., 2010, “Modelling Libor Market Models, Using Different Curves for
Projecting Rates and for Discounting,” Journal of Theoretical and Applied Finance, 13(1), pp
113–37.
Tuckman, B. and P. Porfirio, 2003, “Interest Rate Parity, Money Market Basis Swaps and
Cross-currency Basis Swaps,” Fixed Income Liquid Market Research, Lehman Brothers,
June.
Vasicek, O., 2002, “The Distribution of Loan Portfolio Value,” Risk, December.
xxxviii
Derivatives Pricing
dσ̂ KT s
ST = = T
d ln K F 6 T
For the sake of analytical tractability, we assume that (lTit)0 does not
depend on St. This restricts our analysis to pure stochastic volatility
models with no local volatility component.
sT is given by sT = MT3/(MT2)3/2 where MTi = 〈(xT − 〈xT〉)i 〉 and 〈X〉 denotes
E[X]. Let us denote by δξt the perturbation of the instantaneous
variance at time t at order one in w:
ξ tt = ξ 0t + δξ t
t
⌠ n
δξ t = ωξ 0t ⎮ ∑ (λitτ )0 dWτi
⌡0 i=1
For ω = 0, M3 = 0. At lowest order, M3 is thus of order one in ω. We
then need to compute M3 at order one and M2 at order zero in ω :
T 1 T t 1 T
xT − xT = ∫ 0
ξ 0t + δξ t dZt −
2
∫ 0
(ξ 0 + δξ t ) dt + ∫ 0 ξ 0t dt
2
T 1 T δξ t 1 T
= ∫ ξ 0 dZt + ∫ 0
t
2
dZt − ∫ δξ t dt
2 0
0
ξ 0t
T
M2 = ∫ 0
ξ 0t dt
⎡ 2 ⎛ ⎞⎤
3 δξ t
M3 = E ⎢
2 ⎢⎣ (∫ T
0 ) ⎝
T
ξ 0t dZt ⎜⎜− ∫ 0 δξ t dt + ∫
T
0
ξ 0t
dZt
⎟⎥
⎟⎥
⎠⎦
where ρiS is the correlation between Z and Wi. This expression can
be rewritten as:
T ⎛ t ⎡ dS0 ⎤⎞
M3 = 3 ∫ dt ⎜⎜ ∫ E ⎢ 0τ δξ t ⎥⎟⎟ (1.2)
0
⎝ 0 ⎣ Sτ ⎦⎠
Equation 1.2 shows that M3 is given at first order in the volatility of
volatility by the double integral of the spot/volatility covariance
function. The expression E[dS0t /S0t dxt]) quantifies how much a move
ST =
1 ∫ 0
dt ∫ 0 f (τ ,t ) dτ
(1.4)
3
2 T
( T
∫ 0 ξ 0t dt ) 2
For an illustration of the accuracy of formula 1.4, see Figure 1.1 for
the case of a two-factor lognormal model for forward variances.
1 E ⎡⎣dσ̂ FT d ln S⎤⎦
RT =
dσ̂ KT E ⎡⎣( d ln S) ⎤⎦
2
d ln K F
1 T
dσ̂ tT =
2σ̂ tT (T − t )
∫ t
dξ tu du
T
1 ⌠ ⎡ dSt0 ⎤
= ⎮ E ⎢ 0 δξ u ⎥ du
2σ̂ tT (T − t ) ⌡t ⎣ St ⎦
We now divide by 〈(d ln S)2〉 and evaluate expectations at t = 0,
making use of the definition of f in Equation 1.3 and the expression
of the ATMF skew 1.4 to get:
T T
RT =
∫ 0
ξ 0t dt T ∫ 0 f ( 0, u) du
(1.5)
T t
ξ 00T ∫ dt ∫ f (τ ,t ) dτ
0 0
S0 = lim
1 ∫ 0
dt ∫ f (τ ,t ) dτ
0
=
f ( 0, 0)
(1.6)
3 3
T→0 2 T
( T
) 4 (ξ 00 ) 2
2
∫ 0 ξ dt t
0
The short skew has a finite limit that directly measures the covari-
ance function at the origin. Let us now turn to R. The pre-factor in
Equation 1.5 tends to one and we get:
T
T ∫ du
R0 = lim T
0
t =2 (1.7)
T→0
∫ 0
dt ∫ dτ
0
ST =
∫ 0
(T − t) f (t) dt
, RT =
∫ 0
f (t ) dt
3 T
2 (ξ 0 ) T
2 2
∫ (1− ) f (t) dt
0
t
T
3
%
0
0 0.5 1.0 1.5 2.0
1 ≤ RT ≤ 2
Scaling of ST and RT
Let us investigate the scaling behaviour of ST and RT by assuming
that for large time separations, f decays algebraically with exponent
γ : f(u) ∝ u−γ .
x
As we take the limit x → ∞, the integral ∫ 0 f(u)du either scales like
T 1−γ if γ < 1 or tends to a constant if γ > 1. Working out the limiting
regimes for ST and RT, we get:
❑❑ (Type I) If γ > 1:
1
ST ∝ and lim RT = 1
T T→∞
1
ST ∝ and lim RT = 2 − γ
Tγ T→∞
The connection between the decay of the ATMF skew and the
long-maturity limit of the SSR can be summarised compactly by the
following formula: if the spot/volatility covariance function has
either algebraic or exponential decay, then for long maturities:
1
ST ∝ (1.8)
T 2−R*
with:
R* = lim RT
T→∞
f (τ ) = ω (ξ 0 ) 2 ∑ wi ρSi e −kiτ
n
ω k iT − (1− e −kiT
)
ST = ∑
2 i
wi ρSi 2
( kiT )
RT =
∑ wρ i i Si
1−e− kiT
kiT
(1.9)
(
kiT− 1−e− kiT )
∑wρi i Si ( kiT )
2
10
–2
–2 –1 0 1 2
–3
–4
–5
–6
2.0
1.8
1.6
1.4
1.2
1.0
0 2 4 6 8 10
11
12
We observe that:
Equity volatility markets thus seem to behave like type II. Note,
however, that the short-maturity limit of the SSR (Equation 1.7) is
two whether the model be of type I or type II. This value is very
different from the historical value of the SSR for one-month options
on the Eurostoxx 50. As Figure 1.3 shows, its value is always lower
than two, at times markedly.
It is then natural to ask whether this can be arbitraged: is it
possible to implement an option strategy whose P&L is 2 − R0?
1 d 2Q ⎛⎛ δS ⎞ ⎞
2
13
with the crucial condition that the breakeven levels σS, ν, ρ be strike-
independent – unlike the Black–Scholes implied volatility σ^K – and
such that the market smile is recovered.
The smile is characterised by three quantities: σ0, the skew σ0α (σ0)
and the curvature σ0β (σ0). α and β are functions of σ0 and σ^ (x) has no
explicit dependence on T − t.
The price of an option is then Q(S, K, σ0, α, β, T) = PBS(S, K, σ (x), T)
where PBS(S, σ^ , T) is the Black–Scholes formula. The P&L of a delta-
hedged, σ0-hedged vanilla option reads, at order δ t:
2 2
dQ 1 d 2Q ⎛ δ S ⎞ 1 d 2Q ⎛ δσ 0 ⎞
P&L= δt + S2 2 ⎜ ⎟ + σ 02 ⎜ ⎟
dt 2 dS ⎝ S ⎠ 2 dσ 02 ⎝ σ 0 ⎠
d 2Q δS δσ 0
+Sσ 0 (1.12)
dSdσ 0 S σ 0
Our parameterisation for the smile 1.11, together with the assump-
tion that S, σ0 are the only dynamical quantities in our model, is
consistent only if we are able to find breakeven levels σS, ν, ρ that
make this P&L vanish on average, irrespective of the strike consid-
ered. Since σ^ has no explicit time-dependence, the theta in our
model is the same as the Black–Scholes theta: dQ/dt = dP/dt. Our
consistency requirement can be stated as:
dQ 1 2 d 2 PBSK 2
− = S σ̂ K
dt 2 dS2
1 d 2Q K 2 1 2 d 2Q K 2 d 2Q K
= S2 2
σS + σ0 2
ν + Sσ 0 ρσ Sν (1.13)
2 dS 2 dσ 0 dSdσ 0
14
2
−x + σ 2T
d=
σ T
that make the P&L vanish for all x. Grouping terms by powers of x
we get:
1 N ʹ′ ( d) ⎡
P&L= S
2 σ 0 T ⎣
(−σ 02 + σ S2 ) + (−σ 02α − 3ασ S2 + 2 ρσ Sν ) x
( )
+ −σ 02 β2 + (6α 2 − 25 β ) σ S2 + ν 2 − 2 ( 2α − σ 0α ʹ′) ρσ Sν x 2 ⎤⎦δt
σS =σ0 (1.14)
ρν = 2ασ 0 (1.15)
15
❑❑ The first equation expresses that the breakeven volatility for the
spot is the ATM volatility.
❑❑ The second equation relates the ATM skew to the covariance of S
and σ 0. Using the fact that, in our parameterisation dσ /d ln K|S =
ασ0, it can be rewritten as:
1 δS δσ 0
δt S σ 0
=2
dσ
σ0
d ln K S
d 2Q ⎛ δS δσ 0 dσ ⎞
Sσ 0 ⎜ − 2σ 0 δt ⎟ (1.17)
dSdσ 0 ⎝ S σ 0 d ln K S ⎠
We are free to specify any functional form for α (σ0) and β (σ0)
provided conditions 1.14, 1.15 and 1.16 hold.
Consistency conditions
S, σ0 are allowed to move while functions α, β stay constant. If ρ is
assumed to be constant, then Equations 1.15 and 1.16 show that the
dependence of the ATM skew on σ 0 is related to the dependence of
ν on σ 0.
16
⎛ a b ⎞
σ ( x ) = σ 0 ⎜1+ x + 2 x 2 ⎟ (1.18)
⎝ σ 0 2σ 0 ⎠
where a, b are constant, and ν and ρ are given by:
ρν = 2a (1.19)
ν = 3b + 6a 2 (1.20)
ρµ = 2a µ = 3b + 10a 2
17
1 d 2Q ⎛⎛ δS ⎞ ⎞
2
This was derived in the limit of a short maturity and for strikes near
the money. How reliable are our approximations, for practical
trading purposes? Let us check how accurately Equation 1.13 holds:
how well do our three thetas add up to the Black–Scholes theta and
what are their relative magnitudes?
We consider a one-month maturity and a typical index short-
maturity smile, shown in Figure 1.4. The corresponding values of
σ0, a and b are σ0 = 20%, a = −10% and b = 0.4%.
The top of Figure 1.5 shows:
σ 02 2 d 2Q
S
2 dS2
❑❑ the volatility theta:
3b + 6a 2 2 d 2Q
σ0
2 dσ 02
σ K2 2 d 2 PBS
S
2 dS2
18
30
25
%
20
15
10
80 90 100 110 120
%
and the sum of the three thetas in the top graph. The x-axis in both
graphs is the option’s strike. The lower graph shows acceptable
agreement: the model is usable in practice.
19
Figure 1.5 The spot, volatility and cross theta (top); the sum of the three
thetas compared with the Black–Scholes theta (bottom)
15
Spot
Vol
10 Cross spot/vol
0
80 90 100 110 120
%
–5
–10
16
BS theta
14 Sum of three thetas
12
10
8
6
4
2
0
80 90 100 110 120
%
day, then started again. On top of the third piece in Equation 1.21,
our total P&L comprises: (a) a volatility gamma/theta P&L – the
second piece in expression 1.21; (b) a vega P&L as our position has
some small residual sensitivity to σ0; and (c) additional P&L created
by remarking a, b to market on the next day. Since dP/dσ0 and d2P/
dσ 20 are approximately symmetric around the money, P&Ls (a) and
(b) are expected to be small.
The results of our back-test are illustrated in Figure 1.7. The top
scatter plot shows the daily P&L without portions (b) and (c) as a
function of the P&L calculated using Equation 1.21 while the scatter
plot at the bottom shows the real total daily P&L as a function of the
cross-gamma/theta P&L (Equation 1.7).
The dispersion of points around a straight line in the top graph is
20
10 Realised skew
9 Market skew
8
7
6
5
4
3
2
1
0
May May May May
2002 2004 2006 2008
Conclusion
In time-homogeneous stochastic volatility models at order one in
the volatility of volatility, the SSR and the rate at which the ATMF
skew decays with maturity are structurally related through the
spot/volatility covariance function. Assuming time-homogeneity
and a flat VS volatility term structure implies that the SSR is
restricted to the interval [1, 2]. Inspection of the historical behaviour
of Eurostoxx 50 implied volatilities shows that while for longer
maturities the relationship between the SSR and the ATMF skew
holds approximately, for short maturities the SSR is at
21
0.3
0.2
0.1
0
–0.3 –0.2 –0.1 0 0.1 0.2 0.3
–0.1
–0.2
–0.3
0.3
0.2
0.1
0
–0.3 –0.2 –0.1 0 0.1 0.2 0.3
–0.1
–0.2
–0.3
1 This approach allows for an economical derivation of the ATMF skew at order one in w. We
would get the same result by perturbing the pricing equation at order one in w.
2 For a similar approach based on the spot/realised volatility covariance function, see
Ciliberti, Bouchaud and Potters (2008).
22
references
Ciliberti S., J.-P. Bouchaud and M. Potters, 2008, “Smile Dynamics – A Theory of the
Implied Leverage Effect” (available at http://arxiv.org/pdf/0809.3375v1).
23
25
26
Preliminaries
We start with the risk-free curve for lending, a curve that corre-
sponds to the safest available collateral (cash). We denote the
corresponding short rate at time t by rC(t); C here stands for “CSA”,
as we assume this is the agreed overnight rate paid on collateral
among dealers under CSA. It is convenient to parameterise term
curves in terms of discount factors; we denote corresponding risk-
free discount factors by PC(t, T), 0 ≤ t ≤ T < ∞. Standard
Heath–Jarrow–Morton theory applies, and we specify the following
dynamics for the yield curve:
T
dPC (t,T ) / PC (t,T ) = rC (t ) dt − σ C (t,T ) dWC (t ) (2.1)
27
∂V (t )
Δ (t) =
∂S
Let C(t) be the collateral (cash in the collateral account) held at time
t against the derivative. For flexibility, we allow this amount to be
different1 from V(t).
To replicate the derivative, at time t we hold D(t) units of stock
and g (t) cash. Then the value of the replication portfolio, which we
denote by Π(t), is equal to:
V ( t ) = Π ( t ) = Δ ( t ) S ( t ) + γ (t ) (2.2)
g (t ) dt = ⎡⎣rC (t ) C (t) + rF (t ) (V (t ) − C (t ))
−rR (t ) Δ (t) S (t ) + rD (t ) Δ (t ) S (t)] dt
28
29
We note that:
Et ( dV (t )) = (rF (t ) V (t ) − ( rF (t ) − rC (t )) C (t )) dt
= ( rF (t ) V (t ) − sF (t ) C (t )) dt (2.6)
and the rate of growth is equal to the bank’s unsecured funding rate
or, using credit risk language, adjusted for the possibility of the
bank default. We show later that the case C = V could be handled by
using a measure that corresponds to the risk-free bond PC(t, T) =
Et(e–∫ r (u)du) as a numéraire and, likewise, the case C = 0 could be
T
C
t
30
Forward contract
We now consider a forward contract on S(⋅), where at time t the
bank agrees to deliver the asset at time T, against a cash payment at
time T.
Without CSA
A no-CSA forward contract could be seen as a derivative with the
payout S(T) – FnoCSA(t, T) at time T, where FnoCSA(t, T) is the forward
price at t for delivery at T. As the forward contract is cost-free, we
have by 2.3 that:
⎛ − ∫ T rF (u) du ⎞
0 = Et ⎜ e t (S (T ) − FnoCSA (t,T ))⎟
⎝ ⎠
so we get:
⎛ − ∫ T rF (u) du ⎞
Et ⎜ e t S (T ) ⎟
⎝ ⎠
FnoCSA (t,T ) =
⎛ − ∫ tT rF (u) du ⎞
Et ⎜ e ⎟ (2.9)
⎝ ⎠
31
⎛ − ∫ T rF (u) du ⎞
PF (t,T ) Et ⎜ e t ⎟
⎝ ⎠
Note that this is essentially a credit-risky bond issued by the bank.
Then we can rewrite Equation 2.9 as:
FnoCSA (t,T ) = E tT (S (T ))
~
where the measure P T is defined by the numeraire PF(t, T) as:
−
t
∫ 0 rF (u) du ⎛ − ∫ T rF (u) du ⎞
e PF (t,T ) = Et ⎜ e 0 ⎟
⎝ ⎠
~
is a P-martingale. Finally we see that FnoCSA(t, T) is a P T-
martingale.
We note that the value of an asset under no CSA at time t with
payout V(T) is given, by Equation 2.8, to be:
⎛ − ∫ T rF (u) du ⎞
V (t ) = Et ⎜ e t V (T ) ⎟ = PF (t,T ) E tT (V (T ))
⎝ ⎠
With CSA
Now let us consider a forward contract covered by CSA, where we
assume that the collateral posted C is always equal to the value of
the contract V. Let the CSA forward price FCSA(t, T) be fixed at t, then
the value, from Equation 2.5, is given by:
⎛ − ∫ T rC (u) du ⎞
0 = V (t ) = Et ⎜ e t V (T ) ⎟
⎝ ⎠
⎛ − ∫ rC (u) du
T
⎞
= Et ⎜ e t (S (T ) − FCSA (t,T ))⎟
⎝ ⎠
so that:
⎛ − ∫ T rC (u) du ⎞
Et ⎜ e t S (T ) ⎟
⎝ ⎠
FCSA (t,T ) =
⎛ − ∫ tT rC (u) du ⎞
Et ⎜ e ⎟ (2.10)
⎝ ⎠
32
= Et ⎜ e S (T ) ⎟
PF (t,T ) ⎝ ⎠
⎛ M (T,T ) ⎞
= EtT ⎜ S (T ) ⎟ (2.11)
⎝ M (t,T ) ⎠
where:
PF (t,T ) − ∫ 0t sF (u) du
M (t,T ) e (2.12)
PC (t,T )
is a PT-martingale, as:
⎛ − ∫ T sF (u) du ⎞
M (t,T ) = EtT ⎜ e 0 ⎟
⎝ ⎠
We note that, trivially:
⎛ M (T,T ) ⎞
EtT ⎜ ⎟ = 1
⎝ M (t,T ) ⎠
33
so:
34
European-style options
Consider now a European-style call option on S(T) with strike K.
Depending on the presence or absence of CSA, we get two prices:
⎛ − ∫ T rF (u) du + ⎞
VnoCSA (t) = Et ⎜ e t (S (T ) − K ) ⎟
⎝ ⎠
⎛ − ∫ tT rC (u) du + ⎞
VCSA (t) = Et ⎜ e (S (T ) − K ) ⎟
⎝ ⎠
(where for the CSA case we assumed that the collateral posted, C, is
always equal to the option value, VCSA). By the same measure-
change arguments as in the previous section:
( )
VnoCSA (t ) = PF (t,T ) E tT (S (T ) − K )
+
(t ) = P (t,T ) E ((S (T ) − K ) )
T +
VCSA C t
~
The difference between measures P Tt and PTt not only manifests itself
in the mean of S(T) – as already established in the previous section
– but also shows up in other characteristics of the distribution of
S(⋅), such as its variance and higher moments. We explore these
effects in the next section.
(
VnoCSA (t ) = PF (t,T ) EtT α (t,T,S (T )) (S (T ) − K )
+
) (2.14)
35
α 1 (t,T ) =
EtT ( M(T ,T )
M (t,T ) S (T ) − FCSA (t,T ))
VartT (S (T ))
α 0 (t,T ) = 1− α 1FCSA (t,T )
–10
–20
–30
–40 Credit/rates correlation
Credit/equity correlation
–50
Mar 27, 2006
Dec 6, 2007
Mar 19, 2008
Jul 6, 2006
Oct 7, 2008
Dec 13, 2005
May 1, 2009
Aug 30, 2005
36
Table 2.1 Relative differences between non-CSA and CSA forward prices
with s S = 30%, sF = 1.50%, ℵF = 5.00%
so the PDF of S(T) under the no-CSA measure is obtained from the
density of S(T) under the CSA measure by multiplying it with a
linear function. It is not hard to see that the main impact of such a
transformation is on the slope of the volatility smile of S(⋅). We
demonstrate this impact numerically below.
dS (t ) /S (t ) = O ( dt ) + σ SdWS (t )
and funding spread that follows dynamics inspired by a simple
one-factor Gaussian model of interest rates:2
FCSA (t,T ) = Et (S (T ))
and:
37
where:
1− e −ℵF (T−t)
b (T − t ) =
ℵF
Recall that:
FnoCSA (0,T ) − FCSA (0,T )
⎛ M (T,T ) ⎞
= E ⎜ ( FCSA (T,T ) − FCSA (0,T )) ⎟
M
⎝ ( 0,T ) ⎠
so that:
( T
FnoCSA ( 0,T ) = FCSA ( 0,T ) exp − ∫ 0 σ Sσ F b (T − t ) ρ dt )
⎛ T − b (T ) ⎞
= FCSA ( 0,T ) exp ⎜−σ Sσ F ρ ⎟ (2.16)
⎝ ℵF ⎠
38
31
30
%
29
28
27
26
25
40 60 80 100 120 140 160
Strike
Note: T = 10 years, FCSA(0, T) = 100, σF = 1.50%, ℵF = 5.00%
39
Table 2.2 Absolute differences between non-CSA and CSA forward Libor
rates, using market-implied caplet volatilities and sF = 1.50%, ℵF = 5.00%
Conclusions
In this chapter, we have developed valuation formulas for derivative
contracts that incorporate the modern realities of funding and
collateral agreements that deviate significantly from the textbook
assumptions. We have shown that the pricing of non-collateralised
derivatives needs to be adjusted, as compared with the collateralised
version, with the adjustment essentially driven by the correlation
between market factors for a derivative and the funding spread.
Apart from rather obvious differences in discounting rates used for
CSA and non-CSA versions of the same derivative, we have exposed
the required changes to forward curves and, even, the volatility
information used for options. In a simple model with stochastic
funding spreads we demonstrated the typical sizes of these adjust-
ments and found them significant.
40
1 In what follows we use Equation 2.3, 2.5 with either C = 0 or C = V. However, these formulas,
in their full generality, could be used to obtain, for example, the value of a derivative
covered by one-way (asymmetric) CSA agreement, or a more general case where the collat-
eral amount tracks the value only approximately.
2 While a diffusion process for the funding spread may be unrealistic, the impact of more
complicated dynamics on the convexity adjustment is likely to be muted.
REFERENCES
Antonov A. and M. Arneguy, 2009, “Analytical Formulas for Pricing CMS Products in
the Libor Market Model with the Stochastic Volatility,” SSRN eLibrary.
Barden P., 2009, “Equity Forward Prices in the Presence of Funding Spreads,” ICBI
Conference, Rome, April.
Gregory J., 2009, “Being Two-faced over Counterparty Credit Risk”, Risk, February, pp
86–90.
Hull, J., 2006, Options, Futures and Other Derivatives (6e) (Upper Saddle River, NJ:
Pearson/ Prentice Hall).
Karatzas I. and S. Shreve, 1996, Brownian Motion and Stochastic Calculus (2e) (New York,
NY: Springer).
41
The credit crunch that began in the second half of 2007 has trig-
gered, among many consequences, the explosion of the basis
spreads quoted on the market between single-currency interest rate
instruments (swaps in particular) characterised by different under-
lying rate tenors (Xibor three-month and Xibor six-month, etc,
where Xibor denotes a generic interbank offered rate). In Figure 3.1,
we show a snapshot of the market quotations as of February 16,
2009 for the six basis swap term structures corresponding to the
four Euribor tenors, one month, three months, six months and 12
months. Such very high basis spreads reflect the increased liquidity
risk suffered by financial institutions and the corresponding prefer-
ence for receiving payments with higher frequency (quarterly
instead of semi-annually, etc). Other indicators of changes in the
interest rate markets are the divergence between deposit (Xibor-
based) and overnight indexed swaps (OIS, Eonia based for euro)
rates, and between forward rate agreement (FRA) contracts and the
corresponding forward rates implied by consecutive deposits (see,
for example, Ametrano and Bianchetti, 2009, Mercurio, 2009, and
Morini, 2009).
These frictions reveal that apparently similar interest rate
instruments with different underlying rate tenors are character-
ised, in practice, by different liquidity and credit risk premiums,
reflecting the different views and interests of market players.
Thinking in terms of more fundamental variables, for example, a
short rate, the credit crunch has acted as a sort of “symmetry
breaking mechanism”: from an unstable situation in which a
43
unique short rate process was able to model and explain the whole
term structure of interest rates of all tenors, towards a sort of
“segmentation” into sub-areas corresponding to instruments with
different underlying rate tenors, characterised, in principle, by
distinct dynamics, for example, distinct short rate processes. We
stress that market segmentation was already present (and well
understood) before the credit crunch (see, for example, Tuckman
and Porfirio, 2003), but not effective due to negligible basis
spreads.
Such evolution of the financial markets has had strong effects
on the methodology used to price and hedge interest rate deriva-
tives. In principle, a consistent credit and liquidity theory would
be required to account for the interest rate market segmentation,1
but unfortunately such a framework is not easy to construct (see,
for example, Mercurio, 2009, and Morini, 2009). In practice, an
empirical approach has prevailed among market practitioners,
based on the construction of multiple “forwarding” yield curves
from plain vanilla market instruments homogeneous in the under-
lying rate tenor, used to calculate future cashflows based on
forward interest rates with the corresponding tenor, and of a
“discounting” yield curve, used to calculate discount factors and
cashflows’ present values. Such a “double-curve” approach allows
for an immediate recovery of market prices of quoted instruments
but, unfortunately, it does not fulfil the classic no-arbitrage
constraints of the single-curve pricing approach.
In this chapter, we acknowledge the current market practice,
assuming the existence of a given methodology for bootstrapping
multiple homogeneous forwarding and discounting yield curves,
and focus on the consequences for pricing and hedging interest rate
derivatives. This is a central problem in the interest rate market,
which still lacks attention in the published financial literature. In
particular, Boenkost and Schmidt (2005) discuss two methodologies
for pricing cross-currency basis swaps, the first of which (the actual
pre-crisis common market practice) does coincide, once reduced to
the single-currency case, with the double-curve procedure presently
adopted by the market2 (see also Tuckman and Porfirio, 2003, and
Fruchard, Zammouri and Willems, 1995). Kijima, Tanaka and
Wong (2008) have extended the approach of Boenkost and Schmidt
(2005) to the (cross-currency) case of three curves for discount rates,
44
Figure 3.1 Quotations as of February 16, 2009 for the six euro basis
swap spread curves corresponding to the four Euribor swap curves 1M,
3M, 6M, 12M
80
1M v. 3M
70 1M v. 6M
60 1M v. 12M
Basis spread (bp)
3M v. 6M
50 3M v. 12M
40 6M v. 12M
30
20
10
0
1Y 2Y 3Y 4Y 5Y 6Y 7Y 8Y 9Y 10Y11Y12Y15Y20Y25Y30Y
Source: Reuters
Libor rates and bond rates. Finally, simultaneous with the develop-
ment of this chapter, Morini (2009) has been approaching the
problem in terms of counterparty risk, Mercurio (2009) in terms of
the extended Libor market model and Henrard (2009) using an
axiomatic model.
Here, we follow an alternative route with respect to those cited
above, in the sense that: we adopt a “bottom-up” practitioner’s
perspective, starting from the current market practice of using
multiple yield curves and working out its natural consequences,
looking for a minimal and light generalisation of well-known
frameworks, keeping things as simple as possible; we show how
no-arbitrage can be recovered in the double-curve approach by
taking into account the basis adjustment, whose term structure can
be extracted from available market quotations; and we use a
straightforward foreign currency analogy to derive generalised
double-curve market-like pricing expressions for basic single-
currency interest rate derivatives, such as FRAs, swaps, caps/floors
and swaptions.3
45
by the same currency and by distinct bank accounts Bx, such that
t
Bx(t) = exp ∫0rx(u)du, where rx(t) is the associated short rate. We also
have multiple distinct yield curves Cx in the form of a continuous
term structure of discount factors, Cx = {T → Px(t0, T), T ≥ t0}, where t0
is the reference date (for example, the settlement date, or today) and
Px(t, T) denotes the price at time t ≥ t0 of the Mx-zero-coupon bond
for maturity T, such that Px(T, T) = 1. In each sub-market Mx we
assume the usual no-arbitrage relation:
where Px(t, T1, T2) denotes the Mx-forward discount factor from time
T2 to time T1. By expressing the latter in terms of the corresponding
simple compounded forward rate Fx(t; T1, T2), we obtain from
Equation 3.1 the familiar no-arbitrage expression:
Px (t,T1 ) − Px (t,T2 )
Fx (t;T1 ,T2 ) = (3.2)
τ x (T1 ,T2 ) Px (t,T2 )
where tx(T1, T2) is the year fraction between times T1 and T2 with day
count dcx. Equation 3.2 can be also derived as the fair value condi-
tion of the FRA contract4 with price at time t ≤ T1 < T2, for the
fixed-rate payer side, given by:
{ }
T
= N Px (t,T2 ) τ x (T1 ,T2 ) EQt x ⎡⎣Lx (T1 ,T2 )⎤⎦ − K
2
where N is the nominal amount, Lx(T1, T2) := Fx(T1; T1, T2) is the T1 spot
Xibor rate, K the strike rate (sharing the same compounding and
day count conventions), QxT denotes the Mx-T2- forward measure
2
46
as:
QTi
c (t,Ti , π i ) = E t d [ π i ] (3.5)
47
For instance, the 5.5-year floating swap leg cited above is currently
priced using Euribor one-month forward rates calculated on the C1M
forwarding curve (bootstrapped using Euribor one-month vanillas
only), plus discount factors calculated on the discounting curve Cd.
The delta sensitivity is calculated with respect to the market pillars
of both C1M and Cd curves, and the resulting delta risk is hedged
using the suggested amounts (hedge ratios) of five-year and six-
year Euribor one-month swaps plus the suggested amounts of
five-year and six-year instruments from the discounting curve.8
The static double-curve methodology described above can be
extended, in principle, by adopting multiple distinct models for
the evolution of each underlying interest rate with tenors f1, ... , fn
to calculate the dynamics of yield curves and expected cashflows.
The volatility/correlation dependencies carried by the models
would imply, in principle, the bootstrapping of multiple distinct
variance/covariance matrices and hedging the corresponding
sensitivities using volatility- and correlation-dependent plain
vanilla market instruments. A more general problem has been
approached in Mercurio (2009) in the context of the generalised
Libor market model. In this chapter, we will focus only on the
basic matter of static yield curves and leave out the dynamical
volatility/correlation dimensions. In the following two sections,
we will work out some consequences of the assumptions above in
terms of no-arbitrage.
1 Pf (t,T2 )
Pf (t,T1 ,T2 ) = = (3.8)
1+ Ff (t;T1 ,T2 ) τ f (T1 ,T2 ) Pf (t,T1 )
but clearly cannot hold at the same time. No-arbitrage is recovered
by taking into account the basis adjustment (to be distinguished
from the quoted market basis of Figure 3.1) defined as:
48
1
Pf (t,T1 ,T2 ) :=
1+ ⎡⎣Fd (t;T1 ,T2 ) + BA fd (t;T1 ,T2 )⎤⎦τ d (T1 ,T2 ) (3.9)
49
60
40
20
Basis points
–20
–40 1M v. Disc
3M v. Disc
–60 6M v. Disc
12M v. Disc
–80
Feb May Aug Nov Feb May Aug Nov Feb May Aug Nov Feb
09 09 09 09 10 10 10 10 11 11 11 11 12
4
%
2
Zero rates
Forward rates
1
Feb Feb Feb Feb Feb Feb Feb Feb
09 13 17 21 25 29 33 37
50
2
0
–2
–4 1M v. Disc
–6 3M v. Disc
–8 6M v. Disc
12M v. Disc
–10
Feb Feb Feb Feb Feb Feb Feb Feb Feb Feb
12 15 18 21 24 27 30 33 36 39
0
Basis points
–2
–4
–6 1M v. Disc
3M v. Disc
–8 6M v. Disc
12M v. Disc
–10
Feb Feb Feb Feb Feb Feb Feb Feb Feb Feb
12 15 18 21 24 27 30 33 36 39
Note: Graphs A and C: basis adjustment from Equation 3.10 (basis points) for daily
sampled 3M-tenor forward rates calculated on C1M, C3M, C6M and C12M curves
against Cd taken as reference curve. Graphs A: 0–3-year data; graph C: 3–30-year
data on magnified scales. Graphs B and D: the effect of poor interpolation
schemes (linear on zero rates, see Ametrano and Bianchetti, 2009) on
zero/forward 3M rates (graph B) and on basis adjustment (graph D)
{ }
T2
= N Pd (t,T2 ) τ f (T1 ,T2 ) E Qt x ⎡⎣L f (T1 ,T2 )⎤⎦ − K
51
Obviously the forward rate Ff(t; T1, T2) is not, in general, a martin-
gale under the discounting measure QdT , and the expression in the
2
52
dX fd (t,T2 )
= σ X (t) dWXT2 (t ) , t ≤ T2 (3.14)
X fd (t,T2 )
µ f (t ) = −σ f (t ) σ X (t ) ρ fX (t ) (3.17)
T
EQt d ⎡⎣L f (T1 ,T2 )⎤⎦ = Ff (t;T1 ,T2 ) + QA fd (t,T1 , σ f , σ X , ρ fX )
2
(3.18)
(3.20)
53
QA fd (t, T,S, ν f , ν Y , ρ fY )
where Sf(t, T, S) is the (fair) swap rate on curve Cf, QSd is the
discounting (domestic) swap measure associated with the annuity
Ad(t, S) on curve Cd, nf(t, T, S) is the swap rate volatility, nY(t, S) is the
volatility of the swap forward exchange rate defined as:
A f (t,S)
Yfd (t,S) = x fd (t ) (3.23)
Ad (t,S)
{ }
2 T
= N Pd (t,T2 ) τ f (T1 ,T2 ) EQt d ⎡⎣L f (T1 ,T2 )⎤⎦ − K
where we have used Equation 3.18 and the quanto adjustment term
is given by Equation 3.19.
For a (payer) floating versus fixed swap with payment date
vectors T, S as above, we have the price at time t ≤ T0:
Swap (t; T,S, K, N )
m
= −∑ N j Pd (t,Sj ) τ d (Sj−1 ,Sj ) K j
j=1
n
+∑ N i Pd (t,Ti ) τ f (Ti−1 ,Ti ) ⎡⎣Ff (t;Ti−1 ,Ti )
i=1
54
{ }
= N Pd (t,T2 ) τ f (T1 ,T2 ) Black ⎡⎣Ff (t;T1 ,T2 )
Finally, for swaptions on T0-spot swap rates Sf(T0, T, S), the standard
market-like pricing expression, using the discounting swap
measure QSd associated with the numeraire Ad(t, S) on curve Cd, is
modified as follows at time t ≤ T0:
Swaption (t; T, S, K, ω , N )
= NEQt d Max ⎡⎣ω (S f (T0 , T,S) − K )⎤⎦ Ad (t,S)
S
{ }
= N Ad (t,S) Black ⎡⎣S f (t, T,S)
+ QA fd (t, T,S, ν f , ν Y , ρ fY ) , K, λ f , ν f , ω ⎤⎦ (3.28)
where we have used Equation 3.20 and the quanto adjustment term
is given by Equation 3.21.
The calculations above also show that basic interest rate deriva-
tives prices become, in principle, volatility and correlation
dependent. The volatilities and the correlation in Equations 3.17
and 3.22 can be inferred from market data. In the euro market, the
volatilities sf and nf can be extracted from quoted caps/floors/
swaptions on Euribor six-month, while for sX, rfX and nY, rfY one
must resort to historical estimates.
In Figure 3.3, we show a numerical scenario for the quanto
adjustment in Equation 3.19. We see that, for realistic values
55
60
40
20
0
–20
–40
–60 σf = 10%, σX = 2.5%
σf = 20%, σX = 5%
–80 σf = 30%, σX = 10%
–100
–1.0 –0.8 –0.6 –0.4 –0.2 0 0.2 0.4 0.6 0.8 1.0
Correlation
Note: The time interval is fixed to T1 – t = 10 years and the forward rate entering
Equation 3.19 to 3%
56
1 ⎡ P (t,T ) R (t;t,T , R ) ⎤
1 f
Ff (t;T1 ,T2 ) = ⎢ d 1
− 1⎥ (3.30)
τ f (T1 ,T2 ) ⎢⎣ Pd (t,T2 ) R (t;t,T2 , R f ) ⎥⎦
where qd(T1, T2) = EtQ [1t (t)>T] is the counterparty survival probability
d
1 Pd (t,T1 ) ⎡ R (t;t,T1 , R f ) ⎤
BA fd (t;T1 ,T2 ) = ⎢ − 1⎥ (3.33)
τ d (T1 ,T2 ) Pd (t,T2 ) ⎢⎣ R (t;t,T2 , R f ) ⎥⎦
QA fd (t;T1 ,T2 )
1 Pd (t,T1 ) ⎡ 1 R (t;t,T1 , R f ) ⎤
= ⎢ − ⎥ (3.34)
τ f (T1 ,T2 ) Pd (t,T2 ) ⎢⎣ R (t;T1 ,T2 , R f ) R (t;t,T2 , R f ) ⎥⎦
Thus the basis and the quanto adjustment can be expressed, under
simple credit assumptions, in terms of risk-free zero-coupon bonds,
survival probability and recovery rate. A more complex credit
57
Conclusion
We have shown that after the credit crunch the classical single-curve
no-arbitrage relations are no longer valid and can be recovered by
taking into account the basis adjustment, whose term structure can
be extracted from available market quotations. Our numerical
results show that, once a smooth and robust bootstrapping tech-
nique for yield curve construction is used, the richer term structure
of the basis adjustment curves provides a sensitive indicator of the
tiny, but observable, static differences between different interest rate
market sub-areas in the post-credit crunch interest rate world.
Furthermore, the basis adjustment may also be helpful for a better
understanding of the profit and loss encountered when switching
between single- and double-curve worlds.
Using the foreign currency analogy, we have recalculated gener-
alised, double-curve no-arbitrage market-like pricing formulas for
basic interest rate derivatives, FRAs, swaps, caps/floors and swap-
tions in particular. When the forward exchange rate between the
two curves is stochastic and correlated with the forward rate, these
expressions include a single-currency version of the quanto adjust-
ment typical of cross-currency derivatives, naturally arising from
the change between the numeraires, or probability measures, asso-
ciated with the two yield curves. Numerical scenarios show that the
quanto adjustment can be important, depending on volatilities and
correlation. Unadjusted interest rate derivatives prices are thus, in
principle, not arbitrage-free, but, in practice, at the moment the
market does not trade enough instruments to set up arbitrage
positions.
Both the basis adjustment and the quanto adjustment find a
natural financial explanation in terms of counterparty risk within a
simple credit model including a default-free and a defaultable zero-
coupon bond.
Besides the current lack of information about volatility and corre-
lation, the present framework has the advantage of introducing a
minimal set of parameters with a transparent financial interpreta-
58
1 This would also explain why the frictions cited above do not necessarily lead to arbitrage
opportunities, once counterparty and liquidity risks are taken into account.
2 These authors were concerned with the fact that their first methodology was not consistent
with the pre-crisis single-curve market practice for pricing single-currency swaps.
Actually, it has become consistent with the post-crisis multiple curve practice.
3 Some details have been omitted here for brevity (see Ametrano and Bianchetti, 2009, and
Bianchetti, 2009).
4 See, for example, Brigo and Mercurio (2006), section 1.4. Note that here we are using the
“textbook” representation of the FRA contract, which is slightly different from the market
term sheet (see also Morini, 2009).
5 We refer here to the case of local yield curve bootstrapping methods, for which there is no
sensitivity delocalisation effect (see Hagan and West, 2006 and 2008).
6 This is a description of what really happens inside an investment bank after August 2007.
Even though it is rather familiar to many practitioners in the financial world, we summa-
rise it here in order to keep in touch with a larger audience, and to remark on the changes
induced by the credit crunch.
7 We use the T-forward measure here because it emphasises that the numeraire is associated
with the discounting curve; obviously any other equivalent measure would be fine as well.
8 The construction of the “right” discounting curve at first step in the post-credit crunch
world is a debated question that we do not consider here. See, for example, Piterbarg (2010).
9 This particular discounting curve, Euribor-based, is not considered risk-free in the post-
credit crunch world. Anyway, different choices (for example, an Eonia curve) as well as other
technicalities of the bootstrapping would obviously lead to slightly different numerical
results, but do not alter the conclusions drawn here.
10 Notice the fortunate notation, where d stands either for “discounting” or “domestic” and f
for “forwarding” or “foreign”, respectively.
11 In particular, we will adapt to the present context the discussion found in Brigo and
Mercurio, Chapters 2.9 and 14.4.
12 Frequently the quanto adjustment for cross-currency derivatives is defined as a multipli-
cative factor. Here, we prefer an additive definition to be consistent with the additive basis
adjustment in Equation 3.10)
59
13 In particular, in contrast to Mercurio (2009), we use here the FRA definition of Equation
3.3, leading to Equation 3.31.
REFERENCES
Bianchetti M., 2009, “Two Curves, One Price: Pricing and Hedging Interest Rate
Derivatives Decoupling Forwarding and Discounting Yield Curves,” working paper
(available at http://ssrn. com/abstract=1334356).
Boenkost W. and W. Schmidt, 2005, “Cross Currency Swap Valuation,” working paper,
HfB Business School of Finance & Management.
Brigo D. and F. Mercurio, 2006, Interest Rate Models – Theory and Practice (2e) (Berlin,
Germany: Springer).
Fruchard E., C. Zammouri and E. Willems, 1995, “Basis for Change,” Risk, October, pp
70–75.
Hagan P. and G. West, 2006, “Interpolation Methods for Curve Construction,” Applied
Mathematical Finance, 13(2), June, pp 89–129.
Hagan P. and G. West, 2008, “Methods for Constructing a Yield Curve,” Wilmott
Magazine, May, pp 70–81.
Henrard M., 2009, “The Irony in the Derivatives Discounting – Part II: The Crisis,”
working paper (available at http://ssrn.com/abstract=1433022).
Jamshidian F., 1989, “An Exact Bond Option Formula,” Journal of Finance, 44, pp 205–09.
Kijima M., K. Tanaka and T. Wong, 2008, “A Multi-quality Model of Interest Rates,”
Quantitative Finance, 9(2), pp 133–145.
Mercurio F., 2009, “Post Credit Crunch Interest Rates: Formulas and Market Models,”
working paper, Bloomberg (available at http://ssrn.com/abstract= 1332205).
Morini M., 2009, “Solving the Puzzle in the Interest Rate Market,” working paper
(available at http://ssrn.com/abstract=1506046).
Piterbarg V., 2010, “Funding Beyond Discounting: Collateral Agreements and Derivatives
Pricing,” Risk, February, pp 97–102.
Tuckman B. and P. Porfirio, 2003, “Interest Rate Parity, Money Market Basis Swaps, and
Cross-currency Basis Swaps,” Fixed Income Liquid Markets Research, Lehman Brothers.
60
61
62
PD(0, T) = POIS(0, T), where PD(t, T) denotes the discount factor (zero-
coupon bond) at time t for maturity T. The subscript D stands for
discount curve.
We assume that the tradable assets in our economy, at time t, are
the zero-coupon bonds PD(t, T) and the floating legs of (theoretical)
FRAs setting at times Tk−1 and paying at times Tk, t ≤ Tk−1 < Tk, for a
given time structure T0, ... , TM.
Consider times t, Tk−1 and Tk, t ≤ Tk−1 < Tk. The time-t FRA rate Lk(t)
is defined as the fixed rate to be exchanged at time Tk for the Libor
rate L(Tk−1, Tk) so that the swap has zero value at time t.
As in Kijima, Tanaka and Wong (2009), the pricing measures we
will consider are those associated with the discount curve.2
Denoting by QTD the T-forward measure with numeraire the zero-
coupon bond PD(t, T), we then assume the FRA rate Lk(t) to be
defined by:
where ETD denotes expectation under QTD and Ft denotes the infor-
mation available in the market at time t.
In the classic single-curve valuation, that is, when the forward
Libor curve for the tenor Tk − Tk−1 coincides with the discount curve,
it is well known that the FRA rate Lk(t) coincides with the forward
rate [PD(t, Tk−1)/PD(t, Tk)−1]/(Tk − Tk−1). In our multi-curve setting,
however, this no longer holds, since the simply compounded rates
defined by the discount curve are different, in general, from the
corresponding Libor fixings.
63
❑❑ They reduce to the “old” forward rates when the particular case
of a single-curve framework is assumed.
❑❑ They coincide with the corresponding Libor rates at their reset
times: Lk(Tk−1) = L(Tk−1, Tk).
❑❑ They are martingales, by definition, under the corresponding
forward measures.
❑❑ Their time-0 value Lk(0) can easily be bootstrapped from market
data of swap rates (see Chibane and Sheldon, 2009, Henrard,
2009, Fujii, Shimada and Takahashi, 2009a, and formula 4.11
below).
0.4
0.3
0.2
0.1
0
Oct 7, Feb 19, Jul 3, Nov 15,
2005 2007 2008 2009
Source: Bloomberg
64
1 ⎡ PD (t,Tk−1 ) ⎤
x
Fkx (t ) := FD (t;Tk−1
x
,Tkx ) = ⎢ − 1⎥ , k = 1,..., Mx (4.2)
τ kx ⎢⎣ PD (t,Tkx ) ⎥⎦
where t xk is the year fraction for the interval (T xk−1, T xk], and basis
spreads by:
well.
We define the joint evolution of rates F xk and spreads S xk under the
spot Libor measure QTD associated with times T, and whose numer-
aire is the discretely rebalanced bank account:
BDT (t) =
(
PD t,Tβx(t)−1 )
β (t )−1
∏ PD (T ,Tjx )
x
j−1
j=0
65
Caplet pricing
Let us consider the x-tenor caplet paying out at time T xk:
+ +
τ kx ⎡⎣Lxk (Tk−1
x
) − K ⎤⎦ = τ kx ⎡⎣Fkx (Tk−1x ) + Skx (Tk−1x ) − K ⎤⎦ (4.5)
{
= τ kx PD (t,Tkx ) EDTk ⎡⎣Fkx (Tk−1
x
) + Skx (Tk−1x ) − K ⎤⎦ Ft
+
} (4.6)
tional on Ft) of Sxk(T xk−1) and F xk(T xk−1), respectively, and/or the
associated caplet prices. Thanks to the independence of the random
variables F xk(T xk−1) and Sxk(T xk−1) we equivalently have:
66
{
= τ kx PD (t,Tkx ) ⎮ EDTk ⎡⎣Fkx (Tk−1
x
+
}
) − ( K − z)⎤⎦ Ft fSkx (Tk−1x ) ( z) dz
+∞
⌠
= τ kx PD (t,Tkx ) ⎮ EDTk
⌡−∞
x
{⎡⎣S (T
x
k
x
k−1 ) − ( K − z)⎤⎦ F } f
+
t z dz
( )( )
Fkx Tk−1
x (4.7)
of Sxk and its volatility are the same as those under QTD thanks to our
independence assumption. To this end, we apply the standard
change-of-numeraire result that relates the drifts of a (continuous)
process X under measures QTD and QDT :
x
k
Drift X ;QDTk
( )
x
= Drift ( X ;QDT ) +
(
d X, ln PD (⋅,Tkx ) / PD ⋅,Tβx(t)−1 ( )) t
(4.8)
dt
pricing formula (on F xk). This, along with the assumed tractability of
S xk, will finally allow the calculation of the caplet price by applica-
tion of Equation 4.7 (see also our explicit example below).
Swaption pricing
Let us consider a (payer) swaption, which gives the right to enter at
time T xa = T Sc an interest rate swap with payment times for the
floating and fixed legs given by T xa+1, ... , T xb and T Sc+1, ... , T Sd, respec-
tively, with T xb = T Sd and where the fixed rate is K. We assume that
each TSj belongs to {T xa, ... , T xb}.
67
and write:
b b
Sa,b, c , d (t) = ∑ω k (t ) Lxk (t ) = ∑ω k (t) Fkx (t)
k=a+1 k=a+1
b
68
{ }
+
= ∑τ P (t,TjS ) EDc, d ⎡⎣F (Tax ) + S (Tax ) − K ⎤⎦ Ft
S
j D (4.15)
j=c+1
T x ⊂ ··· ⊂ T x . We set T := T x .
n–1 1 1
The joint evolution of forward OIS rates for all given tenors x can
be defined by modelling the rates with smaller tenor x1. In fact, the
dynamics of rates F xk for tenors x ∈ {x2, ... , xn} can be obtained from
the dynamics of rates F kx by noting that we can write:
1
ik
69
F] and ψFk(V) = V in the general Equation 4.4. The reason for this
modelling choice will be made clear below.7 The stochastic vola-
tility VF is assumed to follow the dynamics in Equation 4.4.
As per spread dynamics, we assume for each tenor x ∈ {x1, ... , xn}
the following one-factor models:
Skx (t ) = Skx ( 0) Mx (t ) , k = 1,..., Mx (4.18)
Mx(0) = 1. The spreads Sxk are thus positive martingales under QTD
and any forward or swap measures. A convenient choice in terms of
model tractability is to assume that Mx are stochastic processes
whose densities or associated option prices are known in closed
form. This will be the case in our explicit example below.
ik
dV F (t) = −V F (t) b F (t,V F (t)) ∑σ x1
h (t ) ρ hx dt
1
h=β (t )
70
From Equation 4.19, we notice that 4.17 are the simplest stochastic
volatility dynamics that are consistent across different tenors. This
means, for example, that if three-month rates follow shifted-
lognormal processes with common stochastic volatility, the same
type of dynamics (modulo the drift correction in the volatility
process) is also followed by six-month rates under the respective
forward measures. Our choice of dynamics in 4.17 is motivated by
this feature, which allows us to price simultaneously in closed form
(with the same type of formula) caps and swaptions with different
underlying tenors.
Caplet prices can then be calculated along the lines suggested
above. Likewise, analytical approximations for swaption prices can
be obtained by applying the procedure described above (“Swaption
pricing”) and by noticing that, thanks to assumption 4.18, formula
4.15 can be simplified as follows:
b
S (t ) = ∑ω k (t ) Skx ( 0) Mx (t )
k=a+1
b
≈ Mx ( t ) ∑ω k (0) Skx (0) = S (0) Mx (t)
k=a+1
Hagan et al (2002) for VF. This leads to the following dynamics for
the x-tenor rate F xk under QDT :
x
k
⎡ 1 ⎤
dFkx (t ) = σ kxV F (t ) ⎢ x + Fkx (t )⎥ dZkk , x (t )
⎣ τ k ⎦
ik
2
dV F (t ) = −ε ⎡⎣V F (t )⎤⎦ ∑σ x1
h ρ hx1 dt + εV F (t ) dW k , x (t )
h=β (t )
F
V ( 0) = 1 (4.21)
71
dM (t ) = σ M (t ) dZ (t ) (4.22)
s is a positive constant.
Caplet prices under this specification can easily be calculated in
closed form as soon as we smartly approximate the drift term of VF.
Some possible choices can be found in Mercurio and Morini (2009).
Applying the first of formulas 4.7, we get:
Cplt (t, K;Tk−1
x
,Tkx )
ax t
⌠ k( ) ⎛ 1 1
= ⎮ Cplt SABR ⎜t, Fkx (t ) + x , K + x
⌡−∞ ⎝ τk τk
1 − 21 z2
− Skx (t ) e
− 21 σ 2Tk−1
x x
+σ Tk−1 z x
;Tk−1 ,Tkx ) 2π
e dz
where:
1
K+ τ x
x
ln Sx (tk) + 21 σ 2 (Tk−1
x
− t)
a (t) :=
k
k
x
σ Tk−1 −t
can be used to price z-based caps again with formula 4.23, this time
setting x = z. This will be done in the following example.
72
0.015
0.010
0.005
0
−0.005
−0.010
−0.015 4y
5y
6y
−0.020 7y
2% 8y
3% 4% 9y
5% 10y
6% 7%
10
8
6
10–3
4
2
0
4y
−2 5y
6y
2% 7y
3% 8y
4% 9y
5% 10y
6%
7%
73
Conclusion
We have shown how to extend the LMM to price interest rate deriv-
atives under distinct yield curves, which are used for generating
future Libor rates and for discounting. We have first modelled the
joint evolution of forward OIS rates and related Libor-OIS spreads
for a given tenor, and then proposed a class of models for the multi-
tenor case. Under assumptions that are standard in the classic LMM
literature, the general dynamics we have considered imply the
possibility of pricing in closed form both caps and swaptions, with
procedures that are only slightly more involved than the corre-
sponding ones in the single-curve case.
Modelling different tenors at the same time has the advantage
of allowing for the valuation of derivatives that are based on
multiple tenors, for example, basis swaps. Another interesting
application involves the pricing of caps or swaptions with a non-
standard underlying tenor, given the market quotes of
standard-tenor options. In both cases, additional constraints on
the model dynamics should be imposed so as to ensure that basis
spreads keep realistic relations between one another as they move
over time.
An issue that needs further investigation is the modelling of
correlations with parametric forms granting the positive definite-
ness of the global correlation matrix. To this end, one may try to
extend to the multi-curve case the parameterisation proposed by
Mercurio and Morini (2007) in the single-curve setting.
74
The author would like to thank Peter Carr, Liuren Wu, Antonio
Castagna, Raffaele Giura and Massimo Morini for stimulating discus-
sions, and Nabyl Belgrade, Marco Bianchetti, Marcelo Piza, Riccardo
Rebonato and two anonymous referees for helpful comments.
1 A similar approach has recently been proposed by Fujii, Shimada and Takahashi (2009b),
who model stochastic basis spreads in a Heath–Jarrow–Morton (HJM) framework both in
single- and multi-currency cases, but without providing examples of dynamics. An alter-
native route is chosen by Henrard (2009), who hints at the modelling of basis swap spreads,
but without addressing typical issues such as the modelling of joint dynamics or the
pricing of plain vanilla derivatives.
2 This is also consistent with the results of Fujii, Shimada and Takahashi (2009a, 2009b) and
Piterbarg (2010), since we assume credit support annex agreements where the collateral
rate to be paid equals the (assumed risk-free) overnight rate
3 The reason for modelling OIS rates in addition to FRA rates is twofold. First, by assump-
tion, our pricing measures are related to the discount (ie, OIS) curve. Second, swap rates
explicitly depend on zero-coupon bonds PD(t, T)
4 For instance, if the tenor is three months, the times T xk must be three-month spaced.
x
5 These functions must be chosen so that Fxk is a martingale under QDT k (see Equation 4.9)
6 We acknowledge that this assumption may lack economic foundation. However, the
historical correlation between OIS rates and spreads in the post credit-crunch period has
been rather unstable. In fact, both positive and negative values have been recorded. The
zero-correlation assumption may thus be regarded as reflecting an average (long-term)
behaviour.
7 Notice also that simply compounded forward rates in a Gaussian short rate model follow
stochastic differential equations analogous to Equation 4.17 with VF ≡ 1.
REFERENCES
Bianchetti M., 2010, “Two Curves, One Price,” Risk, August, pp 66–72.
Brace A., 2010, “Multiple Arbitrage Free Interest Rate Curves”, preprint, National
Australia Bank.
Fujii M., Y. Shimada and A. Takahashi, 2009b, “A Market Model of Interest Rates with
Dynamic Basis Spreads in the Presence of Collateral and Multiple Currencies,” working
paper, University of Tokyo and Shinsei Bank (available at www.e.u-tokyo.ac.jp/cirje/
research/dp/2009/2009cf698.pdf).
Hagan P., D. Kumar, A. Lesniewski and D. Woodward, 2002, “Managing Smile Risk,”
Wilmott Magazine, September, pp 84–108.
75
Henrard M., 2009, “The Irony in the Derivatives Discounting Part II: The Crisis,”
preprint, Dexia Bank, Brussels.
Kijima M., K. Tanaka and T. Wong, 2009, “A Multi-quality Model of Interest Rates,”
Quantitative Finance, 9(2), pp 133–45.
Mercurio F., 2009, “Interest Rates and the Credit Crunch: New Formulas and Market
Models” (available at http://papers.ssrn.com/s013/ papers.cfm?abstract_id=1332205).
Mercurio F., 2010, “Modern Libor Market Models: Using Different Curves for Projecting
Rates and for Discounting,” International Journal of Theoretical and Applied Finance, 13, pp
1–25.
Mercurio F. and M. Morini, 2009, “Joining the SABR and Libor Models Together,” Risk,
March, pp 80–85.
Piterbarg V., 2010, “Funding Beyond Discounting: Collateral Agreements and Derivatives
Pricing,” Risk, February, pp 97–102.
Wu L. and F. Zhang, 2006, “Libor Market Model with Stochastic Volatility,” Journal of
Industrial and Management Optimization, 2(2), pp 199–227.
76
77
Discrete expiries
Given a time grid of expiries 0 = t0 < t1 < ... and a set of volatility
functions {ϑ (k)}i=0,1,..., we construct European-style option prices for
all the discrete expiries, by recursively solving the forward system:
⎡ 1 2 ∂ ⎤
2
+
c ( 0, k ) = ( s ( 0) − k ) , i = 0,1,...
1
δkk f ( k ) =
Δk 2
( f ( k − Δk ) − 2 f ( k ) + f ( k + Δk )) (5.2)
78
For a set of discrete option quotes {c^ (ti, kij)}, the system 5.1 can be
bootstrapped forward, expiry by expiry, to find piecewise constant
functions:
that minimise the pricing error in 5.1. In other words, we solve the
optimisation problems:
2
ϑ i (⋅)
j
(( ) )
inf ∑ c (ti , k ij ) − ĉ (ti , kij ) / wij , wij = ∂ĉ (ti , k ij ) /∂σ̂ (ti , kij ) (5.5)
Note that for expiries that lie between the quoted expiries, the time
stepping is non-standard. Instead of multiple small time steps that
connect all the intermediate time points, we step directly from ti to
all times t ∈ ]ti, ti+1[. The time-stepping scheme is illustrated in Figure
5.1. This methodology is essentially what distinguishes our model-
ling approach from previously presented finite difference-based
algorithms, for example, Coleman, Li and Verma (1999) and
Avellaneda et al (1997).
79
In the appendix, we use this to show that the option prices gener-
ated by 5.1 and 5.6 are consistent with absence of arbitrage, that is,
that ct(t, k) ≥ 0, ckk(t, k) ≥ 0 for all (t, k).
For the discrete space case, we note that with the additional
(absorbing) boundary conditions ckk(t, k0) = ckk(t, kn) = 0, 5.2 can be
written as:
A−1 ≥ 0 (5.11)
This implies that the discrete system 5.2 is stable. As we also have:
80
VOLATILITY INTERPOLATION
120.95 18.67 18.11 18.39 19.90
124.61 18.71 17.85 17.93 19.45 20.54 21.03 21.64 22.51
131.94 19.88 20.54 21.05 21.90
139.27 19.30 20.02 20.54 21.35
146.60 18.49 19.64 20.12
Note: The table shows implied Black volatilities for European-style options on the SX5E index. Expiries range from two weeks to a little under six years
and strikes range from 50–146% of current spot of 2,772.70. Data is as of March 1, 2010
81
11/03/2013 10:11
05 Andreason/Huge PCQF.indd 82
82
Note: The table shows the difference between the model and the target in implied Black volatilities for European-style options on the SX5E index. Data
is as of March 1, 2010
11/03/2013 10:11
VOLATILITY INTERPOLATION
⎡1− 1 Δt ϑ ( x )2 (δ − δ )⎤ c (t , x ) = c (t , x ) ,
⎣ 2 i i xx x ⎦ i+1 i
+
c ( 0, x ) = ( s ( 0) − e x ) , i = 0,1,...
1
δx f ( x ) = ( f ( x + Δx) − f ( x − Δx))
2Δx
1
δxx f ( x) = 2 ( f ( x − Δx ) − 2 f ( x ) + f ( x + Δx )) (5.13)
Δx
⎡ 1 2 ∂ ⎤
2
⎢1− 2 (T (t ) − ti ) ϑ i ( k ) ⎥ c (t, k ) = c (ti , k ) , t ∈ ]ti ,ti+1 ] (5.14)
⎣ ∂k 2 ⎦
where T(ti) = ti and Tʹ(t) > 0. In this case, the local volatility function
5.3 consistent with the model is given by:
2 ct (t, k )
σ (t, k ) = 2
c kk (t, k )
2 ⎡ ∂ln c kk (t, k ) ⎤
= ϑ i ( k ) ⎢T ʹ′ (t ) + (T (t ) − ti ) ⎥ (5.15)
⎣ ∂t ⎦
The introduction of the time-change facilitates the interpolation in
the expiry direction. For example, a choice of piecewise cubic func-
tions T(t) can be used to ensure that implied volatility is roughly
linear in expiry.
Algorithm
In summary, a discrete set of European-style option quotes is inter-
polated into a full continuously parameterised surface of
arbitrage-consistent option quotes by:
83
100
80
60
%
40
20
0
115.8781
0.024641
219.3651
0.519887
415.2729
1.060515
1.617426
786.1397
2.174337
2.722793
1,488.216
3.278013
3.834924
2,817.293
4.391835
5,333.327
4.948746
10,096.35
5.505657
19,113.07
36,182.34
Spot/strike Time/maturity
Note: The graph shows the local volatility surface in the model after it has been
fitted to the SX5E market. Data is as of March 1, 2010
Note that step 2 does not involve any iteration. The process of the
time stepping is shown in Figure 5.1.
Numerical example
Here, we consider fitting the model to the Eurostoxx 50 (SX5E)
equity option market. The number of expiries is 12, with up to 15
strikes per expiry. The target data is given in Table 5.1. We choose to
fit a lognormal version of the model based on a finite difference
solution with 200 grid points. The local volatility function is set up
to be piecewise linear with as many levels as calibration strikes per
expiry. The model fits to the option prices in approximately 0.05
seconds of CPU time on a standard PC. The average number of
84
Conclusion
We have shown how a non-standard application of the fully implicit
finite difference method can be used for arbitrage-free interpolation
of implied volatility quotes. The method is quick and robust, and
can be used both as a pre-pricing step for local volatility models as
well as for market-making in option markets.
Proof of proposition 1
Consider option prices generated by the forward equation:
2
∂g 1 2 ∂ g
0=− + 2 ϑ (k ) (5.17)
∂t ∂k 2
which is solved forward in time t given the initial boundary condi-
tion g(0, k).
As also noted in Andreasen (1996), 5.17 can also be seen as the
backward equation for:
85
⎡ 1 2 ∂ ⎤
2
⎢1− 2 T (u)ϑ ( k ) ⎥ h ( u, k ) = g ( 0, k ) (5.25)
⎣ ∂k 2 ⎦
From 5.20 and 5.23, we conclude that 5.25 defines a positive linear
functional that preserves convexity.
Differentiating 5.25 with respect to u yields:
⎡ 1 2 ∂ ⎤
2
1 2
⎢1− T ( u) ϑ ( k ) ⎥ hu ( u, k ) = T ʹ′ (u) ϑ ( k ) hkk ( u, k ) (5.26)
⎣ 2 ∂k 2 ⎦ 2
hu ( u, k ) ≥ 0 (5.27)
86
Proposition 2
On the discrete non-uniform strike grid k0 < k1 < … < kn, the finite
difference scheme (5.2) produces option prices that are consistent
with absence arbitrage, ie, the generated option prices are increasing
in maturity, decreasing in strike and convex in strike:
c(t, ki ) ≥ c(t, k i+1 )
∂c(t, ki )
≥0
∂t
c(t, ki+1 ) − c(t, ki ) c(t, k i ) − c(t, ki−1 )
≥
k i+1 − ki ki − k i−1
for all i = 1, …, n – 1.
Proof of proposition 2
On the discrete non-uniform strike grid k0 < k1 < … < kn, the finite
difference scheme (5.2) can be written as the matrix equation system
where c(t) = (c(t, k 0 ),…, c(t, k n ))' is a vector of option prices, I is the
identity matrix, Θ is a diagonal matrix and D is proportional to the
discrete second order difference matrix. Specifically:
⎡ θ (k )2 ⎤
⎢ 0 ⎥
⎢ θ (k1 )2 ⎥
⎢ ⎥
⎢ θ (k 2 )2 ⎥
Θ = ⎢ ⎥
⎢ ⎥
⎢ θ (kn−1 )2 ⎥
⎢ ⎥
⎢⎣ θ (kn )2 ⎥⎦
⎡ 0 0 ⎤
⎢ ⎥
⎢ l1 −l1 − u1 u1 ⎥
1 ⎢ l2 −l2 − u2 u2 ⎥
D = t δkk = ⎢ ⎥
2 ⎢ ⎥
⎢ ln−1 −ln−1 − un−1 un−1 ⎥
⎢ ⎥
⎢⎣ 0 0 ⎥⎦
1 1
li = t ⋅
k i+1 − k i−1 ki − ki−1
1 1
ui = t ⋅ (5.29)
k i+1 − ki−1 ki+1 − ki
87
It follows that:
⎡ θ (k )−2 0 ⎤
⎢ 0 ⎥
⎢ −l1 θ (k1 )−2 + l1 + u1 −u1 ⎥
⎢ ⎥
−1 ⎢ −l2 θ (k2 )−2 + l2 + u2 −u2 ⎥
A ≡ [Θ − D] = ⎢ ⎥
⎢ ⎥
⎢ −ln−1 θ (kn−1 )−2 + ln−1 + un−1 −un−1 ⎥
⎢ ⎥
⎢⎣ 0 θ (kn )−2 ⎥⎦ (5.31)
∂c(t) 1
= [Θ−1 − D]−1 ( δkk c(t)) ≥ 0 (5.33)
∂t 2
Since c(t, k0) = (s – k0)+ and c(t, kn) = (s – kn)+, we can further conclude
that c(t) is also monotone decreasing in strike and satisfy:
88
REFERENCES
Carr P., 2008, “Local Variance Gamma,” working paper, Bloomberg, New York.
Coleman T., Y. Li and A. Verma, 1999, “Reconstructing the Unknown Local Volatility
Function,” Journal of Computational Finance, 2(3), pp 77–100.
Fiedler, M., 1986, Special Matrices and their Application in Numerical Mathematics
(Dordrecht, Holland: Martinus Nijhoff Publishers).
JP Morgan, 1999, “Pricing Exotics Under the Smile,” Risk, November, pp 72–75.
Nabben R., 1999, “On Decay Rates of Tridiagonal and Band Matrices,” SIAM Journal on
Matrix Analysis and Applications, 20, pp 820–37.
Sepp A., 2007, “Using SABR Model to Produce Smooth Local Volatility Surfaces,”
working paper, Merrill Lynch.
89
91
92
where d is the Dirac function. The operator pair Dx* , Dy* are the adjoint
operators of Dx, Dy.
The FPDE 6.3 can be seen as the dual of the BPDE 6.2 in the sense
that European-style options satisfy 6.2 but also satisfy:
The BPDE is solved backward in time and the solution describes the
future prices of a particular derivative for all times, spots and vola-
tility levels, whereas the FPDE is solved forward in time. The
solution gives the current marginal densities to all future times,
spot and volatility levels, and thereby the current prices of all
European-style options on spot and volatility.
The European-style call option prices are given by:
c (t, k ) = E ⎡⎣( s (t ) − k ) ⎤⎦ =
+
∫ ∫ ( x − k ) q (t, x, y ) dxdy
+
(6.5)
93
where:
E ⎡⎣z (t ) s (t ) = k ⎤⎦ =
∫ yq (t, k, y ) dy (6.7)
∫ q (t, k, y ) dy
A typical approach for implementing the model 6.1 would be to
discretise 6.3 to find the local volatility function from 6.6 that can
then, in turn, be used in a discretisation of the BPDE or in a Monte
Carlo simulation of a discretisation of the SDE 6.1.
There are several problems with this approach. The approxima-
tions are not mutually consistent. Specifically, direct application of
the same type of finite difference scheme to 6.2 and 6.3 would not
lead to the same results. Further, and more importantly, naive Euler
discretisation of the SDE for Monte Carlo may necessitate very fine
time stepping for the Monte Carlo prices to be close to the finite
difference prices. Finally, application of a finite difference scheme to
the FPDE requires specification of non-trivial boundary conditions
along the edges of the grid. The latter is particularly a problem for
the discretisation in the z direction for parameter choices when the
domain of z is closed and includes z = 0 (see Lucic, 2008).
(1− ΔtD ) v (t ) = v (t )
y h h+1/2
(6.8)
1
Dx = yσ 2δxx
2
1
δxx f ( x) = 2 [ f ( x + Δx ) − 2 f ( x ) + f ( x − Δx )] (6.9)
Δx
and:
1
Dy = θ (1− y ) δy + y 2γ ε 2δyy
2
1y<1 1
δy g ( y ) = ⎡⎣ g ( y + Δy ) − g ( y )⎤⎦ + y>1 ⎡⎣ g ( y ) − g ( y − Δy )⎤⎦
Δy Δy
1 ⎡
δyy g ( y ) = ⎣ g ( y + Δy ) − 2 g ( y ) + g ( y − Δy )⎤⎦ (6.10)
Δy 2
94
95
erty for the Monte Carlo scheme that we present in the next section.
As our schemes for calibration, backward finite difference and
simulation are mutually consistent, low (formal) order of accuracy
and (relatively) poor approximation of the SDE is not critical. We
can say that the basis of our modelling is the BFD scheme 6.8 rather
than the SDE 6.1.
The LOD scheme 6.8 can be rewritten (in transition form) as:
−1 −1
v (th ) = (1− ΔtDy ) (1− ΔtD ) v (t )
x h+1 (6.13)
then:
96
∂v
0= + Lv (6.20)
∂t
for some matrix L, then the associated Green’s function solves the
forward differential equation:
∂p
0=− + Lʹ′p (6.21)
∂t
This can be used for deriving boundary conditions for the FPDE 6.3
(see Andreasen, 2009).
Multiplying the last equation in 6.15 by (x – k)+ and summing
over x, y leads to the following discrete version of the Dupire
Equation 6.6:
1 ⎡
0=−
Δt ⎣
(
c (th+1 , x ) − c th+ 21 , x ⎤⎦ )
1 2
+ σ (th , x ) E ⎣⎡y (th+1 ) x⎤⎦δxx c (th+1 , x )
2
( )
c th+ 21 , k = ∑ ∑ ( x − k ) p th+ 21 , x, y
x y
+
( )
∑ yp (t h+1 , x, y )
E ⎡⎣y (th+1 ) x⎤⎦ = y (6.22)
∑ p (t h+1 , x, y )
y
97
Hence:
Ax−1ι = ι (6.26)
Ax−1 ≥ 0 (6.27)
Inspecting 6.12 reveals that the use of upwind operators for the first
derivative ensures that:
ak ,bk , ck ≥ 0 (6.28)
98
⎧ u v i≤ j
⎪ i j
Aij−1 = ⎨ (6.29)
⎪⎩ rj si i> j
x j −A j+1, j
Aij−1 = Ai,−1j+1 ⋅ ⋅ , j<i
x j+1 A jj
y j −A j−1, j
Aij−1 = Ai,−1j+1 ⋅ ⋅ , j>i (6.30)
y j−1 A jj
we have:
i j
The results in 6.30 can be used for identifying each of the elements
in 6.32 by recursion from the diagonal j = i.
This leads to the following simulation algorithm:
99
100
101
Note: The table reports the implied Black volatilities for five-year European-style options
with strikes given as 50%, 100% and 200% of the initial forward priced in a (Heston)
model with flat parameters. Results are reported for solution by Fourier inversion,
and backward finite difference solution (FD) and Monte Carlo (MC) for different grid
sizes. Here 25 × 100 × 25 refers to a grid with a total of 25 t-steps, 100 x-steps and
25 y-steps. The term “CS” refers to pricing in a second-order accurate Craig–Sneyd
scheme. MC prices are generated by 32 randomly seeded batches of 16,384 Sobol paths
(see Glasserman, 2003). MC pricing error is approximately 0.02% in implied Black
volatility terms. For the FD case, the reported CPU times include one forward sweep
(calibration) and one backward sweep (pricing) of the finite difference grid. For the MC
case, it includes one forward sweep, one decomposition sweep (MC initialisation) and
simulation of 16,384 paths. Hardware is a standard 2.66GHz Intel PC
102
FD 25 × 200 × 50
EPS = 0 36.96% 26.73% 19.19% 0.05s
EPS = 1 36.96% 26.73% 19.19% 0.42s
EPS = 2 36.96% 26.73% 19.19% 0.42s
EPS = 3 36.96% 26.72% 19.18% 0.42s
MC 25 × 200 × 50
EPS = 0 36.94% 26.74% 19.17% 0.68s
EPS = 1 36.90% 26.72% 19.14% 1.20s
EPS = 2 36.95% 26.72% 19.14% 1.20s
EPS = 3 36.99% 26.74% 19.14% 1.20s
FD 1,000 × 200 × 50
EPS = 0 36.98% 26.75% 19.21% 0.41s
EPS = 1 36.98% 26.75% 19.21% 14.14s
EPS = 2 36.98% 26.75% 19.21% 14.14s
EPS = 3 36.98% 26.74% 19.21% 14.14s
Note: The table reports the implied Black volatilities for five-year European-style
options with strikes of 50%, 100% and 200% of the initial forward priced when
the model is calibrated to the SX5E equity option market, for different levels of the
volatility of variance parameter e. Here, we use the same terminology for grid sizes
as in Table 6.1. Monte Carlo (MC) prices are generated the same way as for Table 6.2,
again with an MC pricing error of approximately 0.02% in implied Black volatility
terms. CPU times are also measured in the same way as in Table 6.1. The SX5E equity
option data is as of July 28, 2010
Note: The table reports prices of three different exotics on the SX5E equity index for
different levels of e. Prices are generated from 16,384 paths in a grid of dimensions
25 × 200 × 25. Market data is as of July 28, 2010
103
in this case, roughly 0.8 seconds is spent inside the simulation algo-
rithm described in Equation 6.33. Profiling our code reveals that
roughly 80% of the time spent in the simulation algorithm involves
drawing the random numbers in step 1. So in terms of speed, the
simulation methodology is, step by step, almost as fast as naive Euler
discretisation. Since our algorithm reproduces the exact distribution
of the BFD scheme 6.8, there will, however, be no need for more steps
in the simulation than in the BFD scheme.
Though two models produce the same prices for European-style
options, they can produce markedly different prices for exotics. To
illustrate this, we consider three different exotic options on the
SX5E equity index. Let ti = i/12 and define the returns:
s ( ti )
Ri = − 1, i = 1,..., 36 (6.35)
s (ti−1 )
Variance: U = ∑ Ri2
i
Table 6.3 reports the prices of these exotics for different levels of
volatility of variance e. We see that the variance contract is almost
invariant to the level of e. The intuition here is that if there are no
jumps in the underlying stock, then a contract on the continuously
observed variance can be statically hedged by a contract on the
logarithm of the underlying stock (see Dupire, 1993). Hence, if
European-style option prices are the same in two models with
continuous evolution of the stock, then the value of the variance
contract should be the same. The floored variance contract, on the
other hand, includes an option on the variance and should there-
fore increase with the volatility of variance parameter e. Finally, for
each period, the forward starting straddle payout is the square root
of the realised variance:
Ri = Ri2
104
Extensions
An easy way to extend the model to the multi-asset case is to use a
joint volatility process and correlate the increments of the under-
lying stocks using a Gaussian copula. Specifically, if u~i is the uniform
used for propagating stock i at a given time step, we can set:
u i = Φ ( w i ) , i = 1,...,l, w = ( w i ) = Pξ (6.37)
where x1, ... , xl are independent with xi ~ N(0, 1), and PPʹ is a correla-
tion matrix in Rl×l.
The BFD scheme 6.8 can be extended to include correlation
between stock and volatility, in the following way:
where:
Dxy = σερ y γ +1/2δxy
f ( x + Δx, y + Δy ) − f ( x + Δx, y − Δy )
δxy f ( x, y ) =
4ΔxΔy
f ( x − Δx, y + Δy ) − f ( x − Δx, y − Δy )
− (6.39)
4ΔxΔy
⎛ 1 ⎞ʹ′
p (th+1/2 ) = ⎜1+ ΔtDxy ⎟ p (th+1/4 )
⎝ 2 ⎠
105
The FFD scheme 6.40 is then to be used as the basis for calibration
instead of 6.15.
For simulation, we note that 6.38 can be rearranged as:
−1 ⎡ 1 ⎤
v (th ) = ⎡⎣1− ΔtDy ⎤⎦ ⋅ ⎢1+ ΔtDxy ⎥
⎣ 2 ⎦
−1 ⎡ 1 ⎤
⋅⎡⎣1− ΔtDx ⎤⎦ ⋅ ⎢1+ ΔtDxy ⎥ v (th+1 ) (6.41)
⎣ 2 ⎦
with x or y simulated for the first-order factors, and both for the
second order.
The matrix:
⎡ 1 ⎤
B ≡ ⎢1+ ΔtDxy ⎥ (6.42)
⎣ 2 ⎦
specifies weights that sum to one and link the left-hand side values
at state (x, y) to the right-hand side values at the states:
The trouble is that some of the entries of B are negative. This can be
handled by simulation according to the transition probabilities:
Bij
Pr ⎡⎣( x, y ) = ( xi , y j )⎤⎦ = (6.45)
∑∑ B kl
k l
106
Conclusion
We have presented a methodology that achieves full discrete
consistency between calibration, backward finite difference pricing
and Monte Carlo simulation, in the context of a stochastic local
volatility model. The methods extend to the multi-asset case and to
the case of non-zero correlation between the underlying and the
volatility process, as well as to other model types.
The authors would like to thank two anonymous referees for helpful
comments and suggestions.
REFERENCES
Achdou Y. and O. Pironneau, 2005, “Computational Methods for Option Pricing,” SIAM
Frontiers in Applied Mathematics.
Andersen L., 2006, “Efficient Simulation of the Heston Process,” working paper, Bank of
America.
Craig I. and A. Sneyd, 1988, “An Alternating-direction Implicit Scheme for Parabolic
Equations with Mixed Derivatives,” Computers and Mathematics with Applications, 16(4),
pp 341–50.
Glasserman P., 2003, Monte Carlo Methods in Financial Engineering (New York, NY:
Springer).
Heston S., 1993, “A Closed-form Solution for Options with Stochastic Volatility with
Applications to Bond and Currency Options,” Review of Financial Studies, 6(2), pp 327–43.
Huge B., 2010, “The Inverse of a Tridiagonal Matrix,” working paper, Danske Markets.
Lipton A., 2002, “The Vol Smile Problem,” Risk, February, pp 61–65.
Lucic V., 2008, “Boundary Conditions for Computing Densities in Hybrid Models via
PDE Methods,” working paper, Barclays Capital.
Nabben R., 1999, “On Decay Rates of Tridiagonal and Band Matrices,” SIAM Journal on
Matrix Analysis and Applications, 20, pp 820–37.
Press W., W. Vetterling, S. Teukolsky and B. Flannery, 1988, Numerical Recipes in C: The
Art of Scientific Computing (Cambridge, England: Cambridge University Press).
107
109
McKean SDEs
A McKean equation for an n-dimensional process X is an SDE in
which the drift and volatility depend not only on the current value
Xt of the process, but also on the probability distribution Pt of Xt:
dX t = b (t, X t , Pt ) dt + σ (t, Xt , Pt ) ⋅ dWt , Pt = Law ( Xt ) (7.1)
1 n ⎛ d ⎞
+ ∑ ∂ij ⎜∑σ ki (t, x, Pt ) σ kj (t, x, Pt ) p (t, x ) ⎟ = 0
2 i, j=1 ⎝ k=1 ⎠
lim p (t, x ) = δ ( x − X 0 ) (7.2)
t→0
It is nonlinear because bi(t, x, Pt) and sji (t, x, Pt) depend on the
unknown p.
Particle method
The stochastic simulation of the McKean SDE 7.1 is very natural. It
consists of replacing the law Pt , which appears explicitly in the drift
and diffusion coefficients, by the empirical distribution:
1 N
PtN = ∑δ i ,N
N i=1 Xt
where the Xti,N are the solution to the (Rn)N-dimensional linear SDE:
dX ti,N = b (t, X ti,N , PtN ) dt + σ (t, Xti,N , PtN ) ⋅ dWti , Law (X 0i,N ) = P0
110
σ i
j (t, x, P ) = ∫ σ (t, x, y ) p (t, y ) dy = E ⎡⎣σ (t, x, X )⎤⎦
t
i
j
i
j t
1 N 1 N
dX ti,N = ∑ b (t, X ti,N , X tj,N ) dt + ∑σ (t, X ti,N , X tj,N ) ⋅ dWti
N j=1 N j=1
1 N
∑ f (Xti,N ) ⎯N→∞
N i=1
L1
⎯⎯→∫ Rn
f ( x ) p (t, x ) dx
111
σ Dup (t, ft )
dft = ft T
at dWt (7.4)
E P ⎡⎣at2 ft ⎤⎦
The local volatility function depends on the joint PDF p(t, f, a) of (ft,
at):
112
for large enough volatility of volatility. This may come from numer-
ical error, or from the non-existence of a solution. The problem of
deriving the set of stochastic volatility parameters for which the
LSV model does exist for a given market smile is very challenging
and open (see an illustration in the numerical experiments section).
Local correlation models can similarly be put in McKean form
(see an extended version of this chapter, Guyon and Henry-
Labordère, 2011).
Hybrid models
A hybrid LSV model is defined in a risk-neutral measure P by:
dSt
= rt dt + σ (t,St ) at dWt
St
where the short-term rate rt and the stochastic volatility at are Itô
processes. For simplicity we assume no dividends. We explain how
to include (discrete) dividends in Guyon and Henry-Labordère
(2011).
This model is exactly calibrated to the market smile if and only if
(see Guyon and Henry-Labordère, 2011, for proof):
2 T 2
σ (T, K ) E P ⎡⎣aT2 ST = K ⎤⎦ = σ Dup (T, K ) − P0T (7.6)
1 2
K ∂K C (T, K )
2
113
where:1
⎛ ⎞
E P ⎡⎣PtT−1 ( rt − rt0 ) 1St >K ⎤⎦ ⎟
T
⎜ 2
( t
T
) 2
σ t, K, P = ⎜σ Dup (t, K ) − P0T
1 2 ⎟
⎜ K ∂K C (t, K ) ⎟
⎝ 2 ⎠
PT ⎡ −1 ⎤
E ⎣P St = K ⎦
× PT −1tT 2
E ⎡⎣PtT at St = K ⎤⎦ (7.7)
– –
s PT(t) is the volatility of the bond PtT_, and BtT is the (possibly multidi-
–
mensional) PT -Brownian motion that drives the interest rate curve.
Malliavin representation
We now give another expression of the contribution of stochastic
interest rates to local volatility:
P0T
1 2
K ∂K C (T, K )
2
114
T
E P ⎡⎣σ rT ( s) ⋅ DsB ST ST = K ⎤⎦ ds
T T
P0T = ∫ (7.8)
1 2 K 0
K ∂ K C (T, K )
2
where mr(t, rt) and sr(t, rt) are deterministic functions of the time t
and the short rate rt , and Bt is a one-dimensional P-Brownian motion
with d〈B, W〉t = r dt. Then s TP (t), the volatility of the bond PtT, is also
a deterministic function s TP (t, rt) of the time t and the short rate rt.
Moreover, we assume that the stochastic volatility is not correlated
with the stochastic rate rt. Both assumptions can easily be relaxed
but at the cost of additional straightforward calculations. By explic-
T
itly calculating DsB ST, 7.8 can then be written as (see Guyon and
Henry-Labordère, 2011, for detailed calculations):2
T
P0T = 2E P ⎡⎣VT ( ρUTT + ΘTT ΞTT − Λ TT ) ST = K ⎤⎦
1 2
K ∂ K C (T, K ) (7.9)
2
with:
115
dVt
= St ∂S σ (t,St ) at ( dWt − atσ (t,St ) dt ) , V0 = 1 (7.10)
Vt
σ (t,St )
dU tT = σ tT (t ) at dt, U 0T = 0 (7.11)
Vt
dRtT
= (∂r µ r (t, rt ) + σ r (t, rt ) ∂r σ PT (t, rt )) dt + ∂r σ r (t, rt ) dBt , R0T = 1 (7.12)
RtT
RtT
dΘTt =
Vt
(1+ ρ ∂r σ PT (t, rt )σ (t,St ) at ) dt, ΘT0 = 0 (7.13)
σ rT (t ) σ r (t, rt )
dΞTt = dt, ΞT0 = 0 (7.14)
RtT
and
T −1
σ rT (t) = E tP ⎡⎣RTT ⎤⎦( RtT ) σ r (t, rt ) (7.16)
σ rT (t) = σ r e −κ (T−t)
116
T
P0T = 2σ r e −κT E P ⎡⎣VT ( ρU T + ΘT ΞT − Λ T ) ST = K ⎤⎦ (7.18)
1 2
K ∂K C (T, K )
2
with
dVt
= St ∂S σ (t,St ) at ( dWt − atσ (t,St ) dt ) , V0 = 1
Vt
σ (t,St )
dU t = eκ t at dt, U0 = 0
Vt
e −κ t
dΘt = dt, Θ0 = 0
Vt
⎧ e 2 κ t − 1
⎪σ if κ ≠ 0
Ξt = ⎨ r 2κ
⎪σ t otherwise
⎩ r
2κ t
dΛ t = σ r e Θt dt, Λ0 = 0
117
=
∑ (a ) δ ( f − f )
i=1 t t
i,N
N
∑ δ( f − f )
i=1 t
i,N
σ N (t, f ) = σ Dup
∑ δ (f − f) i=1 t,N t
i,N
(7.20)
(t, f ) N i,N 2
∑ (a ) δ ( f − f ) t t,N t
i,N
i=1
and simulate:
A similar algorithm was used in Jourdain & Sbai (2010) in the case
of the joint calibration of smiles of a basket and its components.
At first sight, 7.20 and 7.21 require O(N2) operations at each
discretisation date: each calculation of sN(t, fti,N) requires O(N) oper-
ations, and there are N such local volatilities to calculate. This naive
method is too slow. First, computing sN(t, fti,N) for all i is useless.
One can save considerable time by calculating sN(t, f) for only a grid
Gf,t of values of f, of a size much smaller than N, say Nf,t , and then
inter- and extrapolate. We use cubic splines, with a flat extrapola-
tion, and Nf,t = max(Nf √t, Nf´); typical values are Nf = 30 and Nf´ = 15.
The range of the grid can be inferred from the prices of digital
options: E[1f >maxG ] = E[1f <minG ] = a. In practice, we take a = 10–3.
t f,t t f,t
Second, in the sums in 7.20, a large number of terms make a
negligible contribution: we can disregard fti,N when it is far from f,
118
Particle algorithm
Let {tk} denote a time discretisation of [0, T]. The particle algorithm
can now be described by the following:
❑❑ 1. Initialise k := 1 and set sN(t, f) = sDup(0, f)/a0 for all t ∈ [t0 = 0, t1].
❑❑ 2. Simulate the N processes {fti,N, ati,N}i=1,…,N from tk–1 to tk according
to 7.21.
❑❑ 3. Sort the particles according to spot value. _ For f ∈ Gf,tk, find the
smallest index _(fi ) and the largest index i (f) for which dt ,N(fti,N – f)
k k
> h, and calculate the local volatility:
i(f)
∑ δ
i= i ( f ) tk ,N
(f i,N
tk − f)
σ N (tk , f ) = σ Dup (tk , f ) i(f) i,N 2
∑ i= i ( f )
(a ) δ ( f
tk tk ,N
i,N
tk − f)
119
40 210 particles,
time = four seconds
35 212 particles,
time = 12 seconds
30 Market
No calibration
25
20
15
10
0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4
Strike
Note: Ho–Lee parameters σr = 6.3bp a day, ρ = 40%. Δ = 1/100, N = 210 on a full
10-year implied volatility surface with a Intel Core Duo, 3GHz, 3 GB of Ram:
four seconds
N −1
σ N (t,S)
2
=
∑ ( P ) δ (S − S)
i=1
i,N
tT t,N
i,N
t
N −1
i,N 2
∑ ( P ) ( a ) δ (S − S)
i=1
i,N
tT t t,N
i,N
t
⎛ 1 N i,N −1 i,N 0 ⎞
⎜ ∑ ( PtT ) ( rt − rt ) 1Si ,N >S ⎟
× ⎜σ Dup (t,S) − P0T N
2 i=1 t
1 2 ⎟
⎜ S∂K C (t,S) ⎟ (7.22)
⎝ 2 ⎠
and simulate:
dfti,N = fti,Nσ N (t, fti,N PtTi,N ) ati,N dWti − fti,Nσ PT ,i,N (t ) .dBti
–
where W i and Bi are PT-Brownian motions.
If we use representation 7.19 of the hybrid local volatility, we
need to add the Malliavin processes to the particle, which means
more processes to simulate, but usually results in a more accurate
estimation of the wings of the local volatility.
120
Numerical tests
Ho–Lee/Dupire hybrid model
We consider a hybrid local volatility model (at ≡ 1) where the short
rate follows a Ho–Lee model, _for which the volatility sr(t,rt) = sr is a
constant. A bond of maturity T is given by:
2
σ 2 T−t t
P0Tmkt r ( 2 ) −σ r (T−t)BtT
PtT = mkt e
P0t
_ _
with a volatility s PT(t) = –sr(T – t). From 7.19, the local volatility is:
T
2 E P ⎡⎣PtT−1VtU t St = K ⎤⎦
(
σ t, K, PtT ) 2
= σ Dup (t, K ) − 2 ρσ r T
E P ⎡⎣PtT−1 St = K ⎤⎦
T
E P ⎡⎣PtT−1Vt (tΘt − Λ t ) St = K ⎤⎦
−2σ r2 T
E P ⎡⎣PtT−1 St = K ⎤⎦
with:
121
Figure 7.3 High σ Bergomi LSV solution may not exist (Dax implied
volatilities T = 4Y: May 30, 2011)
50 Fit of the market smile for T = 4Y. 213 particles
45 VolVol = 350% 215 particles
40 Market
35 Approx
30
25
20
15
10
0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4
Strike
dVt
Vt
( (
= St ∂S σ (t,St ) dWtT + ρσ PT (t ) − σ (t,St ) dt ,V0 = 1 ) )
t σ ( s,Ss ) t ds t
Ut = ∫ 0 Vs
ds, Θt = ∫ 0 Vs
, Λt = ∫ 0
Θs ds
Note that in Benhamou, Gruz and Rivoira (2008), the local volatility
is approximated by:
2 2 t
σ (t, K ) ≈ σ Dup (t, K ) − 2 ρσ r ∫ 0
σ ( s, K ) ds
122
25
20
15
0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4
Strike
Strike 0.5 0.7 0.8 0.9 1.0 1.1 1.2 1.3 1.5 1.8
Note: Errors in basis points using the particle method with N = 210 particles
123
with N = 210 particles. Note that the calibration is also exact using
Equation 7.22, that is, with no use of the Malliavin representation,
but with a larger computational time of eight seconds with N = 210
particles (26 seconds with N = 212). As shown in Table 7.1, the abso-
lute error in implied volatility is a few basis points. For completeness,
we have plotted the smile obtained from the hybrid local volatility
model without any calibration, that is, s (t, K) = sDup(t, K), to materi-
alise the impact of the stochastic rate.
(
xtT = αθ (1− θ ) e −kX (T−t)X t + θ e −kY (T−t)Yt )
−1/2
( 2
αθ = (1− θ ) + θ 2 + 2 ρXYθ (1− θ ) )
X
dX t = −kX X t dt + dW
t
where:
2
h (t,T ) = (1− θ ) e −2 kX (T−t)E ⎡⎣X t2 ⎤⎦ + θ 2 e −2 kY (T−t)E ⎡⎣Yt2 ⎤⎦
+2θ (1− θ ) e −( kX +kY )(T−t)E [ X tYt ]
1− e −2 kX t ⎡ 2 ⎤ 1− e −2 kY t
E ⎡⎣X t2 ⎤⎦ = , E ⎣Yt ⎦ = ,
2k X 2kY
1− e −( kX +kY )t
E [ X tYt ] = ρXY
k X + kY
124
(Dax, May 30, 2011) and the LSV model for maturities of four years
and 10 years. The computational time is four seconds for maturities
up to 10 years with N = 210 particles (11 seconds with N = 212). This
should be compared with the approximate calibration (Henry-
Labordère, 2009), which has a computational time of around 12
seconds. To illustrate that we have used stressed parameters to
check the efficiency of our algorithm, we have plotted the smile
produced by the naked stochastic volatility model, which signifi-
cantly differs from the market smile.
125
Conclusion
We have explained how to calibrate multi-factor hybrid local
stochastic volatility models exactly to market smiles using the
particle algorithm to simulate solutions to McKean SDEs. We have
also provided a Malliavin stochastic representation of the stochastic
interest rate contribution to local volatility. The technique we
proposed proves useful when models incorporating multiple vola-
tility and interest rate risks are needed, typically for long-dated,
forward skew-sensitive payouts. Our algorithm represents, to the
best of our knowledge, the first exact algorithm for the calibration
of multi-factor hybrid local stochastic volatility models. Acceleration
techniques make it efficient in practice. As highlighted in our
numerical experiments, the computation time is excellent and, even
for low-dimensional (hybrid) LSV models, our algorithm outper-
forms PDE implementations.
The analysis of nonlinear (kinetic) PDEs arising in statistical
physics such as the McKean–Vlasov PDE and the Boltzmann equa-
tion has become more popular and drawn attention in part thanks
to the work of Fields medallist Cédric Villani. We hope that this
work will initiate new research and attract the attention of practi-
tioners to the world of nonlinear SDEs.
This research was done at the time when both authors were working
at Société Générale. The authors wish to thank their colleagues at
Société Générale for useful discussions. This chapter is dedicated to
the memory of Paul Malliavin, who pointed out to us the efficiency of
the Clark–Ocone formula as we were working on this project.
126
REFERENCES
Balland P., 2005, “Stoch-vol Model with Interest Rate Volatility,” ICBI conference.
Benhamou E., A. Gruz and A. Rivoira, 2008, “Stochastic Interest Rates for Local Volatility
Hybrids Models,” Wilmott Magazine.
Bergomi L., 2008, “Smile Dynamics III,” Risk, October, pp 90–96 (available at www.risk.
net/1500216).
Guyon J. and P. Henry-Labordère, 2013, “Nonlinear Option Pricing,” Chapman & Hall,
CRC Financial Mathematics Series (forthcoming).
Henry-Labordère P., 2009, Analysis, Geometry, and Modeling in Finance: Advanced Methods
in Option Pricing (Chapman & Hall, CRC Financial Mathematics Series).
Jourdain B. and M. Sbai, 2010, “Coupling Index and Stocks,” Quantitative Finance,
October.
Lipton A., 2002, “The Vol Smile Problem,” Risk, February, pp 61–65 (available at www.
risk.net/1530435).
Piterbarg V., 2006, “Smiling Hybrids,” Risk, May, pp 66–71 (available at www.risk.
net/1500244).
127
129
Collateralised processes
A collateralised derivative has quite a different set of cashflows
from an uncollateralised “traditional” one. At the inception of the
collateralised trade there is no exchange of cashflows – the price
paid for the derivative is immediately returned as collateral. During
the life of a collateralised trade, there is a continuous stream of
payments based on the changes in the trade’s mark-to-market. A
collateralised trade can be terminated at any moment at zero addi-
tional cost. So the notion of a price of a collateralised asset is actually
somewhat misleading – as the trade can be terminated at zero addi-
tional cost, the value of this transaction is always zero. What we
would call a price is nothing but a process that defines the level of
the collateral holding. Or, in the language of the classic asset pricing
theory, a collateralised transaction is an asset with a zero price
process and given cumulative-dividend process (see Duffie, 2001).
A moment’s thought shows that this is very much the same as for a
futures contract. In fact, a futures contract is just a particular type of
collateralised forward contract, with the collateral rate set to zero.
Keeping this in mind helps set the right picture.
Still, given the standard terminology, we would still use the word
“price” for a collateralised trade, but the meaning should always be
clear – this is the level of holding of a collateral at any given time.
Let us start by introducing some notation. Let V(t) be the
price of a collateralised asset between parties A and B. If V(t) is
positive from the point of view of A, party B will post V(t) to A.
Party A will then pay party B a contractually specified collateral
130
with both assets driven by the same Brownian motion. This is the
case when, for example, we have a stock1 and an option on that
stock. At time t, we can enter into a portfolio of two collateralised
transactions to hedge the effect of the randomness of dW(t) on the
cash exchanged at time t + dt. To do that, we go long a notional of
s2 (t)V2(t) in asset 1 and short a notional of s1(t)V1(t) in asset 2. The
cash exchange at time t + dt is then equal to:
σ 2 (t ) V2 (t ) ( dV1 (t) − c (t ) V1 (t ) dt )
−σ 1 (t ) V1 (t) ( dV2 (t) − c (t ) V2 (t ) dt )
131
σ 2 (t ) V1 (t) V2 (t) (µ 1 (t ) − c (t )) dt
−σ 1 (t ) V1 (t) V2 (t) (µ 2 (t) − c (t )) dt
σ 2 (t ) (µ 1 ( t ) − c ( t )) = σ 1 (t ) (µ 2 ( t ) − c ( t ) )
and, in particular:
µ 1 (t ) − c ( t ) µ 2 (t ) − c ( t )
=
σ 1 (t ) σ 2 (t)
(t ) dW (t) + µ 1 (t) − c (t ) dt
dW
σ 1 (t)
(t ) = dW (t ) + µ 2 (t ) − c (t ) dt
dW
σ 2 (t )
~ as:
Hence, we can rewrite 8.1 using the newly defined dW
(t ) , i = 1, 2
dVi (t ) = c (t) Vi (t ) dt + σ i (t) Vi (t ) dW (8.2)
⎛ − ∫ T c(s)ds ⎞
Vi (t ) = EtQ ⎜ e t Vi (T ) ⎟ , i = 1, 2 (8.3)
⎝ ⎠
σ 2 (t ) (µ 1 ( t ) − c 1 (t )) = σ 1 (t ) (µ 2 ( t ) − c 2 ( t ) )
132
wT Σ = 0 (8.5)
Then the cash in the portfolio wTV has no randomness and hence,
by the no-arbitrage arguments used previously, we must have that:
w T (µV − cV ) = 0
So we can write:
dV = cVdt + Σ ( dW + λ dt )
133
Cross-currency model
The previous section gives a flavour of the results one gets for an
asset collateralised with different rates, but probably the main
example when this situation occurs is in cross-currency markets.
134
135
Drift of FX rate
Equations 8.6 and 8.7 are insufficient to determine the drift of the
forex rate X(·). From 8.7 we can only deduce the drift of the
combined quantity XPf,d and the drift of Pf,d is in general not cf (nor is
it cd, for that matter). To understand the drift of X(·), we need to
understand what kind of domestic cashflow we can generate from
holding a unit of foreign currency. So, suppose we have 1F. If it was
a unit of stock, we could repo it out (that is, borrow money secured
by the stock) and pay a repo rate on the stock. What is the equiva-
lent transaction in the forex markets? Having 1F, we can give it to
another dealer and receive its price in domestic currency, X(t)D.
The next instant t + dt we would get back 1F, and pay back X(t) +
rd,f(t)X(t)dt, where rd,f(t) is a rate agreed on this domestic loan collat-
eralised by foreign currency. As we can sell our 1F for X(t + dt)D at
time t + dt the cashflow at t + dt would be:
dX (t ) − rd, f (t ) X (t ) dt
136
⎛ ⎞
⎜ dX / X ⎟
⎜ dPd,d / Pd,d ⎟ = µ dt + ΣdW
⎜ ⎟
⎜ d P X / P X ⎟
⎝ ( f ,d ) ( f ,d )⎠
In particular, we have:
⎛ − ∫ T rd , f (s)ds ⎞
X (t ) = Etd ⎜ e t X (T ) ⎟
⎝ ⎠
⎛ − ∫ cd (s)ds ⎞
T
Note the drift of the first component is the rate –rd,f, which is the rate
on the instantaneous forex swap from the point of view of the
foreign party. In particular:
⎛ − ∫ T c f (s) ds 1 ⎞
Pd, f (t,T ) = X (t ) Etf ⎜ e t ⎟ (8.11)
⎝ X (T ) ⎠
It is not hard to see the connection between Qf and Qd. In
particular:
dQ f ∫ 0 rd , f (s) ds X (t )
t
−
= M (t ) e (8.12)
dQ d Ft
X ( 0)
137
∫ 0 rd , f (s) ds X (t)
t
−
M (t ) = e
X ( 0)
Forward forex
Forward forex contracts are traded among dealers and, as such, are
subject to collateralisation rules. A forward forex contract pays X(T)
– K at T in domestic currency. The price process of the domestic-
currency-collateralised forward contract is given by:
⎛ − ∫ T cd (s) ds ⎞
Etd ⎜ e t (X (T ) − K )⎟ = X (t) Pf ,d (t,T ) − KPd,d (t,T )
⎝ ⎠
and so the forward forex rate, that is, K that makes the price process
have value zero is given by:
X (t) Pf ,d (t,T )
X d (t,T ) = (8.13)
Pd,d (t,T )
X d (t,T ) = Etd,T ( X (T ))
138
⎛ − ∫ T c f (s)ds ⎞
Etf ⎜ e t (1− K / X (T )) ⎟ = Pf , f (t,T ) − KPd, f (t,T ) / X (t)
⎝ ⎠
139
qd, f (t ) c f (t ) + rd, f (t ) − cd (t )
Zero-coupon curves
Before we start, let us discuss time-zero market data that the model
needs to recover. We have the domestic-collateral, domestic-
currency zero-coupon bonds Pd,d(0, T) that can be obtained from the
market on linear instruments in a single currency. We denote corre-
sponding instantaneous forward rates by pd,d(0, T) = –∂log Pd,d(0,
T)/∂T. Similarly we can build the “pure foreign” discounting curve
Pf,f(0, T), pf,f(0, T). From the cross-currency swaps collateralised in
the foreign currency (or from the forex forward market via 8.14), we
can obtain the foreign-collateral domestic zero-coupon bonds Pd,f(0,
T) and corresponding forward rates pd,f(0, T).
Note that we have from 8.11 and the measure change 8.12 that:
Pd, f (t,T )
⎛ − ∫ T c f (s)ds 1 ⎞ ⎛ − ∫ T (c f (s)+rd , f (s)) ds ⎞
= X (t ) Etf ⎜ e t ⎟ = Etd ⎜ e t ⎟
⎝ X (T ) ⎠ ⎝ ⎠
⎛ T T
− ∫ cd ( s) ds − ∫ qd , f ( s)ds ⎞ ⎛ − ∫ qd , f ( s)ds ⎞
T
= Etd ⎜ e t e t d,T
⎟ = Pd,d (t,T ) Et ⎜ e t ⎟
⎝ ⎠ ⎝ ⎠
140
Dynamics
We work under the domestic measure. Let the dynamics for Pd,d(t,
T) be given by:
where rX,f is the correlation between dWX and dWf . Again, the
dynamics for cf will follow from the standard HJM arguments.
Now let us decide on the dynamics of qd,f . Recall:
⎛ − ∫ T qd , f (s)ds ⎞
Qd, f (t,T ) = Etd,T ⎜ e t ⎟
⎝ ⎠
where Qd,f(t, T) Pd,f (t, T)/Pd,d (t, T). Denoting the volatility of Qd,f (t,
T) by Sq(t, T) and using Wq(t) as a driving Brownian motion under
Qd we can write down the dynamics of Qd,f (t, T) as:
141
Observations
We make the following observations regarding the model 8.19.
While writing the dynamics is relatively straightforward, using
such a model in practice presents considerable challenges. There are
a number of parameters that are simply not observable in the
market, such as those for the process qd,f and various correlations.
Even if statistical estimates could be used, hedging these parame-
ters would be very difficult. Moreover, the option to switch collateral
– which is ultimately the application of this model – could disap-
pear for reasons unrelated to the model, such as a move to central
clearing or a standard credit support annex. Moreover, there are
doubts about whether an ability to instantaneously switch collateral
from one currency to another is a good reflection of reality. On the
other hand, it does give a way of getting some estimate for the
option to switch collateral, and is derived in a rigorous way.
Another point worthy of note is that a model 8.19 is a model of
discounting only. If one were to use it to price interest rate deriva-
tives beyond those depending on discounting rates only, additional
dynamics would need to be specified for forecasting curves. This
142
can be done, for example, either for Libor forwards or for short
rates that drive the forecasting curves. These can be specified as
deterministic spreads to the collateral rates or, in full generality,
modelled with their own stochastic drivers, further increasing the
number of unobservable parameters. In the latter case, not only the
discounting curves will depend on the collateralisation used, but
also forecasting curves such as forward Libor curves will as well, in
close analogy to the quanto-type adjustments obtained in Piterbarg
(2010).
⎛ − ∫ T max(qd , f (s),0) ds ⎞
Pd,d ( 0,T ) E d,T ⎜ e 0 V (T )⎟ (8.20)
⎝ ⎠
For an interest rate swap, say, V(T) here will be either a constant
(fixed leg cashflow) or a Libor rate fixing (floating leg cashflow). Let
us consider the fixed leg first. Here we need to calculate:
⎛ − ∫ T max(qd , f (s),0) ds ⎞
E d,T ⎜ e 0 ⎟ (8.21)
⎝ ⎠
143
Given that in the model 8.19 qd,f (s) is Gaussian, the required
Ed,s[max(qd,f (s), 0)] can be readily calculated in closed form. In prac-
tice, we would calculate it for a number of points si and interpolate
in between. While 8.22 is only an approximation, at least for some
values of market parameters it appears to be a good one (see below).
For the floating leg, a pragmatic choice would be to move the
Libor fixing outside of the expected value, that is, replace 8.20 with:
⎛ − ∫ t max(qd , f (s),0)ds ⎞ d,T
Pd,d ( 0,T ) E d,T ⎜ e 0 ⎟ E (V (T ))
⎝ ⎠
Example
Here, we present a numerical example for collateral choice option
valuation. We use data from November 2011 from Barclays. The
domestic currency is the euro and the foreign currency is sterling.
We use the following parameters for the process qd,f (·), estimated
historically: sq = 0.50% and aq = 40%. In Figure 8.1, we plot a number
of forward curves against time t in years. The curve labelled option
forward is pd,f (0, t) – pd,d (0, t), that is, the forward curve for the spread
process qd,f (·). The curve labelled option value (intrinsic) is the curve
max(pd,f (0, t) – pd,d (0, t), 0). This would be the value of the collateral
choice option assuming deterministic evolution of the spread qd,f (·).
The curve labelled option value (exp) is the true value calculated
from 8.21 by Monte Carlo simulation, expressed in instantaneous
forward rate terms:
∂ ⎛ − ∫ t max(qd , f (s),0)ds ⎞
Option value (exp) = − log E d,t ⎜ e 0 ⎟
∂t ⎝ ⎠
Finally, the curve labelled option value (first order) is the first-order
approximation from 8.22, Ed,t[max(qd,f (t), 0)].
We see that the option value is not insignificant. We also see that
the first-order approximation matches the true value of the option
closely, at least for the values of the parameters used.
144
0.6
0.4
0.2
0
?? ?? ?? ?? ?? ?? ?? ?? ??
–0.2
%
–0.4
Option forward
–0.6 Option value (intrinsic)
Option value (exp)
–0.8 Option value (first order)
–1.0
Conclusion
We have developed a framework for asset pricing in an economy
where there is no risk-free rate and all transactions are collateral-
ised. It turns out that much of the machinery of standard risk-neutral
pricing theory can be reused, with a few changes. In the risk-neutral
measure, each collateralised asset grows at the rate at which it is
collateralised. The forex rate drift is not given by the difference of
the risk-free rates in two currencies (as they do not exist in such an
economy), but is given by a rate on an instantaneous forex swap,
which is essentially an overnight repo rate on the sale of one unit of
foreign currency for domestic price. Consequently, the forex rate
drift is not dependent on the collateral rates in the two economies
(domestic and foreign), but the forward forex rates are.
Furthermore, we demonstrated a simple model with stochastic
dynamics for the difference between the forex-adjusted foreign
collateral rate and the domestic collateral rate in which the option
to switch collateral has time value, commented on the practical use
of such a model and presented a numerical example.
The author would like to thank Rashmi Tank and Thomas Roos for
stimulating discussions.
1 Collateralised stock sale is actually a repo transaction. Here we assume that the repo rate
is the same as the collateral rate. We consider different rates later
145
REFERENCES
Andersen L. and V. Piterbarg, 2010, Interest Rate Modeling (London, England: Atlantic
Financial Press).
Black F., 1972, “Capital Market Equilibrium with Restricted Borrowing,” Journal of
Business, 45, pp 444–55.
Duffie D., 2001, Dynamic Asset Pricing Theory (3e) (Princeton, NJ: Princeton University
Press).
Macey G., 2011, “Pricing with Standard CSA Defined by Currency Buckets,” SSRN
eLibrary.
Piterbarg V., 2010, “Funding Beyond Discounting: Collateral Agreements and Derivatives
Pricing,” Risk, February, pp 97–102 (available at www.risk.net/1589992).
146
149
Short-squeezes
A short-squeeze is often defined as a situation in which an imbal-
ance between supply and demand causes the stock to rise abruptly
and a scramble to cover on the part of short sellers. The need to
cover short positions drives the stock even higher. In another
150
Cost of conversions
Converting means selling a call option and buying a put option of
the same strike and 100 shares of stock. According to put–call parity,
for an ordinary (non-dividend paying) stock, the premium-over-
parity of a call (Cpop) should exceed the premium-over-parity of the
corresponding put (Ppop) by an amount approximately equal to the
strike times the spot rate.1 In particular, a converter should receive a
credit for selling the call, buying the put and buying 100 shares.
However, for hard-to-borrow stocks the reverse is often true. In
January 2008, prior to announcing earnings, the stock of VMWare
Corporation (VMW) became extremely hard to borrow. This was
reflected by the unusual cost of converting on the January 2009
at-the-money strike. The difference Cpop – Ppop for the January 2009
US$60 line was –US$8! A converter would therefore need to pay
US$8 (per share) to enter the position, that is, US$800 per contract.
Following the earnings announcement, VMW fell roughly US$28
(see Figure 9.3). At the same time, the cost of the conversion on the
60 strike in January 2009 dropped in absolute value to
35
33
31
29
27
25
1 997
84 167 250 333 416 499 582 665 748 831 914 1,0801,1631,246
1,329
1,412
1,495
1,5781,661
Note: Note the huge spike, which occurred on the closing print of June 19. The
price retreats nearly to the same level as prior to the buy-in
151
120
100
VMW ($)
80
60
40
20
0
Nov Dec Jan Feb Mar Apr May Jun Jul Aug Sep
2007 2008
Note: The large drop in price after an earnings announcement in late January
2008 was accompanied by a reduction in the difficulty to borrow, as seen in
the price of conversions
152
600
500
400
300
200
100
0
Sept Oct Nov
153
The model
We assume that under the physical measure the hard-to-borrow
stock St and buy-in rate lt satisfy the system of coupled equations:
dSt
= σ dWt + γλt dt − γ dN λt (t) (9.1)
St
dSt
dX t = κ dZt + α ( X − X t ) dt + β , X t = ln ( λt / λ0 ) (9.2)
St
where dNl(t) denotes the increment of a standard Poisson process
with intensity lt over the interval (t, t + dt); the parameters s and g
are respectively the volatility of the continuous part and the price
elasticity of demand due to buy-ins; and Wt is a standard Brownian
motion. Equation 9.2 describes the evolution of the logarithm of the
buy-in rate; k is the_volatility of the rate, Zt is a Brownian motion
independent of Wt, X is a long-term equilibrium value for Xt, a is the
speed of mean-reversion and b couples the change in price with the
buy-in rate.
We assume that b > 0; in particular, Xt = ln(lt) is positively
correlated with price changes, introducing a positive feedback
between increases in buy-ins (hence in short interest in the stock)
and price.
Equations 9.1 and 9.2 describe the evolution of the stock price
across an extended period of time. One can think of a diffusion
process for the stock price, which is punctuated by jumps occurring
at the end of the trading day, the magnitude and frequency of the
latter being determined by lt. Fluctuations in lt represent the fact
that a stock may be difficult to borrow one day and easier another.
In this way, the model describes the dynamics of the stock price as
costs for stock-loan vary. Short squeezes can be seen as events asso-
ciated with large values of lt, which will invariably exhibit price
spikes (rallies followed by a steep drop).
154
155
dSt
= σ dWt + rdt − γ dN λt (t ) (9.3)
St
where r is the instantaneous interest rate. The absence of the drift
term ltg in this last equation is due to the fact that, under an arbi-
trage-free pricing measure, the discounted price process is a
martingale.
It follows from Equation 9.3 that the stock price in the risk-neutral
world can be written as:
t
St = S0 e rt Mt (1− γ ) ∫0 λt
dN (t )
(9.4)
⎩⎪ ⎭⎪
⎧ ⎛ T ⎞
k ⎫
⎪ T
⎪ − ∫ λt dt ⎝
⎜ ∫ λ t dt ⎟ ⎪
⎠ k ⎪
= S0 e rT E ⎨e 0 ∑ 0 (1− γ ) ⎬
⎪ k k! ⎪
⎪ ⎪
⎩ ⎭
⎧ −γ ∫ λt dt ⎫
T
⎪ ⎪
= S0 e rT E ⎨e 0 ⎬ (9.5)
⎪⎩ ⎪⎭
∫ dt dt ⎧ −γ T∫ λt dt ⎫
−
⎪ ⎪
e 0
= E ⎨e 0 ⎬ (9.6)
⎪⎩ ⎪⎭
156
where P(K, T), C(K, T) represent respectively the fair values of a put
and a call with strike K and maturity T, S is the spot price and R, D
are respectively the simply discounted interest rate and dividend
rate. It is equivalent to:
may not result in a risk-free profit due to the fact that the crucial
stock hedge (short 100 shares) may be impossible to establish.
We quantify deviations from put-call parity by considering the
function:
157
Figure 9.4 Implied dividend rates as a function of strike price for options
on Dendreon
0.50
0.45
0.40
Implied dividend (%) 0.35
0.30
0.25
0.20
0.15
0.10
0.05
0
0 5 10 15 20 25 30 35 40 45
Strike (US$)
Note: The trade date is January 10, 2008 and the expiry is January 17, 2009.
The stock price is US$5.81. The best fit constant dividend rate is approximately
15%. Dendreon does not pay dividends
158
0.14
0.12
Implied dividend (%)
0.10
0.08
0.06
0.04
0.02
0
0 50 100 150 200 250 300
Strike ($)
Note: The dates are as in Figure 9.4 and the stock price is US$80.30. The best fit
dividend rate (associated with at-the-money options) is 5.5%. VMWare does
not pay dividends
⎡ −γ T∫ λt dt ⎤
T
− ∫ dt dt
1− E ⎢e 0 ⎥
⎢ ⎥
1− e 0 ⎣ ⎦
D * (T ) = = (9.10)
T T
which connects the implied dividend rate obtained from the options
markets to the buy-in rate process.
The option market predicts different borrowing rates over time
for any given stock, through variations in the interest rate quoted
by clearing firms, and by conversion-reversals quoted by option
market-makers. The latter approach suggests different implied
dividends per option series, that is, it contains market expectations
of the varying degree of difficulty of borrowing a stock in the future.
We can use the model 9.1–9.2 and Equation 9.10 to calculate a term
structure of effective dividends (or, equivalently, short rates) that
could be calibrated to any given stock. To generate such a term
structure, we simulate paths of lt, 0 < t < Tmax and calculate the
discount factors by Monte Carlo. Figure 9.6 shows a declining term
structure, which is typical of most stocks. This decay represents the
fact that stocks rarely remain hard to borrow over extremely long
time periods.
159
15
14
D*(T) (%) 13
12
11
10
9
8
0 0.5 1.0 1.5 2.0
T (years)
Note: D*(T) for λ0 = 15, γ = 0.01, β = 30, σ = 0.5
0.12
0.10
0.08
0.06
0.04
0.02
0
50 60 70 80 90 100 110 120 130 140 150
Strike
Note: We assume that the stock price is US$100, σ = 0.50, β = 1.00, λ0 = 50,
T = 0.5yrs, γ = 0.03, r = 10%. The effective dividend rate is dimp(100, T) = 14%.
For low strikes, the drop in value is related to the early exercise of calls, a
feature unique to hard-to-borrow stocks. For high strikes, the broad increase
corresponds to the classical early exercise property of in-the-money puts
160
⎧ T ⎫
Π (n,T ) = Prob ⎨ ∫ dN λt (t ) = n⎬
⎩ 0 ⎭
⎧ ⎞ ⎫
n
⎛ T
⎪ T ⎜ ∫ λt dt ⎟ ⎪
− ∫ λt dt
⎪ ⎝ 0 ⎠ ⎪
= E ⎨e 0 ⎬ (9.11)
⎪ n! ⎪
⎪ ⎪
⎩ ⎭
∫ dN λt (t )
0
161
This suggests that we can use daily data on leveraged ETFs to esti-
mate the cost of borrowing the underlying stock.
For the empirical analysis, we used dividend-adjusted closing
prices from the PowerShares Ultrashort Financial ETF (SKF) and
the PowerShares Ultralong Financial ETF (UYG). The underlying
ETF is the Barclays Dow Jones Financial Index ETF. Using historical
data, we calculated the right-hand side of Equation 9.15, which we
interpret as corresponding to daily sampling, with dt = 1/252, r =
162
Figure 9.8 The cost of borrowing the Barclays Dow Jones Financial
Index
300
250
200
150
100
%
50
0
–50
–100
Jan 2007
Feb 2007
Mar 2007
Apr 2007
May 2007
Jun 2007
Jul 2007
Aug 2007
Sep 2007
Oct 2007
Nov 2007
Dec 2007
Jan 2008
Feb 2008
Mar 2008
Apr 2008
May 2008
Jun 2008
Jul 2008
Aug 2008
Sep 2008
Oct 2008
Nov 2008
Dec 2008
Jan 2009
Feb 2009
Note: The thin line corresponds to the daily values of the cost of borrowing
parameter γλt, in percentage points, estimated from Equation 9.16. The thick
line is a 10-day moving average. Hard-to-borrowness exceeds 100% in
September–October 2008 and remains elevated until March 2009
three-month Libor and f = 0.95%, the expense ratio of SKF and UYG
advertised by Powershares. The results of the simulation are seen in
Figure 9.8.
We see that rt, the cost of borrowing, varies in time and can
change quite dramatically. In Figure 9.8, we consider a 10-day
moving average of rt to smooth out the effect of volatility and end-
of-day marks. The data shows that increases in borrowing costs, as
implied from the leveraged ETFs, began in late summer 2008 and
intensified in mid-September, when Lehman Brothers collapsed
and the US Securities and Exchange Commission (SEC) ban on
shorting 800 financial stocks was implemented (the latter occurred
on September 19, 2008). Note that the implied borrowing costs for
financial stocks remain elevated subsequently, despite the fact that
the SEC ban on shorting was removed in mid-October. This calcula-
tion may be interpreted as exhibiting the variations of lt (or glt) for
a basket of financial stocks. For instance, if we assume that the elas-
ticity remains constant (for example, at 2%), the buy-in rate will
range from a low number (for example, l = 1, or one buy-in a year)
to 50 or 80, corresponding to several buy-ins a week.
163
Conclusion
In the past, attempts have been made to understand option pricing
for hard-to-borrow stocks using models that do not take into
account price dynamics. This approach leads to a view of put–call
parity that is at odds with the functional equilibrium (steady state)
evidenced in the options markets, in which put and call prices are
stable and yet naive put–call parity does not hold. The point of this
chapter has been to show how dynamics and pricing are inter-
twined. The notion of effective dividend is the principal consequence
of our model, which also obtains a term structure of dividend
yields. Reasonable parametric choices lead to a term structure that
is concave down, a shape frequently seen in real option markets.
The model also reproduces the (American-style) early exercise
features, including early exercise of calls, which cannot happen for
non-dividend-paying easy-to-borrow stocks.
The authors would like to thank the referees for many helpful and
insightful comments, Sacha Stanton of Modus Incorporated for assist-
ance with options data and Stanley Zhang for exciting discussions on
leveraged exchange-traded funds.
1 Premium-over-parity (POP) means the difference between the (mid-)market price of the
option and its intrinsic value. Some authors also call the POP the extrinsic value. We use
“approximately equal” because listed options are American-style, so they have an early
exercise premium. Nevertheless, at-the-money options will generally satisfy the put–call
parity equation within narrow bounds.
REFERENCES
Amin K., 1993, “Jump-diffusion Option Valuation in Discrete Time,” Journal of Finance,
48, pp 1,833–63.
Diamond D. and R. Verrecchia, 1987, “Constraints on Short-selling and Asset Price
Adjustment to Private Information,” Journal of Financial Economics, 18, pp 277–312.
Duffie D., N. Garleanu and L. Pedersen, 2002, “Securities Lending, Shorting, and
Pricing,” Journal of Financial Economics, 66, pp 307–39.
Evans R., C. Geczy, D. Musto and A. Reed, 2008, “Failure is an Option: Impediments
to Short Selling and Options Prices,” Review of Financial Studies (available at http://rfs.
oxfordjournals.org/cgi/ content/full/hhm083).
Jones C. and O. Lamont, 2002, “Short Sale Constraints and Stock Returns,” Journal of
Financial Economics, 66, pp 207–39.
Natenberg S., 1998, Option Volatility and Strategies: Advanced Trading Techniques for
Professionals (2e) (Chicago, Ill: Probus).
Nielsen L., 1989, “Asset Market Equilibrium with Short-selling,” Review of Economic
Studies, 56(3), July, pp 467–73.
164
165
is the correlation weight vector for the jth issuer.2 If risk factors
correspond to sectors, for example industrial sectors, and each
issuer corresponds to a unique sector, then we can proceed as
follows: as explained in Martin and Tasche (2007) and Martin (2009),
each issuer’s contribution can be split into a systematic and an
unsystematic part. The systematic parts are grouped by sector, and
are represented as the central segments of the “dartboard” in Figure
10.1, with an area proportional to the associated contribution. The
unsystematic parts, which are necessarily positive, are arranged
around the edge. If a few issuers belong to, say, two sectors (for
example, automotive and financial), then one can legitimately
subdivide their contribution appropriately between the sectors and
still arrive at the same kind of result, with only a small loss in clarity
as to what exactly is meant by sector contribution.
However, factors may not correspond to sectors in a simple way.
For example, the model 10.1 could be a model of credit card loans or
other types of retail loan. In that case, the factors in question might
be interest rates, GDP, unemployment, regional, foreign exchange
(for foreign currency loans), etc. Or, more generally, the portfolio
could be a fund of hedge funds, each with different styles (vola-
tility, momentum, value, etc). Then, each constituent is linked to
many factors, not just one. Again, one wants the sensitivity of risk
to each factor, thereby differentiating with respect to V rather than a.
Whereas simple Euler-type constructions are central to the theory
of position risk contribution, and have a certain appeal, they seem
to cause problems with factor contributions.
The first difficulty with the Euler construction is that, whereas
the derivative of a risk measure with respect to a parameter (such as
asset allocation) is easily defined, the derivative with respect to a
random variable (such as a risk factor) is not. On the other hand, mY|V
can be differentiated with respect to V because it is simply a func-
tion of V. It seems, therefore, that we should attempt to find
“contributions” only to quantities that can be represented as func-
tions of mY|V. Now the systematic part of the ES measure, previously
defined as E[mY|V|Y > y], is one such function, as it is a weighted sum
of mY|V over all values of V (explicitly, the weight is P[V = v|Y > y].
Indeed, some integration over V is inevitable, otherwise one ends
up with a contribution formula that is a random variable). Let us
now denote by R the function of V that we are trying to decompose,
166
and consider what the construction should look like. It is here that
we encounter further problems.
Consider, for example, the derivative with respect to the kth
factor, that is, ∂R/∂Vk. This is unlikely to be useful, because it has
dimensions reciprocal to Vk. In other words, if the units of Vk are
changed, or Vk is rescaled, the risk contribution changes. It is then
impossible to compare the contributions of different factors. If one
examines Vk∂R/∂Vk instead, one obtains a measure that is invariant
to scaling but dependent on the choice of origin of the factor, so it
changes if the factor’s origin is shifted (and by shifting each factor
appropriately, one can arrange for each factor to have zero contri-
bution, which seems fundamentally wrong). An improvement on
both these two ideas is to normalise by the standard deviation of
the factor, that is, something more like s [Vk]∂R/∂Vk. This now has the
required invariance to scaling and shifting, and has some merit,
though the standard deviation might not be the best quantity to
multiply by if one is more interested in tail risks, and it is a little ad
hoc.
There is also the issue of what the contributions add up to, because
whereas risk is a homogeneous function of asset allocations, it is not
in general a homogeneous function of the factors (and indeed it is not
in 10.1, though notably it is linear in CreditRisk+). So the contribu-
tions no longer add up to the risk. Arguably, this is not mandatory.
However, if they do not, it calls into question what the sensitivities
mean and why there is a “bit left over”. Whatever the academic
discussion (which, in view of the large amount of literature on Euler
allocations, is likely to rumble on for some time), there is little doubt
that those given the task of managing firm-wide risk like the contri-
butions to add up to the whole and that a method that does not satisfy
this condition is unlikely to gain general acceptance. None of the
simple constructions above does.
The literature on factor risk contributions is still very small, and
none of it follows the “Euler route”. Cherny and Madan (2006)
propose that the factor contribution to a risk measure be defined as
that risk measure applied to the conditional expectation of the port-
folio on the factor, and Bonti et al (2006) take what might be
described as a cut-down version of this by looking at the impact of
factor-driven stress scenarios. Subsequently, Cherny, Douady and
Molchanov (2009) propose regressing the portfolio return on
167
168
Individual
issuer
contributions
Financials
Industrials Utilities
Telecoms
Note: Sector factors are around the middle, unsystematic parts are around the edge
Individual
issuer
contributions
Rates
GDP EL Forex
Unemployment
Note: Central bullseye is the expected loss, contributions from factors are shown
in the middle ring, unsystematic parts are around the edge
169
m ⎡ 1 ⎤
∑ E ⎢⎣(V − V ) ∫ (∂ f ) (λ (V − V ) + V ) dλ ⎥⎦
°
k k
°
k
° °
k=1 0
⎡ ⎤
= E ⎣ f (V ) Y > y ⎦ − µY (10.4)
170
which decomposes the systematic part of the ES, less the EL (which
it must always exceed), as a sum of factor components.
We can justify that 10.2 is a direct generalisation of the Euler
formula by simply observing that it agrees for any p-homogeneous
function (with p > 0)4:
1 1
∂ ∂
xk ∫ ∂(λ x ) [ f (λ x)] dλ = x ∫ λ k
−1
∂xk
[ f (λ x)] dλ
0 k 0
1
∂ 1
= ∫λ p−1
dλ x k
∂xk
[ f ( x )] = x k ∂k f
p
0
with d[u] denoting the volume element in u space. The kth contribu-
tion of this function is therefore:
m
⎛ 1 ⎞ exp ( iu ⋅ x) − 1
⎜ ⎟
⎝ 2 π ⎠
∫ux k k
u⋅ x
∫ f ( y ) exp (−iu⋅ y ) d [ y ] d [u]
R m
Rm
1
= xk ∫ (∂ f ) (λ x) dλ
k
0
171
⎛ Ψ ( c ⋅V ) − Ψ ( c ⋅V ° ) ⎞
hk (V ) = E° ⎜⎜(Vk − Vk° ) ⎟
⎟ (10.5)
⎝ c ⋅V − c ⋅V ° ⎠
Note that the expected payout of each instrument is zero (that is,
E[hk(V)] = 0, clear from anti-symmetry under the interchange V ↔
V°), and indeed this effect is anticipated in the discussion above: the
factors are there to describe the variation about the EL. It is straight-
forward to check that the position risk contribution of the kth
instrument, that is, the conditional expectation of ckhk(V) in 10.5 on
the loss exceeding the VAR, is simply the factor risk contribution
10.4. The last part of the expression for hk(V) is the gradient of the
chord joining (c⋅V°, Ψ(c⋅V°)) to (c⋅V, Ψ(c⋅V)) and therefore is a “discrete
derivative” of Ψ, and the (Vk – V°k ) term gives the kth component of
the variation. The expectation over V° takes into account the proba-
bility-weighted variation of Ψ between the arbitrary “reference
point” V° and V. The instruments together therefore replicate the
variation of the conditional EL around the EL and each one describes
in a reasonably natural way how much variation is due to the kth
component Vk.
We can now demonstrate 10.4 for two different well-known models.
172
Y = µY + c ⋅V +U, V ~ N ( 0, Σ)
−1
(
c k ( Σc ) k φ Φ ( P )
+
)
+
σY P
with P+ = P[Y > y]. Note that this decomposition is therefore essen-
tially the same as the logical decomposition of the standard
deviation, as the systematic part of the standard deviation is c′Σc/σY,
and the multiplier as the right-hand term of the above expression is
dependent only on the choice of tail probability. The kth contribu-
tion therefore vanishes if the factor is uncorrelated with the
portfolio, as expected.
The equivalence with the standard deviation measure in the
multivariate normal model was obtained for the decomposition of
ES into systematic and unsystematic parts (Martin and Tasche,
2007), so the fact that it carries over to this subdivision of the
systematic part is unsurprising.
with pj|V shorthand for the conditional default probability, that is:
⎛ Φ−1 ( p ) − c ⋅V ⎞
j j
p j V = Φ ⎜⎜ ⎟
⎟
⎝ 1− c ʹ′
j Σc j ⎠
173
1 2 3 4 5 R²
N ⎧⎪ V − V °
∑a c j jk E° ⎨ k k
⎪⎩ c j ⋅ (V − V )
°
j=1
174
1 2 3 4 5
Once this has been done, one has at each loss level the probability of
exceeding it, and the factor contributions. The VAR, ES and contri-
butions for any level can then be found by interpolation.
We showed that in the multivariate normal case, all shortfall
measures are essentially equivalent in that the tail probability has
no effect on the decomposition. In general, and indeed for the
probit model, this no longer holds. To demonstrate this, we take
the grouped portfolio described in Tables 10.1 and 10.2. By
grouping, we mean that we are assuming that each group in the
portfolio consists of many thousands of similar, individual
175
1
2
3
4
5
6
7
8
9
10
Figure 10.4 Shortfall allocation into factors for two percentiles (95%
and 99.5%), numerically (top) and pictorially (bottom)
1
2
3
4
5
Note: The central part of the chart is the expected loss, which of course is the
same in the two percentiles. In the higher percentile, there is a significantly higher
contribution from factor 1
176
single line item. The factor weight vectors (cj) are shown and the
“R-squared”, or proportion of variation explained by the factors,
is c′j Σcj (not simply |cj|2, as the factors are not orthogonal), for each j
from one to 10. The portfolio EL is exactly 3. The portfolio consists
of groups that are linked primarily to one factor each and a collec-
tion of other groups that are linked more generally. Group 1 is a
large exposure to a low-probability default event linked strongly
to factor 1.
The shortfall is calculated at two different tail probabilities,
0.5% and 5%, by Monte Carlo simulation (the grouping allows this
to be made very fast).9 The results are shown in Figures 10.3 and
10.4. Figure 10.3 shows the usual decomposition of shortfall into
constituents,10 and Figure 10.4 shows the decomposition into
factors using the methods shown here. Looking at the portfolio
model, it is reasonably clear that group 1 is a “tail event”, that is,
an unlikely but severe event. At a higher percentile (further into
the tail), we therefore expect it to show a larger risk contribution,
which is seen from Figure 10.3; this much is standard. The new
177
Figure 10.5 Proportion of risk in each factor in the base case (solid
lines) and rebalanced portfolio (dashed lines)
60
GDP
50
Factor contribution (%)
Unempl.
Rates
40
CPI
Housing
30
GDP*
20 Unempl.*
Rates*
10 CPI*
Housing*
0
0 2 4 6 8 10 12 14
Expected shortfall minus expected loss
178
Figure 10.6 Loss distributions for base case and rebalanced portfolio
1
Base case
Improved
Tail probability
0.1
0.01
0.001
0 2 4 6 8 10 12 14
Expected shortfall minus expected loss
179
Conclusions
We have demonstrated how to decompose the systematic part of ES
among its dependent risk factors in arbitrary models for which the
simple Euler formula no longer holds. The decomposition is done
using a direct generalisation of the Euler formula (Equations 10.2
and 10.4), which reduces to the Euler formula for any function that
is homogeneous of positive degree.
In the context of the multivariate normal model, the decompo-
sition we have exhibited is identical to a simple differentiation of
the variance with respect to the risk factors. This identity is lost in
more general models: one can easily find examples in which
contributors to some percentiles are less significant at others and
vice versa.
1 A function f is p-homogeneous if f(θ x) = θ pf(x) for all θ > 0. The Euler formula is Skxk∂k f = pf.
2 For a logit model, one replaces Φ(...) by the function 1/(1 + e–x) and corrects the Φ–1(pj) term
appropriately, so the same remarks will apply to that model too.
3 In compound expressions, we distinguish ∇ and ∂k (differentiate with respect to the kth
argument), which operate on the function, from (∂/∂xk), which operates on a whole expres-
sion containing x. For example, if f(x) = sin(2x1 + 3x2) then (∂1 f)(4x1, 5x2) = 2cos(8x1 + 15x2) and
(∇f )(4x1, 5x2) = (23)cos(8x1 + 15x2), but (∂/∂x1)[f(4x1, 5x2)] = 8cos(8x1 + 15x2). Putting brackets
around the ∇f helps to clarify this.
4 Note again the importance of distinguishing between differentiating with respect to x and
with respect to λx.
5 The argument__in the rest of this section is not used subsequently, so can be skipped.
6 As usual, i = √–1.
7 Note that cj is a vector and its kth component is written cjk.
180
8 Standard techniques such as importance sampling can be used (see, for example,
Glasserman, 2005).
9 One million simulations were used.
10 The ES contribution of a constituent is its EL conditionally on portfolio loss exceeding the
VAR.
REFERENCES
Bank for International Settlements, 2005, “An Explanatory Note on the Basel II IRB Risk
Weight Functions,” BIS, July (available at www.bis.org/bcbs/ irbriskweight.pdf).
Bonti G., M. Kalkbrenner, C. Lotz and G. Stahl, 2006, “Credit Risk Concentrations
Under Stress,” Journal of Credit Risk, 2(3), pp 115–36.
Cherny A., R. Douady and S. Molchanov, 2009, “On Measuring Nonlinear Risk with
Scarce Observations,” Finance & Stochastics, 14(3), pp 375–95.
Glasserman P., 2005, “Importance Sampling for Portfolio Credit Risk,” Management
Science, 51(11), pp 1,643−56.
Martin R., 2009, “Shortfall: Who Contributes and How Much?” Risk, October, pp 84−89.
Martin R. and D. Tasche, 2007, “Shortfall: A Tail of Two Parts,” Risk, February, pp 84−89.
Rosen D. and D. Saunders, 2009, “Analytical Methods for Hedging Systematic Credit
Risk with Linear Factor Portfolios,” Journal of Economic Dynamics & Control, 33(1), pp
37−52.
Tasche D., 2007, “Euler Allocation: Theory and Practice” (available at www.defaultrisk.
com).
Tasche D., 2008, “Capital Allocation to Business Units and Sub-portfolios: The Euler
Principle In Pillar II,” in A. Resti (Ed), The New Basel Accord: The Challenge of Economic
Capital (London: Risk Books): pp 423–53.
181
183
then standard error estimates can report that the result is very accu-
rate, but the result can be completely wrong.
Consider the Black–Scholes model:
dS (t ) = rS (t) dt + σ S (t ) dW (t ) (11.1)
The exact solution of this stochastic differential equation is known
analytically and using an Euler scheme for log(S), we get the time
discretisation scheme:
rΔti − 21 σ 2 Δti +σΔW (ti )
S (ti+1 ) = S (ti ) e
which is the exact solution, that is, the Euler scheme has no discreti-
sation error. Hence, a Monte Carlo simulation using this scheme
bears only the Monte Carlo error, which can be assessed by a simple
standard error estimate.
Because of its functional form, Black–Scholes paths tend to zero
as T → ∞. The convergence is quicker and more easily observed
under extreme volatility scenarios (see Figure 11.1).
For fixed maturity and Brownian sample, the paths also converge
to zero as s → ∞. This may come as a surprise. Intuitively, one might
expect that, since the distribution widens, the paths should widen.
20
18
16
14
12
Value
10
8
6
4
2
0
0 1 2 3 4 5 6 7 8 9 10
Simulation time
Note: All paths share the same volatility but are generated with different
Brownian paths. The paths flatten as maturity increases. For an animation of
the effect, see Fries (2010)
184
converges to zero for s → ∞, while the correct limit for the call
option is V(0) = S(0). In addition, the standard estimate for the
Monte Carlo error:
1 2
ErrorEst := ∑ V (T, ωi ) − V (T0 )
n2 i
( )
converges to zero for s → ∞.
So, applying a stress to the volatility, then re-evaluating the
product, can result in the Monte Carlo error estimator reporting a
185
∂V ∂V 1 ∂2 V
(t,S) + rS (t,S) + σ 2S2 2 (t,S) = rV (t,S) (11.2)
∂t ∂S 2 ∂S
∂2 V
(t,S) = 0
∂S2
on ∂A. Using a space discretisation S(ti) ∈ {S0, S1, ... , Sm–1} to discretise
the PDE by an implicit Euler scheme, we find:
X B
ω1
ω3
ω2
ω4
B
t0 t1 t2 t3 t4 t5
186
~
where V(ti) := (V(ti, S0), ... , V(ti, Sm–1))T is the value vector and L is a
tri-band matrix with:
Si1σ 2 − Si rΔS
Li, i−1 = Li, i+1 =
2ΔS2
and:
Si2σ 2
Li, i = −
ΔS2
187
With this PDE scheme, the valuation of a call with strike K ∈ [Sl,
Su] will converge to:
S ( 0) − Sl S − S ( 0) S −K S −K
(Su − K ) +0 u = S ( 0) u − Sl u
Su − Sl Su − Sl Su − Sl Su − Sl
K
S ( 0) − S ( 0) (11.3)
Su
188
A := {ω X (ti , ω ) ∈ Ai∀i}
for some given sets Ai. Let Bi denote the domain X(ti, Ω)\Ai. The situ-
ation is sketched in Figure 11.2 and we refer to Fries and Kienitz
(2010) on how to construct such a Monte Carlo simulation. The
construction is similar to Fries and Joshi (2011). See also Glasserman
and Staum (2001) and Joshi and Leung (2007).
where:2
⎛ V i+1 (T ) ⎞
V i+1 (Ti ) = N (Ti ) E Q ⎜⎜ i+1
FTi ⎟⎟
⎝ N (Ti+1 ) ⎠
Furthermore, let:
Here V out,i+1 is the value of the paths ending in the outbound domain
in time Ti+1, and V in,i+1 is the value of the paths ending in the inbound
domain in time Ti+1.
189
1.1
1.0
0.9
0.8 Standard Monte Carlo
valuation (light grey)
Option value
0.7 Monte Carlo valuation
0.6 with super-hedge
boundary condition at
0.5 barrier = 6 (black)
0.4 Analytic benchmark (grey)
0.3
0.2
0.1
0.0
0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0
Volatility
inbound domain.3
Then we define the modified valuation algorithm recursively (in
~
a backward algorithm) as V in,i(Ti) = 0 on Bi and on Ai:
V in, i (Ti ) := V out, i+1 (Ti )
N (Ti )
+V in, i+1 (Ti+1 ) Q Ai+1 FTi + C (Ti ) 1 Ai
( )
N (Ti+1 )
~
given some final value V in,n(Tn).
~
The above backward induction defines V in,i(Ti) under the assump-
tion that X(Ti) is inbound. Note that C(Ti) has to be evaluated only
~
on Ai and that given V out,i+1(Ti) (or an approximation thereof), V in,i(Ti)
can be constructed from the modified (inbound) Monte Carlo
simulation.
On {X(Ti) ∈ Ai} we have (by backward induction):
⎛ V in, i (T ) ⎞ 1
E Q ⎜⎜
N ( T )
i
FTi ⎟⎟ =
N (T )
(V out, i+1 (Ti ) + V in, i+1 (Ti ) + C (Ti ))
⎝ i ⎠ i
1 V i (Ti )
=
N (Ti )
( V i+1 (Ti ) + C (Ti )) =
N (Ti )
Given that our Monte Carlo simulation starts within the bounda-
ries, the value of the product at T0 is then:
190
0.2
0.1
0.0
0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0
Volatility
(
V (T0 ) = E Q V 0 (T0 ) FT0 )
If the product has an early exercise or any other payout conditional
on its future value, that can be incorporated in C(Ti).
⎛ G (t , X (T )) ⎞
T+1 i+1
V out, i+1 (Ti ) = N (Ti ) E Q ⎜⎜ X (Ti )⎟⎟
⎝ N (Ti+1 ) ⎠
191
⎛ V out, i+1 (T ) ⎞
EQ ⎜⎜ i+1
X (Ti ) ⎟⎟
⎝ N (Ti+1 ) ⎠
= (b + axl ) DP (X (Ti ) , xl ,Ti+1 − Ti ) − aP (X (Ti ) , xl ,Ti+1 − Ti )
+ ( d + cxu ) DC ( X (Ti ) , xu ,Ti+1 − Ti ) + cC ( X (Ti ) , xl ,Ti+1 − Ti )
determined.
We propose three different variants: analytic calculation, which
is only possible for products with analytic valuation formulas and
is useful for benchmarking; sub-hedge or super-hedge, which is
product-dependent, but easy to find and gives fast, good results;
and numerical calculation, which is product-independent, similar
to PDEs, and gives good results, but may be unstable.
and:
192
where V i+1 denotes the true value of the future cashflows of the
product under consideration. Then, using Gsub in place of G as an
approximation of V out,i+1(Ti+1), we get a lower bound of the true
option price. Using Gsup in place of G as an approximation of
V out,i+1(Ti+1), we get an upper bound of the true option price.
Both methods will only give bounds for the true option prices;
however, the deviation of the value depends on the location of the
boundary. If the boundary is distant, the induced error is small.
The value returned by the Monte Carlo valuation with super-
hedge boundary conditions can be interpreted as the costs of a
corresponding hedge strategy: using dynamic hedging below a
certain barrier and converting to a safe static super-hedge (at addi-
tional costs) once a barrier has been crossed.
Numerical results
We compare a standard Monte Carlo simulation (cross-hatch), the
analytic benchmark (grey) and the Monte Carlo simulation with
boundary conditions (black) for the valuation of a European-style
call under the Black–Scholes model. While this is a simple product
(with an analytic formula), the Monte Carlo algorithm is fully
fledged using fine steps (even where European-style options require
only one time step). We check the behaviour for increasing
193
volatility. The figures show the mean (grey line) and standard error
estimate (black/light grey area).
In Figure 11.3, we see that a Monte Carlo simulation with a super-
hedge boundary condition (black) converges to the analytic value
(grey), while the standard Monte Carlo simulation breaks down
(cross-hatch). Figure 11.4 shows how the error induced by the
super-hedge assumption decreases when the distance from the
barrier increases (from three in Figure 11.3 to six in Figure 11.4).
Figure 11.5 shows a simple sub-hedge boundary condition (V(t,
S) = S – K). The error induced is around 0.1, which is still much less
than that of a corresponding PDE valuation, which would give
1.05/6 = 0.175 (compare 11.3).
Christian Fries would like to thank his co-authors Mark Joshi, Jörg
Kampen and Jörg Kienitz, as well as Peter Landgraf and his colleagues
at DZ Bank. This chapter expresses the views of its authors and does
not represent the opinion of DZ Bank, which is not responsible for
any use that may be made of its contents.
1 In December 2008, volatility was high: 10-year volatility on a 30-year swap rate was
observed to move from 20% to 40%.
2 Here N denotes the numeraire and Q the corresponding equivalent martingale measure.
3 We will derive an approximation to Vout,i+1(Ti) later.
194
REFERENCES
Fries C. and J. Kampen, 2006, “Proxy Simulation Schemes for Generic Robust Monte
Carlo Sensitivities, Process Oriented Importance Sampling and High Accuracy Drift
Approximation,” Journal of Computational Finance, 10(2), pp 97–128 (available at www.
christian-fries.de/finmath/proxyscheme).
Fries C. and J. Kienitz, 2010, “Monte Carlo Simulation with Boundary Conditions,”
(available at www.christian-fries.de/ finmath/montecarloboundaryconditions).
Joshi M. and T. Leung, 2007, “Using Monte Carlo Simulation and Importance Sampling
to Rapidly Obtain Jump-diffusion Prices of Continuous Barrier Options,” Journal of
Computational Finance, 10(4), pp 93–105.
195
197
(CDFs) and their inverses. As a result, the CMA can generate scenarios
for many more copulas than the few parametric families used in the
traditional approach. For instance, it includes large-dimensional,
downside-only panic copulas that can be coupled with, say, extreme
value theory marginals for portfolio stress testing.
An additional benefit of the CMA is that it does not assume that
all the scenarios have equal probabilities.
Finally, the CMA is computationally very efficient even in large
markets, as can be verified in the code available for download.
In Table 12.1, we summarise the main differences between the CMA
and the traditional approach to apply the theory of copulas in practice.
We proceed as follows. First, we review the basics of copula theory.
Then we discuss the traditional approaches to copula implementa-
tion. Next, we introduce the CMA in full generality. Then we present a
first application: we create a panic copula for stress testing that hits
the downside non-symmetrically and is probability-adjusted for risk
premium. Finally, we present a second application of the CMA,
namely how to perform arbitrary transformations of copulas.
Documented code for the general algorithm and for the applications
of the CMA is available at http://symmys.com/node/335.
Table 12.1 Main differences between CMA and the traditional approach to
copulas implementation
198
We call the first step “separation”. This step separates the distribu-
tion FX into the pure “individual” information contained in each
variable Xn, that is, the marginals FX , and the pure “joint” informa-
n
tion of all the entries of X, that is, the copula FU. The copula is the
joint distribution of the grades, that is, the random variables U ≡ (U1,
... , UN) defined by feeding the original variables Xn into their respec-
tive marginal CDF:
copula F U_into a new joint distribution F X for X_. To do so, for each
marginal F X we first compute the inverse CDF F −1
n X , or quantile, and n
then we apply the inverse CDF to the respective grade from the
copula:
FX1 ,..., FX N ⎫
⎪ X1
U1 ⎪⎪
C: ⎬ ( ) ~ FX
( ) ~ FU ⎪
XN
⎪
UN ⎪⎭ (12.5)
199
analytically. Then we draw J joint Monte Carlo scenarios {x1,j, ... , xN,j}
j=1,...,J
from F θX. Next, we compute the marginal CDFs F θX from their n
The grade scenarios {u1,j, ... , uN,j}j=1,...,J now represent simulations from
the parametric copula F θU of X.
To illustrate the traditional implementation of the separation, F θX
can be normal, and the scenarios xn,j can be simulated by twisting N
independent standard normal draws by the Cholesky decomposition
of the covariance and adding the expectations. The marginals of the
joint normal distribution are normal, and the normal CDFs F θX are n
⎪⎪ u1, j
S: {( )} ~ FXθ ⎨
⎪ {( )} ~ FU
θ
xN , j ⎪
⎪⎩ uN , j (12.7)
where for brevity we dropped the subscript j = 1, ... , J from the curly
brackets, displaying only the generic jth joint N-dimensional
scenario.
In the traditional implementation of the combination _ step 12.5,
we first generate scenarios from the desired copula _ F U, typically
θ
200
_
quadratures the inverse CDFs F Xθ . Then we feed as in 12.4 each
–1
_ n
201
grid of scenario pairs {xn,j, un,j} (see Figure 12.1). This function is the
CDF of the generic nth variable:
202
cv U1
u1,3 u3
p3 u1,4
u4
cv I{x ,u } p4
1,j 1,j u1,1 u1
p1 u1,2 u2
p2
u2,1 u2,2 u2,3 u2,4 U2
x1,2 x1,1 x1,4 x1,3
X2 p4
x4 x2,4
0
x3 x2,3 p3
x2,2 p2
x2
x2,1 p1 I{x2,j,u2,j}
x1
cv
X1
FXn ( x ) ≡ I{x }(
x ) , n = 1,..., N (12.11)
n , j , un , j
⎧ I ,..., I{x , u }
⎪ {x1, ju1, j } N ,j N ,j
x1, j ⎪
⎪ u1, j
SCMA : {( ) ; pj} ~ FX ⎨
⎪ {( ) ; p j } ~ FU
xN , j ⎪
⎪ uN , j (12.12)
⎩
Notice that the CMA avoids the parametric CDFs F θX that appear in n
12.6.
Let us now address _ the combination step C. The two inputs are an
arbitrary copula F U and _ arbitrary marginal distributions, repre- _
sented by the CDFs F X . For the copula, we take any copula F U
n
scenario for the copula u n,j and maps it into the desired combined
203
_
scenarios x n,j by interpolation/extrapolation of the copula scenarios
_
u n,j on the grid:
⎫
FX1 ,..., FX N ⎪
⎪ x1, j
u1, j ⎪
CCMA : ⎬ {( ) ; pj} ~ FX
{( ) ; pj} ~ FU ⎪
xN , j
⎪
uN , j ⎪ (12.14)
⎭
204
The variable B ≡ (B1, ... , BN) are panic triggers. More precisely, B
selects the panic downside endogenously as in Merton (1974):
⎧⎪
1 if X n( ) < Φ−1 (b )
p
Bn ≡ ⎨ (12.18)
⎩⎪ 0 otherwise
205
{ p } ≡ arg min ∑ p ln ( Jp )
j j j j
{p j }
such that ∑ p j xn, j ≥ 0, ∑p j ≡ 1, p j ≥ 0 (12.19)
j j
206
u1 x1
_
fU Normal fX
1
1 distribution
: panic
Equally weighted portfolio return PDF
207
FU −−T→ FU
↓C ↑S
T
FZ → FZ (12.24)
208
fU fU~
2 2
~
U ~ fU U ~ fU~
u1
fU fU~
1 1
: panic
209
Conclusion
We have introduced the CMA, a technique to generate new flexible
copulas for risk management and portfolio management. The CMA
generates flexible copulas and glues them with arbitrary marginals
using the scenarios-probabilities representation of a distribution.
The CMA generates many more copulas than the few parametric
families used in traditional copula implementations. For instance,
with the CMA we can generate large-dimensional, downside-only
panic copulas. The CMA also allows us to perform arbitrary trans-
formations of copulas, despite the fact that copulas are only defined
on the unit cube. Finally, unlike in traditional approaches to copula
implementation, the probabilities of the scenarios are not assumed
to be equal. Therefore, the CMA allows us to leverage techniques
such as importance sampling and entropy pooling.
REFERENCES
Brigo D., A. Pallavicini and R. Torresetti, 2010, Credit Models and the Crisis: A Journey into
CDOs, Copulas, Correlations and Dynamic Models (Chichester, England: Wiley).
Cherubini U., E. Luciano and W. Vecchiato, 2004, Copula Methods in Finance (Hoboken,
NJ: Wiley).
Daul S., E. De Giorgi, F. Lindskog and A. McNeil, 2003, “The Grouped t-Copula
with an Application to Credit Risk,” working paper (available at http://ssrn.com/
abstract=1358956).
Durrleman V., A. Nikeghbali and T. Roncalli, 2000, “Which Copula is the Right One?
working paper.
Embrechts P., C. Klueppelberg and T. Mikosch, 1997, Modelling Extremal Events for
Insurance and Finance (New York, NY: Springer).
Embrechts P., F. Lindskog and A. McNeil, 2003, “Modelling Dependence with Copulas
and Applications to Risk Management,” in Handbook of Heavy Tailed Distributions in
Finance (Amsterdam, Holland: Elsevier).
Glasserman P., 2004, Monte Carlo Methods in Financial Engineering (New York, NY:
Springer).
Jaworski P., F. Durante, W. Haerdle and T. Rychlik, 2010, Copula Theory and its
Applications (Heidelberg, Germany: Springer).
210
Li D., 2000, “On Default Correlation: A Copula Function Approach,” Journal of Fixed
Income, 9, pp 43–54.
Merton R., 1974, “On the Pricing of Corporate Debt: The Risk Structure of Interest Rates,”
Journal of Finance, 29, pp 449–70.
Meucci A., 2008, “Fully Flexible Views: Theory and Practice,” Risk, October, pp 97–102
(article and code available at http://symmys.com/node/158).
Meucci A., 2009, “Simulations with Exact Means and Covariances,” Risk, July, pp 89–91
(article and code available at http://symmys.com/node/162).
Meucci A., 2009, Risk and Asset Allocation (New York, NY: Springer), available at at
http://symmys.com/attilio-meucci/book.
Meucci A., 2010, “Historical Scenarios with Fully Flexible Probabilities,” GARP Risk
Professional – The Quant Classroom, December, pp 40–43 (article and code available at
http:// symmys.com/node/150).
Meucci A., 2011, “A Short, Comprehensive, Practical Guide to Copulas,” GARP Risk
Professional, October, pp 40–43 (article and code available at
http://symmys.com/node/351).
Meucci A., Y. Gan, A. Lazanas and B. Phelps, 2007, A Portfolio Managers Guide to Lehman
Brothers Tail Risk Model (Lehman Brothers Publications).
211
213
Hybrid VAR
Suppose the calculation of a historical simulation VaR is to be
undertaken. The P&L of the portfolio for each historical date is
calculated by applying a set of observed changes xi in the market
variables to a function P&L(x1, x2, ...). The term market variable here
applies to any input variable that is required in order to price the
positions in the portfolio, and could include such things as an
interest or foreign exchange rate, the price of a particular equity, the
credit spread of a specific corporate bond issue, etc. For our
purposes, it will be assumed that the P&L function for the portfolio
as a whole can be decomposed into the sum of P&L functions in the
individual market variable changes:
214
215
−∞
change
0
P&L
distribution
–5
–200 0 200
Market variable change
216
cj =
(u j+1 − µ ) P&L ( u j ) − ( u j − µ ) P&L ( u j+1 )
u j+1 − u j
The linear function in the first and last intervals are extrapolated to
–∞ and +∞ respectively.
The required integrals can be calculated by recursively applying
the identities:
b b
217
n ⎛ ⎞
(P&L1 + P&L 2 ) = ∑⎜ n ⎟ P&Ln1 P&Ln−r
n
2
r=0 ⎝ r ⎠
218
where, as before, P&Li(xi) is the P&L function associated with the ith
market variable. As previously, if the P&Li(xi) are all represented as
piecewise continuous functions by interpolating between values on
a discrete grid u1, u2, ... , uj, ... then the range of integration must be
broken up into adjacent intervals within which the functional form
of the term in square brackets remains fixed. The boundaries of the
intervals lie at the points x^ = (uj – m i)/s i for all i, j and the identities
13.2 can then be used to calculate the raw moments. These are
converted to cumulants that are added to those of the rest of the
portfolio, ensuring that there is zero correlation between the resid-
uals of the group and those of all other positions.
Computation of VaR
VaR by historical simulation is calculated by treating the calculated
historical P&L values as random draws from an underlying histor-
ical P&L distribution. Figure 13.2 illustrates the effect of taking
account of the uncertainties, the σ ’s, in daily observations due to
missing data. The dots on the horizontal axis represent P&L values
for the portfolio of interest obtained by applying a set of market
variable changes observed in the same historical window. The
spikes rising from the dots are appropriately weighted stylised
Dirac delta functions, indicating the probability density function
associated with each P&L value when no uncertainty, that is, σ = 0,
is attributed to the daily observations. The effect of associating an
uncertainty with the daily observations is shown by the distribu-
tions plotted in black, with one distribution corresponding to each
219
Probability density
where the weight, wk, gives the probability of drawing the kth
historical observation. Pr(p) is the estimated probability density
220
∑w k =1
k=1
µ nʹ′ = ∫p n
Pr ( p ) dp = ∑ wk ⋅ ∫p n
Prk ( p ) dp
−∞ k=1 −∞
221
Probability density
–6.0 –4.8 –3.6 –2.4 –1.2 0 1.2 2.4 3.6 4.8 6.0
P&L (US$ million)
q= ∫ Pr ( p) dp = ∑ w ⋅ ∫ Pr ( p) dp
k k (13.5)
−∞ k −∞
setting q = 0.01 and q = 0.05 respectively for the 99% and 95% VaR. In
the case where the historical observations are free from uncertainty
and equally weighted, plugging Equation 13.4 into Equation 13.5
yields the VaR to be the (N × q)th worst loss experienced by the port-
folio in the historical period and is commonly used as a conservative
estimate of VaR in historical simulation.
In the case where the Prk(x) are sufficiently close to Gaussian, the
integrals on the right-hand side are just cumulative normal
222
κ̂
= Φ (ν̂ ) − φ ( x̂ ) ∑ n He n−1 (ν̂ )
n=3 n!
∫ Pr ( x) dx = ∫ Pr ( x̂ ) dx̂
k k k k
−∞ −∞
∫ Pr ( x̂) dx̂
−∞
⎛⎧ κ̂ ⎫ ⎧ κ̂ 2 κ̂ ⎫
= Φ (ν̂ ) − φ (ν̂ ) ⎜⎨ 3 He 2 ( v̂ )⎬ + ⎨ 3 He5 (ν̂ ) + 4 He 3 (ν̂ )⎬
⎝ ⎩ 6 ⎭ ⎩ 72 24 ⎭
⎧ κ̂ 3
κ̂ κ̂ κ̂ ⎫⎞
+ ⎨ 3 He 8 (ν̂ ) + 3 4 He6 (ν̂ ) + 5 He 4 (ν̂ )⎬⎟
⎩ 1, 296 144 120 ⎭⎠
223
224
∑ w ⋅ ∫ x Pr ( x) dx
k k
ETL = k=1
N
−∞
ν
∑ w ⋅ ∫ Pr ( x) dx
k k
i=1 −∞
⎧ κ̂ 2 κ̂ ⎫
+ ⎨ 3 ( He6 (ν̂ ) + 6He 4 (ν̂ )) + 4 ( He 4 (ν̂ ) + 4He2 (ν̂ ))⎬
⎩ 72 24 ⎭
⎧ κ̂ 33
+ ⎨ (He9 (ν̂ ) + 9He7 (ν̂ ))
⎩ 1, 296
κ̂ 3κ̂ 4 κ̂ ⎫⎞
+
144
( He7 (ν̂ ) + 7He5 (ν̂ )) + 5 ( He 5 (ν̂ ) + 5He3 (ν̂ ))⎬⎟
120 ⎭⎠
and:
ν ν̂ ν̂
∫ x Pr ( x) dx = µ ∫ Pr ( x̂) dx̂ + σ ∫ x̂ Pr ( x̂) dx̂
k k k k k
−∞ −∞ −∞
225
⎛ x ⎞
He n ( x ) = 2 −n/2 H n ⎜ ⎟
⎝ 2 ⎠
n d nφ ( x )
(−1) = Hen ( x ) ⋅ φ ( x )
dx n
From the above results, the following derivative and identities are easily
obtained:
∫ He n ( x) φ ( x) dx = −Hen−1 (ν ) φ (ν )
−∞
ν
∫ x He
2
n ( x) φ ( x) dx = −{Hen+1 (ν ) + (2n+ 1) Hen−1 (ν ) + n ( n − 1) Hen−3 (ν )} φ (ν )
−∞
226
REFERENCES
Charlier C., 1905, “Über das Fehlergesetz Arkiv för Matematik,” Astronomi och Fysik, 2(8),
pp 1–9.
Cramér H., 1925, “On Some Classes of Series Used in Mathematical Statistics,”
Proceedings of the Sixth Scandinavian Congress of Mathematicians, Copenhagen.
Dempster A., N. Laird and D. Rubin, 1977, “Maximum Likelihood from Incomplete
Data via the EM Algorithm,” Journal of the Royal Statistical Society Series B (Methodological),
39(1), pp 1–38.
Glasserman P., 2004, Monte Carlo Methods in Financial Engineering (New York, NY:
Springer).
Holton G., 2003, Value-at-risk: Theory and Practice (Amsterdam, Holland: Academic Press).
Schafer J., 1997, Analysis of Incomplete Multivariate Data (Boca Raton, Fl: Chapman &
Hall/CRC).
227
229
evident for positions with leverage, that is, when assets are
purchased with borrowed money. As a leveraged position is sold,
the price tends to drop due to market impact. As it is gradually
unwound, the depression in prices due to impact overwhelms the
decrease in position size, and leverage can initially rise rather than
fall. As more of the position is sold, provided the initial leverage
and initial position are not too large, it will eventually come back
down and the position retains some value. However, if the initial
leverage and position are too large, the leverage diverges during
unwinding, and the resulting liquidation value is less than zero,
that is, the debt to the creditors exceeds the resale value of the asset.
The upshot is that, under mark-to-market accounting, a leveraged
position that appears to be worth billions of dollars may predict-
ably be worth less than nothing by the time it is liquidated. Under
firesale conditions or in very illiquid markets, things are even
worse.
From the point of view of a risk manager or regulator, this makes
it clear that an alternative to mark-to-market accounting is badly
needed. Neglecting impact allows huge positions in illiquid instru-
ments to appear profitable when this is not the case. We propose
such an alternative based on the known functional form of market
impact, and that valuation should be based on expected liquidation
value. While mark-to-market valuation only indicates problems
with excessive leverage after they have occurred, this makes them
clear before positions are entered into. At the macro level, this could
be extremely useful for damping the leverage cycle and coping with
pro-cyclical behaviour (Thurner, Geanakoplos and Farmer, 2012,
and Geanakoplos, 2010). An extended discussion of our proposal
that treats extensions to the problem of risky execution can be found
in Caccioli, Bouchaud and Farmer (2012).
230
Q
I (Q) = Yσ (14.1)
V
231
15
10
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
x
Note: Possible deleveraging trajectories, showing the leverage λ(x) based on
mark-to-market accounting as a function of the fraction x of the position that has
been liquidated. We hold the initial leverage λ0 = 9 constant and show four
trajectories for different values of the market impact parameter I = I(Q) = Yσ√Q/V,
that is I = 0 (black dashed line, corresponding to the no-impact case), I = 0.1
(dotted line), 0.15 (black line) and 0.19 (grey dotted-dashed line). If the market
impact is too high, the leverage diverges before the position can be liquidated,
implying that the position is bankrupt
232
⎛ 2 ⎞
p = p0 ⎜1− I (Q )⎟ (14.2)
⎝ 3 ⎠
Qp
λ= (14.3)
Qp − L
233
10
0
0 Q/2 Q Q/2 0
q
30
25
20 Impact-adjusted
Leverage
No-impact
15 Mark-to-market
10
0
0 Q/2 qc Q Q/2 0
q
Note: The x axes denote number of shares held. The position q(t) varies from zero
to Q in the left half of each panel and from Q to zero in the right. The black dashed
line shows the leverage without price; the grey line shows the leverage including
impact under mark-to-market accounting; and the grey dot dashed line shows the
leverage using impact-adjusted valuation. The upper panel is a case in which Q is
small enough that the leverage never becomes critical; the lower panel is a case
where the leverage becomes supercritical. In this case, the impact-adjusted
leverage diverges as the position is entered, warning the manager of the impending
disaster. The vertical grey, dashed line shows the critical position qc.
234
235
0.05
I (Q ) ≈ 2% × ≈ 6% (14.5)
0.005
3 3
λc I = → λc = V /Q (14.6)
2 2Yσ
236
Note: Except as otherwise noted, numbers are based on data for first-quarter 2008
* Impact I1 = I(Q) based on volatility and volume, calculated with Equation 14.1,
with Y = 1 and Q = V for futures and Q = 10V for stocks, roughly 5% of the market
capitalisation
** For futures, we refer to the nearest maturity; the numbers for the 10-year US note
are very similar to those for the Bund
*** Large cap US stocks, Q = 10V
**** Krispy Kreme Doughnuts, a small cap stock, March 2012, with Q = 10V in
US$m
***** ClubMed, a small cap French stock, Q = 10V with V in €m
months to buy a 5.5% share of IBM. The results are given in Table
14.1.
We see that for liquid futures, such as the Bund or S&P 500, the
critical leverage is large enough that the phenomenon we discuss
here is unlikely to ever occur. As soon as we enter the world of
equities, however, the situation looks quite different. For over-the-
counter markets, the effect is certainly very real. Using reasonable
estimates, we find that the impact of deleveraging a position can
easily reach 20% on these markets, corresponding to a critical
leverage lc ≈ 7.5.
Conclusion
Positions need to be based on liquidation prices rather than mark-
to-market prices. For small, unleveraged positions in liquid markets
there is no problem, but as soon as any of these conditions are
violated, the problem can become severe. As we have shown,
standard valuations, which do nothing to take impact into account,
can be wildly overoptimistic.
Impact-adjusted accounting gives a more realistic value by esti-
mating liquidation prices based on recent advances in understanding
market impact. If one believes – as we do – that Equation 14.1 is a
237
238
REFERENCES
Acerbi C. and G. Scandolo, 2008, “Liquidity Risk Theory and Coherent Measures of
Risk,” Quantitative Finance, 8(7), pp 681–92.
Bouchaud J.-P., 2010, “The Endogenous Dynamics of Markets: Price Impact, Feedback
Loops and Instabilities,” in A. Berd (Ed), Lessons From The Financial Crisis (London,
England: Risk Books).
Bouchaud J.-P., D. Farmer and F. Lillo, 2009, “How Markets Slowly Digest Changes
in Supply and Demand,” in T. Hens and K. Schenk-Hoppe (Eds), Handbook of Financial
Markets: Dynamics and Evolution (Amsterdam, Holland: Elsevier).
Caccioli F., J.-P. Bouchaud and D. Farmer, 2012, “A Proposal for Impact Adjusted
Valuation: Critical Leverage and Execution Risk” (available at http://ArXiv:1204.0922).
239
Caccioli F., S. Still, M. Marsili and I. Kondor, 2011, “Optimal Liquidation Strategies
Regularize Portfolio Selection, “ European Journal of Finance, special issue.
Farmer D., A. Gerig, F. Lillo and H. Waelbroeck, 2011, “How Efficiency Shapes Market
Impact,” technical reprint (available at http://arxiv.org/abs/1102.5457).
Gatheral J., 2010, “No-dynamic-arbitrage and Market Impact. Quantitative Finance, 10, pp
749–59.
Geanakoplos J., 2010, “Solving the Present Crisis and Managing the Leverage Cycle,”
FRBNY Economic Policy Review, pp 101–31.
Joulin A., A. Lefevre, D. Grunberg and J.-P. Bouchaud, 2008, “Stock Price Jumps: News
and Volume Play a Minor Role,” Wilmott Magazine, 46, September–October, pp 1–7.
Kyle A., 1985, “Continuous Auctions and Insider Trading,” Econometrica, 53, pp 1,315–35.
Thurner S., G. Geanakoplos and D. Farmer, 2012, “Leverage Causes Fat Tails and
Clustered Volatility,” Quantitative Finance, 12(5), pp 695–707.
Weber P. and B. Rosenow, 2005, “Order Book Approach to Price Impact,” Quantitative
Finance, 5, pp 357–64.
240
243
❑❑ V(s, T) > 0. In this case, since the netted trades are in the institu-
tion’s favour (positive present value), it will close out the position
but retrieve only a recovery value, V(s, T)dC, with dC a percentage
recovery fraction.
❑❑ V(s, T) ≤ 0. In this case, since the netted trades are valued against
the institution, it is still obliged to settle the outstanding amount
(it does not gain from the counterparty defaulting).
We can therefore write the payout3 in default as dCV(tC, T)+ + V(tC, T)–
where tC is the default time of the counterparty. The risky value of a
244
The first term in the expectation is simply the risk-free value condi-
tional upon no default before the final maturity. The second
component 1t ≤TV(t, tC) corresponds to the cashflows paid up to4 the
C
(V (t, τ C ) + δCV (τ C ,T )
+
+ V (τ C ,T ) − V (τ C ,T ) ⎤⎦
+
)
= Et ⎡⎣1τ C >T V (t,T ) + 1τ C ≤T V (t,T )
(
+ 1τ C ≤T δCV (τ C ,T ) − V (τ C ,T ) ⎤⎦
+ +
)
= V (t,T ) − Et ⎡⎣1τ C ≤T (1− δC ) V (τ C ,T ) ⎤⎦
+ (15.2)
This allows us to express the risky value as the risk-free value less
an additional component. This component is often referred to (see,
for example, Pykhtin and Zhu, 2006) as the credit value adjustment
(CVA). As first discussed by Sorensen and Bollier (1994), an analogy
is often made that the counterparty is long a series of options. Let us
denote the standard CVA in this unilateral case as:
CVAunilateral = Et ⎡⎣1τ C ≤T (1− δC ) V (τ ,T ) ⎤⎦
+
(15.3)
245
⎦
(15.4)
246
with t 1 = min(tC, tI, t). The final term corresponds to the fact that in
the event of joint default, the value of the derivatives position is
essentially cancelled, with a recovery value paid to whichever party
is owed money. It can be seen that an overall positive (negative)
CVA will increase (decrease) with increasing joint default
probability.6
We will make the common assumption that the default times and
247
CVAbilateral ≈
m
Example
We now present a simple example8 assuming that the counterparty
and institution default probabilities (conditional on no joint default)
are correlated according to a Gaussian copula. The correlation
parameter is denoted by r . Following the Gaussian correlation
assumption between tC and tI and the independence of t, the above
probabilities can be readily calculated, for example:
Q (τ C ∈ [ti−1 ,ti ] , τ I > ti , τ > ti )
= Q (τ C > ti−1 , τ I > ti , τ > ti ) − Q (τ C > ti , τ I > ti , τ > ti )
⎡N ( N −1 (Q (τ > t )) , N −1 (Q (τ > t )) ; ρ )⎤
2d C i−1 I i
= ⎢ ⎥Q (τ > t )
i
⎢−N ( N (Q (τ > t )) , N (Q (τ > t )) ; ρ ) ⎥
−1 −1 (15.7)
⎣ 2d C i I i ⎦
where N(.) and N2d(.) represent the univariate and bivariate cumula-
tive normal distribution functions.
We assume that the probabilities of default are determined by:
248
Figure 15.1 Expected exposure profiles for case A and case B with
μ = –1%, σ = 10% and μ = 1%, σ = 10%, respectively
20 EPE (case A)
15 ENE (case A)
EPE (case B)
10 ENE (case B)
Exposure (%)
5
0
–5
–10
–15
–20
0 2 4 6 8 10
Time (years)
V ( s,T ) = µ ( s − t ) + σ s − tZ
Case A Case B
249
0.5
–0.5
CVA (%)
–1.0
–1.5
Unilateral
–2.0 Unilateral adjusted
Bilateral
–2.5
0 20 40 60 80 100
Correlation (%)
2.5
2.0
1.5
CVA (%)
1.0
0.5 Unilateral
Unilateral adjusted
Bilateral
0
0 20 40 60 80 100
Correlation (%)
( ) (
Et ⎡⎣V ( s,T ) ⎤⎦ = µΔxN µ Δx / σ + σ Δxϕ µ Δx / σ
+
) (15.9a)
( ) (
Et ⎡⎣V ( s,T ) ⎤⎦ = −µΔxN µ Δx / σ − σ Δxϕ µ Δx / σ
−
) (15.9b)
Δx = s − t
250
Figure 15.3 CVA as a function of the systemic spread intensity with zero
correlation for case A (top) and case B (bottom)
1.0
0.5
0
CVA (%)
–0.5
–1.0
–1.5
Unilateral
–2.0 Unilateral adjusted
Bilateral
–2.5
0 0.5 1.0 1.5 2.0
Joint default intensity (λ)
2.5
2.0
1.5
CVA (%)
1.0
0.5 Unilateral
Unilateral adjusted
Bilateral
0
0 0.5 1.0 1.5 2.0
Joint default intensity (λ)
The (symmetric) exposure for profiles EPE and ENE are shown in
Figure 15.1.
We will consider three distinct CVA measures outlined below:
251
Since this article was originally published there has been a discus-
sion on “closeout conventions” in relation to the above components
and the reader is referred to Brigo and Morini (2011) and Gregory
and German (2013) for further reading. Initially we assume zero
correlation and zero joint default probability, r = l = 0, and show the
three CVA values in Table 15.1.
Case A represents a situation where the bilateral CVA is negative
due to the institution’s higher default probability and the high chance
that they will owe money on the contract (negative exposure due to m
= –1%). Case B is the opposite case and, since the counterparty is more
risky than the institution, the bilateral CVA is reduced by only around
one third compared with the unilateral case. We see that, since case A
and case B represent equal and opposite scenarios for each party, the
sum of the bilateral adjustments is zero.
Now we show the impact of correlation on the CVA. As shown in
Figure 15.2, correlation can have a reasonably significant impact on
both the unilateral and bilateral values. As correlation increases, we
approach comonotonicity, where the more risky credit is sure to
default first. This means that, in case A, the unilateral adjusted CVA
goes to zero (the institution is sure to default first), while in case B it
converges to the pure unilateral value (the counterparty is sure to
default first).
Let us finally consider the impact of joint default in Figure 15.3,
which illustrates the three CVA components versus the joint default
intensity, l ≤ min(lC, lI). We see that joint default plays a similar role
to that of correlation but does not have a significant impact on the
bilateral CVA. This illustrates, importantly, that even with high
joint default probability (systemic component), a substantial
portion of the bilateral benefit comes from the idiosyncratic compo-
nent, a point that is particularly acute in case A.
252
Bilateral or unilateral?
An obvious implication of the bilateral formula is that the overall
CVA may be negative, that is, actually increase the overall value of
the derivatives position(s). Another result of the above symmetry is
that the overall amount of counterparty risk in the market would be
zero.13 While this symmetry or the bilateral risk might seem reason-
able and clean, let us consider the associated hedging issues. While
the default component of the unilateral CVA is often hedged by
buying CDS protection on the counterparty, the additional term in
the bilateral formula would require an institution to sell CDS protec-
tion on themselves (or trading their credit quality in some other way
such as by shorting their own stock). Even using the “adjusted
unilateral” CVA is debatable on hedging grounds since the relevant
hedging instruments do not exist (for example, an institution buying
CDS protection that cancels if they themselves default).
Since hedging arguments do not support the use of a bilateral
CVA, let us consider the ways in which the bilateral reduction to the
CVA could be monetarised.
253
makes money when its credit spread widens (and vice versa).
Our view is that this is problematic, especially since the calcula-
tions earlier showed that the bilateral CVA was not strongly
sensitive to the joint default – an illustration that the idiosyn-
cratic component of the spread constitutes the significant
proportion of the bilateral CVA. We also point out that institu-
tions wishing to sell protection on credits highly correlated with
their own creditworthiness will lead to an increase in the overall
amount of (wrong-way) counterparty risk in the market.
❑❑ As a funding benefit. Since this article was first published in 2008,
it seems that DVA has been increasingly seen as a funding benefit.
Since there is a potential double-counting between DVA and
funding benefits, our view is that DVA should be associated with
“own default” and not as a funding benefit. There has been a
significant amount of discussion in this area since the original
publication of this article and the reader is referred to Gregory
(2012) for a more detailed discussion on DVA and funding and
also Hull and White (2012) and Burgard and Kaer (2011) for more
theoretical discussion around DVA and funding value adjust-
ment (FVA).
Conclusion
We have presented an overview of bilateral counterparty risk
pricing. Using a model that represents a simple extension of
standard counterparty risk pricing approaches, we have illustrated
pricing behaviour and considered the impact of default of both
254
1 Since the publication of the original Risk article, accounting for two-sided counterparty
risk has become mandatory under IFRS13
2 We note that since exposures within netted portfolios are linear then this case is suitably
general.
255
REFERENCES
Arvanitis, A. and J. Gregory, 2001, Credit: The Complete Guide to Pricing, Hedging and Risk
Management (London, England: Risk Books).
Brigo, D., and M. Morini, 2011, “Closeout convention tensions”, Risk, December, pp
86–90.
Burgard, C., and M. Kjaer, 2011, “In the balance”, Risk, October.
Canabarro E., E. Picoult and T. Wilde, 2003, “Analysing Counterparty Risk,” Risk,
September, pp 117–22.
256
Duffie D. and M. Huang, 1996, “Swap Rates and Credit Quality,” Journal of Finance, 6, pp
379–406.
Gregory, J., 2012, “Counterparty credit risk: a continuing challenge for global financial
markets”, Wiley.
Gregory, J., and I. German, 2013, “Closing out DVA”, Risk, January.
Hull, J., and A. White. “CVA, DVA, FVA and the Black-Scholes-Merton Arguments”,
Working paper, September 2012.
Pykhtin M., 2005, Counterparty Credit Risk Modelling (London, England: Risk Books).
Pykhtin M. and S. Zhu, 2006, “Measuring Counterparty Credit Risk for Trading Products
Under Basel II,” in M. Ong (Ed), Basel II Handbook (London, England: Risk Books).
Sorensen E. and T. Bollier, 1994, “Pricing Swap Default Risk,” Financial Analysts Journal,
50, pp 23–33.
United States Tax Court, 2003, “Bank One Corporation, Petitioner, v. Commissioner of
Internal Revenue,” respondent, May 2.
257
259
⎡I (τ c ≤ T ) D ( 0, τ c ) ⎤
VCVA = E ⎢ ⎥
( (
⎢⎣×LGD (τ c ) NPV (τ c ) − C R (τ c− ) ))
⎥⎦ (16.1)
260
NO
⎡
VCVA ∑ E ⎢I (Ti−1 < τ c ≤ Ti ) D ( 0,Ti )
i=1
⎣
)) ⎤⎦⎥
+
(
×LGD (Ti ) NPV (Ti ) − C R (Ti− )( (16.2)
V = E Q [ P ( R, X )] (16.3)
where:
NR
P (Ti , R (Ti ) , X (Ti )) = ∑ Pi ( X (Ti ) ;r ) δr , R(Ti ) (16.5)
r=0
261
~
and X = (X(T1), ... , X(TN ))t are distributed; Pi(⋅; r) is a rating-dependent
0
V ∑
N MC iMC =1
P ( R [ iMC ] , X [ iMC ]) (16.6)
___
with standard error S/√NMC, where S2 = EQ[P(R, X)2] – EQ[P (R, X)]2 is
the variance of the sampled payout.
In the following, we will make minimal assumptions about the
particular model used to describe the dynamics of the market
factors. In particular, we will only assume that for a given MC
sample the value at time Ti of the market factors can be obtained
from their value at time Ti–1 by means of a mapping of the form X(Ti)
= Fi(X(Ti–1), ZX) where ZX is a NX dimensional vector of correlated
standard normal random variates, X(T0) is today’s value of the
market state vector, and Fi is a mapping regular enough for the
pathwise derivatives method to be applicable (Glasserman, 2004),
as is generally the case for practical applications.
As an example of a counterparty rating model generally used in
practice, here we consider the rating transition Markov chain model
of Jarrow, Lando and Turnbull (1997) in which the rating at time Ti
can be simulated as:
NR
R (Ti ) = ∑ I Z iR > Q (Ti , r )
( ) (16.7)
r=1
~
where ZRi is a standard normal variate, and Q(Ti, r) is the quantile-
threshold corresponding to the transition probability from today’s
rating to a rating r at time Ti. Note that the discussion below is not
limited to this particular model, and it could be applied with minor
modifications to other commonly used models describing the
default time of the counterparty and its rating (Schönbucher, 2003).
Here we consider the rating transition model 16.7 for its practical
utility, as well as for the challenges it poses in the application of the
pathwise derivatives method, because of the discreteness of its state
space.
262
∂V (θ ) ∂
= E [ P ( R, X )] (16.8)
∂θ k ∂θ k
∂V (θ ) ⎡ ∂P ( R, X ) ⎤
= E P ⎢ ⎥ (16.9)
∂θ k ⎣ ∂θ k ⎦
263
X → →U → V → → Y (16.12)
264
with i = 1, ... , n.
In the adjoint mode, the cost does not increase with the number
of inputs, but it is linear in the number of (linear combinations of
the) rows of the Jacobian that need to be evaluated independently.
In particular, if the full Jacobian is required, one needs
_ to repeat
the adjoint calculation m times, setting the vector Y equal to each
of the elements of the canonical basis in Rm. Furthermore, since the
partial (branch) derivatives depend on the values of the interme-
diate variables, one generally first has to compute the original
calculation storing the values of all the intermediate variables
such as U and V, before performing the adjoint mode sensitivity
calculation.
One particularly important theoretical result (Griewank, 2000) is
that given a computer program performing some high-level func-
tion 16.11, the execution time of its adjoint counterpart:
X = FUNCTION _ b ( X,Y ) (16.17)
Cost [ FUNCTION _ b]
≤ ωA (16.18)
Cost [ FUNCTION ]
with w A ∈ [3, 4]. Thus, one can obtain the sensitivity of a single
output, or of a linear combination of outputs, to an unlimited
265
_
A few comments are in order. In step 4 , the adjoint of the payout
function is defined while keeping the discrete rating variable
266
_ _
constant. This provides the derivatives _ X l(Ti) = ∂Pθ /∂Xl(Ti), and θ k =
∂Pθ /∂θ k. In defining the adjoint in step 2, we have taken into account
that the propagation rule in step 2 is explicitly dependent on both
X(Ti) and the model parameters θ. As _ a result,
_ its adjoint counter-
part produces contributions to both θ and X(Ti). Both the adjoint of
the payout and of the propagation mapping can be implemented
following the principles of AAD as discussed in Capriotti (2011)
and Capriotti and Giles (2011). In many situations, AD tools can
also be used as an aid or to automate the implementation, espe-
cially for simpler,
_ self-contained
_ functions. In the backward sweep
above, steps 1 and 3 have been skipped because we have assumed
for simplicity of exposition that the parameters θ do not affect the
correlation matrices r i and_ the rating dynamics. If correlation risk is
instead required, step_ 2 also produces the adjoint of the random
variables ZX, and step 1 contains the adjoint of the Cholesky decom-
position, possibly with the support of the binning technique, as
described in Capriotti and Giles (2010).
267
_
where Z* is such that (Z* + Sji –1 ZR)/√i = Q(Ti, r; θ ), and φ (ZXi, r Xi ) is a
=1 j
NX-dimensional standard normal probability density function with
correlation matrix r Xi obtained by removing the first row and
column of r i; here ∂θ Q(Ti, r; θ ) is not stochastic and can be evaluated
k
(for example, using AAD) once per simulation. The final result is
rather intuitive as it is given by the probability-weighted sum of the
discontinuities in the payout.
Results
As a numerical test, we present here results for the calculation of
risk on the CVA of a portfolio of commodity derivatives. For the
purpose of this illustration, we consider a simple one-factor
lognormal model for the futures curve of the form:
dFT (t)
= σ T exp (−β (T − t )) dWt (16.22)
FT (t )
reads:
Ne
j=1
(
NPV (t ) = ∑ D (t,t j ) Ft j (t) − K ) (16.24)
268
Figure 16.1 Speed-up in the calculation of risk for the CVA of a portfolio
of five commodity swaps over a five-year horizon, as a function of the
number of risks calculated (empty dots)
160
°
140
120 °
Speed-up/RCPU
100
80
60 °
40
20
0 •° • • •
0 100 200 300 400 500 600
Nrisks
Note: The full dots are the ratio of the CPU time required for the calculation of
the CVA, and its sensitivities, and the CPU time spent for the calculation of the
CVA alone. Lines are guides for the eye
ters θ . _
For this example, the adjoint propagation rule in step 2 simply
reads:
⎛ 1 ⎞
FT (Ti − 1) + = FT (Ti ) exp ⎜σ i ΔTi Z − σ i2 ΔTi ⎟
⎝ 2 ⎠
σ i = FT (Ti ) F (Ti ) ( ΔTi Z − σ i ΔTi )
_
with s i related to this step’s contribution to the adjoint of the
_
future’s volatility s T by:
σi
σT+ = e −2 βT ( e 2 βTi − e 2 βTi−1 )
2 βΔTi
_ _
At the end of the backward path, F T(0) and s T contain the pathwise
derivatives estimator 16.10 corresponding, respectively, to the
sensitivity with respect to today’s price and volatility of the futures
contract with expiry T.
The remarkable computational efficiency of the AAD implemen-
tation is illustrated in Figure 16.1. Here, we plot the speed-up
269
Table 16.1 Variance reduction (VR) on the sensitivities with respect to the
thresholds Q(1, r) (NR = 3) for a call option with a rating-dependent strike
0.1 24 16 12
0.01 245 165 125
0.001 2,490 1,640 1,350
270
Conclusion
In conclusion, we have shown how AAD allows an extremely effi-
cient calculation of counterparty credit risk valuations in MC. The
scope of this technique is clearly not limited to this important appli-
cation but extends to any valuation performed with MC. For any
number of underlying assets or names in a portfolio, the proposed
method allows the calculation of the complete risk at a computa-
tional cost that is at most four times the cost of calculating the profit
and loss of the portfolio. This results in remarkable computational
savings with respect to standard finite-difference approaches. In
fact, AAD allows one to perform in minutes risk runs that would
take otherwise several hours or could not even be performed over-
night without large parallel computers. AAD therefore makes
possible real-time risk management in MC, allowing investment
firms to hedge their positions more effectively, actively manage
their capital allocation, reduce their infrastructure costs and ulti-
mately attract more business.
271
1 The discussion below applies also to the case in which the payout at time Ti depends on the
history of the market factors X up to time Ti.
2 Here we have used the fact that the payout 16.5 depends on the outturn value of the rating
at time Ti and not on its history.
3 For simplicity of notation, we exclude the case in which θ includes the elements of the
correlation matrix r in φ (Z R, Z X; r ). The extension to this case is straightforward and can be
performed along the lines of Capriotti and Giles (2010).
_
4 Here and in the following we will use the standard AD notation θ k to indicate the sensi-
tivity of the payout with respect to the model parameter θ k.
REFERENCES
Brigo D. and A. Capponi, 2010, “Bilateral Counterparty Risk with Application to CDSs,”
Risk, March, pp 85–90.
Giles M. and P. Glasserman, 2006, “Smoking Adjoints: Fast Monte Carlo Greeks,” Risk,
January, pp 92–96.
Glasserman P., 2004, Monte Carlo Methods in Financial Engineering (New York, NY:
Springer).
Jarrow R., D. Lando and S. Turnbull, 1997, “A Markov Model for the Term Structure of
Credit Spreads,” Review of Financial Studies, 10, pp 481–523.
272
Joshi M. and D. Kainth, 2004, “Rapid Computation of Prices and Deltas of n-th to
Default Swaps in the Li Model,” Quantitative Finance, 4, pp 266–75.
Kallenberg O., 1997, Foundations of Modern Probability (New York, NY: Springer).
Schönbucher P., 2003, Credit Derivatives Pricing Models: Models, Pricing, Implementation
(London, England: Wiley).
273
275
276
where we have flipped the signs of V(t) (the portfolio value from the
counterparty’s perspective is –V(t)) and C(t) (collateral available to
the counterparty is –C(t)). Note that both Ec(t) and Eb(t) for future
time points t are uncertain because neither future portfolio MTM
value nor the amount of collateral the bank will hold in the future
are known at present.
277
where LGDQb is the market-implied LGD for the bank, PDQb(s|t) is the
risk-neutral cumulative probability of the bank’s default between
time t and time s ≥ t, estimated at time t, and EEb* (s|t) is the discounted
risk-neutral expected exposure of the counterparty to the bank at
time s calculated at time t, conditional on the bank defaulting at
time s, given by:
⎡ B ( 0) ⎤
EEb* ( s t ) = EtQ ⎢ Eb ( s) τ b = s⎥ (17.6)
⎢⎣ B ( s) ⎥⎦
where t b is the time of default of the bank. Note that in practice the
bank would often refer to the unilateral CVA calculated from the
counterparty’s perspective as debit valuation adjustment (DVA).
However, neither the bank nor the counterparty is default-risk-
free. If they value counterparty risk for their portfolio unilaterally,
they would never agree on the price, as one would demand a posi-
tive risk premium from the other. The bank and the counterparty
would agree on the price only if they both price counterparty risk
278
where CVAcb(t) is the bilateral CVA at time t from the bank’s perspec-
tive. However, Equation 17.7 is not quite accurate because it ignores
the order in which the bank and the counterparty default.
It is not difficult to account for the default order in calculation of
the bilateral CVA. There are two types of possible default scenario:4
❑❑ Counterparty defaults before the bank does (that is, tc < t b). Under
these scenarios, the loss for the bank is equal to the bank’s
exposure to the counterparty at the time of the counterparty’s
default less the amount the bank is able to recover: LGDQc⋅Ec(tc).
❑❑ Counterparty defaults after the bank does (that is, tc > t b). Under
these scenarios, the counterparty experiences a loss equal to the
counterparty’s exposure to the bank at the time of the bank’s
default less the amount the counterparty is able to recover:
LGDQb⋅Eb(t b). However, from the bank’s perspective, the counter-
party’s loss is the bank’s gain (or negative loss) resulting from the
bank’s option to default.
279
where CF(0, H) is all the trading book cashflows the bank receives
between time zero and time H, discounted to time zero (the bank’s
payments result in negative contributions to CF). These cashflows
may be deterministic or stochastic and include coupon and divi-
dend payments, other periodic payments (for example, swaps),
payments in the event of default, payments at trades’ maturity, and
exercising of options.
Equation 17.11 (or its equivalents) has been used by financial
institutions for many years to quantify trading book losses and
calculate VaR. Until recently, these market risk calculations were
performed without taking into account CCR, as all trade values
were calculated counterparty-risk-free. However, CCR is an
inherent part of a trading book and must be accounted for in
Equation 17.11 by adjusting trade values for CCR and including
cashflows arising from the counterparties’ defaults.
Let us consider trading book losses of the bank on counterparty i.
If the counterparty does not default prior to the horizon H, we only
need to adjust the portfolio value at the beginning and at the end of
the time interval for CCR in Equation 17.11 to arrive at:
280
Li ( H τ i > H )
= [Vi ( 0) − CVAib ( 0)] − [Vi ( H ) − CVAib ( H )] − CFi ( 0, H ) (17.12)
Equation 17.13 is valid if the bank does not take any action after the
settlement with the defaulting counterparty. However, this would
rarely be the case because banks try to maintain a market-neutral
position, and removal of the trades with the defaulting counter-
party from the bank’s trading book exposes the bank to unhedged
market risk. To restore the market-neutral position, the bank would
have to replace the trades it had with the defaulting counterparty
by booking equivalent trades with another counterparty. It is typi-
cally assumed that the bank pays the amount Vi(t i) (if negative, the
bank receives the money) to replace the portfolio.6 Thus, the bank
will have the same portfolio at the horizon as in the case of no
default, and Equation 17.13 transforms to:
Li ( H τ i ≤ H )
= [Vi ( 0) − CVAib ( 0)] − Vi ( H ) + LGDi ⋅ Ei* (τ i ) − CFi ( 0, H ) (17.14)
281
(
LCCR ( H ) = ∑ 1{τ i ≤H} ⋅ ⎡⎣LGDi ⋅ Ei* (τ i ) − CVAib ( 0)⎤⎦
i
282
283
( ) ( )
K i ( H ) = PDi H z1−q ⋅ ⎡⎣LGDi ( z1−q ) ⋅ EADi H z1−q − CVAib ( 0)⎤⎦
( )
+⎡⎣1− PDi H z1−q ⎤⎦ ⋅ ⎡⎣E ⎡⎣CVAib ( H ) Z = z1−q , τ i > H ⎤⎦ − CVAib ( 0)⎤⎦ (17.18)
1− r 2
i 1−q
⎟
⎟ (17.19)
⎝ i ⎠
284
285
The first term in Equation 17.18 is the capital charge covering losses
due to the counterparty’s default, while the second term is the
capital charge covering losses occurring in the event of the counter-
party’s survival. These terms correspond to “default risk” and
“credit migration risk” terms in standard portfolio credit risk
models, but there are important differences:
❑❑ In the default risk term, the default loss is reduced by the amount
of the time zero CVA (since the CVA is bilateral, it can be nega-
tive, so the loss would increase in this case). In loan portfolio
models, no such term is present.
❑❑ In the credit migration risk term, the capital charge covers the
potential increase of the CVA over the time horizon, which can be
caused by changes in the counterparty’s credit spread, in the
bank’s credit spread and in the risk-neutral EE profile (see
Equation 17.8). In loan portfolio models, this capital charge
covers the change in loan value due to potential deterioration of
the obligor’s credit quality.
❑❑ While EL is usually subtracted from credit VaR in loan portfolio
models, no such subtraction occurs in Equation 17.18. The moti-
vation for subtracting EL is that losses up to EL are covered by a
bank’s reserves, funded by the income from the bank’s assets.
However, CCR can be described in terms of virtual trades (see
the previous section) that produce no income that could fund the
reserve. Therefore, subtracting EL from credit VaR is not
appropriate.
286
( )
K i ( H ) = EADi ⋅ LGDiDT ⋅ ⎡⎣PDi H z1−q − PDi ( H )⎤⎦ ⋅ MA ( Mi ) (17.22)
287
If these two issues were resolved, the Basel II framework for CCR
would be very reasonable for banks that do not actively manage or
hedge CCR. However, resolving the MA issue is not trivial because
it would involve calculation of E[CVAib(H)|Z = z1–q, t i > H].
( )
⋅LGDiDT ⋅ ⎡⎣PDi H z1−q − PDi ( H )⎤⎦ ⋅ MA ( Mi ) (17.23)
IMM banks that have approval for specific risk interest rate internal
VaR models that can demonstrate that their specific VaR models
incorporate credit migrations (for brevity, we will call such banks
advanced banks), may set the maturity adjustment equal to one
(that is, MA(Mi) = 1 for any Mi).
One can raise the following issues regarding Equation 17.23.
288
( )
K idef ( H ) = PDi H z1−q ⋅ ⎡⎣LGDiDT ⋅ EADi − CVAib ( 0)⎤⎦ (17.24)
289
risk VaR for CVA separately from the market risk calculations for
actual trades completely ignores dependencies between CVA
and the real trades.
❑❑ Keeping the EE profiles fixed during CVA simulations and exclusion of
market risk hedges – CVA fluctuates due to variability of both the
counterparty credit spread and the EE profile. Keeping the EE
profile fixed ignores a significant portion of CVA risk. Moreover,
banks that actively manage CCR hedge both the EE changes and
the counterparty credit spread changes. EE hedges may include
trades of many types: they offset the sensitivity of the EE to
multiple risk factors: interest rate, foreign exchange, etc.14 While
these trades actually reduce the bank’s trading book risk, they
are not eligible CVA hedges under Basel III and have to be
included in the market risk capital calculations. However, these
trades do not offset any of the real trades – they will appear as
“naked” trades in the market risk calculations, resulting in higher
market risk capital. This could create a perverse incentive for
banks: banks that hedge variability of CVA due to market risk
factors will be punished by higher market risk capital.
❑❑ Unilateral CVA is used in calculations – Bilateral CVA determines
the market value of CCR and, as such, should be used for calcu-
lating the CVA capital charge. Note that only bilateral CVA enters
Equation 17.16.
Thus, the Basel III CVA capital charge does not capture CVA risk
properly and even creates incentives for banks to leave the EE
portion of CVA risk unhedged.
Conclusion
We have proposed a general framework for calculating capital for
CCR that consistently incorporates CVA. We have considered two
applications of this framework: market risk and credit risk. Under
the market risk approach, CCR is modelled and calculated jointly
with the CCR-free trading book. This can be done by extending the
actual trading book via adding to the portfolio one virtual hybrid
defaultable contingent claim per counterparty. VaR calculated for
the extended trading book covers both market risk and CCR. This
approach is appropriate for sophisticated financial institutions that
manage and hedge their CCR dynamically. Under the credit risk
290
291
1 For a comprehensive review of CCR, see, for example, Canabarro and Duffie (2003),
Pykhtin and Zhu (2007) or Gregory (2010).
2 See Pykhtin (2010) for details.
3 The term structure of the risk-neutral PDs is obtained from the CDS spreads quoted in the
market (see, for example, Schönbucher, 2003).
4 There is also a possibility of simultaneous default (tc = tb), but we ignore this scenario as
unlikely. There have been studies that accounted for joint default possibility (see, for
example, Gregory, 2009).
5 See, for example, Pykhtin (2011) for more details.
6 Strictly speaking, the bilateral CVA at the time of default should be subtracted from the
risk-free portfolio value Vi(ti). However, this CVA is usually ignored because it is practi-
cally impossible to estimate and it is likely to be negligible. First, it is not known in advance
what counterparty the bank will use to replace the portfolio, so it is not clear what credit
spread to use in the CVA calculations. Second, the replacement counterparty is likely to be
another bank, and the replacement trades are likely to be a part of a larger netting set.
Their contribution to the netting set CVA will depend on the existing trades in that netting
set. Third, interbank OTC derivatives portfolios are usually well collateralised (with low
threshold), so that the extra exposure and CVA resulting from the replacement trades
should be small. Finally, the replacement trades are not completely equivalent to the orig-
inal trades: they have the same sensitivity to the market risk factors, but may have different
MTM value. For example, regardless of the current market value of an interest rate swap,
the replacement swap’s value will always be zero. Thus, even if a swap portfolio with the
defaulting counterparty is well in-the-money (which would lead to a high bilateral CVA),
the bilateral CVA for the replacement trades will be negligible because the EE profiles for
the bank and the replacement counterparty will be similar.
7 For a rigorous proof, see Gordy (2003).
8 The expectation is taken under the physical probability measure, not risk-neutral.
9 See, for example, Vasicek (2002).
10 While Equation 17.20 does capture wrong-way risk, it relies on the ASRF framework, thus
ignoring the systematic nature of many market risk factors. Thus, the right-hand side of
Equation 17.20 still requires a multiplier alpha, but this alpha would be more stable and
usually have smaller magnitude than alpha in Equation 17.21. ISDA-TBMA-LIBA (2003)
reported alpha calculated by four large banks for their actual portfolios to be in the 1.07–
1.10 range. For theoretical work on alpha, see Canabarro, Picoult and Wilde (2003) and
Wilde (2005).
11 The EL calculation in Equation 17.23 is not quite correct: the unconditional expected LGD
should be used in the EL calculation rather than the downturn LGD.
12 See Basel Committee on Banking Supervision (2004) for details.
13 This issue is addressed only for advanced banks by allowing them to set MA(Mi) = 1 for
any Mi.
14 See Canabarro (2010) for more details.
292
REFERENCES
BCBS, 2004. “An Explanatory Note on the Basel II IRB Risk Weight Functions,” October.
BCBS, 2010. “Basel III: A Global Regulatory Framework for More Resilient Banks and
Banking Systems,” December.
Canabarro E., 2010, “Pricing and Hedging Counterparty Risk: Lessons Relearned?” in E.
Canabarro (Ed), Counterparty Credit Risk (London, England: Risk Books).
Canabarro E., E. Picault and T. Wilde, 2003, “Analysing Counterparty Risk,” Risk,
September, pp 117–22.
Gordy M., 2003, “A Risk-factor Model Foundation for Ratings-based Bank Capital
Rules,” Journal of Financial Intermediation, 12(3), July, pp 199–232.
Gregory J., 2009,” Being Two-faced Over Counterparty Credit Risk.” Risk, February, pp
86–90.
Gregory J., 2010. Counterparty Credit Risk: The New Challenge for Global Financial Markets
(Hoboken, NJ: Wiley).
Li D., 2000, “On Default Correlation: A Copula Approach,” Journal of Fixed Income, 9, pp
43–54.
Pykhtin M. and S. Zhu, 2007, “A Guide to Modeling Counterparty Credit Risk,” GARP
Risk Review, July/August, pp 16–22.
Schönbucher P, 2003, Credit Derivatives Pricing Models: Models, Pricing, and Implementation
(Chichester, England: Wiley).
Wilde T, 2005, “Analytic Methods for Portfolio Counterparty Risk,” in M. Pykhtin (Ed),
Counterparty Credit Risk Modelling (London, England: Risk Books).
293
295
296
MAIN RESULTS
We consider a derivative contract VÈ on an asset S between a seller B
and a counterparty C that may both default. The asset S is not
affected by a default of either B or C, and is assumed to follow a
Markov process with generator At. Similarly, we let V denote the
same derivative between two parties that cannot default. At default
of either the counterparty or the seller, the value of the derivative to
the seller VÈ is determined with a mark-to-market rule M, which
may equal VÈ or V (throughout this chapter we use the convention
that positive derivative values correspond to seller assets and coun-
terparty liabilities).
297
r Risk-free rate
rB Yield on recoveryless bond of seller B
r C Yield on recoveryless bond of
counterparty C
lB lB ≡ rB – r
lC lC ≡ rC – r
rF Seller funding rate for borrowed cash rF = r if derivative can be used as
on seller's derivatives replication cash collateral; rF = r + (1 – RB)lB if
account derivative cannot be used as
collateral
sF sF ≡ rF – r
RB Recovery on derivate mark-to-
market value in case seller B
defaults
R C Recovery on derivate mark-to-
market value in case counterparty
C defaults
298
T
U (t,S) = − (1− RB ) ∫ λBDr+λB +λC E t ⎡⎣V − ( u,S ( u))⎤⎦
t
T
− (1− RC ) ∫ t λBDr+λB +λC E t ⎡⎣V + ( u,S ( u))⎤⎦
T
− ∫ sF ( u)Dr+λB +λC E t ⎡⎣V + ( u,S ( u))⎤⎦ (18.3)
t
where:
Dk (t, u) ≡ exp (∫ u
t
k (υ )dυ )
is the discount factor between times t and u using rate k. If SF = O,
then U is identical to the regular bilateral C VA derived in many of
the papers cited in the first section.
Another important result of the chapter is the justification on
which the seller’s own credit risk can be taken into account. In the
hedging strategy considered, this risk is hedged out by the seller
buying back its own bond. It is shown that the cash needed for
doing so is generated through the replication strategy.
where W(t) is aWiener process, where r(t) > 0, rB(t) > = 0, rC(t) > 0,
s (t) > 0 are deterministic functions of t, and where JB and JC are two
independent point processes that jump from zero to one on default
299
300
Gregory (2009), Li and Tang (2007) and the vast majority of papers
on the valuation of counterparty risk use M(t, S) = V(t, S).
As in the usual Black–Scholes framework, we hedge the deriva-
tive with a self-financing portfolio that covers all the underlying
risk factors of the model. In our case, the portfolio Π that the seller
sets up consists of d (t) units of S, a B(t) units of PB, a C(t) units of PC
and b (t) units of cash, such that the portfolio value at time I hedges
out the value of the derivative contract to the seller, ie, VÈ(t) + Π(t) =
0. Thus:
❑❑ dbSì (t): the share position provides a dividend income of d (t))g (t)
S(t) dt and a financing cost of –d (t)qS(t)S(t) dt, so dbSì (t) = d (t)(gS(t)
– qS(t))S(t) dt. Here the value of qS(t) depends on the risk-free rate
r(t) and the repo rate of S(t).
301
❑❑ dbFì (t): from the above analysis, any surplus cash held by the
seller after the own bonds have been purchased must earn the
risk-free rate r(t) in order not to introduce any further credit risk
for the seller. If borrowing money, the seller needs to pay the rate
rF(t). For this rate, we distinguish two cases. Where the derivative
itself can be used as collateral for the required funding, we
assume no haircut and set rF(t) = r(t). If, however, the derivative
cannot be used as collateral, we set the funding rate to the yield
of the unsecured seller bond with recovery RB, ie, rF(t) = r(t) + (1
– RB)lB. In practice, the latter case is often the more realistic one.
Keeping rF general for now, we have:
{ ( ) } dt
+ −
)
dβ F (t ) = r (t ) −V̂ − α B PB + rF (t ) −V̂ − α B PB (
−
( ) (
= r (t) −V̂ − α B PB dt + sF −V̂ − α B PB dt ) (18.7)
where the funding spread sF ≡ rF – r, ie, sF = 0 if the derivative can
be used as collateral, and sF = (1 — RB)lB if it cannot.
❑❑ dbCì (t): by the arguments above, the seller will short the counter-
party bond through a repurchase agreement and incur financing
costs of dbCì (t) = –a C(t)r(t)PC(t) dt if we assume a zero haircut.
For the remainder of this chapter we will drop the t from our nota-
tion, Where applicable, to improve clarity. From the above analysis
it follows that the change in the cash account is given by:
{( ) } dt − rα P dt
−
) (
dβ = δ (γ S − qS ) Sdt + r −V̂ − α B PB + sF −V̂ − α B PB C C
(3.5)
so 18.6 becomes:
{( ) − α rP − δ (q − γ ) S} dt
−
) (
+ r −V̂ − α B PB + sF −V̂ − α B PB C C S S
= {−rV̂ + s (−V̂ − α P ) + (γ − q ) δS
−
F B B S S
+ ( rB − r ) α B PB + ( rC − r ) α C PC } dt
−α B PBdJ B − α C PC dJC + δ dS (18.9)
On the other hand, by Ito’s lemma for jump diffusions and our
assumption that a simultaneous jump is a zero probability event,
the derivative value moves by:
302
1
dV̂ = ∂t V̂dt + ∂S V̂dS + σ 2 S2 ∂S2 V̂dt + ΔV̂BdJ B + ΔV̂C dJ C (18.10)
2
where:
δ = −δSV̂ (18.13)
ΔV̂B
αB =
PB
V̂ − ( M + + RB M − )
=− (18.14)
PB
ΔV̂C
αC =
PC
V̂ − ( M − + RC M + )
=− (18.15)
PC
1
AtV ≡ σ 2S2 ∂S2 V + ( qS − γ S ) S∂S V (18.17)
2
303
The first, third and fourth terms are related to counterparty risk,
whereas the second term represents the funding cost. From this
interpretation, it follows that the PDE for a so-called extinguisher
trade, whereby it is agreed that no party gets anything at default, is
obtained by removing terms three and four from the PDE 18.19.
In the subsequent sections We will examine the PDE 18.19 in the
following four cases:
304
αB = −
(1− RB ) V̂ − (18.22)
PB
αC = −
(1− RC ) V̂ + (18.23)
PC
305
( T
V̂ (t ) = −exp − ∫ t rB ( s) ds ) (18.27)
( T
V̂ (t) = −exp − ∫ t {r ( s) + (1− RB ) λB ( s)}ds ) (18.29)
306
In this case, if the seller can use the derivative (ie, the loan asset) as
collateral for the funding of its short cash position within its replica-
tion strategy, then (neglecting haircuts) we have rF = r, the risk-free
rate. The net result in this case is then:
∂t V̂ = rCV̂ (18.31)
(
V̂ (t ) = exp − ∫ rC ( s) ds
t
T
) (18.32)
(
V̂ (t ) = exp − ∫
T
t
{r ( s) + (1− RC ) λC ( s)} ds ) (18.33)
as expected.
where:
(
Dk (t,T ) ≡ exp − ∫ t k ( s)ds
T
)
is the discount factor over [t, T] given the rate k.
Alternatively, if, for VÈ ≤ 0, we insert the ansatz3 VÈ = V + U0 into
18.34, apply the Feynman–Kac theorem and finally use that V(t, s) =
Dr(t, u)Et[V(u, S(u))], then we get:
307
{∫ }
T
U (t,S) = −V (t,S) k ( u) Dk (t, u) du (18.40)
t
Comparing 18.37 and 18.40 shows that, when the seller buys an
option from the counterparty, it encounters an additional funding
spread sF = (1 – RB)l B.
∂t V̂ + AtV̂ − ( r + λB + λC ) V̂ = − ( RB λB + λC ) V − ⎫
⎪⎪
− ( λB + RC λC ) V + + sFV + ⎬
⎪
V̂ (T,S) = H (S) ⎪⎭ (18.41)
The PDE 18.41 is linear and has a source term on the right-hand
side. If we write VÈ = V + U, then the hedge ratios become:
U + (1− RB ) V −
αB = − (18.42)
PB
308
U + (1− RC ) V +
αC = − (18.43)
PC
Comparing 18.23 and 18.42 shows that, in the latter case, a default
triggers a windfall cashflow of U that needs to be taken into account
in the hedging strategy.
Writing VÈ = U + V also gives us the following linear PDE for U:
⎪
U (T,S) = 0 ⎪⎭ (18.44)
T
U (t,S) = − (1− RB ) ∫ λB ( u) Dr+λB +λC (t, u) E t ⎡⎣V − ( u,S ( u))⎤⎦ du
t
T
− (1− RC ) ∫ λC ( u) Dr+λB +λC (t, u)E t ⎡⎣V + ( u,S ( u))⎤⎦ du
t
T
− ∫ sF ( u)Dr+λB +λC (t, u) E t ⎡⎣V + ( u,S ( u))⎤⎦ du (18.45)
t
309
The first term of 18.46 now not only contains the bilateral asset
described above, but also the funding liability arising from the fact
that the higher rate rF = (1 – RB)l B is paid when borrowing to fund
the hedging strategy’s cash account.
EXAMPLES
In this section we calculate the total derivative value VÈ for a call
option bought by the seller in the following four cases:
Case 1 : M = V̂, sF = 0
Case 2 : M = V̂, sF = (1− RB ) λB
Case 3 : M = V, sF = 0
Case 4 : M = V, sF = (1− RB ) λB
The results are shown in Figures 18.1–18.3. Since the four CVAs
above are linear in V in all four cases, we have chosen to display
their magnitude as a percentage of V. All CVAs are negative since
the seller faces counterparty risk and funding costs when sF = 0, but
does not have any bilateral asset because of the one-sidedness of
the option payout. From these results we see that the effect of the
funding cost is significantly larger than that of choosing M = VÈ or M
= V for a bought option. For a sold option, the impact of the funding
cost does not have any effect.
310
M = V, s F = 0
20 M = V, s F > 0
15
10
−5
0 1 2 3 4 5
Seller hazard rate (% per annum)
311
15
10
0
0 1 2 3 4 5
Seller hazard rate (% per annum)
312
M = V, s F = 0
20 M = V, s F > 0
15
10
0
0 1 2 3 4 5
Seller hazard rate (% per annum)
correlate them with each other and the other market factors. This
would simply imply that we would not move the discount factors
outside of the expectation operator in 18.25 and 18.45. Also, the
generator A, would incorporate terms corresponding to the new
stochastic state variables.
This chapter represents the views of the authors alone, and not the
views of Barclays Bank Plc. The authors would like to thank Tom
Hulme and Vladimir Piterbarg for useful comments and suggestions.
313
Various versions of this work have been presented and benefited from
discussions with participants at ICBI conferences in Rome (April
2009) and Paris (May 2010). This is a slightly revised version of the
paper that originally appeared in Volume 7, Issue 3 of The Journal of
Credit Risk in September 2011 (submitted November 2009).
1 Note that this growth is the growth in the cash account before rebalancing of the portfolio.
The self-financing condition ensures that, after dt, the rebalancing can happen at zero
overall cost. The original version of this chapter used the notation db, suggesting the total
change in the cash position. This notation has been corrected here. The authors are grateful
to Brigo et al (2012) for pointing this out.
2 For the first term, the cash available to the seller is (–VÈ –), of which a fraction of (1 – R B) is
invested in buying back the recoveryless bond B and the fraction R B is invested risk-free.
This is equivalent to investing the total amount (–VÈ –) into purchasing back a seller bond 13
with recovery R B.
3 We use the zero subscript to indicate that the CVA U0 has been computed with a zero
funding spread sF.
REFERENCES
Brigo, D. and F. Mercurio, 2006, Interest Rate Models: Theory and Practice (2e) (Berlin,
Germany: Springer).
Gregory, J., 2009, “Being Two-faced Over Counterparty Risk,” Risk, 22(2), pp 86–90.
Jarrow, R. and F. Yu, 2001, “Counterparty Risk and the Pricing of Defaultable Securities,”
Journal of Finance, 56(1), pp 1,765–99.
Li, B. and Y. Tang, 2007, Quantitative Analysis, Derivatives Modeling, and Trading Strategies
in the Presence of Counterparty Credit Risk for the Fixed-Income Market (Hackensack, NJ:
World Scientific).
Pykhtin, M. and S. Zhu, 2007, “A Guide to Modelling Counterparty Risk,” GARP Risk
Review, July/August, pp 16–22.
314
315
316
(
+D (t, τ C ) ( NPVC (τ C )) − RECC (−NPVC (τ C )) ⎤⎥⎦
+ +
)}
317
where REC and LGD = 1 – REC denote recoveries and loss given
defaults, D(t, T) is the discount factor between times t and T, and the
expected exposure NPVC(t) = Et[PC(t, T)] is the default risk-free value
of the residual deal at time t, seen by C. The corresponding formula
for I is the classical result from Brigo and Masetti (2005), and can be
seen by changing the lower index – which represents who is doing
the valuation – to I, and switching the order of the recovery and
positive exposure terms in the difference part, remembering that
the sign of the exposure changes.
In the situation where both I and C may default, we have a bilat-
eral valuation adjustment (see, for example, Brigo and Capponi,
2008, or Gregory, 2009, for the general framework, and Brigo,
Pallavicini and Papatheodorou, 2011, for the application to interest
rate instrument portfolios with netting and wrong-way risk). We
define t 1 to be the first-to-default time, t 1 = min(tI, tC).
Inclusion of bilateral default risk leads to the risk-free close-out
adjustment:
{{
NPVIFree (t,T ) = E t 1 τ 1 >T Π I (t,T )
} }
{{
+E t 1 τ 1 =τ
C <T
⎡Π I (t, τ C ) + D (t, τ C )
} ⎣
(REC (NPV (τ
C I C
+
)) − (−NPVI (τ C ))
+
)⎤⎦⎥}
{{
+E t 1 τ 1 =τ
I <T
⎡Π I (t, τ I ) + D (t, τ I )
} ⎣
where we use the risk-free NPV upon the first default to close the
deal, in keeping with a risk-free close-out. But, as we saw in the
introduction, this choice is not obvious in a bilateral setting because
the surviving party is not default risk-free, and even Isda docu-
mentation considers a replacement close-out taking into account
the credit quality of the surviving party. So, we consider the
substitutions:
NPVI (τ C ) → NPVII (τ C )
NPVC (τ I ) → NPVCC (τ I )
with the counterparties valuing the NPV account for the risk of
their own default, as denoted by the superscript. The final formula
318
319
NPVIRepl ( 0, K,T )
= NPVI ( 0, K,T ) ⎡⎣Q (τ C > T ) + RECC Q (τ C ≤ T )⎤⎦
NPVIFree ( 0, K,T )
= NPVI ( 0, K,T )
⎡⎣Q (τ C > T ) + Q (τ I < τ C < T ) + RECC Q (τ C ≤ min (τ I ,T ))⎤⎦
= NPVI ( 0, K,T )
⎡⎣Q (τ C > T ) + RECC Q (τ C ≤ T ) + LGDC Q (τ I < τ C < T )⎤⎦
100
80
60
%
40
20
0
100
0
90
10
80
20
70
30
60
40
50
50
40
60
70
30
20
90
10
320
100
80
60
%
40
20
0
100
0
90
10
80
20
70
30
60
40
50
50
40
60
30
20
80
10
321
Figure 19.3 Loss for the borrower at default of the lender under risk-free
close-out
τLen
0
D
–Pr(τBor > T)e–rT E
Replacement close-out F
A
U
L
T
–Pr(τBor > T)e–rT – PrLen (τBor > T)e–r(T–τLen)
τ
–Pr(τLen < τBor < T)e–rT Replacement close-out
Risk-free close-out
–e–r(T–τLen)
Risk-free close-out
factory whose only client is car producer I). In this case, the two
formulas become:
NPVIRepl ( 0, K,T )
= NPVI ( 0, K,T ) ⎡⎣Q (τ C > T ) + RECC Q (τ C ≤ T )⎤⎦
NPVIFree ( 0, K,T )
= NPVI ( 0, K,T ) ⎡⎣Q (τ C > T ) + Q (τ C < T )⎤⎦ = NPVI ( 0, K,T )
322
τLen
0
–Pr(τBor > T)e–rT D
Replacement close-out E Replacement close-out
F
A when lender and borrower
U have strong links
L
T
–Pr(τBor > T)e–rT
–Pr(τLen < τBor < T)e–rT
Risk-free close-out
–e–r(T–τLen)
Risk-free close-out
NPVIRepl (t,T )
= NPVI (t,T ) ⎡⎣Q t (τ C > T ) + RECC Q t (t < τ C ≤ T )⎤⎦
NPVIFree (t,T )
= NPVIRepl (t,T ) + NPVI (t,T ) LGDC Q t (t < τ I < τ C < T ) (19.2)
to:
323
324
P ( 0, 5y ) = 860.7 million
while taking into account the default probability of the two parties,
which are assumed to be independent, we have:
⎪
⎪⎩ 5y with prob 30%
The two formulas disagree only when the lender defaults first. Let
us analyse in detail what happens in this case. Suppose the exact
day when default happens is t I = 2.5y. Just before default, at 2.5
years less one day, we have for the borrower C the following book
value of the liability produced by the above deal, depending on the
assumed close-out:
NPVCFree (τ I − 1d, 5y ) = −578.9million
NPVCRepl (τ I − 1d, 5y ) = −562.7 million
The borrower, which has not defaulted, must pay this amount
entirely – and soon. He has a sudden loss of 348.8 million due to
default of the lender. With the substitution close-out, we have
instead:
325
Conclusion
We have analysed the effect of the assumptions about the computa-
tion of the close-out amount on the counterparty risk adjustments
of derivatives. We have compared the risk-free close-out assumed
in the earlier literature with the replacement close-out we introduce
here, inspired by Isda documentation on the subject.
We have provided a formula for bilateral counterparty risk when
a replacement close-out is used at default. We reckon that the
replacement close-out is consistent with counterparty risk adjust-
ments for standard and consolidated financial products, such as
bonds and loans. On the contrary, the risk-free close-out introduces
at time zero a dependence on the risk of default of the party with no
future obligations.
We have also shown that in case of the risk-free close-out, a party
that is a net debtor of a company will have a sudden loss at the
default of the latter, and this loss is higher the higher the debtor’s
credit spreads. This does not happen when a replacement close-out
is considered.
Thus, the risk-free close-out increases the number of operators
subject to contagion from a default, including parties that currently
seem to think they are not exposed, and this is certainly a negative
fact. On the other hand, it spreads the default losses to a higher
number of parties and reduces the classic contagion channel
affecting creditors. For the creditors, this is a positive fact because
it brings more money to the liquidators of the defaulted company.
We think the close-out issue should be considered carefully by
market operators and Isda. For example, if the risk-free close-out
introduced in the previous literature had to be recognised as a
326
Close-out ↓
Risk-free Negatively affects borrower No contagion
Substitution No contagion Further negatively affects lender
327
REFERENCES
Brigo D. and I. Bakkar, 2009, “Accurate Counterparty Risk Valuation for Energy-
commodities Swaps,” Energy Risk, March, pp 106–11.
Brigo D. and A. Capponi, 2008, “Bilateral Counterparty Risk Valuation with Stochastic
Dynamical Models and Application to Credit Default Swaps,” working paper (available
at http://arxiv. org/abs/0812.3705). A short updated version appeared in Risk, March
2010, pp 85–90.
Brigo D. and K. Chourdakis, 2009, “Counterparty Risk for Credit Default Swaps: Impact
of Spread Volatility and Default Correlation,” International Journal of Theoretical and
Applied Finance, 12(7), pp 1,007–26.
Gregory J., 2009, “Being Two-faced Over Counterparty Credit Risk,” Risk, February, pp
86–90.
Morini M., 2009, “One More Model Risk When Using Gaussian Copula for Risk
Management,” April (available at http://ssrn.com/ abstract=1520670).
Morini M., 2010, “Can the Default of a Bank Cause the Default of its Debtors? The
Destabilizing Consequences of the Standard Definition of Bilateral Counterparty Risk,”
working paper, March.
Morini M., 2011, Understanding and Managing Model Risk: A Practical Guide for Quants,
Traders and Validators (Hoboken, NJ: Wiley).
Morini M. and A. Prampolini, 2011, “Risky Funding with Counterparty and Liquidity
Charges,” Risk, March, pp 70–75.
328
329
and:
Derivation
We assume the issuer is allowed to dynamically trade d underlying
assets independent of the counterparty’s default process in a
complete market. Counterparty quantities are denoted with a
superscript C. To hedge their credit risk on the counterparty name,
the issuer can trade a default-risky bond, denoted PtC, and the
default process is modelled by a Poisson jump process, with
constant intensity lC, although this assumption can be easily
relaxed, for instance to have the intensity follow an Itô diffusion.
We consider the case of a long position in a single derivative whose
330
∂t u + Lu + λC ( u − u) − ru = 0
u = RM + − M −
331
( )
∂t V̂ + LV̂ + β V̂ + − V̂ = 0, V̂ (T, x ) = ψ ( x ) : PDE2 (20.3)
β
∂t V + LV +
1− R
(
(1− R ) Et,x [ψ ] + RE t,x [ψ ] − V = 0,
+
)
V (T, x ) = ψ ( x ) : PDE1 (20.4)
332
333
with qb Dti < 1 and q ∈ [0, 1]. This requires the calculation of the
conditional expectation Eti–1[Yti] (in practice by regression methods),
which could be quite difficult and time-consuming, especially for
multi-asset portfolios. A careful analysis of the convergence in
terms of the number of simulations, the function regressors and the
time step Dti is achieved in Gobet, Lemor and Warin (2006).
334
where Et,x[⋅] = E[⋅⎪Nt = 1, z1t = x]. Note that as NT can become infinite
when m = Sk=0 ∞
kpk > 1 (super-critical regime, see Durrett, 2004), a
sufficient condition on ψ in order to have a well-behaved product is
⎥ψ⎥ < 1. Then û solves the semi-linear PDE 20.9. This stochastic
representation can be understood as follows: mathematically, by
conditioning on t , the first time to jump of a Poisson process with
intensity b (t), we get from 20.11:
⎡ − ∫ T β (s) ds ⎤
= E t,x ⎢e t ψ ( zT1 )⎥
⎣ ⎦
T
∞
⎡ s
− ∫ β ( u) du k ⎤
+∫ ∑p E k t,x ⎢β ( s) e
t
û ( s, zs1 )⎥ ds
t
k=0 ⎣ ⎦
335
PDE2
+
PDE2
+
Note: The grey (respectively black) vertex corresponds to the weight a/p1 (respectively
b/p2). The diagram with two grey vertices has the weights (ω1 = 2, ω2 = 0)
∂t u + Lu + β ( F ( u) − u) = 0 (20.12)
336
Assumption (comparison)
To have uniqueness in the viscosity sense, we assume PDE 20.12
satisfies a comparison principle for sub- and super-solutions (see
Fleming and Soner, 1993).
For each Galton–Watson tree, we denote by wk ∈ N the number of
particles that branch into k descendants with k ∈ {0, …, M}. SMk=0wk(k
– 1) + 1 gives the total number of individuals produced by the
branching w ≡ (w 0, …, w M). The descendants are drawn with an arbi-
trary distribution pk – for example, we can take a uniform
distribution pk = 1/(M+1) (see another choice below). In Figure 20.1,
we have drawn the diagrams for the nonlinearity F(u) = (a/p2)p2u2 +
(b/p3)p3u3 up to two defaults. We then define the multiplicative
functional:
Main formula
⎡ NT M ⎛ wk ⎤
a ⎞
û (t, x ) = E t,x ⎢∏ψ ( zTi ) ∏ ⎜ k ⎟ ⎥ (20.13)
⎢⎣ i=1 k=0 ⎝ p k ⎠ ⎥⎦
We state our main result (the proof is reported in the Appendix,
available in an extended version of this chapter, Henry-Labordère,
2012).
Theorem
Assume that û ∈ L∞([0, T] × Rd) and the comparison holds. The func-
tion û(t, x) is the unique viscosity solution of 20.12.
Diagrammatic interpretation
From Feynman–Kac’s formula, we have:
( ) )
+Eτ 0 ⎡⎣F Eτ 1 ⎡⎣1τ 2 ≥T ψ ( XT )⎤⎦ 1τ 1 <T ⎤⎦ 1τ 0 <T ⎤⎥ +…
⎦
(20.15)
337
By assuming that the series 20.15 is convergent, one can guess that
the solution is given by our multiplicative functional 20.13.
Next, we focus on convergence issues and deduce a sufficient
condition to ensure that û ∈ L∞([0, T] × Rd) if ψ is bounded (the proof
is reported in the Appendix), so that our algorithm is certain to
converge.
Proposition 1
Let us assume that ψ ∈ L∞ (Rd). Set q(s) = SM ⎥a ⎥⎥⎥ψ⎥⎥∞k–1sk.
k=0 k
ak ψ ∞
k−1
(1− e − βT ( k−1)
)<1
❑❑ Case q(1) ≤ 1. û ∈ L∞ ([0, T] × Rd) for all T.
Note that our blow-up criterion – the converse of the above
inequality – does not depend on the probabilities pk as expected.
PDE 20.4
We assume that the function (1 – R)x+ + Rx can be well approxi-
mated by a polynomial F(x) (see the next section) and consider the
PDE:
β
∂t u (t, x) + Lu (t, x) +
1− R
(
F (E t,x ⎡⎣ψ ( XT )⎤⎦) − u (t, x) = 0, )
u (T, x ) = ψ ( x )
u (t, x ) = E t,x ⎡⎣1τ ≥T ψ ( XT )⎤⎦ + E t,x ⎡⎣F ( Eτ ⎡⎣ψ ( XT )⎤⎦) 1τ <T ⎤⎦
338
Optimal probabilities pk
Is there a better choice than a uniform distribution pk = 1/(M+1) for
improving the convergence?
For the PDE 20.3, the variance of the algorithm (depending on
the probabilities pk) is bounded by (see the Appendix):
⎛ a k−1
⎞
ψ ∞ P̂ ⎜T,−2 ln k − 2 ln ψ ⎟ (20.16)
⎝ p k
∞
⎠
Similarly, for the PDE 20.4, the variance (depending on the proba-
bilities pk) is bounded by:
M
ak2 2k
∑p ψ ∞
βTe − βT
k=0 k
339
> 0.
∂t v + Lv + β ( v+ − v ) = 0, v (T,⋅) ≤ 1 (20.19)
The proposition above gives that the solution does not blow up if
b T < 0.50829 (take X = ∞ with ⎥⎥ψ⎥⎥∞ = 1). Moreover, as a numerical
check of 20.17, we have calculated using a PDE solver the solution
~
of (20.20) with ψ (x) = 1, F(u) = 0.0589 + 0.5u + 0.8164u2 + 0.4043u4, b
= 0.05 and T = 10 years. The solution X is given by P̂(T, –ln (⎥ak⎥/pk))
and should satisfy (see Equation (20.17)):
X ds
∫ 1 −s + 0.0589 + 0.5u + 0.8164u2 + 0.4043u 4
= 0.5 (20.22)
340
We found X = 4.497 (PDE solver) and the reader can check that this
value satisfies the above identity 20.22 as expected.
∏ψ ( z )∏ ⎜ pk ⎟ , PDE2
i
T
i=1 k=0 ⎝ k ⎠
NT ∈[0,M ] ω1 M ωk
⎛ a 1− R + R ⎞ ⎛ a 1− R ⎞
∏ ψ ( zTi ) ⎜ 1 ( p ) ⎟ ∏ ⎜ k ( p ) ⎟
i ⎝ 1 ⎠ k≠1 ⎝ k ⎠
⎛ M ⎞
⎜ here, ∑ω k = 0 or 1⎟, PDE1
⎝ k=0 ⎠
Extensions
In the case of collateralised positions, the nonlinearity u+t should be
substituted with (ut – ut+D)+ where D is a delay. Using our poly
nomial approximation, we get F(ut – ut+D). By expanding this
function, we get monomials of the form {upt uqt+D}. Our algorithm can
then be easily extended to handle this case. At each default time t,
we produce p descendants starting at (t, Xt ) and q descendants
starting at (t + D, Xt +D).
In the case of bilateral counterparty risk, the PDE 20.1 reads:
341
90%
70%
50%
30%
10%
⎛ λ (1− R1 ) − r ⎞
∂t V̂ + LV̂ + λ2 (1− R2 ) ⎜⎜−V̂ + + 1 V̂ − V̂ ⎟⎟ = 0
⎝ λ 2( 1− R2) λ 2( 1− R2) ⎠
Proposition 2
Let us assume that F(v) and F(v) are two polynomials satisfying
(comparison), the sufficient condition in Proposition 1 for a matu-
rity T and:
F ( x ) ≤ x+ ≤ F ( x )
342
v≤v≤v
A similar result can be found for PDE 20.4. In the case of American-
style options, our algorithm gives robust lower and upper bounds.
Complexity
By approximating u+ with a high-order – say N2 – polynomial our
algorithm converges towards the brute force nested Monte Carlo
method with a complexity O(N1 × N2). By comparison, with our
choice 20.21, the complexity is at most O(4N1) for PDE type 20.4.
In comparison, the regression-based method has a complexity of
O(N1 × b) with b the number of regressors. Here, the accuracy
depends on the choice of basis functions, which may require
Note: PDE pricer (PDE 1) = 21.50; PDE pricer (PDE 2) = 21.82; non-
linearity F(u) = 1/2(u3 – u2)
Note: PDE pricer (PDE 1) = 20.39; PDE pricer (PDE 2) = 21.37; non-
linearity F(u) = 1/3(u3 – u2 – u4)
343
344
Numerical examples
Before applying our algorithm to the problem of credit valuation
adjustment, we check it on polynomials that do not belong to the
classes defined by 20.10.
Experiment 1
We have implemented our algorithm for the two PDE types:
∂t u + Lu + β ( F ( u) − u) = 0, u (T, x) = 1x>1 : PDE2
and:
( ( ) )
∂t u+ Lu + β F Et,x ⎡⎣1XT >1 ⎤⎦ − u = 0, u (T, x ) = 1x>1 : PDE1
345
Experiment 2
Same test with F(u) = 1/3(u3 – u2 – u4) (see Table 20.2).
Experiment 3: Blow-up
It is well known that the semi-linear PDE in Rd:
∂t u + Lu+ u2 = 0
Tmax ψ ∞
<1
∂t v + Lv + e (T−t) v 2 − v = 0, v (T, x ) = ψ ( x )
and this can be interpreted as a binary tree with a weight e(T–t). Our
stochastic representation then gives:
⎡ NT ∑
# branching
(T−τ i ) ⎤
u (t, x ) = e T−t E t,x ⎢∏ψ ( zTi ) e i=1 ⎥ (20.23)
⎣ i=1 ⎦
where t i is the time where the ith branching appears. This represen-
tation (20.23) appears in López-Mimbela and Wakolbinger (1998)
and was used to reproduce Sugitani’s blow-up criteria (1975).
1
2
∂t u + x 2σ BS ∂2x u + β ( u+ − u) = 0, u (T, x) = 1− 2.1x>1 : PDE1
2
346
and:
1 β
∂t u + x 2σ BS
2
2
∂2x u +
1− R (
(1− R ) E t,x ⎡⎣1− 2.1XT >1 ⎤⎦
+
)
+RE t,x ⎡⎣1− 2.1X >1 ⎤⎦ − u = 0, PDE2
T
Conclusion
CVA is now an important quantitative issue that needs to receive
special attention. Brute force nested Monte Carlo or the BSDE
approach are not a good solution for large, multi-asset portfolios.
The algorithm presented here, based on marked branching diffu-
sions, reduces the complexity by an order of magnitude, as
illustrated by our numerical examples. This method can also be
used for semi-linear PDEs with polynomial nonlinearities and
extended to fully nonlinear PDEs by including in the branching
process Malliavin weights for derivatives. We leave this investiga-
tion for future research.
347
The author would like to thank the other members of the team for
REFERENCES
their comments. He is also grateful to Jean-François Delmas and
Benth F., K. Karlsen and K. Reikvam, 2003, “A Semilinear Black and Scholes Partial
Denis Talay for useful discussions.
Differential Equation for Valuing American Options,” Finance and Stochastics, 7(3), pp
277–98.
Brigo D. and A. Pallavicini, 2008, “Counterparty Risk and CCDSs under Correlation,”
Risk, February, pp 84–88 (available at www.risk.net/1500236).
Durrett R, 2005, Probability: Theory and Examples (3e) (Belmont, CA: Thomson Brooks/
Cole).
Fahim A., N. Touzi and X. Warin, 2011, “A Probabilistic Numerical Method for Fully
Nonlinear Parabolic PDEs,” Annals of Applied Probability, 21(4), pp 1,322–64.
Fleming W. and H. Soner, 1993, Controlled Markov Processes and Viscosity Solutions (New
York, NY: Springer).
Gobet E., J.-P. Lemor and X. Warin, 2006, “Rate of Convergence of an Empirical
Regression Method for Solving Generalized Backward Stochastic Differential Equations,”
Bernoulli, 12(5), pp 889–916.
Sugitani S., 1975, “On Non-existence of Global Solutions for Some Nonlinear Integral
Equations,” Osaka Journal of Mathematics, 12, pp 45–51.
Touzi N., 2010, “Optimal Stochastic Control, Stochastic Target Problems, and Backward
SDE,” lecture notes.
348
349
Chapter 16: Luca Capriotti, Jacky Lee and Matthew Peacock, 2011, “Real-
time Counterparty Credit Risk Management in Monte Carlo,” Risk, June.
Chapter 17: Michael Pykhtin, 2011, “Counterparty Risk Capital and CVA,”
Risk, August.
Chapter 18: Christoph Burgard and Mats Kjaer, 2011, “Partial Differential
Equation Representations of Derivatives with Bilateral Counterparty Risk
and Funding Costs,” Journal of Credit Risk, 7(3).
Chapter 19: Damiano Brigo and Massimo Morini, 2011, “Close-out
Convention Tensions,” Risk, December.
Chapter 20: Pierre Henry-Labordère, 2012, “Cutting CVA’s Complexity,”
Risk, July.
350
351
352
353
354
355
356
short maturities, arbitraging SSR smile dynamics 3–22, 9, 11, 12, 19,
for 13–21 20, 21, 22
arbitraging the 95/105 one- and scaling behaviour 8–13
month skew on Eurostoxx Type II behaviour with
50 19–21 Eurostoxx 50 index 10–12
and consistency conditions Type II, in a t wo-factor model
16–17 10–12
and one-month-maturity and short maturities, arbitraging
options, application to SSR for 13–21
18–19 arbitraging the 95/105 one-
and short near-the-money month skew on Eurostoxx
options, model for 13–21 50 19–21
short-maturity limit of AT MF and consistency conditions
skew and SSR 7–8 16–17
shortfall factor contributions and o ne-month-maturity
165–81, 169, 170–1, 174, 175, options, application to
176, 177, 178, 179 18–19
and factor decomposition short near-the-money options,
formula for ES less EL model for 13–21
170–1 short-maturity limit of AT MF
and hedges, interpretation of skew and SSR 7–8
formula via 171–2 and skew stickiness ratio 4–8
multivariate normal model 173 short-maturity limit of AT MF
Vasicek (probit) model and skew and SSR 7–8
numerical examples 173–4 stochastic differential equation
skew stickiness ratio (SSR): (SDE):
arbitraging, for short maturities discrete approximations of 91
13–21 McKean 109–11, 112, 113
arbitraging the 95/105 one- stress-testing 183–94
month skew on Eurostoxx and Monte Carlo:
50 19–21 failure of 183–6
and consistency conditions and PDE’s greater robustness
16–17 186–8
and one-month-maturity Monte Carlo simulation, with
options, application to boundary conditions
18–19 188–93
short n ear-the-money options, boundary assumptions 192–3
model for 13–21 definition of boundary value
vanilla ATM skew 4–6 process 191–2
357
358