Academia.eduAcademia.edu

The Estimation of Copulas: Theory and Practice

2007, Copulas: From theory to …

INTRODUCTION Copulas are a way of formalising dependence structures of random vectors. Although they have been known about for a long time (Sklar (1959)), they have been rediscovered relatively recently in applied sciences (biostatistics, reliability, biology, etc). ...

01 02 03 2 04 05 07 08 09 The Estimation of Copulas: Theory and Practice ofs 06 10 12 13 14 15 Arthur Charpentier; Jean-David Fermanian; Olivier Scaillet1 Ensae-Crest and Katholieke Universiteit Leuven; BNP-Paribas and Crest; HEC Genève and Swiss Finance Institute pro 11 16 17 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 INTRODUCTION Copulas are a way of formalising dependence structures of random vectors. Although they have been known about for a long time (Sklar (1959)), they have been rediscovered relatively recently in applied sciences (biostatistics, reliability, biology, etc). In finance, they have become a standard tool with broad applications: multiasset pricing (especially complex credit derivatives), credit portfolio modelling, risk management, etc. For example, see Li (1999), Patton (2001) and Longin and Solnik (1995). Although the concept of copulas is well understood, it is now recognised that their empirical estimation is a harder and trickier task. Many traps and technical difficulties are present, and these are, most of the time, ignored or underestimated by practitioners. The problem is that the estimation of copulas implies usually that every marginal distribution of the underlying random vectors must be evaluated and plugged into an estimated multivariate distribution. Such a procedure produces unexpected and unusual effects with respect to the usual statistical procedures: non-standard limiting behaviours, noisy estimations, etc (eg, see the discussion in Fermanian and Scaillet, 2005). In this chapter, we focus on the practical issues practitioners are faced with, in particular concerning estimation and visualisation. Rev ised 18 1 Revised Proof Ref: 33259e September 29, 2006 COPULAS 01 02 03 04 05 In the first section, we give a general setting for the estimation of copulas. Such a framework embraces most of the available techniques. In the second section, we deal with the estimation of the copula density itself, with a particular focus on estimation near the boundaries of the unit square. 06 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 ofs 09 A GENERAL APPROACH FOR THE ESTIMATION OF COPULA FUNCTIONS Copulas involve several underlying functions: the marginal cumulative distribution functions (CDF) and a joint CDF. To estimate copula functions, the first issue consists in specifying how to estimate separately the margins and the joint law. Moreover, some of these functions can be fully known. Depending on the assumptions made, some quantities have to be estimated parametrically, or semior even non-parametrically. In the latter case, the practitioner has to choose between the usual methodology of using “empirical counterparts” and invoking smoothing methods well-known in statistics: kernels, wavelets, orthogonal polynomials, nearest neighbours, etc. Obviously, the estimation precision and the graphical results are functions of all these choices. A true known marginal can greatly improve the results under well-specification, but the reverse is true under misspecification (even under a light one). Without any valuable prior information, non-parametric estimation should be favoured, especially for marginal estimation. To illustrate this point Figure 2.1 shows the graphical behaviour of the exceeding probability function pro 08 Rev ised 07 χ : p → P(X > FX−1 (p), Y > FY−1 (p)) If the true underlying model is a multivariate Student vector (X, Y), the associated probability is the upper line. If either marginal distributions are misspecified (eg, Gaussian marginal distributions), or the dependence structure is misspecified (eg, joint Gaussian distribution), these probabilities are always underestimated, especially in the tails. Now, let us introduce our framework formally. Consider the estimation of a d-dimensional copula C, that can be written C(u) = F(F1−1 (u1 ), . . . , Fd−1 (ud )) 2 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 01 02 03 04 05 Figure 2.1 (a) The function χ when (X, Y) is a Student random vector, and when either margins or the dependence structure are misspecified. (b) The associated ratios of exceeding probability corresponding to the χ function obtained for the misspecified model versus the true χ (for the true Student model). 06 0.25 10 Misfitting dependence structure Misfitting margins Misfit margins and dependence 0.8 Student dependence structure, Student margins Gaussian dependence structure, Student margins Student dependence structure, Gaussian margins Gaussian dependence structure, Gaussian margins 09 ofs 1.0 0.30 08 (b) (a) 07 0.6 0.10 15 0.4 14 19 20 21 0.05 0.0 18 0.00 17 0.2 16 pro 13 0.15 12 0.20 11 0.0 0.5 1.0 1.5 2.0 2.5 3.0 1 2 Levels 3 4 5 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 Rev ised 22 Obviously, all the marginal CDFs have been denoted by Fk , k = 1, . . . , d, when the joint CDF is F. Throughout this chapter, the inverse operator −1 should be understood to be a generalised inverse; namely that for every function G, G −1 (x) = inf{y | G(y) ≥ x} Assume we have observed a T-sample (Xi )i=1,...,T . These are some realisations of the d-random vector X = (X1 , . . . , Xd ). Note that we do not assume that Xi = (X1i , . . . , Xdi ) are mutually independent (at least for the moment). Every marginal CDF, say the kth, can be estimated empirically by (1) Fk (x) = 1 T T ∑ 1(Xki ≤ x) i=1 3 Revised Proof Ref: 33259e September 29, 2006 COPULAS 02 03 04 05 06 07 08 09 10 (1) and [Fk ]−1 (uk ) is simply the empirical quantile corresponding to uk ∈ [0, 1]. Another means of estimation is to smooth such CDFs, and the simplest way is to invoke the kernel method (eg, see Härdle and Linton (1994) or Pagan and Ullah (1999) for an introduction):  consider a univariate kernel function K : R −→ R, K = 1, and a bandwidth sequence h T (or simply h hereafter), h T > 0 and h T −→ 0 when T → ∞. Then, Fk (x) can be estimated by (2) Fk (x) 1 = T  T  x − Xki ∑K h i=1 ofs 01 11 13 14 15 16 17 18 19 20 21 for every real number x, by denoting K the primitive function of K: x K(x) = −∞ K. There exists another common case: assume that an underlying parametric model has been fitted previously for the kth margin. (3) Then, the natural estimator for Fk (x) is some CDF Fk (x, θ̂k ) that depends on the relevant estimated parameter θ̂k . When such a model is well-specified, θ̂k is tending almost surely to a value θk (3) such that Fk (·) = Fk (·, θk ). The last limiting case is the knowledge pro 12 (0) of the true CDF Fk . Formally, we will set Fk = Fk . Similarly, the joint CDF F can be estimated empirically by 22 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 F (1) (x) = 1 T T Rev ised 23 ∑ 1(Xi ≤ x) i=1 or by the kernel method F (2) (x) = 1 T T ∑K i=1  x − Xi h with a d-dimensional kernel K, so that   x1 K(x) = ... −∞ xd  K −∞ for every x = (x1 , . . . , xd ) ∈ R d . Besides, there may exist an underlying parametric model for X: F is assumed to belong to a set of multivariate CDFs indexed by a parameter τ. A consistent estimation τ̂ for the “true” value τ allows setting F (3) (·) = F(·, τ̂). Finally, we can denote F (0) = F. 4 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 03 04 05 06 07 08 09 10 11 12 13 14 15 16 17 18 19 for each of the indexes j, j1 , j2 , . . . , jd that belong to {0, 1, 2, 3}. Thus, it is not so obvious to discriminate between all these competitors, especially without any parametric assumption. Every estimation method has its own advantages and drawbacks. The full empirical method (j = j1 = · · · = jd = 1 with the notations of Equation (2.1)) has been introduced in Deheuvels (1979, 1981a, 1981b) and studied more recently by Fermanian et al (2004), in the independent setting, and by Doukhan et al (2004) in a dependent framework. It provides a robust and universal way for estimation purposes. Nonetheless, its discontinuous feature induces some difficulties: the graphical representations of the copula can be not very nice from a visual point of view and not intuitive. Moreover, there is no unique choice for building the inverse (1) function of Fk . In particular, if Xk1 ≤ · · · ≤ XkT is the ordered ofs 02 Therefore, generally speaking, a d-dimensional copula C can be estimated by   (j ) (j ) Ĉ(u) = F (j) [F1 1 ]−1 (u1 ), . . . , [Fd d ]−1 (ud ) (2.1) pro 01 (1) 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 sample on the k-axis, the inverse function of Fk at some point i/T may be chosen arbitrarily between Xki and Xk(i+1). Finally, since the copula estimator is not differentiable when only one empirical CDF is involved in Equation (2.1), it cannot, for example, be used straightforwardly to derive an estimate of the associated copula density (by differentiation of Ĉ(u) with respect to all its arguments) or for optimisation purposes. Smooth estimators are better suited to graphical usage, and can provide more easily the intuition to achieve the “true” underlying parametric distribution. However, they depend on an auxiliary smoothing parameter (eg, h in the case of the kernel method), and suffer from the well-known “curse of dimensionality”: the higher the dimension (d with our notations), the worse the performance in terms of convergence rates. In other words, as the dimension increases, the complexity of the problem increases exponentially.2 Such methods can be invoked safely in practice when d ≤ 3 and for sample sizes larger than, say, two hundred observations (which is usual in finance). The theory of fully smoothed copulas (j = j1 = · · · = jd = 2 with the notation of Equation (2.1)) can be found in Fermanian and Scaillet (2003) in a strongly dependent framework. Rev ised 20 5 Revised Proof Ref: 33259e September 29, 2006 COPULAS 02 03 04 05 06 07 08 09 10 11 12 13 A more comfortable situation exists when “good” parametric assumptions are put into (2.1) for the marginal CDFs and/or the joint CDF F. The former case is relatively usual because there exist a great many univariate models for financial variables (eg, see Alexander (2002)). Nevertheless, for a lot of dynamic models (eg, stochastic volatility models), their (unconditional) marginal CDFs cannot be written explicitly. Obviously, we are under the threat of a misspecification, which can have disastrous effects (see Fermanian and Scaillet (2005)). Concerning a parametric assumption for F itself, our opinion is balanced. At first glance, we are absolutely free to choose an “interesting” parametric family F of d-dimensional CDFs that would contain the true law F. But, by setting for every real number x and every k = 1, . . . , d 14 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 pro 16 (3) Fk (x) = F(+∞, . . . , +∞, x, +∞, . . . |τ̂) where x is the kth argument of F, we should have found the “right” marginal distributions too, to be self-coherent. Indeed, the joint law contains the marginal ones. Then, the estimated copula should be   (3) (3) (3) Ĉ(u) = F (3) [F1 ]−1 (u1 ), [F2 ]−1 (u2 ), . . . , [Fd ]−1 (ud ) In reality, the problem is finding a sufficiently rich family F ex ante that might generate all empirical features. What people do is more clever. They choose a parametric family F ∗ and other marginal parametric families Fk∗ , k = 1, . . . , d, and set Rev ised 15 ofs 01   C(u) = F̂ ∗ [ F̂1∗ ]−1 (u1 ), . . . , [ F̂d∗ ]−1 (ud ) for some F̂ ∗ ∈ F , and F̂k∗ ∈ Fk for every k = 1, . . . , d. Note that the choice of all the parametric families is absolutely free of constraints, and that these families are not related to each other (they can be arbitrary and independently chosen). This is the usual way of generating new copula families. The price to be paid is that the true joint law F does not belong to F ∗ generally speaking. Similarly, the true marginal laws Fk do not belong to the sets Fk∗ in general. If a parametric assumption is made in such a case, the standard estimation procedure is semi-parametric: the copula is a function of some parameter θ = (τ, θ1 , . . . , θd ). Recall that the copula density c 6 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 01 is the derivative of C with respect to each of its arguments: 02 cθ (u) = 03 04 05 06 07 ∂d C(u) ∂1 . . . ∂ d Here, the copula density cθ itself can be calculated under a full parametric assumption. Thus, we get an estimator of θ by maximizing the log-likelihood 08 i=1 10 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 for some T-convergent estimates Fk (Xki ) of the marginal CDFs. (1) (2) Obviously, we may choose Fk = Fk or Fk . Note that such an estimator is called an “omnibus estimator”, and it can be seen as a maximum-likelihood estimator of θ after replacing the unobservable ranks Fk (Xki ) by the pseudoobservations. The asymptotic distribution of the estimator has been studied in Genest et al (1995) and Shi and Louis (1995). The main aim of semi-parametric estimation is to avoid possible misspecification of marginal distributions, which may overestimate the degree of dependence in the data (eg, see Silvapulle et al (2004)). Note finally that Chen and Fan (2004a, 2004b) have developed the theory of this semi-parametric estimator in a time-series context. Thus, depending on the degree of assumptions about the joint and marginal models, there exists a wide range of possibilities for estimating copula functions as provided by Equation (2.1). The only trap to avoid is to be sure that the assumptions made for margins are consistent with those drawn for the joint law. The statistical properties of all these estimators are the usual ones, namely consistency and asymptotic normality. pro 12 √ Rev ised 11 ∑ log cθ ( F1 (X1i ), . . . , Fd (Xdi )) ofs T 09 THE ESTIMATION OF COPULA DENSITIES After the estimation of C by Ĉ as in Equation (2.1), it is tempting to define an estimate of the copula density c at every u ∈ [0, 1]d by ĉ(u) = ∂d Ĉ(u) ∂1 · · · ∂ d Unfortunately, this works only when Ĉ is differentiable. Most of the time, this is the case when the marginal and joint CDFs are 7 Revised Proof Ref: 33259e September 29, 2006 COPULAS 03 04 05 06 07 08 09 10 11 12 13 14 15 16 17 18 19 20 21 ofs 02 parametric or nonparametrically smoothed (by the kernel method, for instance). In the latter case and when d is “large” (more than 3), the estimation of c can be relatively poor because of the curse of dimensionality. Nonparametric estimation procedures for the density of a copula function have already been proposed by Behnen et al (1985) and Gijbels and Mielniczuk (1990). These procedures rely on symmetric kernels, and have been detailed in the context of uncensored data. Unfortunately, such techniques are not consistent on the boundaries of [0, 1]d . They suffer from the so-called boundary bias. Such bias can be significant in the neighbourhood of the boundaries too, depending on the size of the bandwidth. Hereafter, we will propose some solutions to cope with such issues. To ease notation and without a lack of generality, we will restrict ourselves to the bivariate case (d = 2). Thus, our random vector will be denoted by (X, Y) instead of (X1 , X2 ). In the following sections, we will study some properties of some kernel-based estimators, and illustrate some of these by simulations. The benchmark will be a simulated sample, whose size is T = 1,000 and that will be generated by a Frank copula with copula density pro 01 22 c Fr (u, v, θ) = 24 25 26 27 30 31 32 33 Nonparametric density estimation for distributions with finite support An initial approach relies on a kernel-based estimation of the density based on the pseudo-observations (FX,T (Xi ), FY,T (Yi )), where FX,T and FY,T are the empirical distribution functions 34 35 36 37 38 39 ([1 − e−θ ] and Kendall’s tau equal to 0.5. Hence, the copula parameter is θ = 5.74. This density can be seen in Figure 2.2 together with its contour plot on the right. 28 29 θ[1 − e−θ ]e−θ(u+v) − (1 − e−θu )(1 − e−θv ))2 Rev ised 23 FX,T (x) = 1 T+1 T ∑ 1(Xi ≤ x) and i=1 FY,T (y) = 1 T+1 T ∑ 1(Yi ≤ y) i=1 where the factor T + 1 (instead of standard T, as in Deheuvels (1979) for instance) allows the avoidance of boundary problems: the 8 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 02 03 04 05 06 07 08 quantities FX,T (Xi ) and FY,T (Yi ) are the ranks of the Xi ’s and the Yi ’s divided by T + 1, and therefore take values  2 T 1 , ,..., T+1 T+1 T+1 Standard kernel-based estimators of the density of pseudoobservations yield, using diagonal bandwidth (see Wand and Jones (1995)) 09 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 35 36 37 38 39  u − FX,T (Xi ) v − FY,T (Yi ) , ∑K h h i=1   for a bivariate kernel K : R2 −→ R, K = 1. The variance of the estimator can be derived, and is O((Th2 )−1 ). Moreover, it is asymptotically normal at every point (u, v) ∈ (0, 1):  ch (u, v) − E( ch (u, v)) L − → N (0, 1) Var( ch (u, v)) As a benchmark, Figure 2.2 shows the theoretical density of a Frank copula. In Figure 2.3 we plot the standard Gaussian kernel  i, V i ) ≡ estimator based on the sample of pseudo-observations (U (FX,T (Xi ), FY,T (Yi )). Recall that even if kernel estimates are consistent for distributions with unbounded support and the support is bounded, the boundary bias can yield some “ill” underestimation (even if the distribution is twice differentiable in the interior of its support). We can explain this phenomenon easily in the univariate case. Consider a T sample X1 , . . . , XT of a positive random variable with density f . The support of their density is then R + . Let K denote a symmetric kernel, whose support is [−1, +1]. Then, for all x ≥ 0, using a Taylor expansion, we get 33 34 T pro 11 1  ch (u, v) = Th2 E( fh (x)) = Rev ised 10 ofs 01  x/h −1 = f (x) · K(y) f (x − hy) dy  ′ x/h K(y) dy −1 − h · f (x) ·  x/h −1 yK(y) dy + O(h2 ) 9 Revised Proof Ref: 33259e September 29, 2006 COPULAS 01 02 Figure 2.2 Density of the Frank copula with a Kendall tau equal to 0.5. 03 05 1.0 04 5 4 0.6 07 0.8 06 08 3 0.4 ofs 09 1.0 0.8 0.6 2 11 0.4 13 0.0 0.2 00.0 0.0 14 0.2 0.4 0.6 0.8 0.0 1.0 15 16 17 18 1.0 3.0 24 30 31 32 33 34 35 36 37 38 39 0.4 0.5 0.2 0.0 0.0 0.0 0.2 27 0.0 1.0 0.8 0.6 1.0 29 1.0 0.4 1.5 26 28 0.8 0.6 2.0 Rev ised 2.5 23 25 0.6 0.8 21 22 0.4 Figure 2.3 Estimation of the copula density using a Gaussian kernel based on 1,000 observations drawn from a Frank copula. 19 20 0.2 pro 1 12 0.2 10 0.2 0.4 0.6 0.8 1.0 0.0 Hence, since the kernel is symmetric, x = 0, and therefore 0.2  x/h −1 0.4 0.6 0.8 1.0 h→0 K(y) dy −−→ 1/2 when E( fh (0)) = 12 f (0) + O(h)  x/h Note that, if x > 0, the expression −1 K(y) dy is 1 when h is sufficiently small (when x > h to be specific). Thus, this integral cannot be one, uniformly, with respect to every x ∈ (0, 1]. And for 10 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 01 02 03 Figure 2.4 Estimation of the uniform density on [0, 1] using a Gaussian kernel and different bandwidth with (a) n = 100 and (b) n = 1,000 observations. 04 (b) (a) 1.2 06 1.2 05 1.0 1.0 07 0.2 0.0 0.2 0.4 0.6 0.8 1.0 16 17 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 0.2 0.4 0.6 0.8 1.0 more general kernels, it has no reason to be equal to 1. In the latter case, since this expression can be calculated, normalising  x/h fh (x) by dividing by −1 K(z) dz (at each x) achieves consistency. Nonetheless, it remains a bias that is of the order of O(h). Using some boundary kernels (see Gasser and Müller (1979)), it is possible to achieve O(h2 ) everywhere in the interior of the support. Consider the case of variables uniformly distributed on [0, 1], U1 , . . . , Un . Figure 2.4 shows kernel-based estimators of the uniform density, with Gaussian kernel and different bandwidths, with n = 100 and 1,000 simulated variables. In that case, for any h > 0,  2   1  1 f (0) 1 y h→0 1 E( fh (0)) = Kh (y) dy = √ exp − 2 dy −−→ = 2 2 2h h 2π 0 0 Rev ised 18 0.0 pro 15 0.0 14 0.0 13 0.2 12 ofs 0.4 0.6 Density 0.8 0.6 0.4 11 Density 09 10 0.8 08 and in the interior, ie, x ∈ (0, 1),  1  E( f h (x)) = Kh (y − x) dy 0    1 1 (y − x)2 h→0 = √ exp − dy −−→ 1 = f (x) 2h2 h 2π 0 Dealing with bivariate copula densities, we observe the same phenomenon. On boundaries, we obtain some “multiplicative 11 Revised Proof Ref: 33259e September 29, 2006 COPULAS 01 02 03 Figure 2.5 Estimation of the copula density on the diagonal using a (standard) Gaussian kernel with (a) 100 and (b) 1,0000 observations drawn from a Frank copula. 04 (a) 05 (b) 13 15 0.2 0.4 0.6 0.8 1.0 17 18 21 32 33 E( ch (u, v)) = c(u, v) + O(h2 ) The bias can be observed in Figure 2.5, which represents the diagonal of the estimated density for several samples. Several techniques have been introduced to obtain a better estimation on the borders for univariate densities: 34 35 • 36 37 38 39 1.0 and in the interior ((u, v) ∈ (0, 1) × (0, 1)) 29 31 0.8 E( ch (0, v)) = 12 c(u, v) + O(h) 26 30 0.6 on the interior of the borders (eg, u = 0 and v ∈ (0, 1)) 25 28 0.4 E( ch (0, 0)) = 14 c(u, v) + O(h) 23 27 0.2 bias”, 1/4 in corners and 1/2 in the interior of borders. The additional bias is of the order of O(h) on the frontier, and standard O(h2 ) in the interior. More precisely, in any corners (eg, (0, 0)) 22 24 0.0 Rev ised 20 3 4 0.0 16 19 2 0 0 14 ofs 12 pro 11 1 10 Estimation of the density on the diagonal 09 3 08 2 07 1 Estimation of the density on the diagonal 4 06 • mirror image modification (Schuster (1985); Deheuvels and Hominal (1979)), where artificial data are obtained, using symmetric (mirror) transformations on the borders; transformed kernels (Devroye and Györfi (1985); Wand et al (1991)), where the idea is to transform the data Xi using a 12 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 02 03 04 05 06 07 08 09 10 • bijective mapping φ so that the φ(Xi ) have support R. Efficient kernel-based estimation of the density of the φ(Xi ) can be derived, and, by the inverse transformation, we get back the density estimation of the Xi themselves; boundary kernels (Gasser and Müller (1979); Rice (1984); Müller (1991)), where a smooth distortion is considered near the border, so that the bandwidth and the kernel shape can be modified (the closer to the border, the smaller). Finally, the last section will briefly mention the impact of pseudoobservations, ie, working on samples 11 13 14 15 16 17 {(FX,T (X1 ), FY,T (Y1 )), . . . , (FX,T (XT ), FY,T (YT ))} instead of {(FX (X1 ), FY (Y1 )), . . . , (FX (XT ), FY (YT ))} as if we know the true marginal distributions. 18 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 Mirror image The idea of this method, developed by Deheuvels and Hominal (1979) and Schuster (1985), is to add some “missing mass” by reflecting the sample with respect to the boundaries. They focus on the case where variables are positive, ie, whose support is [0, ∞). Formally and in its simplest form, it means replacing Kh (x − Xi ) by Kh (x − Xi ) + Kh (x + Xi ). The estimator of the density is then      x − Xi 1 T x + Xi K fh (x) = + K ∑ Th i=1 h h Rev ised 19 pro 12 ofs 01 In the case of densities whose support is [0, 1] × [0, 1], the nonconsistency can be corrected on the boundaries, but the convergence rate of the bias will remain O(h) on the boundaries, which is larger than the usual rate O(h2 ) obtained in the interior if h → 0. The only case where the usual rate of convergence is obtained on boundaries is when the derivative of the density is zero on such subsets. Note that the variance is 4 times higher in corners and 2 times higher in the interior of borders. For copulas, instead of using only the “pseudo-observations”  i ) ≡ (FX,T (Xi ), FY,T (Yi )), the mirror image consists in reflect( Ui , V ing each data point with respect to all edges and corners of the 13 Revised Proof Ref: 33259e September 29, 2006 COPULAS 01 02 03 Figure 2.6 Estimation of the copula density using a Gaussian kernel and the mirror reflection principle with 1,000 observations from the Frank copula. 04 05 1.0 3.0 07 0.8 06 2.5 2.0 10 0.4 1.5 11 13 0.8 1.0 0.6 0.2 12 0.4 0.5 0.0 0.2 15 0.4 0.6 0.8 16 17 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 0.2 0.4 0.6 0.8 1.0 unit square [0, 1] × [0, 1]. Hence, additional observations can be  i , ±V i ), the (±U  i, 2 − V i ), the (2 − U  i , ±V i ) considered; ie, the (±U  i, 2 − V i ), so that one can consider and the (2 − U ĉh (u, v)          i i i i u−U v−V u+U v−V +K K ∑ K h K h h h i=1         i i i i v+V u+U v+V u−U K +K K +K h h h h         i i i i u−U v−2+V u+U v−2+V +K K +K K h h h h            i u − 2 + Ui v − Vi u − 2 + Ui v+V +K K +K K h h h h     i i u−2+U v−2+V +K K h h 1 = Th2 T Rev ised 18 0.0 pro 14 0.0 0.2 ofs 09 0.6 08 Figure 2.6 has been obtained using the reflection principle. We can check that the fit is far better than in Figure 2.3. Transformed kernels Recall that c is the density of (U, V), U = FX (X) and V = FY (Y). The two latter random variables (RVs) follow uniform distributions 14 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 03 04 f (x, y) = g(x)g(y)c[G(x), G(y)] 05 06 07 08 09 10 11 12 This density is twice continuously differentiable on R2 , and the standard kernel approach applies. Since we do not observe a sample of (U, V) but instead make pseudo-observations (Ûi , V̂i ), we build an “approximated sample” of the transformed variables ( X̃1 , Ỹ1 ), . . . , ( X̃T , ỸT ) by setting X̃i = G −1 (Ûi ) and Ỹi = G −1 (V̂i ). Thus, the kernel estimator of f is 13 1 fˆ(x, y) = Th2 14 15 16 17 T  x − X̃ y − Ỹ ∑ K h i, h i i=1 c(u, v) = f (G −1 (u), G −1 (v)) , g(G −1 (u))g(G −1 (v)) (2.3) (u, v) ∈ [0, 1] × [0, 1] 20 21  The associated estimator of c is then deduced by inverting (2.2), 18 19 (2.2) ofs 02 (marginally). Consider a distribution function G of a continuous distribution on R, with differentiable strictly positive density g. We build new RVs X̃ = G −1 (U) and Ỹ = G −1 (V). Then, the density of ( X̃, Ỹ) is pro 01 and therefore we get 22 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39  ch (u, v) = 1 Rev ised 23 Th2 g(G −1 (u)) T × ∑ i=1 K  · g(G −1 (v)) G −1 (u) − G −1 (Ûi ) G −1 (v) − G −1 (V̂i ) , h h  Note that this approach can be extended by considering different transformations GX and GY , different kernels K X and KY , or different bandwidths h X and hY , for the two marginal random variables. Figure 2.7 was obtained using the transformed kernel, where K was a Gaussian kernel and G was respectively the CDF of the N (0, 1) distribution. The absence of a multiplicative bias on the borders can be observed in Figure 2.8, where the diagonal of the copula density is plotted, based on several samples. The copula density estimator obtained with transformed samples has no bias, is asymptotically normal, etc. Actually, we get all the usual properties of the multivariate kernel density estimators. 15 Revised Proof Ref: 33259e September 29, 2006 COPULAS 01 02 03 04 05 Figure 2.7 Estimation of the copula density using a Gaussian kernel and Gaussian transformations with 1,000 observations drawn from the Frank copula. 06 07 5 0.8 08 0.6 10 3 0.4 11 12 2 0.8 13 0.2 0.6 14 0.4 1 pro 0.2 15 0 16 0.2 0.4 0.6 0.8 0.2 17 18 19 20 21 Figure 2.8 Estimation of the copula density on the diagonal using a Gaussian kernel and Gaussian transformations with (a) 100 and (b) 1,0000 observations drawn from a Frank copula. 25 (a) 4 32 33 2 3 Estimation of the density on the diagonal 31 1 30 Estimation of the density on the diagonal 27 29 0 0 34 35 0.0 36 37 38 39 (b) 4 26 28 0.8 3 24 0.6 2 23 0.4 Rev ised 22 ofs 4 1 09 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 16 Revised Proof Ref: 33259e September 29, 2006 0.8 1.0 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 03 04 05 06 07 08 09 10 1 fh (x) = T 11 12 13 14 15   1−x x + 1, + 1 K X , i ∑ h h i=1 T where K(·, α, β) denotes the density of the beta distribution with parameters α and β, 16 K(x, α, β) = 17 x α (1 − x) β , B(α, β) x ∈ [0, 1] 18 19 20 where B(α, β) = Γ(α + β) Γ(α)Γ(β) 21 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 The main difficulty when working with this estimator is the lack of a simple “rule of thumb” for choosing the smoothing parameter h. The beta kernel has two leading advantages. First it can match the compact support of the object to be estimated. Secondly it has a flexible form and changes the smoothness in a natural way as we move away from the boundaries. As a consequence, beta kernel estimators are naturally free of boundary bias and can produce estimates with a smaller variance. Indeed we can benefit from a larger effective sample size since we can pool more data. Monte Carlo results available in these papers show that they have better performance compared to other estimators which are free of boundary bias, such as local linear (Jones (1993)) or boundary kernel (Müller (1991)) estimators. Renault and Scaillet (2004) also report better performance compared to transformation kernel estimators (Silverman (1986)). In addition, Bouezmarni and Rolin (2001, 2003) show that the beta kernel density estimator is consistent even if the true density is unbounded at the boundaries. This feature may also arise in our situation. For example the density of a Rev ised 22 ofs 02 Beta kernels In this section we examine the use of the beta kernel introduced by Brown and Chen (1999), and Chen (1999, 2000) for nonparametric estimation of regression curves and univariate densities with compact support, respectively. Following an idea by Harrell and Davis (1982), Chen (1999, 2000) introduced the beta kernel estimator as an estimator of a density function with known compact support [0, 1], to remove the boundary bias of the standard kernel estimator: pro 01 17 Revised Proof Ref: 33259e September 29, 2006 COPULAS 02 03 04 05 06 bivariate Gaussian copula is unbounded at the corners (0, 0) and (1, 1). Therefore beta kernels are appropriate candidates to build well-behaved nonparametric estimators of the density of a copula function. The beta-kernel based estimator of the copula density at point (u, v) is obtained using product beta kernels, which yields 07 08 09 10  ch (u, v) = 11   u 1−u K X , + 1, + 1 i ∑ h h i=1   1−v v +1 × K Yi , + 1, h h 1 Th2 T ofs 01 12 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 pro 14 Figure 2.9 shows that the shape of the product beta kernels for different values of u and v is clearly adaptive. For convenience, the bandwidths are here assumed to be equal, but, more generally, one can consider one bandwidth per component. See Figure 2.10 for an example of an estimation based on beta kernels and a bandwidth h = 0.05. Let (u, v) ∈ [0, 1] × [0, 1]. The bias of  c(u, v) is of the order of h,  ch (u, v) = c(u, v) + O(h). The absence of a multiplicative bias on the boundaries can be observed on Figure 2.11, where the diagonal of the copula density is plotted, based on several samples. On the other hand, note that the variance depends on the location. More precisely, Var( ch (u, v)) is O((Thκ )−1 ), where κ = 2 in corners, κ = 3/2 in borders, and κ = 1 in the interior of [0, 1] × [0, 1]. Moreover, as well as “standard” kernel estimates,  ch (u, v) is asymptotically normally distributed: Rev ised 13 L ′ ch (u, v) − c(u, v)] − → N (0, σ(u, v)2 ) Thκ [ ′ as Thκ → ∞ and h → 0 where κ ′ depends on the location, and where σ(u, v)2 is proportional to c(u, v). Working with pseudo-observations As we know, most of the time the marginal distributions of random vectors are unknown, as recalled in the first section. Hence, the associated copula density should be estimated not on samples (FX (Xi ), FY (Yi )) but on pseudo-samples (FX,T (Xi ), FY,T (Yi )). 18 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 01 02 03 04 05 Figure 2.9 Shape of bivariate beta kernels for different values of u and v. (a) u = 0.0, v = 0.0; (b) u = 0.2, v = 0.0; (c) u = 0.5, v = 0.0; (d) u = 0.0, v = 0.2; (e) u = 0.2, v = 0.2; (f) u = 0.5, v = 0.2; (g) u = 0.0, v = 0.5; (h) u = 0.2, v = 0.5; (i) u = 0.5, v = 0.5. 06 07 (a) (b) (d) (e) (c) 08 ofs 09 10 11 12 13 15 16 17 18 19 20 21 pro 14 (f) 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 (g) Rev ised 22 (h) (i) 19 Revised Proof Ref: 33259e September 29, 2006 COPULAS 01 02 Figure 2.10 Estimation of the copula density using beta kernels (u = 0.05) with 1,000 observations drawn from a Frank copula. 04 1.0 03 3.0 2.5 07 0.6 06 0.8 05 2.0 08 ofs 0.8 1.0 0.6 11 0.2 10 0.4 1.5 09 0.4 0.5 0.0 0.2 12 0.0 13 0.2 0.4 0.6 0.8 0.0 0.2 0.4 15 Figure 2.11 Estimation of the copula density on the diagonal using beta kernels with (a) 100 and (b) 1,000 observations drawn from a Frank copula. 19 (a) 28 29 30 31 32 33 34 35 36 37 38 39 4 3 Estimation of the density on the diagonal 27 0 26 Rev ised 25 2 24 1 23 Estimation of the density on the diagonal 22 0 21 (b) 4 20 3 18 1.0 2 17 0.8 1 16 0.6 pro 14 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Figure 2.12 shows some scatterplots when the margins are known (ie, we know (FX (Xi ), FY (Yi ))), and when margins are estimated (ie, (FX,T (Xi ), FY,T (Yi )). Note that the pseudo-sample is more “uniform”, in the sense of a lower discrepancy (as in quasi Monte Carlo techniques; eg, see Niederreiter (1992)). Here, by mapping every point of the sample on the marginal axis, we get uniform 20 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 01 02 03 04 05 Figure 2.12 Scatterplots of observations and pseudo-observations and histograms of margins. (a) 100 observations (Xi , Yi ) drawn from a Frank copula and (b) associated pseudo-sample (Ui , Vi ) = ( F̂X (Xi ), F̂Y (Yi )). (b) (a) 1.0 1.0 06 07 ofs 0.6 0.4 0.8 0.6 0.0 0.0 15 16 0.0 17 0.2 19 10 15 0.4 0.6 34 35 36 37 38 39 1.0 1.0 0.0 0.2 0.4 0.6 0.8 1.0 10 15 Histogram of the observations Yi 5 10 5 0 0 33 0.8 Histogram of the observations Yi 27 32 0.8 Histogram of the observations Xi 5 26 31 0.6 0 0.2 25 30 0.4 Rev ised 0.0 24 29 0.2 PseudoŦobservations Ui Histogram of the observations Xi 23 28 0.0 1.0 0 22 0.8 10 15 21 0.6 5 20 0.4 Observations Xi 18 pro 0.2 14 PseudoŦobservations Vi 13 0.2 12 0.4 11 Observations Yi 09 10 0.8 08 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 grids, which is a type of “Latin hypercube” property (eg, see Jäckel (2002)). Because samples are more “uniform” using ranks and pseudoobservations, the variance of the estimator of the density, at some given point (u, v) ∈ (0, 1) × (0, 1), is usually smaller. For instance, 21 Revised Proof Ref: 33259e September 29, 2006 COPULAS 01 02 03 04 05 Figure 2.13 The impact of estimating from pseudo-observations (n = 100). The dashed line is the distribution of ĉ(u, v) from sample (FX (Xi ), FY (Yi )), and the solid line is from pseudo-sample (FX,T (Xi ), FY,T (Yi )). 2.0 06 07 13 14 15 1.5 16 18 0.0 17 19 0 1 2 20 21 pro 12 1.0 11 Density of the estimator 10 0.5 09 ofs 08 3 4 Distribution of the estimation of the density c(u,v) 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 Rev ised 22 Figure 2.13 shows the impact of considering pseudo-observations, ie, substituting FX,T and FY,T into unknown marginal distributions FX and FY . The dashed line shows the density of ĉ(u, v) from 100 observations (Ui , Vi ) (drawn from the same Frank copula), and the solid line shows the density of ĉ(u, v) from the sample of pseudoobservations (ie, the ranks of the observations). A heuristic interpretation can be obtained from Figure 2.12. Consider the standard kernel-based estimator of the density, with a rectangular kernel. Consider a point (u, v) in the interior, and a bandwidth h such that the square [u − h, u + h] × [v − h, v + h] lies in the interior of the unit square. Given a T sample, an estimation of the density at point (u, v) involves the number of points located in the small square around (u, v). Such a number will be denoted by N, and it is a random variable. Larger N provides more precise estimations. 22 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 03 04 05 06 p1 = P((U, V) ∈ [u − h, u + h] × [v − h, v + h]) 07 = C(u + h, v + h) + C(u − h, v − h) 08 − C(u − h, v + h) − C(u + h, v − h) 09 10 and therefore 11 Var(N1 ) = T p1 (1 − p1 ) 12 14 15 16 17 18 19 20 21 22 On the other hand, assume that margins are unknown, or equivalently that we are dealing with a sample of pseudo-observations  1, V 1 ), . . . , (U T, V T ). By construction of pseudo-observations, we (U have  i ∈ [u − h, u + h]} = ⌊2hT⌋ #{U pro 13 where ⌊·⌋ denotes the integer part. As previously, the number of points in the small square N2 satisfies N2 ∼ B(⌊2hT⌋, p2 ) where  V)  ∈ [u − h, u + h] × [v − h, v + h] | U  ∈ [u − h, u + h]) p2 = P((U, = 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 ofs 02 Assume that the margins are known, or equivalently, let (U1 , V1 ), . . . , (UT , VT ) denote a sample with distribution function C. The number of points in the small square, say N1 , is random and follows a binomial law with size T and some parameter p1 . Thus, we have N1 ∼ B(T, p1 ) with  V)  ∈ [u − h, u + h] × [v − h, v + h]) P((U,  ∈ [u − h, u + h]) P(U Rev ised 01 ≈ C(u + h, v + h) + C(u − h, v − h) − C(u − h, v + h) − C(u + h, v − h) /2h p1 = 2h Therefore the expected number of observations is the same for both methods (E[N1 ] ≃ E[N2 ] ≃ T p1 ), but Var(N2 ) ≈ 2hT p2 (1 − p2 ) = 2hT Thus p T p1 (1 − 1 ) = p (2h − p1 ) 2h 2h 2h 1 Var(N2 ) T p1 (2h − p1 ) 2h − p1 = = ≤1 Var(N1 ) 2hT p1 (1 − p1 ) 2h − 2hp1 since h ≤ 1/2 and thus 2hp1 ≤ p1 . So finally, the variance of the number of observations in the small square around (u, v) is larger than the variance of the number of pseudo-observations in the same square. Therefore, this 23 Revised Proof Ref: 33259e September 29, 2006 COPULAS 01 02 03 04 05 06 07 larger uncertainty concerning the relevant sub-sample used in the neighbourhood of (u, v) in the former case implies a loss of efficiency. The consequence of this result is largely counterintuitive. By working with pseudo-observations instead of “true” ones, we would expect an additional noise, which should induce more noisy estimated copula densities. This is not in fact the case as we have just shown. 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 CONCLUDING REMARKS We have discussed how various estimation procedures impact the estimation of tail probabilities in a copula framework. Parametric estimation may lead to severe underestimation when the parametric model of the margins and/or the copula is misspecified. Nonparametric estimation may also lead to severe underestimation when the smoothing method does not take into account potential boundary biases in the corner of the density support. Since the primary focus of most risk management procedures is to gauge these tail probabilities, we think that the methods analysed above might help to better understand the occurrence of extreme risks in stand-alone positions (single asset) or inside a portfolio (multiple assets). In particular we have shown that nonparametric methods are simple, powerful visualisation tools that enable the detection of dependencies among various risks. A clear assessment of these dependencies should help in the design of better risk measurement tools within a VAR or an expected shortfall framework. pro 10 27 28 29 30 31 1 The third author acknowledges financial support by the National Centre of Competence in Research “Financial Valuation and Risk Management” (NCCR FINRISK). 2 For example, the number of histogram grid cells increases exponentially. This effect cannot be avoided, even using other estimation methods. Under smoothness assumptions on the density, the amount of training data required for nonparametric estimators increases exponentially with the dimension (eg, see Stone (1980)). 32 33 34 35 36 37 38 39 Rev ised 09 ofs 08 REFERENCES Alexander, C., 2002, Market Models (New York: John Wiley and Sons). Behnen, K., M. Huskova, and G. Neuhaus, 1985, “Rank Estimators of Scores for Testing Independence”, Statistics and Decision, 3, pp 239–262. Bouezmarni, T. and J. M. Rolin, 2001, “Consistency of Beta Kernel Density Function Estimator”, The Canadian Journal of Statistics, 31, pp 89–98. 24 Revised Proof Ref: 33259e September 29, 2006 THE ESTIMATION OF COPULAS: THEORY AND PRACTICE 01 02 03 Bouezmarni, T. and J. M. Rolin, 2003, “Bernstein Estimator for Unbounded Density Function”, Université Catholique de Louvain-la-Neuve. Brown, B. M. and S. X. Chen, 1999, “Beta-Bernstein Smoothing for Regression Curves with Compact Supports”, Scandinavian Journal of Statistics, 26, pp 47–59. 04 06 07 08 09 10 11 12 13 14 Chen, S. X., 1999, “A Beta Kernel Estimation for Density Functions”, Computational Statistics and Data analysis, 31, pp 131–135. Chen, S. X., 2000, “Beta Kernel for Regression Curve”, Statistica Sinica, 10, pp 73–92. Chen, S. X. and Y. Fan, 2004a, “Efficient Semiparametric Estimation of Copulas”, Working Paper. ofs 05 Chen, S. X. and Y. Fan, 2004b, “Copula-based Tests for Dynamic Models”, Working Paper. Deheuvels, P., 1979, “La fonction de dépendance empirique et ses propriétés”, Acad. Roy. Belg., Bull. C1 Sci. 5ième sér., 65, pp 274–292. Deheuvels, P., 1981a, “A Kolmogorov–Smirnov Type Test for Independence and Multivariate Samples”, Rev. Roum. Math. Pures et Appl., 26(2), pp 213–226. Deheuvels, P., 1981b, “A Nonparametric Test for Independence, Pub. Inst. Stat. Univ. Paris, 26(2), pp 29–50. 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 Deheuvels, P. and P. Hominal, 1979, “Estimation non paramétrique de la densité compte tenu d’informations sur le support”, Revue de Statistique Appliquée, 27, pp 47–68. Devroye, L. and L. Györfi, 1985, Nonparametric Density Estimation: The L1 View (New York: John Wiley and Sons). Doukhan, P., J.-D. Fermanian and G. Lang, 2004, “Copulas of a Vector-valued Stationary Weakly Dependent Process”, Working Paper. Fermanian, J.-D., D. Radulovic and M. Wegkamp, 2004, “Weak Convergence of Empirical Copula Processes”, Bernoulli, 10, pp 847–860. Fermanian, J.-D. and O. Scaillet, 2003, “Nonparametric Estimation of Copulas for Time Series”, Journal of Risk, 95, pp 25–54. Rev ised 16 pro 15 Fermanian, J.-D. and O. Scaillet, 2005, “Some Statistical Pitfalls in Copula Modelling for Financial Applications”, in Klein, E. (ed), Capital Formation, Governance and Banking (New York: Nova Science Publishers), pp 59–74. Gasser, T. and H. G. Müller, 1979, “Kernel Estimation of Regression Functions”, in Gasser, T. and M. Rosenblatt (eds), Smoothing Techniques for Curve Estimation, Lecture Notes in Mathematics, 757 (Springer), pp 23–68. Genest, C., K. Ghoudi and L. P. Rivest, 1995, “A Semiparametric Estimation Procedure of Dependence Parameters in Multivariate Families of Distributions”, Biometrika, 82, pp 543–552. Gijbels, I. and J. Mielniczuk, 1990, “Estimating the Density of a Copula Function”, Communications in Statistics: Theory and Methods, 19, pp 445–464. Härdle, W. and O. Linton, 1994, “Applied Nonparametric Methods”, in Engle, R. and D. McFadden (eds), Handbook of Econometrics, IV (Amsterdam: North-Holland). Harrell, F. E. and C. E. Davis, 1982, “A New Distribution-free Quantile Estimator”, Biometrika, 69, pp 635–640. Jäckel, P., 2002, Monte Carlo Methods in Finance (New York: John Wiley and Sons). 37 38 39 Jones, M. C., 1993, “Simple Boundary Correction for Kernel Density Estimation”, Statistics and Computing, 3, pp 135–146. 25 Revised Proof Ref: 33259e September 29, 2006 COPULAS 01 02 03 Li, D. X., 1999, “On Default Correlation: a Copula Function Approach”, RiskMetrics Group, Working Paper. Longin, F. and B. Solnik, 1995, “Is the Correlation in International Equity Returns Constant: 1960–1990?”, Journal of International Money and Finance, 14, pp 3–26. 04 06 07 08 09 10 Müller, H. G., 1991, “Smooth Optimum Kernel Estimators near Endpoints”, Biometrika, 78, pp 521–530. Niederreiter, H., 1992, “Random Number Generation and Quasi-Monte Carlo Methods”, CBMS-SIAM, 63, pp 86–112. Pagan, A. and A. Ullah, 1999, Nonparametric Econometrics (Cambridge: Cambridge University Press). ofs 05 Patton, A., 2001, “Modelling Time-Varying Exchange Rate Dependence Using the Conditional Copula”, University of California, San Diego, Discussion Paper, 01-09. 11 13 14 15 16 17 18 19 Renault, O. and O. Scaillet, 2004, “On the Way to Recovery: a Nonparametric Bias Free Estimation of Recovery Rate Densities”, Journal of Banking & Finance, 28, pp 2915–2931. Rice, J. A., 1984, “Boundary Modification for Kernel Regression”, Communication in Statistics, 12, pp 899–900. Schuster, E., 1985, “Incorporating Support Constraints into Nonparametric Estimators of Densities”, Communications in Statistics: Theory and Methods, 14, pp 1123–1136. pro 12 Shih, J. H. and T. A. Louis, 1995, “Inferences on the Association Parameter in Copula Models for Bivariate Survival Data”, Biometrics, 55, pp 1384–1399. Silvapulle, P., G. Kim and M. J. Silvapulle, 2004, “Robustness of a Semiparametric Estimator of a Copula”, Econometric Society 2004 Australasian Meetings, 317. 20 22 23 24 Silverman, B., 1986, Density Estimation for Statistics and Data Analysis (New York: Chapman and Hall). Sklar, A., 1959, “Fonctions de répartition à n dimensions et leurs marges”, Publ. Inst. Statist. Univ. Paris, 8, pp 229–231. Rev ised 21 25 Stone, C. J., 1980, “Optimal Rates of Convergence for Nonparametric Estimators”, Annals of Probability, 12, pp 361–379. 26 Wand, M. P. and M. C. Jones, 1995, Kernel Smoothing (New York: Chapman and Hall). 27 28 29 30 31 32 33 34 35 36 37 38 39 Wand, M. P., J. S. Marron and D. Ruppert, 1991, “Transformations in Density Estimation: Rejoinder (in Theory and Methods)”, Journal of the American Statistical Association, 86, pp 360–361. 26 Revised Proof Ref: 33259e September 29, 2006