This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Copula-based measures of asymmetry between the lower and upper tail probabilities

Shogo Kato ,a,  Toshinao Yoshibab,c and Shinto Eguchi a
a Institute of Statistical Mathematics
b Tokyo Metropolitan University
c Bank of Japan
Address for correspondence: Shogo Kato, Institute of Statistical Mathematics, 10-3 Midori-cho, Tachikawa, Tokyo 190-8562, Japan. E-mail: [email protected]
(August 4, 2020)
Abstract

We propose a copula-based measure of asymmetry between the lower and upper tail probabilities of bivariate distributions. The proposed measure has a simple form and possesses some desirable properties as a measure of asymmetry. The limit of the proposed measure as the index goes to the boundary of its domain can be expressed in a simple form under certain conditions on copulas. A sample analogue of the proposed measure for a sample from a copula is presented and its weak convergence to a Gaussian process is shown. Another sample analogue of the presented measure, which is based on a sample from a distribution on 2\mathbb{R}^{2}, is given. Simple methods for interval estimation and nonparametric testing based on the two sample analogues are presented. As an example, the presented measure is applied to daily returns of S&P500 and Nikkei225.

Keywords: Asymptotic theory; Bootstrap; Extreme value theory; Gaussian process; Stock daily return.

1 Introduction

In statistical analysis of multivariate data, it is often the case that data have complex dependence structure among variables. As a statistical tool for analyzing such data, copulas have gained their popularity in various academic fields, especially, finance, actuarial science and survival analysis (see, e.g., Joe, 1997, 2014; Nelsen, 2006; McNeil et al., 2015).

A copula is a multivariate cumulative distribution function with uniform [0,1][0,1] margins. The bivariate case of Sklar’s theorem states that, for a bivariate cumulative distribution function FF with margins F1F_{1} and F2F_{2}, there exists a copula CC such that F(x1,x2)=C(F1(x1),F2(x2))F(x_{1},x_{2})=C(F_{1}(x_{1}),F_{2}(x_{2})). Hence a copula can be used as a model for dependence structure and is applicable for flexible modeling. Another important advantage of using copulas is that copulas are useful as measures of dependence. For example, the tail dependence coefficient is well-known as a measure of dependence in a tail (see, e.g., Section 2.13 of Joe, 2014).

One important problem in copula-based modeling is to decide which should be fitted to data of interest, a copula with symmetric tails or a copula with asymmetric tails. An additional problem arising from this is that if a copula with asymmetric tails is appropriate for the data, how much degree of tail asymmetry the copula should have. These problems are important because the lack of fit in tails of copulas leads to erroneous results in statistical analysis. For example, it is said that widespread applications of Gaussian copula, which has symmetric light tails, to financial products have contributed to the global financial crisis of 2008–2009 (see Donnelly and Embrechts, 2010). Therefore, in order to carry out decent statistical analysis, it is essential to evaluate the degree of tail asymmetry of copula appropriately. Given the stock market spooked by the outbreak of COVID-19, these problems would be even more important.

Some copula-based measures of tail asymmetry have been proposed in the literature. Nikoloulopoulos et al. (2012) and Dobrić et al. (2013) discussed a measure of tail asymmetry based on the difference between the conditional Spearman’s rhos for truncated data. Krupskii (2017) proposed an extension of their measure, which can regulate weights of tails. Rosco and Joe (2013) proposed three measures of tail asymmetry; two of them are based on moments or quantiles of a transformed univariate random variable and one of them is based on a difference between a copula and its reflected copula. As related works, measures of radial symmetry for the entire domain, not for tails, have been proposed, for example, by Dehgani et al. (2013) and Genest and Nešlehová (2014). See Joe (2014, Section 2.14) for the book treatment on this topic.

In this paper we propose a new copula-based measure of asymmetry between the lower and upper tails of bivariate distributions. The proposed measure and its sample analogues have various tractable properties; the proposed measure has a simple form and its calculation is fast; the proposed measure possesses desirable properties which as a measure of tail asymmetry; the limits of the proposed measure as the index goes to the boundary of its domain can be easily evaluated under mild conditions on copulas; sample analogues of the proposed measure converge weakly to a Gaussian process or its mixture; simple methods for interval estimation and hypothesis testing based on the sample analogues are available; a multivariate extension of the proposed measure is straightforward.

The paper is organized as follows. In Section 2 we propose a new copula-based measure of tail asymmetry and present its basic properties. Section 3 considers the limits of our measure as the index goes to the boundary of its domain. Values of the proposed measure for some well-known copulas are discussed in Section 4. In Section 5 two sample analogues of the proposed measure are presented and their asymptotic properties are investigated. Also statistical inference for our measure such as interval estimation and hypothesis tests is discussed and a simulation study is carried out to demonstrate the results. In Section 6 the proposed measure is compared with other copula-based measures of tail asymmetry. In Section 7 the proposed measure is applied to daily returns of S&P500 and Nikkei225. Finally, a multivariate extension of the proposed measure is briefly considered in Section 8.

Throughout this paper, a ‘copula’ refers to the bivariate case of a copula, namely, a bivariate cumulative distribution function with uniform [0,1][0,1] margins. Let 𝒞\mathscr{C} be a set of all the bivariate copulas. Let C¯\overline{C} denote the survival copula associated with CC, which defined by C¯(u1,u2)=1u1u2+C(u1,u2)\overline{C}(u_{1},u_{2})=1-u_{1}-u_{2}+C(u_{1},u_{2}). Define u¯\bar{u} by u¯=1u\bar{u}=1-u.

2 Definition and basic properties

In this section we propose a measure for comparing the probabilities of the lower and upper tails of bivariate distributions. The proposed measure is defined as follows.

Definition 1.

Let (X1,X2)(X_{1},X_{2}) be an 2\mathbb{R}^{2}-valued random vector. Assume X1X_{1} and X2X_{2} have continuous margins F1F_{1} and F2F_{2}, respectively. Then a measure of comparison between the lower-left and upper-right tail probabilities of (X1,X2)(X_{1},X_{2}) is defined by

α(u)=log((F1(X1)>1u,F2(X2)>1u)(F1(X1)u,F2(X2)u)),0<u0.5.\alpha(u)=\log\left(\frac{\mathbb{P}(F_{1}(X_{1})>1-u,F_{2}(X_{2})>1-u)}{\mathbb{P}(F_{1}(X_{1})\leq u,F_{2}(X_{2})\leq u)}\right),\quad 0<u\leq 0.5.

Here the definition of the logarithm function is extended to be log(x/y)=\log(x/y)=-\infty if x=0x=0 and y>0y>0, log(x/y)=\log(x/y)=\infty if x>0x>0 and y=0y=0, and log(x/y)=0\log(x/y)=0 if x=y=0x=y=0.

Similarly it is possible to define a measure to compare the lower-right and upper-left tail probabilities of bivariate distributions. Properties of this measure immediately follow from those of α(u)\alpha(u), which will be given hereafter, by replacing (X1,X2)(X_{1},X_{2}) by (X1,X2)(X_{1},-X_{2}).

The calculation of α(u)\alpha(u) can be simplified if the distribution of (X1,X2)(X_{1},X_{2}) is represented in terms of copula. The proof is straightforward and therefore omitted.

Proposition 1.

Let CC denote a copula of (X1,X2)(X_{1},X_{2}) given by C(u1,u2)=(F1C(u_{1},u_{2})=\mathbb{P}(F_{1}
(X1)u1,F2(X2)u2)(X_{1})\leq u_{1},F_{2}(X_{2})\leq u_{2}). Then α(u)\alpha(u) defined in Definition 1 can be expressed as

α(u)=log(2u1+C(1u,1u)C(u,u)).\alpha(u)=\log\left(\frac{2u-1+C(1-u,1-u)}{C(u,u)}\right). (1)

Note that, using the survival copula C¯\overline{C} associated with CC with u¯=1u\bar{u}=1-u, the proposed measure (1) has the simpler expression

α(u)=log(C¯(u¯,u¯)C(u,u)).\alpha(u)=\log\left(\frac{\overline{C}(\bar{u},\bar{u})}{C(u,u)}\right).

Throughout this paper, the lower [0,u]2[0,u]^{2} tail and the upper [1u,1]2[1-u,1]^{2} tail of the copula CC are said to be symmetric if C(u,u)=C¯(u¯,u¯)C(u,u)=\overline{C}(\bar{u},\bar{u}).

Unlike many existing measures, the proposed measure (1) is not a global measure but a local one in the sense that this measure focuses on the probability of a subdomain of the copula regulated by the index uu. Setting a particular value of uu or looking at the behavior of α(u)\alpha(u) for multiple choices of uu, the proposed measure (1) provides a different insight from the global measure. For more details on the comparison between the proposed measure and existing ones, see Section 6.

It is straightforward to see that the following basic properties hold for α(u)\alpha(u).

Proposition 2.

Let 𝒞\mathscr{C} be a set of all bivariate copulas. Denote the measure α(u)\alpha(u) for the copula C𝒞C\in\mathscr{C} by αC(u)\alpha_{C}(u). Assume that pL=C(u,u)p_{L}=C(u,u), pU=C¯(u¯,u¯)p_{U}=\overline{C}(\bar{u},\bar{u}), and CP(u,v)=C(v,u)C_{P}(u,v)=C(v,u) is the permutated copula of CC. Then, for 0<u0.50<u\leq 0.5, we have that:

  • (i)

    αC(u)-\infty\leq\alpha_{C}(u)\leq\infty for every C𝒞C\in\mathscr{C}; the equality holds only when either pU=0p_{U}=0 or pL=0p_{L}=0;

  • (ii)

    αC(u)=0\alpha_{C}(u)=0 if and only if pL=pUp_{L}=p_{U};

  • (iii)

    for fixed pUp_{U}, αC(u)\alpha_{C}(u) is monotonically non-increasing with respect to pLp_{L}; similarly, for fixed pLp_{L}, αC(u)\alpha_{C}(u) is monotonically non-decreasing with respect to pUp_{U};

  • (iv)

    αC(u)=αC¯(u)\alpha_{C}(u)=-\alpha_{\overline{C}}(u) for every C𝒞C\in\mathscr{C};

  • (v)

    αCP(u)=αC(u)\alpha_{C_{P}}(u)=\alpha_{C}(u) for every C𝒞C\in\mathscr{C};

  • (vi)

    if C𝒞C\in\mathscr{C} and {Cn}n\{C_{n}\}_{n\in\mathbb{N}} is a sequence of copulas such that CnCC_{n}\rightarrow C uniformly, then α(Cn)α(C)\alpha(C_{n})\rightarrow\alpha(C).

Property (i) implies that the proposed measure is potentially unbounded although it is bounded except for the unusual case pU=0p_{U}=0 or pL=0p_{L}=0. Compared with a similar measure based on the difference between pUp_{U} and pLp_{L}, our measure is advantageous in the sensitivity of detecting the asymmetry of tail probabilities for small uu; see Section 6.2 for details. Property (ii) implies that αC(u)=0\alpha_{C}(u)=0 for any 0<u0.50<u\leq 0.5 if the copula CC is radially symmetric, namely, CC¯C\equiv\overline{C}. Property (ii) is the same as an axiom of Dehgani et al. (2013) and is an extended property of Rosco and Joe (2013). Properties (iv)–(vi) are the same as the axioms of tail asymmetry presented in Section 2 of Rosco and Joe (2013). It is possible to use any function of pU/pLp_{U}/p_{L} other than the logarithm function as a measure of tail asymmetry. However one nice property of the proposed measure is property (iv) which other functions of pU/pLp_{U}/p_{L} do not have in general.

3 Limits of the proposed measure

We consider limits of the proposed measure (1) the index goes to the boundary of its domain. It follows from the expression (1) that α(0.5)=0,\alpha(0.5)=0, for any copula C𝒞C\in\mathscr{C}. Therefore we have

limu0.5α(u)=0.\lim_{u\uparrow 0.5}\alpha(u)=0. (2)

The limiting behavior of α(u)\alpha(u) as u0u\rightarrow 0 is much more intricate. To consider this problem, define

α(0)=limu0α(u),\alpha(0)=\lim_{u\downarrow 0}\alpha(u), (3)

given that the limit exists. Here we present three expressions for the limit (3).

The first expression is based on the tail dependence coefficients. Tail dependence coefficients are often used as local dependence measures of bivariate distributions. The lower-left and upper-right tail dependence coefficients of the random variables X1X_{1} and X2X_{2} are defined by

λL=limu0(F1(X1)u,F2(X2)u)uandλU=limu1(F1(X1)>u,F2(X2)>u)1u,\lambda_{L}=\lim_{u\downarrow 0}\frac{\mathbb{P}(F_{1}(X_{1})\leq u,F_{2}(X_{2})\leq u)}{u}\quad\mbox{and}\quad\lambda_{U}=\lim_{u\uparrow 1}\frac{\mathbb{P}(F_{1}(X_{1})>u,F_{2}(X_{2})>u)}{1-u},

respectively, given the limits exist. If (X1,X2)(X_{1},X_{2}) has the copula CC, the expressions for λL\lambda_{L} and λU\lambda_{U} are simplified as

λL=limu0C(u,u)uandλU=limu1C¯(u,u)1u,\lambda_{L}=\lim_{u\downarrow 0}\frac{C(u,u)}{u}\quad\mbox{and}\quad\lambda_{U}=\lim_{u\uparrow 1}\frac{\overline{C}(u,u)}{1-u}, (4)

respectively (see, e.g., Joe, 2014, Section 2.13).

Theorem 1.

Let (X1,X2)(X_{1},X_{2}) be an 2\mathbb{R}^{2}-valued random vector with the copula CC. Assume that the lower-left and upper-right tail dependence coefficients of X1X_{1} and X2X_{2} exist and are given by λL\lambda_{L} and λU\lambda_{U}, respectively. Suppose that either λL\lambda_{L} or λU\lambda_{U} is not equal to zero. Then

α(0)=log(λUλL).\alpha(0)=\log\left(\frac{\lambda_{U}}{\lambda_{L}}\right).

See Supplementary Material for the proof. Theorem 1 can be generalized by utilizing the concepts of tail orders and tail order parameters. If there exists a constant κL>0\kappa_{L}>0 and a slowly varying function L(u)\ell_{L}(u) such that C(u,u)uκLL(u)C(u,u)\sim u^{{\color[rgb]{0,0,0}\kappa_{L}}}\ell_{L}(u) (u0)(u\rightarrow 0), then κL\kappa_{L} is called the lower tail order of CC and ΥL=limu0L(u)\Upsilon_{L}=\lim_{u\downarrow 0}\ell_{L}(u) is called the lower tail order parameter of CC, where f(u)g(u)f(u)\sim g(u) (u0)(u\rightarrow 0) is defined by limu0f(u)/g(u)=1\lim_{u\downarrow 0}f(u)/g(u)=1. Similarly, the upper tail order and the upper tail order parameter of CC are defined by the lower tail order and the lower tail order parameter of the survival copula C¯\overline{C}, respectively. See Joe (2014, Section 2.16) for more details on the tail orders and tail order parameters. Using the tail orders and tail order parameters, we have the following result. The proof is given in Supplementary Material.

Theorem 2.

Let κL\kappa_{L} and κU\kappa_{U} be the lower and upper tail orders of the copula CC, respectively. Then α(0)=\alpha(0)=\infty if κL<κU\kappa_{L}<\kappa_{U} and α(0)=\alpha(0)=-\infty if κL>κU\kappa_{L}>\kappa_{U}. If κL=κU\kappa_{L}=\kappa_{U} and either of the lower tail order parameter ΥL\Upsilon_{L} or the upper tail order parameter ΥU\Upsilon_{U} of CC is not equal to zero, then α(0)=log(ΥU/ΥL)\alpha(0)={\color[rgb]{0,0,0}\log(\Upsilon_{U}/\Upsilon_{L})}.

Note that Theorem 2 with κL=κU=1\kappa_{L}=\kappa_{U}=1 reduces to Theorem 1. Theorems 1 and 2 are useful to evaluate α(0)\alpha(0) if we already know the tail dependence coefficients or tail orders and tail order parameters of a copula. If those values are not known, the following third expression for α(0)\alpha(0) could be useful.

Theorem 3.

Let (X1,X2)(X_{1},X_{2}) be an 2\mathbb{R}^{2}-valued random vector with the copula CC. Suppose that there exists ε>0\varepsilon>0 such that c(u)=d2C(t,t)/dt2|t=uc(u)=d^{2}C(t,t)/dt^{2}|_{t=u} exists in (0,ε)(1ε,1)(0,\varepsilon)\cup(1-\varepsilon,1). Assume that limu0dC(u,u)/du=limu0dC¯(u¯,u¯)/du=0\lim_{u\downarrow 0}dC(u,u)/du=\lim_{u\downarrow 0}d\overline{C}(\bar{u},\bar{u})/du=0. Then

α(0)=log(limu0c(1u)c(u)),\alpha(0)=\log\left(\lim_{u\downarrow 0}\frac{c(1-u)}{c(u)}\right),

given the limit exists.

See Supplementary Material for the proof. As will be seen in the next section, Theorem 3 can be utilized to calculate α(0)\alpha(0) for Clayton copula and Ali-Mikhail-Haq copula.

4 Values of the proposed measure for some existing copulas

In this section we discuss the values of the proposed measure α(u)\alpha(u) for some existing copulas. It is seen to be useful to plot α(u)\alpha(u) with respect to uu for comparing the probabilities of the lower [0,u]2[0,u]^{2} tail and upper [1u,1]2[1-u,1]^{2} one for the whole range of u(0,0.5]u\in(0,0.5]. See, e.g., Joe (2014) for the definitions of the existing copulas discussed in this section.

4.1 Copulas with symmetric tails

Proposition 2 implies that α(u)=0\alpha(u)=0 for any u[0,0.5]u\in[0,0.5] if C(u,u)=C¯(u¯,u¯)C(u,u)=\overline{C}(\bar{u},\bar{u}) for any u[0,0.5]u\in[0,0.5]. Such copulas include the independence copula, Gaussian copula, tt-copula, Plackett copula and FGM copula. Among well-known Archimedean copulas, Frank copula has a radially symmetric shape and therefore α(u)=0\alpha(u)=0 for any uu.

4.2 Copulas with asymmetric tails

There exist various copulas for which α(u)\alpha(u) is not equal to zero in general. Many Archimedean copulas have asymmetric tails, including Clayton copula, Gumbel copula, Ali-Mikhail-Haq copula and two-parameter BB copulas. In addition, some asymmetric extensions of Gaussian copula and tt-copula have been proposed recently. Such extensions include the skew-normal copulas and skew-tt copulas discussed in Joe (2006) and Yoshiba (2018), for which α(u)\alpha(u) is not equal to zero in general. As examples of copulas with asymmetric tails, here we discuss the values of α(u)\alpha(u) for the three well-known copulas, namely, Clayton copula, Ali-Mikhail-Haq copula and BB7 copula.

Clayton copula: Clayton copula is defined by

Ccl(u1,u2;θ)=max{u1θ+u2θ1,0}1/θ,C_{cl}(u_{1},u_{2};\theta)=\max\left\{u_{1}^{-\theta}+u_{2}^{-\theta}-1,0\right\}^{-1/\theta}, (5)

where θ[1,){0}\theta\in[-1,\infty)\setminus\{0\}.

Refer to caption Refer to caption Refer to caption
         (a)          (b)          (c)
Figure 1: Plots of α(u)\alpha(u) for: (a) Clayton copula (5) with respect to uu for θ=1\theta=1 (solid), θ=5\theta=5 (dashed), θ=10\theta=10 (dotted), and θ=20\theta=20 (dotdashed), (b) Ali-Mikhail-Haq copula (6) with respect to uu for θ=0.1\theta=0.1 (solid), θ=0.4\theta=0.4 (dashed), θ=0.7\theta=0.7 (dotted), and θ=1\theta=1 (dotdashed), and (c) Ali-Mikhail-Haq copula (6) with respect to θ\theta for u=0.01u=0.01 (solid), u=0.05u=0.05 (dashed), u=0.1u=0.1 (dotted), and u=0.3u=0.3 (dotdashed).

Figure 1(a) plots the values of α(u)\alpha(u) as a function of uu for four positive values of θ\theta. (For an intuitive understanding of the distributions of Clayton copula, see Figure S1(a) and (b) of Supplementary Material which plot random variates from Clayton copula with the two values of the parameters used in Figure 1.) As is clear from equation (2), α(0.5)=0\alpha(0.5)=0 for any θ\theta. The smaller the value of uu, the smaller the value of α(u)\alpha(u). The figure also suggests that, for a fixed value of uu, as θ\theta increases, the value of α(u)\alpha(u) approaches zero. The upper tail dependence coefficient of Clayton copula is 0 and the lower tail dependence coefficient is 21/θ2^{-1/\theta} for θ>0\theta>0 and 0 for θ0\theta\leq 0. Therefore, for θ>0\theta>0, Theorem 1 implies that α(0)=\alpha(0)=-\infty, meaning that the lower tail dependence is considerably stronger than the upper one. If θ[1,0)\theta\in[-1,0), it follows from Theorem 3 that α(0)=\alpha(0)=\infty.

Ali-Mikhail-Haq copula: Ali-Mikhail-Haq copula is of the form

C(u1,u2)=u1u21θ(1u1)(1u2),C(u_{1},u_{2})=\frac{u_{1}u_{2}}{1-\theta(1-u_{1})(1-u_{2})}, (6)

where θ[1,1]\theta\in[-1,1]. The values of α(u)\alpha(u) as a function of uu and θ\theta are exhibited in Figure 1(b) and (c), respectively. (See Figure S1(d) and (e) of Supplementary Material for plots of random variates generated from Ali-Mikhail-Haq copula with the two values of the parameters used in Figure 1(b).) Figure 1(b) suggests that α(u)\alpha(u) decreases with uu. Also it appears that, for a fixed value of uu, the greater the value of θ\theta, the smaller the value of α(u)\alpha(u). This observation can be seen more clearly in Figure 1(c) which plots the values of α(u)\alpha(u) as a function of θ\theta. Since both the lower and upper tail dependence coefficients of this copula are equal to zero, one can not apply Theorem 1 for the calculation of α(0)\alpha(0). However Theorems 2 and 3 are applicable in this case and we have a simple form α(0)=log(1θ2)\alpha(0)=\log(1-\theta^{2}).

BB7 copula: Finally, consider the BB7 copula of Joe and Hu (1996) defined by

C(u1,u2)=1[1{(1u¯1θ)δ+(1u¯2θ)δ1}1/δ]1/θ,C(u_{1},u_{2})=1-\left[1-\left\{\left(1-\overline{u}_{1}^{\theta}\right)^{-\delta}+\left(1-\overline{u}_{2}^{\theta}\right)^{-\delta}-1\right\}^{-1/\delta}\right]^{1/\theta}, (7)

where δ>0\delta>0 and θ1\theta\geq 1. Unlike the last two copulas, this model has two parameters. The parameter δ\delta controls the lower tail dependence coefficient, while θ\theta regulates the upper one. Indeed, the lower and upper tail dependence coefficients are known to be 21/δ2^{-1/\delta} and 221/θ2-2^{1/\theta}, respectively.

It follows from Theorem 1 that α(0)=log(221/θ)δ1log2\alpha(0)=\log(2-2^{1/\theta})-\delta^{-1}\log 2.

Refer to caption Refer to caption
         (a)               (b)
Figure 2: (a) Plot of α(u)\alpha(u) for BB7 copula (7) with respect to uu for (δ,θ)=(1,1.71)(\delta,\theta)=(1,1.71) (solid), (δ,θ)=(1.94,1.71)(\delta,\theta)=(1.94,1.71) (dashed), (δ,θ)=(1,7.27)(\delta,\theta)=(1,7.27) (dotted), and (δ,θ)=(1.94,7.27)(\delta,\theta)=(1.94,7.27) (dotdashed). (b) Contour plot of α(0.01)\alpha(0.01) (solid) and plot of α(0)=0\alpha(0)=0 (dashed) with respect to (δ,θ)(\delta,\theta).

Figure 2 displays a plot of α(u)\alpha(u) with respect to uu for four selected values of (δ,θ)(\delta,\theta) and that of α(u)\alpha(u) with respect to (δ,θ)(\delta,\theta) for u=0.01u=0.01. Note that, in Figure 2(a), δ=1\delta=1 and δ=1.94\delta=1.94 imply that the lower tail dependence coefficients are around 0.5 and 0.9, respectively, while θ=1.71\theta=1.71 and θ=7.27\theta=7.27 suggest that the upper tail dependence coefficients are about 0.5 and 0.7, respectively. (See also Figure S1(g)–(h) of Supplementary Material for plots of random variates from BB7 copula (7) with the three combinations of the parameters in Figure 2(a).) Figure 2(a) suggests that, when both the lower and upper tail dependence coefficients are around 0.5, the values of α(u)\alpha(u) are close to zero for any uu. When the difference between the lower and upper tail dependence coefficients is large, α(u)\alpha(u) appears to be monotonic with respect to uu. It can be seen from Figure 2(b) that the values of α(0.01)\alpha(0.01) monotonically decreases as δ\delta increases. Also, α(0.01)\alpha(0.01) monotonically increases with θ\theta. The two contours α(0.01)=0\alpha(0.01)=0 and α(0)=0\alpha(0)=0 show somewhat similar shapes, implying that α(0.01)=0\alpha(0.01)=0 is a reasonable approximation to α(0)=0\alpha(0)=0.

5 Two sample analogues of α(u)\alpha(u)

In practice, it is often the case that the form of the copula C(u1,u2)C(u_{1},u_{2}) underlying data is not known. In such a case, we need to estimate α(u)\alpha(u) based on the data. In Sections 5.1S1.6 we propose a sample analogue of α(u)\alpha(u) based on a sample from the copula. Section 5.4 presents a sample analogue of α(u)\alpha(u) based on a sample from a distribution on 2\mathbb{R}^{2}. A comparison between the two proposed sample analogues of α(u)\alpha(u) is discussed via a simulation study in Section 5.5.

5.1 A sample analogue of α(u)\alpha(u) based on a sample from a copula

A sample analogue of α(u)\alpha(u) based on a sample from a copula is defined as follows.

Definition 2.

Let (U11,U21),,(U1n,U2n)(U_{11},U_{21}),\ldots,(U_{1n},U_{2n}) be a random sample from a copula. Then we define a sample analogue of α(u)\alpha(u) by

α^(u)=log(TU(u)TL(u)),\hat{\alpha}(u)=\log\left(\frac{T_{U}(u)}{T_{L}(u)}\right),

where

TL(u)=1ni=1n𝟏(U1iu,U2iu),T_{L}(u)=\frac{1}{n}\sum_{i=1}^{n}\bm{1}(U_{1i}\leq u,U_{2i}\leq u),
TU(u)=1ni=1n𝟏(U1i1u,U2i1u),T_{U}(u)=\frac{1}{n}\sum_{i=1}^{n}\bm{1}(U_{1i}\geq 1-u,U_{2i}\geq 1-u),

and 𝟏()\bm{1}(\cdot) is an indicator function, i.e., 𝟏(A)=1\bm{1}(A)=1 if AA is true and 𝟏(A)=0\bm{1}(A)=0 otherwise.

In Sections 5.1S1.6, we assume that (U11,U21),(U_{11},U_{21}),,\ldots,(U1n,U2n)(U_{1n},U_{2n}) is an iid sample from the copula C(u1,u2)C(u_{1},u_{2}). For iid 2\mathbb{R}^{2}-valued random vectors (X11,X21)(X_{11},X_{21}),,,\ldots, (X1n,X2n)(X_{1n},X_{2n}), if the margins of X11X_{11} and X21X_{21} are known to be F1F_{1} and F2F_{2}, respectively, then α^(u)\hat{\alpha}(u) can be obtained by replacing (U1j,U2j)(U_{1j},U_{2j}) by (F1(X1j),F2(X2j))(F_{1}(X_{1j}),F_{2}(X_{2j})) (j=1,,n)(j=1,\ldots,n).

The goal of this subsection is to investigate some properties of α^(u)\hat{\alpha}(u). To achieve this, we first show the following lemma. See Supplementary Material for the proof.

Lemma 1.

For 0<u,v0.5,0<u,v\leq 0.5, we have the following:

𝔼[TL(u)]=Cu,𝔼[TU(u)]=C¯u¯,var[TL(u)]=1nCu(1Cu),\mathbb{E}\left[T_{L}(u)\right]=C_{u},\quad\mathbb{E}\left[T_{U}(u)\right]=\overline{C}_{\bar{u}},\quad{\rm var}\left[T_{L}(u)\right]=\frac{1}{n}C_{u}(1-C_{u}),
var[TU(u)]=1nC¯u¯(1C¯u¯),cov[TL(u),TL(v)]=1nCuv(1Cuv),{\rm var}\left[T_{U}(u)\right]=\frac{1}{n}\overline{C}_{\bar{u}}(1-\overline{C}_{\bar{u}}),\quad{\rm cov}\left[T_{L}(u),T_{L}(v)\right]=\frac{1}{n}C_{u\wedge v}(1-C_{u\vee v}),
cov[TL(u),TU(v)]=1nCuvC¯u¯v¯,cov[TU(u),TU(v)]=1nC¯u¯v¯(1C¯u¯v¯),{\rm cov}\left[T_{L}(u),T_{U}(v)\right]=-\frac{1}{n}C_{u\wedge v}\overline{C}_{\bar{u}\wedge\bar{v}},\quad{\rm cov}\left[T_{U}(u),T_{U}(v)\right]=\frac{1}{n}\overline{C}_{\bar{u}\vee\bar{v}}(1-\overline{C}_{\bar{u}\wedge\bar{v}}),

where uv=min(u,v)u\wedge v=\min(u,v), uv=max(u,v)u\vee v=\max(u,v), and Cw=C(w,w)C_{w}=C(w,w).

This lemma implies that α^(u)\hat{\alpha}(u) is a consistent estimator of α(u)\alpha(u). Applying this lemma, we obtain the following asymptotic result. The proof is given in Supplementary Material.

Theorem 4.

Define

𝔸n(u)=n{α^(u)α(u)},0<u0.5.\mathbb{A}_{n}(u)=\sqrt{n}\left\{\hat{\alpha}(u)-\alpha(u)\right\},\quad 0<u\leq 0.5.

Then, as nn\rightarrow\infty, {𝔸n(u)| 0<u0.5}\{\mathbb{A}_{n}(u)\,|\,0<u\leq 0.5\} converges weakly to a centered Gaussian process with covariance function

σ(u,v)𝔼[𝔸n(u)𝔸n(v)]=C(uv,uv)+C¯(u¯v¯,u¯v¯)C(uv,uv)C¯(u¯v¯,u¯v¯).{\color[rgb]{0,0,0}\sigma(u,v)\equiv\mathbb{E}[\mathbb{A}_{n}(u)\mathbb{A}_{n}(v)]=\frac{C(u\vee v,u\vee v)+\overline{C}(\bar{u}\wedge\bar{v},\bar{u}\wedge\bar{v})}{C(u\vee v,u\vee v)\cdot\overline{C}(\bar{u}\wedge\bar{v},\bar{u}\wedge\bar{v})}.} (8)

We note that the covariance function (8) is monotonically decreasing with respect to uvu\vee v. If uv=0.5u\vee v=0.5, then the covariance function (8) reaches the minimum value 2/C(0.5,0.5)2/C(0.5,0.5).

5.2 Interval estimation based on α^(u)\hat{\alpha}(u)

An asymptotic interval estimator of α(u)\alpha(u) can be obtained by applying the asymptotic results obtained in the previous subsection. Theorem 4 implies that, for fixed u(0,0.5]u\in(0,0.5],

n{α^(u)α(u)}𝑑N(0,σ2(u))(n),\sqrt{n}\left\{\hat{\alpha}(u)-\alpha(u)\right\}\xrightarrow{d}N(0,\sigma^{2}(u))\quad(n\rightarrow\infty),

where σ2(u)=σ(u,u)\sigma^{2}(u)=\sigma(u,u) and σ(u,u)\sigma(u,u) is defined as in equation (8). Since σ(u)\sigma(u) includes the copula CC which is usually not known in practice, we use an estimator of σ(u)\sigma(u) defined by

σ^(u)=TL(u)+TU(u)TL(u)TU(u).\hat{\sigma}(u)=\sqrt{\frac{T_{L}(u)+T_{U}(u)}{T_{L}(u)\cdot T_{U}(u)}}.

It follows from Lemma 1 that σ^(u)a.s.σ(u)\hat{\sigma}(u)\xrightarrow{a.s.}\sigma(u) as nn\rightarrow\infty. Then we have n{α^(u)α(u)}/σ^(u)𝑑N(0,1)\sqrt{n}\{\hat{\alpha}(u)-\alpha(u)\}/\hat{\sigma}(u)\xrightarrow{d}N(0,1) as nn\rightarrow\infty. Hence a 100(1p)100(1-p)% nonparametric asymptotic confidence interval for α(u)\alpha(u) is

α^(u)zp/2σ^(u)nα(u)α^(u)+zp/2σ^(u)n,\hat{\alpha}(u)-\frac{z_{p/2}\hat{\sigma}(u)}{\sqrt{n}}\leq\alpha(u)\leq\hat{\alpha}(u)+\frac{z_{p/2}\hat{\sigma}(u)}{\sqrt{n}}, (9)

where zp/2z_{p/2} satisfies (Zzp/2)=p/2\mathbb{P}(Z\geq z_{p/2})=p/2, where ZN(0,1)Z\sim N(0,1) and 0<p<10<p<1.

Notice that the asymptotic confidence interval (9) is a pointwise one for a fixed value of uu. If the interest of statistical analysis is to construct an asymptotic confidence band for α(u)\alpha(u) for a range of uu, one can adopt Bonferroni correction. This can be done by replacing pp by p/np/n in the equation (9) for u𝕌u\in\mathbb{U}, where 𝕌={min(max(u1i,u2i),max(1u1i,1u2i))}i=1n\mathbb{U}=\{\min(\max(u_{1i},u_{2i}),\max(1-u_{1i},1-u_{2i}))\}_{i=1}^{n}. Therefore the asymptotic confidence band for α(u)\alpha(u) with Bonferroni correction is

{[α^(u)zp/(2n)σ^(u)n,α^(u)+zp/(2n)σ^(u)n]|uminuumax},\left\{\left[\hat{\alpha}(u)-\frac{z_{p/(2n)}\hat{\sigma}(u)}{\sqrt{n}},\hat{\alpha}(u)+\frac{z_{p/(2n)}\hat{\sigma}(u)}{\sqrt{n}}\right]\,\Biggl{|}\,u_{\min}\leq u\leq u_{\max}\right\}, (10)

where umin=min𝕌u_{\min}=\min\mathbb{U} and umax=max𝕌u_{\max}=\max\mathbb{U}. Since α^(u)\hat{\alpha}(u) and σ^(u)\hat{\sigma}(u) are step functions, it suffices to evaluate the bounds of confidence intervals only for u𝕌u\in\mathbb{U}.

5.3 Hypothesis testing based on α^(u)\hat{\alpha}(u)

Some hypothesis tests can be established based on α^(u)\hat{\alpha}(u). For a given value of u=u0u=u_{0}, one can carry out a hypothesis test to test H0:α(u0)=α0H_{0}:\alpha(u_{0})=\alpha_{0} against H1:α(u0)α0H_{1}:\alpha(u_{0})\neq\alpha_{0} by using the asymptotic confidence interval (9). Similarly, an one-sided test for the alternative hypothesis H1:α(u0)>α0H_{1}:\alpha(u_{0})>\alpha_{0} or H1:α(u0)<α0H_{1}:\alpha(u_{0})<\alpha_{0} can be derived by modifying the asymptotic confidence interval (9).

If the interest of analysis is to evaluate the values of α^(u)\hat{\alpha}(u) for multiple values of uu, one can consider the test H0:α(u)=α0(u)H_{0}:\alpha(u)=\alpha_{0}(u) against H1:α(u)α0(u)H_{1}:\alpha(u)\neq\alpha_{0}(u) for {u;u=u1,,um}\{u;u=u_{1},\ldots,u_{m}\}. One example of such tests is based on a asymptotic confidence band based on Bonferroni correction (10). However, since the confidence band based on Bonferroni correction is known to be conservative, especially, for dependent hypotheses, the test based on Bonferroni correction (10) is not powerful in general. Alternatively, the following result can be used to present a test for the multiple values of uu. See Supplementary Material for the proof.

Theorem 5.

Let 𝐚=n{α^(u1)α(u1),,α^(um)α(um)}T\bm{a}=\sqrt{n}\{\hat{\alpha}(u_{1})-\alpha(u_{1}),\ldots,\hat{\alpha}(u_{m})-\alpha(u_{m})\}^{T} and u1<<umu_{1}<\cdots<u_{m}. Suppose

𝚺^=(σ^2(u1)σ^(u1,u2)σ^(u1,um)σ^(u1,u2)σ^2(u2)σ^(u2,um)σ^(u1,um)σ^(u2,um)σ^2(um)),\hat{\bm{\Sigma}}=\left(\begin{array}[]{cccc}\hat{\sigma}^{2}(u_{1})&\hat{\sigma}(u_{1},u_{2})&\ldots&\hat{\sigma}(u_{1},u_{m})\\ \hat{\sigma}(u_{1},u_{2})&\hat{\sigma}^{2}(u_{2})&\ldots&\hat{\sigma}(u_{2},u_{m})\\ \vdots&\vdots&\ddots&\vdots\\ \hat{\sigma}(u_{1},u_{m})&\hat{\sigma}(u_{2},u_{m})&\ldots&\hat{\sigma}^{2}(u_{m})\end{array}\right),

where σ^2(ui)=σ^(ui,ui)\hat{\sigma}^{2}(u_{i})=\hat{\sigma}(u_{i},u_{i}) and σ^(ui,uj)={TL(uj)+TU(uj)}/{TL(uj)TU(uj)}\hat{\sigma}(u_{i},u_{j})=\{T_{L}(u_{j})+T_{U}(u_{j})\}/\{T_{L}(u_{j})T_{U}(u_{j})\} (ij)(i\leq j). Assume that 𝚺^\hat{\bm{\Sigma}} is invertible. Then

𝒂T𝚺^1𝒂𝑑χ2(m)as n,\bm{a}^{T}\hat{\bm{\Sigma}}^{-1}\bm{a}\xrightarrow{d}\chi^{2}(m)\quad\mbox{as }n\rightarrow\infty,

where χ2(m)\chi^{2}(m) denotes the chi-squared distribution with mm degrees of freedom.

Substituting α0(u)\alpha_{0}(u) into α(u)\alpha(u) in the test statistic 𝒂T𝚺^1𝒂\bm{a}^{T}\hat{\bm{\Sigma}}^{-1}\bm{a}, the null hypothesis α(u)=α0(u)\alpha(u)=\alpha_{0}(u) is rejected for a large value of 𝒂T𝚺^1𝒂\bm{a}^{T}\hat{\bm{\Sigma}}^{-1}\bm{a}. In order that 𝚺^\hat{\bm{\Sigma}} becomes invertible, the values of (u1,,um)(u_{1},\ldots,u_{m}) need to be selected such that TU(ui)<TU(ui+1)T_{U}(u_{i})<T_{U}(u_{i+1}) and/or TL(ui)<TL(ui+1)T_{L}(u_{i})<T_{L}(u_{i+1}) hold for any 1im11\leq i\leq m-1.

5.4 A sample analogue of α(u)\alpha(u) based on a sample from a distribution on 2\mathbb{R}^{2}

The sample analogue of α(u)\alpha(u) given in Definition 2 can be calculated on the assumption that the margins of the 2\mathbb{R}^{2}-valued random vector are known. Here we discuss the case in which margins are unknown and empirical distributions are adopted as the margins.

Definition 3.

Let (X11,X21),,(X1n,X2n)(X_{11},X_{21}),\ldots,(X_{1n},X_{2n}) be 2\mathbb{R}^{2}-valued random vectors. Then we define a sample analogue of α(u)\alpha(u) by

α^(u)=log(TU(u)TL(u)),\hat{\alpha}^{*}(u)=\log\left(\frac{T^{*}_{U}(u)}{T^{*}_{L}(u)}\right),

where

TL(u)=1ni=1n𝟏(F^1(X1i)u,F^2(X2i)u),T^{*}_{L}(u)=\frac{1}{n}\sum_{i=1}^{n}\bm{1}(\hat{F}_{1}(X_{1i})\leq u,\hat{F}_{2}(X_{2i})\leq u),
TU(u)=1ni=1n𝟏(F^1(X1i)1u,F^2(X2i)1u),T^{*}_{U}(u)=\frac{1}{n}\sum_{i=1}^{n}\bm{1}(\hat{F}_{1}(X_{1i})\geq 1-u,\hat{F}_{2}(X_{2i})\geq 1-u),
F^j(Xji)=1n+1k=1n𝟏(XjkXji),j=1,2.\hat{F}_{j}(X_{ji})=\frac{1}{n+1}{\color[rgb]{0,0,0}\sum_{k=1}^{n}}\bm{1}(X_{jk}\leq X_{ji}),\quad j=1,2. (11)

Note that the denominator of the empirical distribution function (11) is defined by n+1n+1 rather than nn in order to avoid positive bias of α^(u)\hat{\alpha}(u).

The authors have not yet obtained the asymptotic distribution for α^(u)\hat{\alpha}^{*}(u). However the following results are available regarding TU(u)T_{U}^{*}(u) and TL(u)T_{L}^{*}(u). The proof is straightforward from Fermanian et al. (2004), Tsukahara (2005) and Segers (2012) and therefore omitted.

Proposition 3.

Let (X11,X21),,(X1n,X2n)(X_{11},X_{21}),\ldots,(X_{1n},X_{2n}) be iid random vectors with the copula C(u,v)C(u,v) and the continuous margins. Assume that C(u,v)C(u,v) is differentiable with continuous ii-th partial derivatives (i=1,2)(i=1,2). Then, as nn\rightarrow\infty,

n{TL(u)C(u,u)}𝑑DC(u),\sqrt{n}\left\{T^{*}_{L}(u)-C(u,u)\right\}\xrightarrow{d}D^{C}(u),
n{TU(u)C¯(u¯,u¯)}𝑑DC¯(u¯),\sqrt{n}\left\{T^{*}_{U}(u)-\overline{C}(\bar{u},\bar{u})\right\}\xrightarrow{d}D^{\overline{C}}(\bar{u}),

where

DC(u)=U(u,u)C(u1,u)u1|u1=uU(u,1)C(u,u2)u2|u2=uU(1,u),D^{C}(u)=U(u,u)-\frac{\partial C(u_{1},u)}{\partial u_{1}}\biggr{|}_{u_{1}=u}U(u,1)-\frac{\partial C(u,u_{2})}{\partial u_{2}}\biggr{|}_{u_{2}=u}U(1,u),

UU is a centered Gaussian process with covariance function

𝔼[U(u1,u2)U(v1,v2)]=C(u1v1,u2v2)C(u1,u2)C(v1,v2),\mathbb{E}[U(u_{1},u_{2})U(v_{1},v_{2})]=C(u_{1}\wedge v_{1},u_{2}\wedge v_{2})-C(u_{1},u_{2})C(v_{1},v_{2}),

and uvu\wedge v is defined as in Lemma 1.

Confidence intervals for α(u)\alpha^{*}(u) can be numerically constructed using the bootstrap method. Hypothesis tests can also be established based on the bootstrap confidence intervals. It should be noted that, in order to calculate α(u)\alpha^{*}(u) based on bootstrap samples, F^1\hat{F}_{1} and F^2\hat{F}_{2} in (11) should be calculated based on each bootstrap sample. If F^1\hat{F}_{1} and F^2\hat{F}_{2} are calculated from the original data, the bootstrap confidence intervals become similar to the asymptotic confidence intervals (9) for large nn.

5.5 Simulation study

In order to compare the performance of the two proposed sample analogues of α(u)\alpha(u) for a large sample size, we consider the following cumulative distribution function

F(x1,x2)=Ccl(F1(x1),F2(x2);20),<x1,x2<,F(x_{1},x_{2})=C_{cl}(F_{1}(x_{1}),F_{2}(x_{2});20),\quad-\infty<x_{1},x_{2}<\infty, (12)

where Fj(x)F_{j}(x) is the cumulative distribution function of the standard Cauchy distribution, i.e., Fj(x)=0.5+π1arctanxF_{j}(x)=0.5+\pi^{-1}\arctan x, and Ccl(u1,u2;θ)C_{cl}(u_{1},u_{2};\theta) denotes the Clayton copula (5).

Refer to caption
Figure 3: Plot of α^(u)α(u)\hat{\alpha}(u)-\alpha(u) (solid, red), α^(u)α(u)\hat{\alpha}^{*}(u)-\alpha(u) (dashed, purple), the lower and upper bounds of 90%\% asymptotic confidence intervals of α^(u)α(u)\hat{\alpha}(u)-\alpha(u) (dotdashed, black), and the lower and upper bounds of 90% bootstrap confidence intervals α^(u)α(u)\hat{\alpha}^{*}(u)-\alpha(u) (dotted, blue) obtained from a sample of size 10000 from the distribution (12).

Figure 3 plots the values of α^(u)α(u)\hat{\alpha}(u)-\alpha(u), α^(u)α(u)\hat{\alpha}^{*}(u)-\alpha(u) and their bounds of 90% confidence intervals for a sample of size n=10000n=10000 from the distribution (12). See also Figure 1(a) for the plot of α(u)\alpha(u). For the calculations of α^(u)\hat{\alpha}(u) and its confidence intervals (9), the sample {(x1i,x2i)}\{(x_{1i},x_{2i})\} is transformed into the copula sample {(u1i,u2i)}\{(u_{1i},u_{2i})\} via uji=Fj(xji)u_{ji}=F_{j}(x_{ji}), where FjF_{j} is the true margin (j=1,2)(j=1,2). The confidence intervals of α(u)\alpha^{*}(u) are calculated using the basic bootstrap method based on 999999 resamples of size 1000010000; see the last paragraph of Section 5.4 for details. The minimum value of uu in the plot is defined as umin=min{u(0,0.5];TL(u),TU(u),TL(u),u_{\rm min}=\min\{u\in(0,0.5];T_{L}(u),T_{U}(u),T_{L}^{*}(u), TU(u)30}0.01T_{U}^{*}(u)\geq 30\}\simeq 0.01 in order that the asymptotic theory is applicable.

The figure suggests that when uu is around 0.2 or greater, the performance of both α^(u)\hat{\alpha}(u) and α^(u)\hat{\alpha}^{*}(u) seems satisfactory. For u0.2u\leq 0.2, the difference between the sample analogues and the true value increases with uu in general. It appears that the 90% confidence intervals of both α^(u)\hat{\alpha}(u) and α^(u)\hat{\alpha}^{*}(u) are generally narrow if uu is around 0.2 or more. For u0.2u\leq 0.2, the smaller the value of uu, the wider the ranges of the confidence intervals of both α^(u)\hat{\alpha}(u) and α^(u)\hat{\alpha}^{*}(u). Interestingly, the confidence intervals of α^(u)\hat{\alpha}^{*}(u) are narrower than those of α^(u)\hat{\alpha}(u) in most of the plotted range of uu. This tendency is particularly obvious in the range [0.2,0.5][0.2,0.5], where the confidence intervals of α^(u)\hat{\alpha}^{*}(u) are much narrower than those of α^(u)\hat{\alpha}(u).

6 Comparison with other measures

6.1 Comparison with existing measures

In this section we compare our measure with other copula-based measures of tail asymmetry. Rosco and Joe (2013) proposed three measures of tail asymmetry. One of their measures based on the distance between a copula CC and its survival copula is defined by

ς3=sup(u1,u2)[0,1]2{|C(u1,u2)C¯(u¯1,u¯2)|}.\varsigma_{3}=\sup_{(u_{1},u_{2})\in[0,1]^{2}}\left\{\left|C(u_{1},u_{2})-\overline{C}(\bar{u}_{1},\bar{u}_{2})\right|\right\}. (13)

This measure has also been proposed by Dehgani et al. (2013) as a limiting case of a measure of radial asymmetry for bivariate random variables.

Our measure (1) has some similarities to and differences from the measure (13). Similarities include that both are functions of a copula CC and its survival function. Also, both measures satisfy Properties (ii), (v) and (vi) of Proposition 2.

However there are considerable differences between the two measures (1) and (13). First, the domains of a copula the two measures evaluate are different. The measure (13) is a global measure in the sense that the whole domain of the copula is taken into account to evaluate the value of the measure, while our measure (1) is a local measure which focuses on squared subdomains of the copula. By choosing the value of the index uu, our measure (1) enables us to choose the subdomain of a copula which analysts are interested in. However the prescription for selecting the value of uu is not always straightforward and the choice of the index uu could influence the results of analysis. The index-free measure (13) does not have such a problem. However the supremum value of this measure is not necessarily attained in the tails of the distribution and the value of the measure might not reflect the tail probabilities if u10.5u_{1}\geq 0.5 or u20.5u_{2}\geq 0.5. Also, because of its locality, computations associated with our measure (1) are very fast.

Also there are differences between the two measures (1) and (13) in terms of properties. Our measure (1) satisfies all the properties of (i)–(vi) of Proposition 2 which include four (out of five) axioms of Rosco and Joe (2013). However this measure does not satisfy one of the axioms, i.e., axiom (i), of Rosco and Joe (2013) and therefore the value of the measure could be unbounded for special cases. The measure (13) also satisfies four axioms of Rosco and Joe (2013) including the axiom (i). On the other hand, the measure (13) does not satisfy their axiom (iii) which is equivalent to Property (iv) of Proposition 2, implying that the measure (13) does not distinguish which tail probability is greater than the other one.

The other two measures of Rosco and Joe (2013) are derived through different approaches. For a bivariate random vector (U1,U2)(U_{1},U_{2}) from a copula, the two measures are based on the moments or quantile function of the univariate random variable U1+U21U_{1}+U_{2}-1. Therefore these measures are essentially different from ours which is based on the joint distribution of the bivariate random vector (U1,U2)(U_{1},U_{2}).

Another copula-based measure for tail asymmetry has been proposed by Krupskii (2017). It is defined by

ϱK(a,u)=ϱL(a,u)ϱU(a,u),\varrho_{K}(a,u)=\varrho_{L}(a,u)-\varrho_{U}(a,u), (14)

where 0<u0.50<u\leq 0.5, aa is a weighting function,

ϱL(a,u)=cor[a(1U1u),a(1U2u)|U1<u,U2<u],\varrho_{L}(a,u)={\rm cor}\left[\left.a\left(1-\frac{U_{1}}{u}\right),a\left(1-\frac{U_{2}}{u}\right)\right|U_{1}<u,U_{2}<u\right],
ϱU(a,u)=cor[a(11U1u),a(11U2u)|U1>1u,U2>1u].\varrho_{U}(a,u)={\rm cor}\left[\left.a\left(1-\frac{1-U_{1}}{u}\right),a\left(1-\frac{1-U_{2}}{u}\right)\right|U_{1}>1-u,U_{2}>1-u\right]{\color[rgb]{0,0,0}.}

If a(x)=xa(x)=x, the measure (14) reduces to the measure discussed by Nikoloulopoulos et al. (2012) and Dobrić et al. (2013). Properties of each term of the measure (14) have been investigated by Krupskii and Joe (2015).

The measure (14) is related to ours in the sense that the values of their measures are calculated from the subdomain of a copula indexed by the truncation parameter. However the measure (14) is based on Spearman’s rhos or correlation coefficients of a truncated copula, and therefore the interpretation of the values of the measure (14) is essentially different from ours. A nice property of the measure (14) is that the weights of tails can be controlled through the weight function aa. Therefore this measure can be a useful measure of tail asymmetry if the weight function is appropriately defined.

6.2 An alternative measure based on tail probabilities

The proposed measure α(u)\alpha(u) is a function of the lower and upper tail probabilities. Here we briefly consider another measure of comparison between the two tail probabilities.

Definition 4.

Let (X1,X2)(X_{1},X_{2}), F1F_{1} and F2F_{2} be defined as in Definition 1. Then we define a measure of comparison between the lower-left and upper-right tail probabilities of (X1,X2)(X_{1},X_{2}) by

β(u)=\displaystyle\beta(u)= uκ{(F1(X1)>1u,F2(X2)>1u)\displaystyle\ u^{-\kappa}\bigl{\{}\mathbb{P}(F_{1}(X_{1})>1-u,F_{2}(X_{2})>1-u)
(F1(X1)u,F2(X2)u)},0<u0.5,\displaystyle-\mathbb{P}(F_{1}(X_{1})\leq u,F_{2}(X_{2})\leq u)\bigr{\}},\quad 0<u\leq 0.5,

where the index κ\kappa is given by κ1\kappa\geq 1.

If (X1,X2)(X_{1},X_{2}) has the copula CC, the expression of β(u)\beta(u) can be simplified to

β(u)=C¯(u¯,u¯)C(u,u)uκ.\beta(u)=\frac{\overline{C}({\color[rgb]{0,0,0}\bar{u},\bar{u}})-C(u,u)}{u^{\kappa}}.

Hence this measure is based on the difference between the lower and upper tail probabilities of the copula as well as the value of κ\kappa which could be decided based on the lower and upper tail orders.

The measure β(u)\beta(u) is another simple measure to compare the lower-right and upper-left tail probabilities. Properties (ii)–(vi) of Proposition 2 hold for β(u)\beta(u). As for the range of the measure β(u)\beta(u) related to Property (i), it can be seen 1β(u)1-1\leq\beta(u)\leq 1 if κ=1\kappa=1 and β(u)-\infty\leq\beta(u)\leq\infty otherwise.

One big difference between α(u)\alpha(u) and β(u)\beta(u) is that, when considering the values of the measures for u0u\simeq 0, α(u)\alpha(u) and β(u)\beta(u) possibly lead to different conclusions. As an example of this, consider the Clayton copula (5) with θ<0\theta<0, which is said to have asymmetric tails (see Figure S1 of Supplementary Material for a plot of random variates from Clayton copula with θ=0.3\theta=-0.3). For this model, the measure α(u)\alpha(u) for u0u\simeq 0 takes large positive values because α(0)=\alpha(0)=\infty. However, the values of β(u)\beta(u) with κ=1\kappa=1 for u0u\simeq 0 are close to zero since β(0)=0\beta(0)=0. This fact about β(u)\beta(u) is reasonable in one sense, but one might argue that β(u)\beta(u) does not capture the asymmetry of tail probabilities appropriately. Although this problem can be solved by selecting a different value of κ\kappa, the selection of κ\kappa, which influences the conclusion of analysis, appears difficult in practical situations where CC is unknown.

7 Example

As an application of the proposed measure, we consider a dataset of daily returns of two stock indices. The dataset is taken from historical data in Yahoo Finance, available at https://finance.yahoo.com/quote/%5EGSPC/history/ and https://finance.yahoo.
com/quote/%5EN225/history/
. We consider stock daily returns of S&P500 and Nikkei225 observed from the 1st of April, 2008 until the 31st of March, 2019, inclusive. We fit the autoregressive-generalized autoregressive conditional heteroscedastic model AR(1)-GARCH(1,1) to each of the stock daily returns using ugarchfit in ‘rugarch’ package in R (R Core Team, 2020; Ghalanos, 2020). The Student tt-distribution is used as the conditional density for the innovations. We consider the residuals {(x1i,x2i)}i=1n\{(x_{1i},x_{2i})\}_{i=1}^{n} (n=2605)(n=2605) of the fitted AR(1)-GARCH(1,1), where x1ix_{1i} and x2ix_{2i} are the residuals of S&P500 and Nikkei225, respectively. The residuals show unexpected changes in daily return which are not explained by the model; if the joint plunging probability is higher than the joint soaring probability, then the proposed measures α(u)\alpha(u) is supposed to be negative.

We discuss α^(u)\hat{\alpha}(u) defined in Definition 1 and α^(u)\hat{\alpha}^{*}(u) defined in Definition 3. In order to obtain the copula sample {(u1i,u2i)}\{(u_{1i},u_{2i})\} for α^(u)\hat{\alpha}(u), we transform the residuals {(x1i,x2i)}\{(x_{1i},x_{2i})\} via (u1i,u2i)=(F1(x1i),F2(x2i))(u_{1i},u_{2i})=(F_{1}(x_{1i}),F_{2}(x_{2i})), where F1F_{1} and F2F_{2} are the cumulative distribution functions of Student tt-distribution estimated using the maximum likelihood method. We assume, though not mathematically precise, that F1F_{1} and F2F_{2} are known.

Refer to caption Refer to caption Refer to caption
       (a)        (b)           (c)
Refer to caption Refer to caption
         (d)          (e)
Figure 4: Plots of the sample {(u1i,u2i)}i=12605\{(u_{1i},u_{2i})\}_{i=1}^{2605} which the residuals are transformed into via the cumulative distribution functions of: (a) Student tt-distribution and (b) empirical distribution. (c) Plot of ϱK(a,u)-\varrho_{K}(a,u), a modified version of the measure (14) of Krupskii (2017), with: a(x)=xa(x)=x (solid, black), a(x)=x2a(x)=x^{2} (dashed, blue), and a(x)=x4a(x)=x^{4} (dotted, red). Plots of the proposed measure (black) and its asymptotic or bootstrap 90% confidence intervals (gray) for: (d) α^(u)\hat{\alpha}(u) and (e) α^(u)\hat{\alpha}^{*}(u).

Figure 4(a) plots the sample {(u1i,u2i)}\{(u_{1i},u_{2i})\} which the residuals are transformed into via the cumulative distribution functions of Student tt. The values of α^(u)\hat{\alpha}(u) calculated from the sample and their 90% asymptotic confidence intervals (9) are displayed in Figure 4(d). In Figure 4(d), the minimum value of uu is defined as umin=min{u(0,0.5];TL(u),TU(u)30}0.0526u_{\rm min}=\min\{u\in(0,0.5];T_{L}(u),T_{U}(u)\geq 30\}\simeq 0.0526 in order that the asymptotic theory is applicable.

For the calculation of α^(u)\hat{\alpha}^{*}(u), we use the empirical distribution functions (11) to transform the residuals {(x1i,x2i)}\{(x_{1i},x_{2i})\} into the copula sample {(u1i,u2i)}\{(u_{1i},u_{2i})\}. The transformed sample is displayed in Figure 4(b). The values of α^(u)\hat{\alpha}^{*}(u) calculated from the sample are plotted in Figure 4(d). The same frame also plots the 90% confidence intervals based on 999999 resamples of size 26052605 using the basic bootstrap method. The minimum value of uu in the plot is umin=min{u(0,0.5];TL(u),TU(u)30}0.0453u^{*}_{\rm min}=\min\{u\in(0,0.5];T_{L}^{*}(u),T_{U}^{*}(u)\geq 30\}\simeq 0.0453.

Figure 4(a) and (b) suggest that there are more observations in the lower-left [0,u]2[0,u]^{2} tail than the upper-right [1u,1]2[1-u,1]^{2} one for u0.1u\simeq 0.1. However it does not seem immediately clear from these data plots whether there is significant difference between the two tail probabilities for u0.1u\simeq 0.1 as well as for general uu. To solve this problem, Figure 4(d) and (e) showing the values of α(u)\alpha(u) and α(u)\alpha^{*}(u), respectively, are helpful. Indeed Figure 4(d) and (e) suggest that α(u)\alpha(u) and α(u)\alpha^{*}(u) are negative in most areas of the domain of uu, suggesting that the lower tail probability is greater than the upper one for most values of u(umin,0.5]u\in(u_{\rm min},0.5]. In particular the two frames imply the general tendency that, for u0.2u\leq 0.2, α(u)\alpha(u) and α(u)\alpha^{*}(u) decrease with uu. The asymptotic and bootstrap 90% confidence intervals do not include 0 for u0.19u\leq 0.19 in Figure 4(d) and for u0.14u\leq 0.14 in Figure 4(e). Hence, when considering the tests H0:α(u0)=0H_{0}:\alpha({\color[rgb]{0,0,0}u_{0}})=0 against H1:α(u0)0H_{1}:\alpha({\color[rgb]{0,0,0}u_{0}})\neq 0 for a fixed value of u0[umin,0.14]{\color[rgb]{0,0,0}u_{0}}\in[u_{\rm min},0.14] based on the two 90% confidence intervals, both tests reject the null hypothesis at a significance level of 0.1. This implies that the lower [0,u0)2[0,{\color[rgb]{0,0,0}u_{0}})^{2} tail probability is significantly greater than the upper (1u0,1]2(1-{\color[rgb]{0,0,0}u_{0}},1]^{2} one for u0[umin,0.14]{\color[rgb]{0,0,0}u_{0}}\in[u_{\rm min},0.14]. On the other hand, when u0{\color[rgb]{0,0,0}u_{0}} is greater than 0.21, both 90% confidence intervals include 0 and therefore each of the tests for a nominal size of 0.1 accepts the null hypothesis H0H_{0}. There is disagreement in conclusions between the tests based on the two 90% confidence intervals in some areas of uu in (0.14,0.21](0.14,0.21].

As seen in the discussion above, both α^(u)\hat{\alpha}(u) and α^(u)\hat{\alpha}^{*}(u) show similar tendencies in general. Actually, the two data plots given in Figure 4(a) and (b) look similar at the first glance. However Figure 4(d) and (e) reveal that there are some differences between α^(u)\hat{\alpha}(u) and α^(u)\hat{\alpha}^{*}(u). For example, the values of α^(u)\hat{\alpha}(u) are generally smaller than those of α^(u)\hat{\alpha}^{*}(u) for u(0,0.1]u\in(0,0.1]. Also the bootstrap confidence intervals of α^(u)\hat{\alpha}^{*}(u) are narrower than the asymptotic confidence intervals of α^(u)\hat{\alpha}(u) for large uu.

Apart from the tests based on pointwise confidence intervals given in Figure 4(d) and (e), we carry out a different test for a nominal size of 0.1 based on the test statistic in Theorem 5. We test H0:α(u)=0H_{0}:\alpha(u)=0 against H1:α(u)0H_{1}:\alpha(u)\neq 0 for {u;u=umin+(0.1umin)j/10,j=0,,10}\{u;u=u_{\rm min}+(0.1-u_{\rm min})j/10,\ j=0,\ldots,10\}. The test statistic is T=𝒂T𝚺^1𝒂26.64T=\bm{a}^{T}\hat{\bm{\Sigma}}^{-1}\bm{a}\simeq 26.64 with (T>26.64)0.00520.1\mathbb{P}(T>26.64)\simeq 0.0052\ll 0.1. Therefore we reject the null hypothesis that the lower [0,u]2[0,u]^{2} tail and upper [1u,1]2[1-u,1]^{2} tail are symmetric for the 11 equally spaced points of uu in [umin,0.1][u_{\rm min},0.1].

We apply other measures of tail asymmetry to the copula sample displayed in Figure 4(a). The measure (13) of Rosco and Joe (2013) is calculated as ς^30.0326\hat{\varsigma}_{3}\simeq 0.0326. Another measure we consider here is a modified version of Krupskii’s (2017) measure (14), namely, ϱK(a,u)-\varrho_{K}(a,u). This modification is made to interpret the sign of the measure in the same manner as in that of ours. Figure 4(c) displays the estimates of ϱK(a,u)-\varrho_{K}(a,u) with respect to uu for the three specific functions of aa. The three curves of the modified measure ϱK(a,u)-\varrho_{K}(a,u) agree that there is stronger correlation in the lower [0,u]2[0,u]^{2} tail than the upper [1u,1]2[1-u,1]^{2} one for u0.12u\geq 0.12. This is somewhat similar to the result based on our measure as well. For a(x)=x4a(x)=x^{4}, the correlation coefficient in the lower [0,u]2[0,u]^{2} tail is greater than that in the upper [1u,1]2[1-u,1]^{2} tail for any u[0,0.5]u\in[0,0.5].

Finally, we summarize the results of the analysis of stock daily return data. The results based on the proposed measures suggest that the lower [0,u]2[0,u]^{2} tail probability is greater than the upper [1u,1]2[1-u,1]^{2} one for most values of u(umin,0.5]u\in(u_{\rm min},0.5], where umin0.0526u_{\rm min}\simeq 0.0526. In particular, there is significant difference between the lower and upper tail probabilities for u<0.14u<0.14. From the economic perspective, this result implies that the joint plunging probability is higher than the joint soaring probability with the threshold u<0.14u<0.14. Therefore it is recommended to use a copula with asymmetric tails for an appropriate modeling of the residuals of the daily return data appropriately. The three cases of the measure of Krupskii (2017) agree that, for u>0.12u>0.12, there is stronger correlation in the lower [0,u]2[0,u]^{2} tail than in the upper [1u,1]2[1-u,1]^{2} one.

8 Discussion

In this paper we have proposed a copula-based measure of asymmetry between the lower and upper tail probabilities. It has been seen that the proposed measure has some properties which are desirable as a measure of tail asymmetry. Sample analogues of the proposed measure have been presented, and statistical inference based on them, including point estimation, interval estimation and hypothesis testing, has been shown to be very simple. The practical importance of the proposed measure has been demonstrated through statistical analysis of stock return data.

This paper discusses a measure for bivariate data. However it is straightforward to extend the proposed bivariate measure to a multivariate one in a similar manner as in Embrechts et al. (2016) and Hofert and Koike (2019). Let (X1,,Xd)(X_{1},\ldots,X_{d}) be an d\mathbb{R}^{d}-valued random vector with continuous univariate margins. Then an extended measure of tail asymmetry for (X1,,Xd)(X_{1},\ldots,X_{d}) is defined by

𝑨(u)=(α11(u)α12(u)α1d(u)α21(u)α22(u)α2d(u)αd1(u)αd2(u)αdd(u)),{\bm{A}}(u)=\left(\begin{array}[]{cccc}\alpha_{11}(u)&\alpha_{12}(u)&\cdots&\alpha_{1d}(u)\\ \alpha_{21}(u)&\alpha_{22}(u)&\cdots&\alpha_{2d}(u)\\ \vdots&\vdots&\ddots&\vdots\\ \alpha_{d1}(u)&\alpha_{d2}(u)&\ldots&\alpha_{dd}(u)\end{array}\right),

where αij\alpha_{ij} is the proposed measure (1) of the random vector (Xi,Xj)(i,j=1,,d)(X_{i},X_{j})\ (i,j=1,\ldots,d). The properties of each element of the measure 𝑨(u){\bm{A}}(u) are straightforward from the results of this paper. It would be a possible topic for future work to investigate properties of this extended measure as a matrix and evaluate the values of the measure for multivariate copulas such as some examples of the vine copulas (Aas et al., 2009; Czado, 2010).

Supplementary material

Supplementary material contains the proofs of Lemma 1 and Theorems 15 and plots of random variates from the copulas discussed in Sections 4.2 and 6.2.

Acknowledgements

The authors are grateful to Hideatsu Tsukahara for his valuable comments on the work. Kato’s research was supported by JSPS KAKENHI Grant Numbers JP17K05379 and JP20K03759.

References

  • (1)
  • Aas et al. (2009) Aas K, Czado C, Frigessi A, Bakken H (2009) Pair-copula constructions of multiple dependence. Insur Math Econ 44(2):182–198.
  • (3)
  • Czado (2010) Czado C (2010) Pair-copula constructions of multivariate copulas. In Copula Theory and its Applications, Springer, Dordrecht, pp 93–109
  • Dehgani et al. (2013) Dehgani A, Dolati A, Úbeda-Flores M (2013) Measures of radial asymmetry for bivariate random vectors. Stat Pap 54:271–286
  • Dobrić et al. (2013) Dobrić J, Frahm G, Schmid F (2013) Dependence of stock returns in bull and bear markets. Dependence Model 1:94–110
  • Donnelly and Embrechts (2010) Donnelly C, Embrechts P (2010) The devil is in the tails: actuarial mathematics and the subprime mortgage crisis. ASTIN Bulletin 40(1):1–33
  • Embrechts et al. (2016) Embrechts P, Hofert M, Wang R (2016) Bernoulli and tail-dependence compatibility. Ann Appl Probab 26(3):1636–1658
  • Fermanian et al. (2004) Fermanian JD, Radulovic D, Wegkamp M (2004) Weak convergence of empirical copula processes. Bernoulli 10(5):847–860
  • Genest and Nešlehová (2014) Genest C, Nešlehová JG (2014) On tests of radial symmetry for bivariate copulas. Stat Pap 55:1107–1119
  • Ghalanos (2020) Ghalanos, A (2020) rugarch: univariate GARCH models, R package version 1.4-2
  • Hofert and Koike (2019) Hofert M, Koike T (2019) Compatibility and attainability of matrices of correlation-based measures of concordance. ASTIN Bulletin 49:885–918
  • Joe (1997) Joe H (1997) Multivariate models and dependence concepts. Chapman & Hall, London
  • Joe (2006) Joe H (2006) Discussion of “copulas: tales and facts” by Thomas Mikosch. Extremes 9(1):37–41
  • Joe (2014) Joe H (2014) Dependence modeling with copulas. Chapman & Hall/CRC, Boca Raton, FL
  • Joe and Hu (1996) Joe H, Hu T (1996) Multivariate distributions from mixtures of max-infinitely divisible distributions. J Multivar Anal 57:240–265
  • Krupskii (2017) Krupskii P (2017) Copula-based measures of reflection and permutation asymmetry and statistical tests. Stat Pap 58:1165–1187
  • Krupskii and Joe (2015) Krupskii P, Joe H (2015) Tail-weighted measures of dependence. J Appl Stat 42:614–629
  • Lee et al. (2018) Lee D, Joe H, Krupskii P (2018) Tail-weighted dependence measures with limit being the tail dependence coefficient. J Nonparametr Stat 30:262–290
  • McNeil et al. (2015) McNeil AJ, Frey R, Embrechts P (2015) Quantitative risk management: concepts, techniques, and tools (revised ed.). Princeton University Press, Princeton, NJ
  • Nelsen (2006) Nelsen RB (2006) An introduction to copulas (2nd ed.) Springer, New York
  • Nikoloulopoulos et al. (2012) Nikoloulopoulos AK, Joe H, Li H (2012) Vine copulas with asymmetric tail dependence and applications to financial return data. Comput Stat Data Anal 56:3659–3673
  • R Core Team (2020) R Core Team (2020) R: a language and environment for statistical computing, R Foundation for Statistical Computing, Vienna, Austria. https://www.R-project.org/
  • Rosco and Joe (2013) Rosco JF, Joe H (2013) Measures of tail asymmetry for bivariate copulas. Stat Pap 54:709–726
  • Segers (2012) Segers J (2012) Asymptotics of empirical copula processes under non-restrictive smoothness assumptions. Bernoulli 18(3):764–782
  • Tsukahara (2005) Tsukahara H (2005) Semiparametric estimation in copula models. Can J Stat 33:357–375
  • Yoshiba (2018) Yoshiba T (2018) Maximum likelihood estimation of skew-tt copulas with its applications to stock returns. J Stat Comput Simul 88(13):2489–2506

Supplementary Material for “Copula-based measures of asymmetry between the lower and upper tail probabilities”
Shogo Kato∗,a, Toshinao Yoshibab,c and Shinto Eguchi a
a Institute of Statistical Mathematics
b Tokyo Metropolitan University
c Bank of Japan August 4, 2020

The Supplementary Material is organized as follows. Section S1 presents the proofs of Lemma 1 and Theorems 15 of the article. Section S2 displays plots of random variates from Clayton copula, Ali-Mikhail-Haq copula and BB7 copula discussed in Sections 4.2 and 6.2 of the article.

S1 Proofs

S1.1 Proof of Theorem 1

Proof.

It follows from the expression (4) that

λU=limu1C¯(u,u)1u=limu0C¯(u¯,u¯)u.\lambda_{U}=\lim_{u\uparrow 1}\frac{\overline{C}(u,u)}{1-u}=\lim_{u\downarrow 0}\frac{\overline{C}(\bar{u},\bar{u})}{u}.

This result and Proposition 1 imply that

α(0)\displaystyle\alpha(0) =limu0log(C¯(u¯,u¯)C(u,u))=log(limu0C¯(u¯,u¯)/uC(u,u)/u)=log(λUλL).\displaystyle=\lim_{u\downarrow 0}\log\left(\frac{\overline{C}(\bar{u},\bar{u})}{C(u,u)}\right)=\log\left(\lim_{u\downarrow 0}\frac{\overline{C}(\bar{u},\bar{u})/u}{C(u,u)/u}\right){\color[rgb]{0,0,0}=\log\left(\frac{\lambda_{U}}{\lambda_{L}}\right).}

The last equality holds because λU\lambda_{U} and λL\lambda_{L} exist and either of λU\lambda_{U} and λL\lambda_{L} is not equal to zero. ∎

S1.2 Proof of Theorem 2

Proof.

If follows from the assumption that there exists a slowly varying function L(u)\ell_{L}(u) such that C(u,u)uκLL(u)C(u,u)\sim u^{\kappa_{L}}\ell_{L}(u) as u0u\rightarrow 0. Similarly, there exists a slowly varying function U(u)\ell_{U}(u) such that C¯(u¯,u¯)uκUU(u)\overline{C}(\bar{u},\bar{u})\sim u^{\kappa_{U}}\ell_{U}(u) as u0.u\rightarrow 0. Therefore

α(0)\displaystyle\alpha(0) =limu0log(C¯(u¯,u¯)C(u,u))=limu0log(uκUκLU(u)L(u))\displaystyle=\lim_{u\downarrow 0}\log\left(\frac{\overline{C}(\bar{u},\bar{u})}{C(u,u)}\right)=\lim_{u\downarrow 0}\log\left(u^{\kappa_{U}-\kappa_{L}}\frac{\ell_{U}(u)}{\ell_{L}(u)}\right)
={,κU>κL,,κU<κL.\displaystyle=\left\{\begin{array}[]{ll}\infty,&\kappa_{U}>\kappa_{L},\\ -\infty,&\kappa_{U}<\kappa_{L}.\end{array}\right.

The last equality holds because U(u)/L(u)\ell_{U}(u)/\ell_{L}(u) is slowly varying. If κU=κL\kappa_{U}=\kappa_{L} and either ΥU0\Upsilon_{U}\neq 0 or ΥL0\Upsilon_{L}\neq 0, then

α(0)=limu0log(U(u)L(u))=log(limu0U(u)L(u))=log(ΥUΥL).\alpha(0)=\lim_{u\downarrow 0}\log\left(\frac{\ell_{U}(u)}{\ell_{L}(u)}\right)=\log\left(\lim_{u\downarrow 0}\frac{\ell_{U}(u)}{\ell_{L}(u)}\right)=\log\left(\frac{\Upsilon_{U}}{\Upsilon_{L}}\right).

S1.3 Proof of Theorem 3

Proof.

Proposition 1 implies that α(0)\alpha(0) can be expressed as

α(0)=limu0log(C¯(u¯,u¯)C(u,u))=log(limu0C¯(u¯,u¯)C(u,u)).\alpha(0)=\lim_{u\downarrow 0}\log\left(\frac{\overline{C}(\bar{u},\bar{u})}{C(u,u)}\right)=\log\left(\lim_{u\downarrow 0}\frac{\overline{C}(\bar{u},\bar{u})}{C(u,u)}\right).

Since limu0dC(u,u)/du=limu0dC¯(u¯,u¯)/du=0\lim_{u\downarrow 0}dC(u,u)/du=\lim_{u\downarrow 0}d\overline{C}(\bar{u},\bar{u})/du=0, the l’Hôpital’s rule is applicable to the last expression of the equation above. Hence we have

α(0)\displaystyle\alpha(0) =log(limu0d2C¯(u¯,u¯)/du2d2C(u,u)/du2)=log(limu0c(1u)c(u))\displaystyle=\log\left(\lim_{u\downarrow 0}\frac{d^{2}\overline{C}(\bar{u},\bar{u})/du^{2}}{d^{2}C(u,u)/du^{2}}\right)=\log\left(\lim_{u\downarrow 0}\frac{c(1-u)}{c(u)}\right)

as required. ∎

S1.4 Proof of Lemma 1

Proof.

It is straightforward to see that 𝔼[TL(u)]\mathbb{E}\left[T_{L}(u)\right] and var[TL(u)]{\rm var}\left[T_{L}(u)\right] can be calculated as

𝔼[TL(u)]\displaystyle\mathbb{E}\left[T_{L}(u)\right] =𝔼[1ni=1n𝟏(U1iu,U2iu)]=1ni=1n𝔼[𝟏(U1iu,U2iu)]\displaystyle=\mathbb{E}\left[\frac{1}{n}\sum_{i=1}^{n}\bm{1}(U_{1i}\leq u,U_{2i}\leq u)\right]=\frac{1}{n}\sum_{i=1}^{n}\mathbb{E}\left[\bm{1}(U_{1i}\leq u,U_{2i}\leq u)\right]
=1nnCu=Cu,\displaystyle=\frac{1}{n}\cdot nC_{u}=C_{u},
var[TL(u)]\displaystyle{\rm var}\left[T_{L}(u)\right] =var[1ni=1n𝟏(U1iu,U2iu)]\displaystyle={\rm var}\left[\frac{1}{n}\sum_{i=1}^{n}\bm{1}(U_{1i}\leq u,U_{2i}\leq u)\right]
=nn2var[𝟏(U11u,U21u)]\displaystyle=\frac{n}{n^{2}}{\rm var}\left[\bm{1}(U_{11}\leq u,U_{21}\leq u)\right]
=1n(𝔼[𝟏2(U11u,U21u)]{𝔼[𝟏(U11u,U21u)]}2)\displaystyle=\frac{1}{n}\left(\mathbb{E}\left[\bm{1}^{2}(U_{11}\leq u,U_{21}\leq u)\right]-\left\{\mathbb{E}\left[\bm{1}(U_{11}\leq u,U_{21}\leq u)\right]\right\}^{2}\right)
=1n(CuCu2)=1nCu(1Cu).\displaystyle=\frac{1}{n}\left(C_{u}-C_{u}^{2}\right)=\frac{1}{n}C_{u}(1-C_{u}).

Noting that 𝔼[𝟏(U11>1u,U21>1u)]=C¯u¯\mathbb{E}\left[\bm{1}(U_{11}>1-u,U_{21}>1-u)\right]=\overline{C}_{\bar{u}}, the other expectation and variance, namely, 𝔼[TU(u)]\mathbb{E}\left[T_{U}(u)\right] and var[TU(u)]{\rm var}\left[T_{U}(u)\right], can be calculated in a similar manner.

Consider

cov[TL(u),TL(v)]=𝔼[TL(u)TL(v)]𝔼[TL(u)]𝔼[TL(v)].{\rm cov}\left[T_{L}(u),T_{L}(v)\right]=\mathbb{E}\left[T_{L}(u)T_{L}(v)\right]-\mathbb{E}\left[T_{L}(u)\right]\mathbb{E}\left[T_{L}(v)\right].

The first term of the left-hand side of the equation above is

𝔼[TL(u)TL(v)]=\displaystyle\mathbb{E}\left[T_{L}(u)T_{L}(v)\right]= 1n2𝔼[i=1n𝟏(U1iu,U2iu)j=1n𝟏(U1jv,U2jv)]\displaystyle\ \frac{1}{n^{2}}\mathbb{E}\left[\sum_{i=1}^{n}\bm{1}(U_{1i}\leq u,U_{2i}\leq u)\sum_{j=1}^{n}\bm{1}(U_{1j}\leq v,U_{2j}\leq v)\right]
=\displaystyle= 1n2i,j=1n𝔼[𝟏(U1iu,U2iu)𝟏(U1jv,U2jv)]\displaystyle\ \frac{1}{n^{2}}\sum_{i,j=1}^{n}\mathbb{E}\left[\bm{1}(U_{1i}\leq u,U_{2i}\leq u)\bm{1}(U_{1j}\leq v,U_{2j}\leq v)\right]
=\displaystyle= 1n2i=1n𝔼[𝟏(U1iu,U2iu)𝟏(U1iv,U2iv)]\displaystyle\ \frac{1}{n^{2}}\sum_{i=1}^{n}\mathbb{E}\left[\bm{1}(U_{1i}\leq u,U_{2i}\leq u)\bm{1}(U_{1i}\leq v,U_{2i}\leq v)\right]
+1n2ij𝔼[𝟏(U1iu,U2iu)𝟏(U1jv,U2jv)]\displaystyle+\frac{1}{n^{2}}\sum_{i\neq j}\mathbb{E}\left[\bm{1}(U_{1i}\leq u,U_{2i}\leq u)\bm{1}(U_{1j}\leq v,U_{2j}\leq v)\right]
=\displaystyle= 1n2i=1n𝔼[𝟏(U1iuv,U2iuv)]\displaystyle\ \frac{1}{n^{2}}\sum_{i=1}^{n}\mathbb{E}\left[\bm{1}(U_{1i}\leq u\wedge v,U_{2i}\leq u\wedge v)\right]
+1n2ij𝔼[𝟏(U1iu,U2iu)]𝔼[𝟏(U1jv,U2jv)]\displaystyle+\frac{1}{n^{2}}\sum_{i\neq j}\mathbb{E}\left[\bm{1}(U_{1i}\leq u,U_{2i}\leq u)\right]\mathbb{E}\left[\bm{1}(U_{1j}\leq v,U_{2j}\leq v)\right]
=\displaystyle= 1n2{nCuv+n(n1)CuCv}=1nCuv{1+(n1)Cuv}.\displaystyle\ \frac{1}{n^{2}}\left\{nC_{u\wedge v}+n(n-1)C_{u}C_{v}\right\}=\frac{1}{n}C_{u\wedge v}\left\{1+(n-1)C_{u\vee v}\right\}.

Therefore we have

cov[TL(u),TL(v)]=1nCuv{1+(n1)Cuv}CuCv=1nCuv(1Cuv).{\rm cov}\left[T_{L}(u),T_{L}(v)\right]=\frac{1}{n}C_{u\wedge v}\left\{1+(n-1)C_{u\vee v}\right\}-C_{u}C_{v}=\frac{1}{n}C_{u\wedge v}(1-C_{u\vee v}).

Similarly, cov[TU(u),TU(v)]{\rm cov}\left[T_{U}(u),T_{U}(v)\right] can be calculated. The other covariance cov[TL(u),TU(v)]{\rm cov}[T_{L}(u),T_{U}(v)] can also be obtained via a similar approach, but notice that

𝔼[TL(u)TU(v)]=\displaystyle\mathbb{E}\left[T_{L}(u)T_{U}(v)\right]= 1n2i=1n𝔼[𝟏(U1iu,U2iu)𝟏(U1i>1v,U2i>1v)]\displaystyle\ \frac{1}{n^{2}}\sum_{i=1}^{n}\mathbb{E}\left[\bm{1}(U_{1i}\leq u,U_{2i}\leq u)\bm{1}(U_{1i}>1-v,U_{2i}>1-v)\right]
+1n2ij𝔼[𝟏(U1iu,U2iu)𝟏(U1j>1v,U2j>1v)]\displaystyle+\frac{1}{n^{2}}\sum_{i\neq j}\mathbb{E}\left[\bm{1}(U_{1i}\leq u,U_{2i}\leq u)\bm{1}(U_{1j}>1-v,U_{2j}>1-v)\right]
=\displaystyle= 0+n(n1)n2CuC¯v¯=n1nCuC¯v¯.\displaystyle\ 0+\frac{n(n-1)}{n^{2}}C_{u}\overline{C}_{\bar{v}}=\frac{n-1}{n}C_{u}\overline{C}_{\bar{v}}.

The second equality holds because 0<u,v0.50<u,v\leq 0.5. Thus

cov[TL(u),TU(v)]=𝔼[TL(u)TU(v)]𝔼[TL(u)]𝔼[TU(v)]=1nCuC¯v¯.{\rm cov}\left[T_{L}(u),T_{U}(v)\right]=\mathbb{E}\left[T_{L}(u)T_{U}(v)\right]-\mathbb{E}\left[T_{L}(u)\right]\mathbb{E}\left[T_{U}(v)\right]=-\frac{1}{n}C_{u}\overline{C}_{\bar{v}}.

S1.5 Proof of Theorem 4

Proof.

Without loss of generality, assume 0<uv0.50<u\leq v\leq 0.5. Let

𝜷^=(β^1,β^2,β^3,β^4)T=(TL(u),TU(u),TL(v),TU(v))T,\hat{\bm{\beta}}=(\hat{\beta}_{1},\hat{\beta}_{2},\hat{\beta}_{3},\hat{\beta}_{4})^{T}=(T_{L}(u),T_{U}(u),T_{L}(v),T_{U}(v))^{T},
𝜷=(β1,β2,β3,β4)T=(C(u,u),C¯(u¯,u¯),C(v,v),C¯(v¯,v¯))T,\bm{\beta}=(\beta_{1},\beta_{2},\beta_{3},\beta_{4})^{T}=(C(u,u),\overline{C}(\bar{u},\bar{u}),C(v,v),\overline{C}(\bar{v},\bar{v}))^{T},
𝚺𝜷=(σβij)i,j,σβij=ncov(β^i,β^j).\bm{\Sigma_{\beta}}=(\sigma_{\beta ij})_{i,j},\quad\sigma_{\beta ij}=n\cdot{\rm cov}(\hat{\beta}_{i},\hat{\beta}_{j}).

Then it follows from Lemma 1 and the central limit theorem that

n(𝜷^𝜷)𝑑N(𝟎,𝚺𝜷)as n.\sqrt{n}(\hat{\bm{\beta}}-\bm{\beta})\xrightarrow{d}N(\bm{0},\bm{\Sigma_{\beta}})\quad\mbox{as }n\rightarrow\infty.

Define

h(𝜷)=(log(β2/β1)log(β4/β3))=(α(u)α(v)).h(\bm{\beta})=\left(\begin{array}[]{c}\log(\beta_{2}/\beta_{1})\\ \log(\beta_{4}/\beta_{3})\end{array}\right)=\left(\begin{array}[]{c}\alpha(u)\\ \alpha(v)\end{array}\right).

Applying the delta method, we have

n{h(𝜷)h(𝜷^)}𝑑N(0,h(𝜷)T𝚺𝜷h(𝜷))as n,\sqrt{n}\left\{h(\bm{\beta})-h(\hat{\bm{\beta}})\right\}\xrightarrow{d}N\left(0,\nabla h(\bm{\beta})^{T}\bm{\Sigma_{\beta}}\nabla h(\bm{\beta})\right)\quad\mbox{as }n\rightarrow\infty,

where

h(𝜷)=(β1log(β2/β1)β1log(β4/β3)β4log(β2/β1)β4log(β4/β3))=(1/β101/β2001/β301/β4).\nabla h(\bm{\beta})=\left(\begin{array}[]{cc}\frac{\partial}{\partial\beta_{1}}\log(\beta_{2}/\beta_{1})&\frac{\partial}{\partial\beta_{1}}\log(\beta_{4}/\beta_{3})\\ \vdots&\vdots\\ \frac{\partial}{\partial\beta_{4}}\log(\beta_{2}/\beta_{1})&\frac{\partial}{\partial\beta_{4}}\log(\beta_{4}/\beta_{3})\\ \end{array}\right)=\left(\begin{array}[]{cc}-1/\beta_{1}&0\\ 1/\beta_{2}&0\\ 0&-1/\beta_{3}\\ 0&1/\beta_{4}\end{array}\right).

The asymptotic variance can be calculated as

h(𝜷)T𝚺𝜷h(𝜷)\displaystyle\nabla h(\bm{\beta})^{T}\bm{\Sigma_{\beta}}\nabla h(\bm{\beta}) =\displaystyle= (σ11β122σ12β1β2+σ22β22σ13β1β3σ23β2β3σ14β1β4+σ24β2β4σ13β1β3σ23β2β3σ14β1β4+σ24β2β4σ33β322σ34β3β4+σ44β42)\displaystyle\left(\begin{array}[]{cc}\frac{\sigma_{11}}{\beta_{1}^{2}}-\frac{2\sigma_{12}}{\beta_{1}\beta_{2}}+\frac{\sigma_{22}}{\beta_{2}^{2}}&\frac{\sigma_{13}}{\beta_{1}\beta_{3}}-\frac{\sigma_{23}}{\beta_{2}\beta_{3}}-\frac{\sigma_{14}}{\beta_{1}\beta_{4}}+\frac{\sigma_{24}}{\beta_{2}\beta_{4}}\\ \frac{\sigma_{13}}{\beta_{1}\beta_{3}}-\frac{\sigma_{23}}{\beta_{2}\beta_{3}}-\frac{\sigma_{14}}{\beta_{1}\beta_{4}}+\frac{\sigma_{24}}{\beta_{2}\beta_{4}}&\frac{\sigma_{33}}{\beta_{3}^{2}}-\frac{2\sigma_{34}}{\beta_{3}\beta_{4}}+\frac{\sigma_{44}}{\beta_{4}^{2}}\end{array}\right) (S3)
=\displaystyle= (C(u,u)+C¯(u¯,u¯)C(u,u)C¯(u¯,u¯)C(v,v)+C¯(v¯,v¯)C(v,v)C¯(v¯,v¯)C(v,v)+C¯(v¯,v¯)C(v,v)C¯(v¯,v¯)C(v,v)+C¯(v¯,v¯)C(v,v)C¯(v¯,v¯))\displaystyle\left(\begin{array}[]{cc}\frac{C(u,u)+\overline{C}(\bar{u},\bar{u})}{C(u,u)\cdot\overline{C}(\bar{u},\bar{u})}&\frac{C(v,v)+\overline{C}(\bar{v},\bar{v})}{C(v,v)\cdot\overline{C}(\bar{v},\bar{v})}\\ \frac{C(v,v)+\overline{C}(\bar{v},\bar{v})}{C(v,v)\cdot\overline{C}(\bar{v},\bar{v})}&\frac{C(v,v)+\overline{C}(\bar{v},\bar{v})}{C(v,v)\cdot\overline{C}(\bar{v},\bar{v})}\end{array}\right) (S6)
=\displaystyle= (σ(u,u)σ(u,v)σ(u,v)σ(v,v)).\displaystyle{\color[rgb]{0,0,0}\left(\begin{array}[]{cc}\sigma(u,u)&\sigma(u,v)\\ \sigma(u,v)&\sigma(v,v)\end{array}\right).} (S9)

The case 0<v<u0.50<v<u\leq 0.5 can be calculated in the same manner. Then, for any 0<u,v0.50<u,v\leq 0.5, it follows that, as nn\rightarrow\infty, (𝔸n(u),𝔸n(v))(=n{h(𝜷)h(𝜷^)})(\mathbb{A}_{n}(u),\mathbb{A}_{n}(v))(=\sqrt{n}\{h(\bm{\beta})-h(\hat{\bm{\beta}})\}) converges weakly to the two-dimensional Gaussian distribution with mean 0 and the covariance matrix (S9). Weak convergence of (𝔸n(u1),𝔸n(um))(\mathbb{A}_{n}(u_{1}),\ldots\mathbb{A}_{n}(u_{m})) to an mm-dimensional centered Gaussian distribution for u1,,um(0,0.5](uiuj,ij)u_{1},\ldots,u_{m}\in(0,0.5]\ (u_{i}\neq u_{j},\ i\neq j) can be shown in a similar manner. Therefore {𝔸n(u)| 0<u0.5}\{\mathbb{A}_{n}(u)\,|\,0<u\leq 0.5\} converges weakly to a centered Gaussian process with covariance function σ(u,v)\sigma(u,v) as nn\rightarrow\infty. ∎

S1.6 Proof of Theorem 5

Proof.

Theorem 4 implies that 𝒂\bm{a} converges weakly to an mm-dimensional normal distribution N(𝟎,𝚺)N(\bm{0},\bm{\Sigma}) as nn tends to infinity, where

𝚺=(σ2(u1)σ(u1,u2)σ(u1,um)σ(u1,u2)σ2(u2)σ(u2,um)σ(u1,um)σ(u2,um)σ2(um)),\bm{\Sigma}=\left(\begin{array}[]{cccc}\sigma^{2}(u_{1})&\sigma(u_{1},u_{2})&\ldots&\sigma(u_{1},u_{m})\\ \sigma(u_{1},u_{2})&\sigma^{2}(u_{2})&\ldots&\sigma(u_{2},u_{m})\\ \vdots&\vdots&\ddots&\vdots\\ \sigma(u_{1},u_{m})&\sigma(u_{2},u_{m})&\ldots&\sigma^{2}(u_{m})\end{array}\right),

σ2(ui)=σ(ui,ui)\sigma^{2}(u_{i})=\sigma(u_{i},u_{i}), and σ(ui,uj)\sigma(u_{i},u_{j}) is defined as in Theorem 4. Then we have 𝒂T𝚺1𝒂𝑑χ2(m)\bm{a}^{T}\bm{\Sigma}^{-1}\bm{a}\xrightarrow{d}\chi^{2}(m) as nn\rightarrow\infty. Since TL(uj)T_{L}(u_{j}) and TU(uj)T_{U}(u_{j}) are consistent estimators of C(uj,uj)C(u_{j},u_{j}) and C¯(u¯j,u¯j)\overline{C}(\bar{u}_{j},\bar{u}_{j}), respectively, it holds that, for any (i,j)(i,j), σ^(ui,uj)\hat{\sigma}(u_{i},u_{j}) converges in probability to σ(ui,uj)\sigma(u_{i},u_{j}) as nn\rightarrow\infty. It then follows from Slutsky’s theorem that 𝒂T𝚺^1𝒂𝑑χ2(m)\bm{a}^{T}\hat{\bm{\Sigma}}^{-1}\bm{a}\xrightarrow{d}\chi^{2}(m) as nn\rightarrow\infty. ∎

S2 Plots of random variates from some existing copulas

Figure S1 plots random variates from Clayton copula (5), Ali-Mikhail-Haq copula (6) and BB7 copula (7) with some selected values of the parameter(s). This figure is given to help an intuitive understanding of the distributions of those copulas discussed in the paper.

Refer to caption Refer to caption Refer to caption
       (a)        (b)        (c)
Refer to caption Refer to caption
       (d)        (e)
Refer to caption Refer to caption Refer to caption
       (f)        (g)        (h)
Figure S1: Plots of 5000 random variates from: Clayton copula (5) with (a) θ=1\theta=1, (b) θ=20\theta=20 and (c) θ=0.3\theta=-0.3; Ali-Mikhail-Haq copula (6) with (d) θ=0.1\theta=0.1 and (e)(e) θ=1\theta=1; and BB7 copula (7) with (f) (δ,θ)=(1,1.71)(\delta,\theta)=(1,1.71), (g) (δ,θ)=(1.94,1.71)(\delta,\theta)=(1.94,1.71) and (h) (δ,θ)=(1,7.27)(\delta,\theta)=(1,7.27).