High-Dimensional Extreme Quantile Regression
Abstract
The estimation of conditional quantiles at extreme tails is of great interest in numerous applications. Various methods that integrate regression analysis with an extrapolation strategy derived from extreme value theory have been proposed to estimate extreme conditional quantiles in scenarios with a fixed number of covariates. However, these methods prove ineffective in high-dimensional settings, where the number of covariates increases with the sample size. In this article, we develop new estimation methods tailored for extreme conditional quantiles with high-dimensional covariates. We establish the asymptotic properties of the proposed estimators and demonstrate their superior performance through simulation studies, particularly in scenarios of growing dimension and high dimension where existing methods may fail. Furthermore, the analysis of auto insurance data validates the efficacy of our methods in estimating extreme conditional insurance claims and selecting important variables.
Keywords: Extrapolation; Extreme value; High-dimensional data; Regression analysis.
1 Introduction
Quantile regression has become a widely recognized and useful alternative to classical least-squares regression for analyzing heterogeneous data. Since its introduction by Koenker and Bassett (1978), quantile regression has gradually been extended to a wide variety of data analytic settings; for a comprehensive review, see Koenker (2005); Koenker et al. (2017).
While traditional quantile regression allows exploration of a wide range of conditional quantiles for with , there is often interest in the extreme tails where is close to 0 or 1. Without loss of generality, we focus the discussion on close to 1. The inherent challenge in estimating tail quantiles lies in the fact that the number of observations in the tails, that is, above the th quantile, is often small. In various contexts, methods have been developed to study extreme quantile regression by leveraging extreme value theory. Chernozhukov (2005), Wang, Li and He (2012), and Wang and Li (2013) studied extreme conditional quantiles under the linear quantile regression framework, and Li and Wang (2019) focused on the linear quantile autoregressive model. Daouia et al. (2013) and Gardes and Stupfler (2019) considered nonparametric smoothing methods. Wang and Tsai (2009) and Youngman (2019) applied generalized Pareto models to analyze the exceedance at a high threshold. Velthoen et al. (2023) and Gnecco et al. (2024) explored tree-based methods such as gradient boosting and random forest.
With advancements in data collection techniques such as genomics, economics, finance, and imaging studies, the dimension of covariates is becoming larger and can grow with the sample size . Existing methodology and theory in quantile regression for high-dimensional covariates have primarily focused on a central quantile level or compact quantile sets in . In situations where grows with , various studies have investigated asymptotic behaviors of quantile regression estimators, e.g., Welsh (1989), He and Shao (2000), Belloni et al. (2019), Pan and Zhou (2021), and He et al. (2023). Assuming the sparsity in regression coefficients, researchers have explored the -penalized quantile regression estimator and its generalizations (Belloni and Chernozhukov, 2011; Fan et al., 2014), concave penalties (Wang, Wu and Li, 2012), and smoothed quantile regression with concave penalties (Tan et al., 2022). Additionally, the debiased estimator of high-dimensional quantile regression has been studied for inference at central quantiles (Zhao et al., 2014; Bradic and Kolar, 2017).
In this article, we address the challenge of estimating extreme conditional quantiles of given a set of predictors in a high-dimensional context. The challenge lies in the scarcity of efficient samples at the tail and the sparsity of samples in high-dimensional space. As noted by Gnecco et al. (2024), a lack of samples exceeding the corresponding conditional th quantile leads to empirical estimators with large bias and variance. The high dimensionality of the predictor space introduces additional bias, as noise covariates can obscure true signals.
Current research on extreme quantiles primarily focuses on estimation within fixed-dimensional settings and is either not applicable or less competitive in high-dimensional settings. Although tree-based methods (Velthoen et al., 2023; Gnecco et al., 2024) demonstrate effectiveness in relatively large and fixed covariate dimensions, addressing scenarios where the covariate dimension grows with the sample size, comparable to the sample size itself, remains inadequately explored. Recently, Sasaki et al. (2024) proposed a high-dimensional tail index regression model, assuming that the dimension grows with and is comparable to . However, this approach imposes constraints such as a Pareto tail and specific link functions between the extreme value index and covariates.
In this article, we propose novel estimators for extreme quantiles in high dimensions by integrating concepts from extreme value theory with regularized estimators. Building upon a linear conditional quantile model, we employ regularized quantile regression to estimate intermediate conditional quantiles and extrapolate them to extreme quantiles. Here, “intermediate quantile” refers to a quantile level that approaches one at a moderate rate such that , whereas “extreme quantile” denotes a quantile level that approaches one more rapidly.
The theoretical analysis of extreme quantiles in high dimensions is challenging. The limiting distribution of a quantile estimator could be intractable in high dimensions, even at central quantile levels. Existing work on central quantiles within often requires a common assumption, namely , where is the conditional density of given , and is the th conditional quantile of ; see for instance Belloni and Chernozhukov (2011), and Wang, Wu and Li (2012). However, this assumption is violated in the tails as approaches . It is even more challenging to achieve a uniform result of the regularized estimator over the entire tail, which is crucial for constructing an effective extreme value index estimator. This paper addresses the gap and establishes the uniform error rate for quantile regression estimator in high dimensions and at intermediate quantiles. This rate includes terms analogous to those in high-dimensional quantile regression at a central quantile level, augmented by an inflationary component attributed to escalating quantile levels or diminishing effective sample sizes. In addition, we propose a refined Hill approach for estimating the extreme value index, which is based on a fixed number of intermediate high-dimensional quantile estimates. This framework enables us to analyze the uniform behavior of regularized quantile regression in the tail region and assess the rate at which tends to zero. The error rates of the proposed refined Hill and extreme quantile estimators offer theoretical insights into threshold selection in high-dimensional contexts.
The rest of the article is organized as follows. In Section 2, we present the proposed estimators for the extreme value index and extreme conditional quantiles, and derive theoretical results for the proposed estimators using techniques from both extreme value theory and high-dimensional statistics theory. We assess the finite sample performance of the proposed methods through a simulation study in Section 3 and the analysis of auto insurance data in Section 4. Technical details and additional information for Sections 3 and 4 are provided in the online Supplementary Material.
2 Extreme Quantile Estimation in High Dimensions
Suppose we observe a random sample of the random vector , where is the univariate response variable and is the -dimensional centralized design vector. This article considers the high dimensional case: as . Let denote the th conditional quantile of given , where is the cumulative distribution function (CDF) of given . Denote as the support of .
Throughout the article, we assume that is in the maximum domain of attraction of an extreme value distribution with the extreme value index (EVI) , denoted by . That is, for a given random sample from , there exist constants and such that
as , for . In this paper, we assume , which means that has heavy-tailed distributions as commonly seen in many applications, such as stock market returns, insurance claims, earthquake magnitudes, river flows during floods.
The main objective of this paper is to estimate the conditional quantile when and as . We address two distinct regimes for the quantile level as it approaches 1: intermediate order quantile levels such that , and extreme order quantile levels with , where is some constant.
We focus on the following tail linear quantile regression model:
(1) |
where , as , and the quantile slope coefficients may vary across . The linear quantile model in (1) is specifically assumed at the upper tail. Similar tail model specifications have also been considered in Wang, Li and He (2012), Li and Wang (2019), Xu et al. (2022). In high-dimensional settings, we extend this assumption and additionally impose a sparsity condition to ensure model identifiability. We assume there are tail-relevant variables, that is, , and . A tail-relevant variable may influence the tail quantiles of , even if it does not affect the central quantiles. Without loss of generality, we assume that the first slope coefficients are nonzero, i.e., .
We propose a three-step procedure to estimate the extreme conditional quantile. In the first step, we obtain -penalized quantile estimators at a sequence of intermediate quantile levels. For heavy-tailed distributions, the Hill estimator (Hill, 1975) is the most commonly used method to estimate the EVI. However, based on the average of log excesses, the Hill estimator requires analysis of the tail quantile process (de Haan and Ferreira, 2006), which poses both computational and theoretical challenges in high dimensions. To address these challenges, we develop a refined Hill estimator in the second step, which relies on quantile estimates at a fixed number of intermediate quantile levels that approach to one at the same rate. In the third step, we develop an extrapolation estimator for the extreme conditional quantile by leveraging extreme value theory and results from the first two steps.
Before presenting the proposed estimators and their theoretical properties, we first introduce the notations used throughout the paper. For a sequence of random variables , we denote its order statistics as . Given a random sample , let and . We denote the , , and norms by , , and , respectively. Let denote the weighted -norm with . Given a vector , and a set of indices , we denote by the vector where if if . We use to denote , meaning for some constant that does not depend on ; and to denote and . We use to denote . Additionally, we use and . For notational simplicity, let be the conditional distribution function of given , and denote . Finally, let , and .
2.1 -penalized Intermediate Quantile Estimator
Define where is an intermediate quantile level such that and , and are constants. For any such that , we define the -penalized quantile estimator of as
(2) |
where is the penalization parameter.
For high dimensional settings, the choice of is crucial for achieving estimation consistency. As suggested by Bickel et al. (2009), should exceed a suitably rescaled (sub)gradient of the sample criterion function evaluated at the true parameter value. While the asymptotic scale of has been derived for central quantiles (Belloni and Chernozhukov, 2011; Zheng et al., 2013), this scale differs for intermediate quantiles as approaches one. Specifically, a subgradient of the penalized quantile regression objective function at is given by , where and is a subgradient of with for . The infinity norm of the first term is of the same order for both central and intermediate quantiles; however, the second term varies in order due to the factor . Lemma 1 in the Supplementary Material provides an asymptotic uniform lower bound of over , and a practical choice for is discussed in Section 2.4.
One major challenge in establishing the theoretical properties of arise from the violation of the condition at intermediate quantiles. This violation complicates the establishment of a quadratic lower bound for the Taylor expansion of the expected quantile loss function. Consequently, the general framework of Negahban et al. (2012) cannot be applied to obtain the convergence rate through the restricted strong convexity property of the empirical loss function. However, leveraging condition C5 and extreme value theory, we assess the rate at which the conditional quantile and density converge, determining how approaches zero as goes to one. This allows us to establish an asymptotic quadratic lower bound for the expected intermediate quantile loss after appropriate standardization (see Lemma 4 in the Supplementary Material).
To establish the theoretical properties of over , we impose Conditions C1-C6. For brevity, the conditions are detailed in Supplementary Material, but we highlight one key condition, C5, below.
Condition C5. There exists an auxiliary line with and a bounded vector such that for and some heavy-tailed distribution function with extreme value index and density , the following hold for some positive continuous functions on .
-
(i)
There exists positive sequences such that hold for all .
-
(ii)
For any sequence , if and as , then
holds uniformly for , where is a constant.
-
(iii)
satisfies the second-order condition with , and with . Additionally, is regularly varying at infinity with index .
Remark 1.
Condition C5 presents a novel framework tailored for high-dimensional settings, where , , and may be unbounded and affect the tail behavior of . Unlike fixed-dimensional cases (Wang, Li and He, 2012; Chernozhukov, 2005), this condition addresses the combined complexities of high dimensionality and tail behavior. The conditions and as ensure that the tail effect of outweighs the diverging scale effect of in high dimensions. This requirement allows us to apply extreme value theory to infer the tail behavior of , and is mild, as it is automatically satisfied when the number of important variables is finite (see Example 1). The second-order condition of implies that , which can be interpreted as a first-order condition (see Condition C6 for the definition). The first-order condition is a necessary and sufficient condition for to belong to the maximum domain of attraction for heavy-tailed distributions (de Haan and Ferreira, 2006). Most commonly used families of continuous distributions satisfies the second-order condition. Although both the extreme value index estimator and the extreme quantile estimator are based on the first-order condition, the second-order condition is essential for deriving their theoretical results. Additionally, Lemma 3 in the Supplementary Material demonstrates that under Condition C5, a second-order condition of still holds, similar to the fixed-dimensional case.
Define . The following Theorem 1 establishes a uniform bound of for .
Theorem 1.
Assume Conditions C1-C6 hold, along with , , , and . Then, for any and some positive constant , there exist a positive integer such that, for all ,
with probability at least .
Theorem 1 establishes bounds similar in spirit to those found in Chernozhukov (2005) but with an additional term, , which inflates the upper bound. This term reflects the decay of the conditional density in the tails. Specifically, as shown in Remark S.2 in the Supplementary Material, under certain conditions, as , which describes the rate at which the conditional density approaches zero at intermediate quantiles. Theorem 1 poses both upper and lower bounds conditions on . The condition prevents from approaching one too quickly, aligning with the requirement for restricted strong convexity of the objective function (see Lemma 4 in the Supplementary Material). Restricted strong convexity refers to maintaining strong convexity within a restricted set (Negahban et al., 2012). In the high-dimensional context, this restricted set is defined such that the error vector lies within it with high probability, given a suitably chosen penalty and certain conditions. Intuitively, restricted strong convexity holds only if the prediction error of the penalized estimator, , grows slower than the quantile of . The conditions and ensure that the tail effect dominates the high-dimensional effect. In Section 2.3, we provide an example illustrating the feasibility of these conditions.
2.2 Refined Hill Estimator
Let . For simplicity, we omit the rounding notation, as it does not impact the theoretical discussion that follows. Define quantile levels , where . For each , we estimate by the -penalized quantile regression estimator in (2). Suppose that we are interested in estimating the extreme conditional quantile of given . Let , and we can define the refined Hill estimator as
where is some positive measurable function.
The refined Hill estimator is a weighted variant of the Hill-type estimators found in the literature but with the following key differences. First, utilizes a fixed number of log excesses at intermediate quantile levels with the same rate, whereas the Hill estimator typically relies on quantiles approaching one at varying rates, as seen in (Wang, Li and He, 2012). Second, allows for flexible weighting through , while the Hill estimator commonly applies equal weights to upper quantiles (Hill, 1975; Wang, Li and He, 2012; Daouia et al., 2023). We will discuss the choice of the weight function , and the tuning parameters and in Section 2.4. Similar weighting schemes have been explored by other researchers to generalize Hill and Pickands estimators (Drees, 1995; Daouia et al., 2013; He et al., 2022) and for estimating the Weibull tail coefficient (Gardes and Girard, 2016; He et al., 2020). These methods were developed in fixed-dimensional settings where the theoretical guarantees of the refined estimators could be established using the tractable asymptotic normality of intermediate quantile estimates. While this strategy is not feasible in high-dimensional settings, we can leverage the uniform convergence results from Theorem 1 to establish the convergence rate of the refined Hill estimator. Suppose there exist a positive sequence such that, hold for all .
Theorem 2.
Assume that belongs to a compact set in . Assume the conditions of Theorem 1, and that , . Then we have
Remark 2.
Theorem 2 suggests that if , then . In Theorem 2, the conditions on are posed to account for the error in the second-order term in the tail expansion of . These conditions can be simplified for specific models, as discussed in Section 2.3. Note that the convergence rate of is , which is slower than the typical rate achieved in the fixed-dimensional case (Wang and Tsai, 2009; Daouia et al., 2013). The difference arises from the slower convergence of the intermediate quantile estimators in high-dimensional settings. Furthermore, to ensure consistency, must satisfy , which is stricter than the requirement in the fixed dimensional case (Wang, Li and He, 2012), where with being a small positive constant. Finally, while the error bound of the -penalized intermediate quantile estimator depends on , the convergence rate of remains invariant across .
2.3 Extreme Conditional Quantile Estimator
Let be an extreme quantile level such that . Suppose that we are interested in estimating the th conditional quantile of given . Direct estimation using the -penalized quantile estimator in (2) is often inaccurate or unstable due to limited data in the extreme tail. However, we can leverage extreme value theory and extrapolate from intermediate to extreme quantiles by using the relationship between quantiles with levels approaching one at different rates.
Define for , the ()th quantile of . According to Corollary 1.2.10 in de Haan and Ferreira (2006), for a heavy-tailed distribution , we have
Motivated by this, for a given , we estimate by
where .
Theorem 3.
Assume the conditions of Theorem 2. Let be a quantile level such that , then we have
Additionally, if , .
According to our estimation procedure, the error in estimating extreme quantiles primarily arises from the EVI estimator and the -penalized intermediate quantile estimator. Specifically, the error rate of the EVI estimator and the relative error rate of the intermediate quantile estimator are both given by . Since the EVI estimator contributes to the error rate at the exponent, the final error rate for the extreme quantile estimation is . In the fixed dimension case, the rate of the relative error of extreme quantile estimator is . Comparing this with our results, we observe that the inflated term arises from the EVI estimator, which has an error rate of in the fixed-dimensional setting.
Remark 3.
In this paper, we assume that , which represents a first-order condition on . To derive the properties of , a second-order condition is needed. Condition C5 assumes the existence of an auxiliary line so that satisfies the second-order condition. Under this condition, we can show that also satisfies the second-order condition even for high-dimensional (see Lemma 3), and it has the same EVI across . This implies that the EVI can be technically estimated based on the intermediate conditional quantiles at any . Theorem 2 establishes the convergence rate for obtained at belonging to a compact set, which we now denote as to emphasize its dependence on . Careful selection of is important, particularly in high-dimensional cases where data points may be sparse due to the curse of dimensionality. In practice, we recommend estimating at , as this leads to more stable numerical performance and a higher convergence rate than , whose convergence rate is presented in Remark 2. Specifically, ; see the proof of Theorem 2 in the Supplementary Material for more details. A pooled estimator that averages could also improve estimation efficiency; however, this approach is computationally intensive and less practical in high-dimensional settings.
The regularity conditions outlined in our paper encompass a broad range of conventional regression settings. To illustrate these conditions, we consider the location-scale shift model as a representative example.
Example 1. Consider the location-scale shift model:
where for , and with satisfying the second order condition for some and . Obviously, .
Suppose that the scale effect variable is finite, i.e., . Under this model, Condition C5 holds with , , where is some constant and . Details of the derivation can be found in the Supplementary Material. Let , where . Then , , and . Thus the conditions of Theorem 2 are , , , , for any , which implies The error rate is . Next, we consider two special cases to simplify the conditions on and discuss the feasibility.
-
(I).
The non-zero components of and are at the same positions, i.e., , where .
Theorem 1 requires the condition . The addtional conditions on in Theorems 2 and 3 are and , which can be simplified to and , respectively. To obtain consistency of , the condition on is . In summary, the condition on is To ensure the existence of an appropriate , the additional sparsity conditions are and .
-
(II).
Suppose that is finite, that is, .
2.4 Computational Issues
The proposed method relies on several tuning parameters/functions, including the penalization parameter , the integer associated with the intermediate quantile level , the weight function and the number of intermediate quantile levels in the refined Hill estimator. In this section, we will discuss practical methods for selecting these quantities.
Selection of . In high-dimension settings, the optimal choice of depends on many factors, including , the design matrix and error distribution. Bickel et al. (2009) suggested that should exceed the supremum norm of a suitably rescaled (sub)gradient of the sample criterion function, Belloni and Chernozhukov (2011) proposed a similar approach for selecting at central quantiles. However, applying this method directly to intermediate quantiles often results in over-shrinkage. Our theoretical study suggests that at intermediate quantiles, should scale as , but accurately estimating the constant is challenging, as noted in Homrighausen and McDonald (2017); Chetverikov et al. (2021). In our implementation, we conduct 10-fold cross-validations to select for each . This involves splitting the data into ten folds and using each fold to evaluate the estimator based on the remaining folds. For each , we calculate the average quantile loss at across all folds and select the that minimizes this loss. While time-intensive, our numerical studies show that cross-validation performs well across various settings considered. Developing an efficient and data-driven selection method tailored to intermediate quantiles is an interesting direction for future research.
Selection of . The intermediate quantile level, or equivalently , plays an important role of balancing between bias and variance. A small increases variance, while a large leads to higher bias. A common approach is to select at the first stable point in the plot of the EVI estimator versus ; see Neves et al. (2015) for a heuristic algorithm. However, this method is computationally intensive in high-dimensional cases, as the estimator needs to be calculated over a sequence of values, and the EVI may exhibit local variations across . To reduce the computational cost, we propose a rule of thumb for selecting . Based on Theorem 2 and Example 1, should satisfy . Motivated by this, we set , where is a constant, and and are small positive constants. The sensitivity analysis in the Supplementary Material shows that our proposed method remains stable for , , and , across various scenarios and sample sizes. Throughout our numerical studies, we use , and .
Choice of weights. The refined Hill estimator assigns different weights to intermediate quantiles. In our implementation, we use and , where , as recommended by He et al. (2022) for extreme analysis in an autoregressive model. The authors showed that, in their setup, the asymptotic variance of the refined Hill estimator is proportional to a term depending only on . Minimizing this term yields the optimal choice . While the limiting distribution is intractable without debiasing in high-dimensional settings, we find that the same parameters continue to perform well, resulting in more efficient estimation than methods based on the unweighted Hill estimator.
3 Simulation Study
We conduct a simulation study to investigate the performance of the proposed method for estimating extreme quantiles in high-dimensional settings. The data are generated from
where for and are independent and identically distributed student t random variables, and is a constant controlling the degree of heteroscedasticity. Under this model, the th conditional quantile of is where , , , and is the th quantile of . We consider six different cases. In Cases 1, 3, and 5, and in Cases 2, 4 and 6, . Cases 1-2 have homoscedastic errors with , while in Cases 3-4, and in Cases 5-6. Note that, for , the EVI is . We consider two settings for the sample size and dimension : (1) grows with but is smaller than , with and ; (2) is larger than or equal to , with and . For each scenario, the simulation is repeated times.
For each simulation, we estimate at using three methods: the extreme quantile regression method (EQR) without assuming common slope in Wang, Li and He (2012); the high-dimensional quantile regression (HQR) of Belloni and Chernozhukov (2011), and the proposed high-dimensional extreme quantile regression (HEQR). For EQR, the number of upper-order statistics is set to , as recommended in Wang, Li and He (2012).
For each simulation, we calculate integrated squared error (ISE), defined as
where are evaluation points of the covariates, and we define the mean integrated squared error (MISE) as the average ISE across simulations. In our simulation, we set , and let be random replicates of .
HEQR | EQR | HQR | |||
---|---|---|---|---|---|
0 | 1000 | 0.995 | 2.9(0.05) | 3.6(0.05) | 5.7(0.13) |
0.999 | 4.7(0.09) | 5.2(0.09) | 12.5(0.12) | ||
5000 | 0.995 | 1.6(0.02) | 2.7(0.02) | 3.7(0.04) | |
0.999 | 2.4(0.03) | 4.2(0.05) | 6.1(0.06) | ||
0.5 | 1000 | 0.995 | 3.3(0.05) | 4.4(0.06) | 7.5(0.21) |
0.999 | 4.8(0.10) | 5.8(0.10) | 14.6(0.14) | ||
5000 | 0.995 | 2.0(0.02) | 3.4(0.03) | 5.0(0.06) | |
0.999 | 2.4(0.03) | 4.8(0.06) | 6.9(0.08) | ||
0.9 | 1000 | 0.995 | 3.7(0.05) | 5.2(0.07) | 9.0(0.29) |
0.999 | 5.0(0.09) | 6.5(0.12) | 15.8(0.17) | ||
5000 | 0.995 | 2.3(0.02) | 4.0(0.04) | 6.0(0.08) | |
0.999 | 2.6(0.03) | 5.1(0.06) | 7.8(0.12) |
HEQR | EQR | HQR | |||
---|---|---|---|---|---|
0 | 1000 | 0.995 | 7.4(0.26) | 10.7(0.56) | 24.6(3.33) |
0.999 | 15.0(0.68) | 19.9(2.11) | 69.7(16.05) | ||
5000 | 0.995 | 5.3(0.11) | 7.7(0.12) | 12.7(0.28) | |
0.999 | 7.2(0.23) | 10.5(0.26) | 28.5(2.78) | ||
0.5 | 1000 | 0.995 | 8.5(0.31) | 13.5(0.70) | 32.3(3.70) |
0.999 | 15.4(0.84) | 23.8(2.63) | 96.8(28.76) | ||
5000 | 0.995 | 6.3(0.14) | 9.7(0.18) | 17.1(0.42) | |
0.999 | 8.1(0.33) | 11.9(0.39) | 36.5(3.93) | ||
0.9 | 1000 | 0.995 | 9.6(0.33) | 16.1(0.80) | 40.5(4.25) |
0.999 | 16.0(0.86) | 27.3(2.86) | 119.9(38.54) | ||
5000 | 0.995 | 7.3(0.17) | 11.5(0.23) | 21.5(0.58) | |
0.999 | 9.4(0.39) | 13.5(0.49) | 46.9(5.11) |
HEQR | HQR | ||||
---|---|---|---|---|---|
0 | 0.2 | 1000 | 0.995 | 0.23(0.03) | 0.19(0.02) |
0.999 | 0.23(0.03) | 0.43(0.03) | |||
1200 | 0.995 | 0.20(0.02) | 0.16(0.02) | ||
0.999 | 0.22(0.03) | 0.30(0.03) | |||
0.5 | 0.2 | 1000 | 0.995 | 0.26(0.03) | 0.23(0.02) |
0.999 | 0.26(0.03) | 0.50(0.03) | |||
1200 | 0.995 | 0.23(0.03) | 0.20(0.02) | ||
0.999 | 0.24(0.03) | 0.35(0.03) | |||
0 | 0.5 | 1000 | 0.995 | 0.32(0.04) | 0.44(0.33) |
0.999 | 0.41(0.06) | 0.63(0.13) | |||
1200 | 0.995 | 0.41(0.05) | 0.36(0.23) | ||
0.999 | 0.51(0.05) | 0.55(0.14) | |||
0.5 | 0.5 | 1000 | 0.995 | 0.34(0.06) | 0.52(0.71) |
0.999 | 0.39(0.07) | 0.67(0.17) | |||
1200 | 0.995 | 0.34(0.05) | 0.41(0.41) | ||
0.999 | 0.44(0.08) | 0.58(0.20) |
-
•
HEQR: the proposed estimator; HQR: the high-dimensional qunatile regression estimator in Belloni and Chernozhukov (2011).
We begin by examining the setting. Tables 1 and 2 summarize the MISE of three estimators for the extreme conditional quantile at and . Results show that for all three methods, both MISE and its standard error increase as approaches 1, and as the distribution becomes more heavy-tailed or heteroscedastic. In all scenarios, the proposed method HEQR outperforms the others in terms of both MISE and standard error. Both EQR and HEQR rely on extrapolation using extreme value theory. In this setting, since increases with but remains smaller than , ERQ method is still applicable, and both EQR and HEQR perform better than HQR. The advantages of HEQR over EQR come from its penalized estimation of and its use of the refined Hill estimator, compared to EQR’s unweighted Hill estimator.
We next examine the setting where . Table 3 summarizes the results for two methods HEQR and HQR, with and , as EQR is not applicable in these cases. While HQR performs reasonably well for small at , it is significantly less efficient than HEQR for all other scenarios, regarding both MISE and its standard error. Overall, the results suggest that HEQR demonstrates superior accuracy and stability, particularly for estimating extremely high quantiles in heavy-tailed distributions.
We perform a sensitivity analysis to assess the stability of the proposed method with respect to the choices of , , and in the rule of thumb . For brevity, we present results only for Case 1. The other cases exhibit similar behavior, and the results are provided in the Supplementary file. Figure 1 shows the MISE of EQR, HQR and HEQR for the extreme conditional quantiles at against , with fixed at for and . The shaded area denotes the 95% pointwise confidence band of the MISE. The two horizontal lines represent the MISE of HQR and EQR, respectively, while the shaded area corresponds to the 95% pointwise confidence band of the MISE. For , the MISE of HEQR is consistently smaller than that of HQR and EQR, indicating that in this range is a suitable choice. Figures 2 and 3 plot the MISE against and , respectively, with the other two constants fixed at their recommended values. The horizontal line represents the MISE associated with the recommended values , , and , while the shaded area corresponds to the 95% pointwise confidence band of the MISE. The results indicate that the method is insensitive to variations in and .






4 Analysis of Auto Insurance Claims
In this section, we analyze an auto insurance claims dataset to investigate the effects of various factors on the higher quantiles of claim amounts. The data is available on Kaggle at https://www.kaggle.com/datasets/xiaomengsun/car-insurance-claim-data. Insurance claims data are typically heavy-tailed and heterogeneous, reflecting the diverse nature of risks they embody and the array of factors that can influence claim sizes and frequencies. These factors not only complicate the analysis but also provide a rich vein of insights. We consider the following quantile regression model:
where represents the auto claim amount (in USD), and are covariates included after preprocessing. To account for nonlinearity, we include quadratic terms for “Age” and “Income,” as well as quadratic and cubic terms for “Vehicle Value” in the model, with both “Income” and “Vehicle Value” on the log-transformed scale. A detailed description of the variables can be found in Section S.4 of the Supplementary Materials. We focus on high quantiles () to study the impact of various factors on large claims.
We conduct a cross-validation study to compare the performance of different methods for predicting the extremely high conditional quantiles of auto claims. We randomly divide the data set into a training set (20% of the data set, observations) and a testing set (the remaining 80% of the data set, observations). We apply HEQR, EQR, and HQR to analyze the training set and predict the extreme quantiles of auto claims conditional on the covariates in the testing set. For HEQR, we set and estimate at a central data point, with continuous covariates set to their mean values and binary variables set to their modes. Since has a mean of and a variance of , this motivates us to consider the following prediction error (PE) metric:
where is the index set of the testing set. The cross-validation process is repeated 500 times. Table 4 summarizes the median absolute PE for different methods at , , and , with values in parentheses representing the corresponding median absolute deviation. The results indicate that HEQR yields significantly smaller prediction errors than the other two methods across all three quantile levels, demonstrating higher efficiency.
Next, we analyze the full dataset to estimate extreme quantiles and identify important variables using HEQR. Out of the 43 covariates, our method identified 34 to have nonzero effects at the intermediate high quantile level at with and . Notably, one of the most significant variables identified by the HEQR model is whether an individual resides in a rural district. This finding aligns with prior research by Clemente et al. (2023), which links rural residences to lower auto insurance claims due to reduced claim frequency and severity. Additionally, our results show that an increase in vehicle age corresponds to a decrease in claims, consistent with Clemente et al. (2023), who also found vehicle age negatively impacts claim frequency. Figure 4 shows the -penalized quantile coefficient estimates for selected covariates at upper quantiles. The results indicate that being a commercial vehicle positively affects auto claims, with the effect increasing at higher quantiles. In contrast, “Rural Population” and “Time in Force” (the duration the insurance policy has been in effect, abbreviated as TIF) have negative effects, which become stronger at higher quantiles. The variable “Bachelor’s Degree” shows little impact on auto claims. These findings highlight the heterogeneity of covariate effects, particularly in the upper tail of the claim distribution.

We also include the full-data analysis results from HQR for comparison. Figure 5 presents the extreme quantile estimates of auto claims at and 0.999 as a function of , log-transformed Vehicle Value, for two groups of individuals with different TIF from both HQR and HEQR, with other continuous covariates set to their mean values and categorical variables to their modes. The two groups correspond to individuals with TIF of 1 year and 7 years, representing the 0.25 and 0.75 quantiles, respectively. Results from HEQR suggest that, for both groups, higher vehicle values lead to larger claims, likely due to higher repair costs for more expensive vehicles, which aligns with common expectations and prior findings (Clemente et al., 2023). In contrast, HQR shows little variation in high quantiles of claims across vehicle values. Both methods indicate that individuals with a shorter TIF tend to have higher claims at . This finding aligns with our common understanding, as shorter TIFs may reflect less experienced or stable policyholders, potentially leading to higher-risk profiles and more frequent or severe claims. HEQR consistently shows the effect of TIF across high quantiles, while HQR becomes unstable at , displaying almost no difference between the two groups.


SUPPLEMENTARY MATERIAL
- Supplementary Material for High-dimensional Extreme Quantile Regression:
-
The supplementary material includes all the technical conditions, proofs of theorems, additional simulation results, and further details on the analysis of the auto insurance claims data. (.pdf file)
- Auto Insurance Claims Data Set:
-
Data set used in the illustration of the proposed methods in Section 4. (.csv file)
References
- (1)
- Belloni and Chernozhukov (2011) Belloni, A. and Chernozhukov, V. (2011), ‘-penalized quantile regression in high-dimensional sparse models’, Annals of Statistics 39(1), 82–130.
- Belloni et al. (2019) Belloni, A., Chernozhukov, V., Chetverikov, D. and Fernández-Val, I. (2019), ‘Conditional quantile processes based on series or many regressors’, Journal of Econometrics 213(1), 4–29.
- Bickel et al. (2009) Bickel, P. J., Ritov, Y. and Tsybakov, A. B. (2009), ‘Simultaneous analysis of lasso and dantzig selector’, Annals of Statistics 37(4), 1705–1732.
- Bradic and Kolar (2017) Bradic, J. and Kolar, M. (2017), ‘Uniform inference for high-dimensional quantile regression: linear functionals and regression rank scores’, arXiv:1702.06209.
- Chernozhukov (2005) Chernozhukov, V. (2005), ‘Extramal quantile regression’, Annals of Statistics 33(2), 806–839.
- Chetverikov et al. (2021) Chetverikov, D., Liao, Z. and Chernozhukov, V. (2021), ‘On cross-validated lasso in high dimensions’, Annals of Statistics 49(3), 1300–1317.
- Clemente et al. (2023) Clemente, C., Guerreiro, G. R. and Bravo, J. M. (2023), ‘Modelling motor insurance claim frequency and severity using gradient boosting’, Risks 11(9), 1–20.
- Daouia et al. (2013) Daouia, A., Gardes, L. and Girard, S. (2013), ‘On kernel smoothing for extremal quantile regression’, Bernoulli 19(5B), 2557–2589.
- Daouia et al. (2023) Daouia, A., Stupfler, G. and Usseglio-Carleve, A. (2023), ‘Inference for extremal regression with dependent heavy-tailed data’, Annals of Statistics 51(5), 2040–2066.
- de Haan and Ferreira (2006) de Haan, L. and Ferreira, A. (2006), Extreme Value Theory: An Introduction, Springer Science & Business Media.
- Drees (1995) Drees, H. (1995), ‘Refined pickands estimators of the extreme value index’, Annals of Statistics 23(6), 2059–2080.
- Fan et al. (2014) Fan, J., Fan, Y. and Barut, E. (2014), ‘Adaptive robust variable selection’, Annals of Statistics 42(1), 324–351.
- Gardes and Girard (2016) Gardes, L. and Girard, S. (2016), ‘On the estimation of the functional weibull tail-coefficient’, Journal of Multivariate Analysis 146, 29–45.
- Gardes and Stupfler (2019) Gardes, L. and Stupfler, G. (2019), ‘An integrated functional weissman estimator for conditional extreme quantiles’, REVSTAT-Statistical Journal 17(1), 109–144.
- Gnecco et al. (2024) Gnecco, N., Terefe, E. M. and Engelke, S. (2024), ‘Extremal random forests’, DOI: https://doi.org/10.1080/01621459.2023.2300522.
- He et al. (2020) He, F., Wang, H. J. and Tong, T. (2020), ‘Extremal linear quantile regression with weibull-type tails’, Statistica Sinica 30(3), 1357–1377.
- He et al. (2022) He, F., Wang, H. J. and Zhou, Y. (2022), ‘Extremal quantile autoregression for heavy-tailed time series’, Computational Statistics & Data Analysis 176, 107563.
- He et al. (2023) He, X., Pan, X., Tan, K. M. and Zhou, W.-X. (2023), ‘Smoothed quantile regression with large-scale inference’, Journal of Econometrics 232(2), 367–388.
- He and Shao (2000) He, X. and Shao, Q.-M. (2000), ‘On parameters of increasing dimensions’, Journal of Multivariate Analysis 73(1), 120–135.
- Hill (1975) Hill, B. M. (1975), ‘A simple general approach to inference about the tail of a distribution’, Annals of Statistics 3(5), 1163–1174.
- Homrighausen and McDonald (2017) Homrighausen, D. and McDonald, D. J. (2017), ‘Risk consistency of cross-validation with lasso-type procedures’, Statistica Sinica 49(3), 1017–1036.
- Koenker (2005) Koenker, R. (2005), Quantile Regression, Cambridge University Press.
- Koenker and Bassett (1978) Koenker, R. and Bassett, J. G. (1978), ‘Regression quantiles’, Econometrica 46(1), 33–50.
- Koenker et al. (2017) Koenker, R., Chernozhukov, V., He, X. and Peng, L. (2017), Handbook of quantile regression, CRC press.
- Li and Wang (2019) Li, D. and Wang, H. J. (2019), ‘Extreme quantile estimation for autoregressive models’, Journal of Business & Economic Statistics 37(4), 661–670.
- Negahban et al. (2012) Negahban, S. N., Ravikumar, P., Wainwright, M. J. and Yu, B. (2012), ‘A unified framework for high-dimensional analysis of m-estimators with decomposable regularizers’, Statistical Science 27(4), 538–557.
- Neves et al. (2015) Neves, M. M., Gomes, M. I., Figueiredo, F. and Prata Gomes, D. (2015), ‘Modeling extreme events: sample fraction adaptive choice in parameter estimation’, Journal of Statistical Theory and Practice 9(1), 184–199.
- Pan and Zhou (2021) Pan, X. and Zhou, W.-X. (2021), ‘Multiplier bootstrap for quantile regression: non-asymptotic theory under random design’, Information and Inference 10(3), 813–861.
- Sasaki et al. (2024) Sasaki, Y., Tao, J. and Wang, Y. (2024), ‘High-dimensional tail index regression: with an application to text analyses of viral posts in social media’, arXiv:2403.01318.
- Tan et al. (2022) Tan, K. M., Wang, L. and Zhou, W.-X. (2022), ‘High-dimensional quantile regression: Convolution smoothing and concave regularization’, Journal of the Royal Statistical Society Series B: Statistical Methodology 84(1), 205–233.
- Velthoen et al. (2023) Velthoen, J., Dombry, C., Cai, J.-J. and Engelke, S. (2023), ‘Gradient boosting for extreme quantile regression’, Extremes 26(4), 639–667.
- Wang and Li (2013) Wang, H. J. and Li, D. (2013), ‘Estimation of extreme conditional quantiles through power transformation’, Journal of the American Statistical Association 108(503), 1062–1074.
- Wang, Li and He (2012) Wang, H. J., Li, D. and He, X. (2012), ‘Estimation of high conditional quantiles for heavy-tailed distributions’, Journal of the American Statistical Association 107(500), 1453–1464.
- Wang and Tsai (2009) Wang, H. and Tsai, C.-L. (2009), ‘Tail index regression’, Journal of the American Statistical Association 104(487), 1233–1240.
- Wang, Wu and Li (2012) Wang, L., Wu, Y. and Li, R. (2012), ‘Quantile regression for analyzing heterogeneity in ultra-high dimension’, Journal of the American Statistical Association 107(497), 214–222.
- Welsh (1989) Welsh, A. (1989), ‘On m-processes and m-estimation’, Annals of Statistics 17(1), 337–361.
- Xu et al. (2022) Xu, W., Hou, Y. and Li, D. (2022), ‘Prediction of extremal expectile based on regression models with heteroscedastic extremes’, Journal of Business & Economic Statistics 40(2), 522–536.
- Youngman (2019) Youngman, B. D. (2019), ‘Generalized additive models for exceedances of high thresholds with an application to return level estimation for us wind gusts’, Journal of the American Statistical Association 114(528), 1865–1879.
- Zhao et al. (2014) Zhao, T., Kolar, M. and Liu, H. (2014), ‘A general framework for robust testing and confidence regions in high-dimensional quantile regression’, arXiv:1412.8724.
- Zheng et al. (2013) Zheng, Q., Gallagher, C. and Kulasekera, K. (2013), ‘Adaptive penalized quantile regression for high dimensional data’, Journal of Statistical Planning and Inference 143(6), 1029–1038.