Multi-step ahead prediction intervals for non-parametric autoregressions via bootstrap: consistency, debiasing and pertinence
Abstract
To address the difficult problem of multi-step ahead prediction of non-parametric autoregressions, we consider a forward bootstrap approach. Employing a local constant estimator, we can analyze a general type of non-parametric time series model, and show that the proposed point predictions are consistent with the true optimal predictor. We construct a quantile prediction interval that is asymptotically valid. Moreover, using a debiasing technique, we can asymptotically approximate the distribution of multi-step ahead non-parametric estimation by bootstrap. As a result, we can build bootstrap prediction intervals that are pertinent, i.e., can capture the model estimation variability, thus improving upon the standard quantile prediction intervals. Simulation studies are given to illustrate the performance of our point predictions and pertinent prediction intervals for finite samples.
1 Introduction
To model the asymmetry in financial returns, volatility of stock markets, switching regimes, etc., non-linear time series models have attracted attention since the 1980s. Compared to linear time series models, non-linear models possess more capabilities to depict the underlying data-generating mechanism; see the review of politis2009financial for example. However, unlike linear models where the one-step ahead predictor can be iterated, multi-step ahead prediction of non-linear models is cumbersome, since the innovation influences the forecasting value severely.
In this paper, by combining the forward bootstrap of politis2015model with non-parametric estimation, we develop multi-step ahead (conditional) predictive inference for the general model:
(1) |
here, the are assumed to be independent, identically distributed (i.i.d.) with mean 0 and variance 1, and the and are some functions that satisfy some smoothness conditions. We will also assume that the time series satisfying Eq. 1 is geometrically ergodic and causal, i.e., that for any , is independent of .
In Eq. 1, we have the trend/regression function depending on the last data points, while the standard deviation/volatility function depends on the last data points; in many situations, and are taken to be equal for simplicity. Some special cases deserve mention, e.g., if (constant), Eq. 1 yields a non-linear/non-parametric autoregressive model with homoscedastic innovations. The well-known ARCH/GARCH models are a special case of Eq. 1 with .
Although the optimal one-step ahead prediction of Eq. 1 is trivial when we know the regression function or have a consistent estimator of it, the multi-step ahead prediction is not easy to obtain. In addition, it is non-trivial to find the optimal prediction even for the one-step ahead forecasting. In several applied areas, e.g. econometrics, climate modeling, water resources management, etc., data might not possess a finite 2nd moment in which case optimizing loss is vacuous; For all such cases—but also of independent interest—prediction that is optimal with respect to loss should receive more attention in practice; see detailed discussions from Ch. 10 of politis2015model. Later, we will show our method is compatible with both and optimal multi-step ahead predictions.
The efforts to overcome the difficulty of forecasting non-linear time series could be traced back to the work of pemberton1987exact, where a numerical approach was proposed to explore the exact conditional -step ahead optimal prediction of for a homoscedastic Eq. 1. However, this method is intractable computationally with long-horizon prediction, and requires knowledge of the distribution of innovations and the regression function which is not realistic in practice.
Consequently, practitioners started to investigate some suboptimal methods to perform multi-step ahead prediction. Generally speaking, these methods take one of two avenues: (1) Direct prediction or (2) Iterative prediction. The first idea involved working with a different (‘direct’) model, specific to -step ahead prediction, namely:
(2) |
Even though and are unknown to us, we can construct non-parametric estimators and , and plug them in Eq. 2 to perform -step ahead prediction. lee2003new give a review of this approach. However, as pointed out by chen2004nonparametric, a drawback of this approach is that the information of intermediate observations is disregarded. Furthermore, if the of Eq. 1 are i.i.d., then the of Eq. 2 can not be i.i.d. In other words, a practitioner must employ the (estimated) dependence structure of the of Eq. 2 in order to perform the prediction in an optimal fashion
The second idea is “iterative prediction” which employs one-step ahead predictors in a sequential way, to perform a multi-step ahead forecast. For example, consider a 2-step ahead prediction under model Eq. 1; first note that the optimal predictor of is . The optimal predictor of but since is unknown, it is tempting to plug-in in its place. This plug-in idea can be extended to multi-step ahead forecasts but it does not lead to the optimal predictor except in the special case where the function is linear, e.g., in the case of a linear auto-regressive (LAR) model.
Remark.
Since neither of the above two approaches is satisfactory, we propose to approximate the distribution of the future value via a particular type of simulation when the model is known or, more generally, by bootstrap. To describe this approach, we rewrite Eq. 1 as
where is a vector which represents and is some appropriate function. Then, when model and innovation information are known to us, we can create a pseudo value . Take a three-step ahead prediction as an example, the pseudo value can be defined as below:
(3) |
here are simulated as i.i.d. from . Repeating this process to pseudo , the optimal prediction of can be estimated by the mean of . As already discussed, constructing the optimal predictor may also be required since sometimes loss is not well-defined; in our simulation framework, we can construct the optimal prediction by taking the median of . Moreover, we can even build a prediction interval (PI) to measure the forecasting accuracy based on quantile values of simulated pseudo values. The extension of this algorithm to longer step ahead prediction is illustrated in Section 2.
Realistically, practitioners would not know , and . In this situation, the first step is to estimate these quantities and plug them into the above simulation which then turns into a bootstrap method. In the spirit of this idea, some studies were done by adopting different bootstrap techniques. thombs1990bootstrap proposed a backward bootstrap trick to predict model. The advantage of the backward method is that each bootstrap prediction is naturally conditional on the latest observations which coincide with the conditional prediction in the real world. However, this method can not handle non-linear time series whose backward representation may not exist. Later, pascual2004bootstrap proposed a strategy to generate bootstrap series forward. For resolving the conditional prediction issues, they fixed the last bootstrap values to be the true observations and compute predictions iteratively in the bootstrap world starting from there. They then extended this procedure to forecast the GARCH model in pascual2006bootstrap.
Sharing a similar idea, pan2016bootstrap defined the forward bootstrap to do prediction, but they proposed a different PI format which empirically has better performance according to the coverage rate (CVR) and the length (LEN), compared to the PI of pascual2004bootstrap. Although pan2016bootstrap covered the forecasting of a non-linear and/or non-parametric time series model, only one-step ahead prediction was considered. The case of multi-step ahead prediction of non-linear (but parametric) time series models was recently addressed in wu2023bootstrap In the paper at hand, we address the case of multi-step ahead prediction of non-parametric time series models as in Eq. 1. Beyond discussing optimal and point predictions, we consider two types of PI—Quantile PI (QPI) and Pertinent PI (PPI). As already mentioned, the former can be approximated by taking quantile values of the future value’s distribution in the bootstrap world. The PPI requires a more complicated and computationally-heavy procedure to be built as it attempts to capture the variability of parameter estimation. This additional effort results in improved finite-sample coverage as compared to the QPI.
As in most non-parametric estimation problems, the issue of bias becomes important. We will show that debiasing on the inherent bias-type terms of non-parametric estimation is necessary to guarantee the pertinence of a PI when multi-step ahead predictions are required. Although QPI and PPI are asymptotically equivalent, the PPI renders better CVR in finite sample cases; see the formal definition of PPI in the work of politis2015model and pan2016bootstrap. Analogously to the successful construction of PIs in the work of politis2013model, we may employ predictive—as opposed to fitted—residuals in the bootstrap process to further alleviate the finite-sample undercoverage of bootstrap PIs in practice.
The paper is organized as follows. In Section 2, forward bootstrap prediction algorithms with local constant estimators will be given. The asymptotic properties of point predictions and PIs will be discussed in Section 3. Simulations are given in LABEL:Sec:simulation to substantiate the finite-sample performance of our methods. Conclusions are given in LABEL:Sec:conclusion. All proofs can be found in Appendix A. Discussions on the debiasing and pertinence related to building PIs are presented in Appendix B to Appendix D.
2 Non-parametric forward bootstrap prediction
As discussed in Remark Remark, we can apply the simulation or bootstrap technique to approximate the distribution of a future value. In general, this idea works for any geometrically ergodic auto-regressive model no matter in a linear or non-linear format. For example, if we have a known general model at hand, we can do -step ahead predictions according to the same logic of the three-step ahead prediction example of Remark Remark.
To elaborate, we need to simulate as i.i.d. from and then compute pseudo value iteratively with simulated innovations as below:
(4) |
Repeating this procedure times, we can make prediction inference with the empirical distribution of . Similarly, if the model and innovation distribution are unknown to us, we can do the estimation first to get and . Then, the above simulation-based algorithm turns out to be a bootstrap-based algorithm. More specifically, we bootstrap from and calculate pseudo value iteratively with . The prediction inference can also be conducted with the empirical distribution of .
The simulation/bootstrap idea of Remark Remark was recently implemented by wu2023bootstrap in the case where the model is either known or parametrically specified. In what follows, we will focus on the case of a non-parametric model Eq. 1 and will analyze the asymptotic properties of the point predictor and prediction interval. For the sake of simplicity, we consider only the case ; the general case can be handled similarly but the notation is quite more cumbersome. Assume we observe datapoints and we denote them as ; our goal is prediction inference of for some . If we know , and , we can take a simulation approach to develop prediction inference as we explained in Section 1. When , and are unknown, we start by estimating and ; we then estimate based on the empirical distribution of residuals. Subsequently, we can deploy a bootstrap-based method to approximate the distribution of future values. Several algorithms are given for this purpose in the later context.
2.1 Bootstrap algorithm for point prediction and QPI
For concreteness, we focus on local constant estimators, i.e., kernel-smoothed estimators of Nadaraya-Watson type; other estimators can be applied similarly. The local constant estimators of and are respectively defined as:
(5) |
here, is a non-negative kernel function that satisfies some regularity assumptions; see Section 3 for details. We use to represent the bandwidth of kernel functions but may take a different value for mean and variance estimators. Due to the theoretical and practical issues, we need to truncate the above local constant estimators as follows:
(6) |
here, and are large enough and is small enough.
Using and on Eq. 1, we can obtain the fitted residuals which is defined as:
(7) |
Later in Section 3, we will show that the innovation distribution can be consistently estimated by the centered empirical distribution of , i.e., , under some standard assumptions. We now have all the ingredients to perform the bootstrap-based Algorithm 1 to yield the point prediction and QPI of .
Step 1 | With data , construct the estimators and by formula Eq. 6. |
Step 2 | Compute fitted residuals based on Eq. 7, and let . Denote the empirical distribution of the centered residuals for . |
Step 3 | Generate i.i.d. from . Then, construct bootstrap pseudo-values , iteratively, i.e., (8) For example, and . |
Step 4 | Repeating Step 3 times, we obtain pseudo-value replicates of that we denote . Then, and optimal predictors can be approximated by and , respectively. Furthermore, a % QPI can be built as where and denote the and sample quantiles of values . |
Remark.
To construct the QPI of Algorithm 1, we may employ the optimal bandwidth rate, i.e., . However, in practice with small sample size, the QPI has a better empirical CVR for multi-step ahead predictions by adopting an under-smoothing bandwidth; see Appendix B for related discussions and see LABEL:Sec:simulation for simulation comparisons between applying optimal and under-smoothing bandwidths on QPI.
In the next section, we will show the conditional asymptotic consistency of our optimal point predictions and QPI. In particular, we verify that our point predictions converge to oracle optimal point predictors in probability –conditional on . In addition, we look for an asymptotically valid PI with CVR to measure the prediction accuracy conditional on the latest observed data, which is defined as:
(9) |
where and are lower and higher PI bounds, respectively. Although not explicitly denoted, the probability should be understood as conditional probability given . Later, based on a sequence of sets that contains the observed sample with a probability tending to 1, we will show how to build a prediction interval that is asymptotically valid by the bootstrap technique even for model information being unknown.
Although asymptotically correct, in finite samples the QPI typically suffers from undercoverage; see the discussion in politis2015model and pan2016bootstrap. To improve the CVR in practice, we consider taking the predictive residuals to boost the bootstrap process. To derive such predictive residuals, we need to estimate the model based on the delete- dataset, i.e., the available data for the scatter plot of vs. for , i.e., excludes the single point at . More specifically, we define the delete- local constant estimator as:
(10) |
Similarly, the truncated delete- local estimator and can be defined according to Eq. 6. We now construct the so-called predictive residuals as:
(11) |
The -step ahead prediction of with predictive residuals is depicted in Algorithm 2. Although Algorithms 1 and 2 are asymptotically equivalent, Algorithm 2 gives a QPI with better CVR for finite samples; see the simulation comparisons of these two approaches in LABEL:Sec:simulation.
Step 1 | Same with Step 1 of Algorithm 1. |
Step 2 | Compute predictive residuals based on Eq. 11. denote the empirical distribution of the centered predictive residuals . |
Steps 3-4 | Replace by in Algorithm 1. All the rest are the same. |
2.2 Bootstrap algorithm for PPI
To improve the CVR of a PI, we can try to take the variability of the model estimation into account when we build the PI, i.e., we need to mimic the estimation process in the bootstrap world. Employing this idea results in a Pertinent PI (PPI) as discussed in Section 1; see also wang2021model.
Algorithm 3 outlines the procedure to build a PPI. Although this algorithm is more computationally heavy, the advantage is that PPI gives better CVR compared to QPI in practice, i.e., with finite samples; see the examples in LABEL:Sec:simulation.
Step 1 | With data , construct the estimators and by formula Eq. 6. Furthermore, compute fitted residuals based on Eq. 7. Denote the empirical distribution of centered residuals by . |
Step 2 | Construct the or prediction using Algorithm 1. |
Step 3 | (a) Resample (with replacement) the residuals from to create pseudo-errors and . |
(b) Let where is generated as a discrete random variable uniformly on the values . Then, create bootstrap pseudo-data in a recursive manner from the formula (12) | |
(c) Based on the bootstrap data , re-estimate the regression and variance functions according to Eq. 6 and get and ; we use the same bandwidth as the original estimator . | |
(d) Guided by the idea of forward bootstrap, re-define the latest value to match the original, i.e., re-define . | |
(e) With estimators and , the bootstrap data , and the pseudo-errors , use Eq. 12 to generate recursively the future bootstrap data . | |
(f) With bootstrap data and estimators and , utilize Algorithm 1 to compute the optimal bootstrap prediction which is denoted by ; to generate bootstrap innovations, we still use . | |
(g) Determine the bootstrap predictive root: . | |
Step 4 | Repeat Step 3 times; the bootstrap root replicates are collected in the form of an empirical distribution whose -quantile is denoted . The equal-tailed prediction interval for centered at is then estimated by |
Remark 2.1 (Bandwidth choices).
In Step 3 (b) of Algorithm 3, we may use an optimal bandwidth and an over-smoothing bandwidth to generate bootstrap time series so that we can capture the asymptotically non-random bias-type term of non-parametric estimation by the forward bootstrap; see the application in franke2002bootstrap. We can also apply an under-smoothing bandwidth (and then use ) to render the bias term negligible. It turns out that both approaches work well for one-step ahead prediction, although applying the over-smoothing bandwidth may be slightly better. However, taking under-smoothing bandwidth(s) is notably better for multi-step ahead prediction. The reason for this is that the bias term can not be captured appropriately for multi-step ahead estimation with over-smoothing bandwidth. On the other hand, with under-smoothing bandwidth the bias term is negligible; see LABEL:Subsec:PPI for more discussion—also see politisstudentization for related discussion. The simulation studies in Appendix C explore the differences between these two bandwidth strategies.
As Algorithm 2 was a version of Algorithm 1 using predictive (as opposed to fitted) residuals, we now propose Algorithm 4 that constructs a PPI with predictive residuals.
Step 1 | With data , construct the estimators and by formula Eq. 6. Furthermore, compute predictive residuals based on Eq. 11. Denote the empirical distribution of centered residuals by . |
Steps 3-4 | Same as in Algorithm 3 but change the residual distribution from to , and change the application of Algorithm 1 to Algorithm 2. |
3 Asymptotic properties
In this section, we provide the theoretical substantiation for our non-parametric bootstrap prediction methods—Algorithms 1, 2, 3 and 4. We start by analyzing optimal point predictions and QPI based on Algorithms 1 and 2.
Remark.
Since the effect of leaving out one data pair is asymptotically negligible for large , the delete- estimator and are asymptotically equal to and , respectively. Then, the predictive residual is asymptotically the same as the fitted residual ; see Lemma 5.5 of pan2016bootstrap for a formal comparison of these two types of estimators and residuals. Thus, we just give theorems to guarantee the asymptotic properties of point predictions and PIs with fitted residuals. The asymptotic properties for variants with predictive residuals also stand true.