Testing exogeneity in the functional linear regression model
Abstract
We propose a novel test statistic for testing exogeneity in the functional linear regression model. In contrast to Hausman-type tests in finite dimensional linear regression setups, a direct extension to the functional linear regression model is not possible. Instead, we propose a test statistic based on the sum of the squared difference of projections of the two estimators for testing the null hypothesis of exogeneity in the functional linear regression model. We derive asymptotic normality under the null and consistency under general alternatives. Moreover, we prove bootstrap consistency results for residual-based bootstraps. In simulations, we investigate the finite sample performance of the proposed testing approach and illustrate the superiority of bootstrap-based approaches. In particular, the bootstrap approaches turn out to be much more robust with respect to the choice of the regularization parameter.
AMS 2010 Classification: Primary: 62R10, 62F05, 62E20, 62J20; Secondary: 62F40
Keywords and Phrases: asymptotic theory, bootstrap infe
1 Introduction
In functional linear regression models, goodness-of-fit tests are much more complicated to construct then e.g. in the multiple linear setting. This, among others stems to the fact that in functional linear regression models the -distance of the slope function estimator to the true function has no proper limiting distribution. This was shown in [Cardot et al. (2006)] and [Ruymgaart et al. (2000)] for two estimators in the classical functional linear regression model under exogeneity. It turns out, that the lack of a proper limiting distribution also applies for other estimators using different model assumptions. This phenomenon inherent to functional data setups is probably one of the reasons why goodness-of-fit testing is generally not that widely developed for functional regression models yet. In particular, desirable counterparts of standard tests that are well-established in the multiple linear model are still missing in the functional linear setting.
In functional data settings, existing goodness-of-fit tests are described in [Müller and Stadtmüller (2005)], who use a suitable scalar product to transform the functions to a different space using the autocovariance operator to obtain a test statistic having a proper limiting distribution. Further approaches are given in [Cuesta-Albertos et al. (2019)] and [García-Portugués et al. (2014), García-Portugués et al. (2020)], who use random projections together with empirical process techniques.
In practice, one important model assumption is the exogeneity of the regressor. Especially in economics, this assumption is often violated such that the regressors are correlated with the error terms which leads to endogeneity issues. Estimating in such a model is an inverse problem. Neglecting endogeneity generally results in inconsistent estimators. Hence, it is important to test the data for exogeneity first. If the null hypothesis of exogeneity is rejected, different estimators such as e.g. instrumental variable (IV) estimators are required to achieve consistent estimation. See [Johannes (2016)], [Florens et al. (2011)] or [Florens and Van Bellegem (2014)], who consider such IV estimators in functional regression setups and derive asymptotic theory. While in the multiple linear regression model the Hausman test (see [Hausman (1978)] und [Wu (1973)]) is a standard and natural approach for testing exogeneity, this method cannot be transferred directly to the functional linear model since it is based on the -distance of two slope function estimators due to the following proposition which transfers the results in [Cardot et al. (2006)] and [Ruymgaart et al. (2000)] to the present setting.
In the following, let denote the estimator of the slope function in the exogeneous model described in [Johannes (2013)] and [Joh13add], which is consistent under exogeneity, but inconsistent under endogeneity, and by the IV estimator in the endogeneous functional linear model given in [Johannes (2016)], which is consistent in both cases. Then, we get the following result.
Proposition 1.1
In the functional linear regression model (2.1) defined below, even under exogeneity, there is no random variable with non-degenerate distribution, such that
for some real sequence with , where denotes the norm of the Hilbert space.
The proof of this result mainly goes along the lines of the one in [Cardot et al. (2006)], see also [Dorn (2021)] for further details. This is why we just state the result here and use it as motivation for a different approach in the following. Motivated by the fact, that in contrast to the -distance, the projection error typically has an asymptotic distribution (see e.g. [Müller and Stadtmüller (2005)] and [Florens and Van Bellegem (2014)]), we propose to use the sum of the squared difference of projections of the two estimators as test statistic.
The rest of the paper is organized as follows. In Section 2, we state the model assumptions, construct the test statistic and derive its asymptotic distribution. As the limiting distribution turns out to depend on unknown functional nuisance parameters, which are difficult to estimate, we propose residual-based bootstrap methods in Section 3 and prove their consistency. The finite sample performance of all discussed tests is investigated in Section 5. All longer proofs are deferred to the Appendix and additional auxiliary results to a supplement.
2 Model and test statistic
We consider the functional linear regression model
(2.1) |
where is a real-valued random variable, is a real-valued error term with and , is a functional random variable with values in such that . In this setup, the error variance is unknown, and is an unknown function from the Sobolev space of periodically extendable square integrable functions denoted by
(2.2) |
where is the Fourier basis of , and , , see e.g. [Neuba, Neubb], [MR96] or [Tsybakov (2004)]. In the setup of (2.1), we will speak of exogeneity (and call an exogenous regressor), if
(2.3) |
Otherwise, we will speak of endogeneity (and call an endogeneous regressor), if
(2.4) |
For consistent estimation in the endogeneous case, we assume to additionally have a functional instrumental variable with values in such that and for all . For the sake of simplicity, it is often assumed in the literature, that holds for all . However, the general case can be handled along the same lines by centering with the sample mean in a first step and our results are stated for the general case. For estimating the cross-covariance operator, we also assume that is second-order stationary, see [Johannes (2016)].
Assumption 1
There exist functions , such that
for all , respectively, where is assumed to be continuous.
By imposing continuity of , we have that whenever (2.4) holds for one , this immediately implies on some set with positive Lebesgue measure. This condition ensures, that the test statistic proposed in the following can be used to consistently test for the null hypothesis (2.3) against alternatives (2.4).
Note, that and define the kernels of the covariance operators of and of , respectively, and is the kernel of the cross covariance operator of and . The (joint) weak stationarity of ensures, that both covariance operators as well as the cross covariance operator have the same exponential system of eigenfunctions, which we denote by . Hence, let be the eigensystem of , the eigensystem of and the eigensystem of . Furthermore, denote .
Assumption 2
Throughout the article, we assume that all eigenvalues are strictly positive and that
Furthermore, we denote by and the expectations of and , respectively. Additionally, we assume that there exists some such that
(2.5) |
The last assumptions ensures, that the linear prediction of with respect to is well defined.
In principle, if they were available, IV estimation would be based on the optimal instrument defined by
and the eigenvalues of the corresponding cross covariance operator . However, this is usuall not the case and the optimal instrument respectively the corresponding eigenvalues of the cross covariance operator have to be estimated. Note, that could be exactly computed from and , if the (cross) covariance operators were known and remark that for all .
In the following, let be independent and identically distributed (i.i.d.) copies of and suppose (2.3) is valid. Then, we can consistently estimate the unknown slope function due to [Johannes (2013)] and [Johannes (2016)] in two different ways. For this purpose, let be a sequence of regularization parameters such that for all and . To simplify notation, we will write for the regularization keeping in mind that it still depends on . Since the covariance operators and therefore the corresponding eigenvalues are unknown, they have to be estimated in a first step. Further, let denote the empirical versions of and , respectively, defined by
for . These estimators as well as the deduced estimators
for the eigenvalues , , and , respectively, are consistent for all . Hence, observations of the optimal linear instrument can be estimated by
and the corresponding cross covariance operator by
(2.6) |
This allows to construct the IV-based estimator of the slope function defined by
(2.7) |
where
As shown in [Johannes (2016)], under Assumptions 1 and 2 the estimator is consistent under the exogeneity assumption (2.3) as well as under endogeneity of (2.4). In contrast, again under Assumptions 1 and 2, the estimator
(2.8) |
is only consistent under the exogeneity assumption (2.3) (see [Johannes (2013)]) and inconsistent under endogeneity of (2.4). Note, that in comparison to the original definition in [Johannes (2013)], for , we use the same indicator function as in . It turned out, that the tests perform better if the same regularization is used in both estimators although it might not be the best choice for estimating by under assumption (2.3).
Based on the two estimators (2.7) and (2.8), we construct the test statistic as
(2.9) |
The last representation above corresponds to the idea used in [Müller and Stadtmüller (2005)] to construct a goodness-of-fit test. The equivalence of both approaches can be seen by using the singular value decomposition for the estimators and for the covariance operator.
Assumption 3
For the sequence of regularization parameters, we assume
For the next results, different moment conditions for , and are required. To simplify the notation, we introduce the following sets. In doing so, we assume, that all conditions on and mentioned above are fulfilled and define
(2.10) | ||||
(2.11) |
In the following, for an operator , we denote by the regularized inverse of the operator, that is
and we define
(2.12) |
where denotes the Hilbert-Schmidt norm and we set
Now, we are in a position to state an asymptotic result for the test statistic.
Theorem 2.1
Proof. For the sake of simplicity, we assume, that is centered. If not, the additional bias term has to be taken into account as well as stated in the assertion of the theorem. We give a short overview of the proof. The used propositions and lemmas are stated and proven in the appendix. For the employed decomposition of the test statistic, we need several (modified) correlation operators of the instruments and . We define , by
and set
For the test statistic, the following decomposition holds
(2.13) |
where
and
(2.14) |
When subtracting , the last term in (2.13) can be further decomposed to get
where , are defined in the appendix. There, we will also see that
converges weakly due to Theorem A.1 to a normal distribution with mean 0 and variance , while all remaining terms are discussed to be asymptotically negligible using Proposition A.5, A.2, A.3 and A.4 together with standard estimation techniques for the mixed terms. With the Lemma of Slutsky, the assertion follows.
To apply the above result for testing, the bias and variance term have to be estimated. To this end, note that can be consistently estimated by
(2.15) |
due to the law of large numbers and since by similar calculations as in the derivation of the asymptotic distribution of .
Corollary 2.2
Using Corollary 2.2, it is possible to construct a test for the null hypothesis
(2.16) |
against
(2.17) |
For given size , we can reject if
(2.18) |
where denotes the -quantile of the standard normal distribution. That is, we get a one-sided test for against . In the special case we can neglect the additional bias term which avoids the use of its plug-in estimator such that the test has the simpler structure .
Corollary 2.3
Proof. We only consider the special case here. The general case can be proven by similar arguments. Under , is not consistently estimating such that it converges in probability to for some with
which is in general not equal to under endogeneity (by the continuity imposed in Assumption 1). Hence, we have
The standardized version of the first part converges in distribution to a standard normal distribution by similar arguments as in Theorem 2.1 and Corollary 2.2 while the sum of the remainder terms multiplied with goes to infinity for . Consequently, we have
for .
In practice, we do not know, if such that a naive application of the asymptotic test without estimating could result in wrong decisions. In addition, asymptotic tests based on plug-in methods as above usually exhibit a smaller power compared to other methods. This is due to the additional estimation step. The bootstrap version of the test discussed in the next section is expected to have better finite sample behavior, since it is not required to estimate the unknown bias and variance. This has additionally the effect, that we need not distinguish between the cases and which is a clear advantage of the bootstrap test.
3 Bootstrap Consistency
In this section, we use residual-based bootstrap procedures to estimate the distribution of
under the null of exogeneity . To this end, we first estimate the residuals from the original data set and define
where we use the IV-based estimator, because it is consistent under the null hypothesis as well as under the alternative. However, using the classical estimator would also result in a proper bootstrap scheme to approximate the distribution of the test statistics under the null of exogeneity, since the independence of error and regressor in the bootstrap sample is achieved by the (fixed-design) bootstrap procedure itself. However, to get bootstrap data that mimics the true distribution under the null hypothesis of exogeneity given the original sample as close as possible, the IV-based estimator turns out to be more natural and performs better in simulations. In the sequel, different versions of residual-based bootstraps are considered. All bootstrap methods will follow these steps
- Step 1.)
-
Given i.i.d. observations , , we generate a bootstrap sample , , by
where the bootstrap errors are generated from the residuals in such a way that the conditional independence of and is ensured. A thorough discussion, which types of bootstrap are appropriate in this sense follows in the next subsection.
- Step 2.)
-
From , , a bootstrap test statistic is calculated.
- Step 3.)
-
Repeat Steps 1.) and 2.) times, where is large, to get bootstrap realizations of the test statistic and denote by the corresponding empirical -quantile.
As the bootstrap errors are generated such that conditional independence of and is ensured, the bootstrap automatically adopts the exogeneity assumption. For the naive (Efron-type)residual bootstrap, this is trivially the case, because the bootstrap errors are drawn independently with replacement from the residuals, and for the wild bootstrap, since suitable bootstrap multiplier variables will also be drawn independently from and .
Theorem 3.1
Based on this result, we can again construct a one-sided test for the hypotheses (2.17) which rejects the null hypothesis if from Step 3 since and both asymptotically have the same bias and variance.
4 Generalization to other estimators and measuring goodness-of-fit
While the above results are stated for the spectral-cut-off estimators as proposed in [Johannes (2013)] and [Johannes (2016)], it is also possible to derive analogue results for other types of estimators like cut-off as in [Müller and Stadtmüller (2005)] or ones based on Tikhonov or ridge-type regularization. A quite general approach is given in [Cardot et al. (2006)] with a sequence of regularization functions such that is decreasing on where are the eigenvalues of the relevant covariance operator and is a decreasing sequence of positive values with . Furthermore and is differentiable on which replaces Assumption 3. While the estimator from (2.8) above does not completely fit this situation it is not neccessary to consider this modification if one is only interested in testing goodness-of-fit in an exogeneous model (2.1). For the sake of shorter notation we assume here . If denotes the estimator proposed in [Cardot et al. (2006)] we obtain under Assumption 1 and the moment conditions in Theorem 2.1 the following result
with , as in Theorem 2.1 and
If is straight forward to also generalize the instrumental variable estimator to other regularization schemes. We get an estimator
and, if we are willing to assume exogeneity here, derive by the same arguments as above
with , as in Theorem 2.1 and
The assumption of exegeneity is in this case not realistic because one would only use the instrumental variable estimator under endogeneity. Proving an analogue result under endogeneity is in principle possible but the proof differs in several points from the one presented here.
Using the estimators and we can construct a test statistic similar to the one above. To this end we need a similar regularization scheme for both estimators. If we allow for a second argument in the estimators involved in the test above can also be written with for and for and it is straight forward to generalize them at least to regularisation functions of type respectively . Under Assumption 1, the moment assumptions of Theorem 2.1 and certain regularity conditions we derive under the null hypthesis
For all results presented in this section it is again straight forward to derive empirical versions and bootstrap results.
5 Finite sample properties
In this section, we investigate the finite sample behavior of the tests proposed above under several degrees of endogeneity and for different slope functions. We generate our data from the model
and
for some large enough to approximate the integral sufficiently well. To control all correlations in the model, we generate i.i.d. copies of
with , see [Wong (1996)]. The random variable is uniformly distributed on and independent of . The parameter controls the severity of endogeneity (if we are in the exogenous case, i.e. under the null ) and the strength of the instrument . The standard deviation is assumed to be . In the following, as illustrated in Figure 1, we will use three different slope functions and defined by
(5.1) |

where in , and with
and . For all simulations, we generate 1000 Monte Carlo realizations and use bootstrap replications.
Besides an Efron-type residual-based bootstrap, which draws the bootstrap errors , independently with replacement from the residuals , we consider also several versions of a residual-based wild bootstrap, where
and the ’s are i.i.d. with and and independent of . We consider different choices for the distribution of the ’s as commonly used in the literature, see e.g. [Mammen (1993)],
(5.2) | ||||
(5.3) | ||||
(5.4) |

In a first step we try to get an idea how to choose and, in a next step how to choose . To this end, we fix the degree of endogeneity with and the strength of the instrument with . In Figure 2, the results for the asymptotic test using as slope parameter and different choices of are shown. We see that the best results are obtained for between and . For smaller , the test does not hold the prescribed level, while for larger the power is comparably small up to biased tests for larger than . Based on Figure 2, we can find a sequence of good choices for depending on the sample size varying from for to for larger sample sizes up to . We see that the asymptotic test has only moderate power even for larger sample sizes. This is a well known effect with asymptotic tests using plug-in estimators.
The way out is typically a bootstrap-based test. The results for the residual-based bootstraps proposed in Section 3 and again are shown in Figure 3.

It turns out, that the regularization parameter can be chosen considerably smaller than for the asymptotic test and the procedure is much more robust in choosing . Nearly all tests hold the size of for larger sample sizes and the power increases with sample size for most choices of up to a value close to 1 already for . Again we can get an idea of choosing a good depending on the sample size which varies from for to for and .
Apparently all bootstrap procedures discussed in Section 3 perform comparably good which can be seen in Figure 4 for a choice of .

Comparing the performance of the bootstrap test for different slope functions, we discover that in all models the bootstrap test holds the size while we see in Table 1 that the power is similarly good for all settings with only sligh disadvantages for the smoothed indicator function .
25 | 50 | 75 | 100 | 125 | 150 | 175 | 200 | 225 | 250 | 275 | 300 | |
---|---|---|---|---|---|---|---|---|---|---|---|---|
0.111 | 0.507 | 0.773 | 0.901 | 0.960 | 0.980 | 0.992 | 0.997 | 0.998 | 0.998 | 1 | 1 | |
0.164 | 0.568 | 0.798 | 0.912 | 0.958 | 0.979 | 0.992 | 0.997 | 0.999 | 0.998 | 1 | 1 | |
0.255 | 0.560 | 0.733 | 0.853 | 0.904 | 0.961 | 0.978 | 0.990 | 0.993 | 0.994 | 0.997 | 0.998 |
Finally, we inspect the influence of the degree of endogeneity and the strength of the instrument on the performance of the test. In Figure 5, we see that the power of the bootstrap test increases with increasing degree of endogeneity being already acceptable for .

Figure 6 shows, that the performance of the test is highly dependent on the strength of the instrument. If the instrument is too weak, the power is too low and the test does not hold the size. It turns out, that for the setting with slope function , and , the bootstrap test performs best for a strength of the instrument around .

6 Concluding remarks
The underlying work is the first approach of testing for endogeneity in a functional regression setup by introducing a modified approach of the classical Hausman test in a multiple linear regression model. This modification is required, because the -distance of two slope function estimators in functional linear regression models are shown to have no proper limiting distribution. We prove asymptotic normality for the proposed modified Hausman-type test statistic, which allows for the construction of asymptotic tests for exogeneity. As the asymptotic test has several drawbacks such as many nuisance parameters, which are cumbersome to estimate, an additional bias term, which diverges when multiplied with the rate of convergence, and a high sensitivity to the choice of the regularization parameter, we propose suitable bootstrap versions of the test to approximate the null distribution. This avoids the additional estimation of nuisance parameters and turns out to be much more robust to the choice of the regularization parameter. This behavior is demonstrated in a detailed simulation study. Topics of ongoing work are the choice of the instrument, a data driven choice of the regularization parameter and the transfer to other regression models.
Appendix A Auxiliary Results for the Proof of Theorem 2.1
We assume for the sake of simplicity for all and remember from Section 2 the decomposition of the test statistic with
(A.1) | ||||
(A.2) |
and define
(A.3) | ||||
(A.4) | ||||
(A.5) |
The first result establishes the asymptotic distribution of the test statistic.
Theorem A.1
Under the assumptions of Theorem 2.1, under the null hypothesis, and for and for all , we have
The remaining results are to show, that the remainder terms are negligible.
Proposition A.2
Let and . Under the assumptions of Theorem 2.1, we have
Proposition A.3
Under the assumptions of Theorem 2.1 and if and , we have
Proposition A.4
Under the assumptions of Theorem 2.1, and if and , we have
Proposition A.5
Under the assumptions of Theorem 2.1, and if and , we have
Appendix B Auxiliary results
The results in this section are used at several places in the proofs. They follow from Lemma A.1 in [Johannes (2016)].
Lemma B.1
Let and have finite second moments and . Then we have and . If additionally und , we have
Lemma B.2
Let be fixed and suppose and . Then, there is a positive Konstant such that for , we have
(B.1) |
and
(B.2) |
Appendix C Proof of Theorem A.1
The proof follows by using a central limit theorem for martingal difference sequences with respect to , where and , see [Hall und Heyde (1980)], Theorem 3.2 and Corollary 3.1, for
where
and
In a first step, we consider the conditional variance of the maringale difference scheme.
Proposition C.1
Under the assumptions of Theorem 2.1, under the null hypothesis and for , we have
Proof. Using that is independent of , we can decompose
We define
and show
by proving the corresponding -convergence. Afterwards we show that converges in probability to . Writing for and
and, observing that for some constant , we get
with
We have
(C.1) |
because and are uncorrelated for all and . With Lemma B.1 and (E.6), for all and , we have
(C.2) |
as well as
(C.3) |
For the mixed terms with and and for all , we get
(C.4) |
Using this, we have
with some constant . With similar arguments we obtain
and
which altogether results in
The stochastic convergence of follows by
for . For proving, that converges stochastically to 0 we show again the corresponding -convergence. To this end, we bound for all und the term by a constant using the centeredness of and Lemma B.1, to obtain
The detailed arguments can be found in the supplementary material.
The second step is to show the conditional Lindeberg condition by verifying an unconditional Ljapunov condition.
Proposition C.2
Under the assumptions of Theorem 2.1, under the null hypothesis, and with , we have
(C.5) |
Proof. It is shown in [Alj et al. (2014)] and [Gänssler et al. (1978)] that the conditional Lindeberg condition follows from the unconditional Ljapunov condition. We will show in the following, that
and decompose
where
For we use, that for all , are stochastically independent of and are uncorrelated with . Furthermore, the fourth absolute moment of is due to the centredness of and Lemma B.1 uniformly bounded. The fourth absolute moment of can be estimated using Assumption 3 and as
(C.6) |
Again using similar arguments, we obtain
(C.7) |
This results in
(C.8) |
Putting these results together, we get
where the first series converges due to Lemma B.1 and the second series either also converges or, if not, can be bounded by .
Considering , we use the stochastic independence of and for all , which results in
The rest of the argumentation is just calculating the expectations using that for all , and are uncorellated with for all and stochastically independent of . Finally,
(C.9) |
and, in the same way, , which gives .
With similar arguments as above, which can be found in the supplementary material we get
and
All remainder terms can be estimated with similar techniques. We exemplarily show the idea for Proposition A.5, that is for , in the supplementary material.
Appendix D Proof of Theorem 3.1
Let denote the distribution function of the normal distribution with mean zero and variance , the distribution function of and the distribution function of the conditional distribution of given . By bounding
similar to the example in Section 29 of [DasGupta (2008)], it is enough to show the convergence of and . Due to the continuity of , the convergence of follows directly from Theorem 2.1 and Polya’s Theorem, as stated in Section 1.5.3 of [Serfling (1980)]. Again, using Polya’s Theorem, it is enough to show for , that for all
(D.1) |
For this we just immitate the proof of Theorem 2.1. Analogously to (2.13), we decompose
where, similar to the proof of Theorem 2.1, we get
Then, converges weakly in probability to along the lines of Theorem A.1. The remainder terms can be discussed to be negligible with the same arguments as for the remainder terms in Theorem 2.1.
Appendix E Supplementary Material
E.1 Proof of Proposition A.5
We give only the proof for . We have
The terms quadratic in can be estimated by Lemma B.1 und (E.6), while the other terms except the one coming from vanish
Using the Cauchy-Schwarz inequality (E.3), leads to
The expectations with , und , can be estimated by (E.4). This finally yields
The second part can be shown by using
(E.1) |
for all together with Lemma B.1
All the other parts of Proposition A.5 as well as Lemmas A.2-A.4 follow by very similar techniques. For details we refer to [5] in the main article.
E.2 Details for the proof of Proposition C.1
Using that is independent of , we can decompose
We define
and show
by proving the corresponding -convergence. Afterwards we show that converges in probability to . Writing for and
and, observing that for some constant , we get
with
We have
(E.2) |
because and are uncorrelated for all and . With Lemma B.1 and (E.6), for all and , we have
(E.3) |
as well as
(E.4) |
For the mixed terms with and and for all , we get
(E.5) |
Using this, we have
with some constant . With similar arguments, we obtain
which can be further bounded using the Cauchy-Schwarz inequality to get
Using similar arguments as for the first two terms, can also be bounded to get
Altogether, we have
The stochastic convergence of follows by
for . For proving, that converges stochastically to 0 we show again the corresponding -convergence. To this end we bound for all und the term by a constant using the centredness of and Lemma B.1, to obtain
Since and are stochastically independent for , only the quadratic terms for are relevant
Under the assumptions of Theorem 2.1, this leads to
and therefore
E.3 Details for the proof of Proposition C.2
It is shown in [1] and [10] that the conditional Lindeberg condition follows from the unconditional Ljapunov condition. We will show in the following, that
and decompose
where
For we use, that for all , are stochastically independent of and are uncorrelated with . Furthermore, the fourth absolute moment of is due to the centredness of and Lemma B.1 uniformly bounded. The fourth absolute moment of can be estimated using Assumption 3 and as
(E.6) |
Again using similar arguments, we obtain
(E.7) |
This results in
(E.8) |
Putting these results together, for , we get
where the first series converges due to Lemma B.1 and the second series either also converges or, if not, can be bounded by .
Considering , we use the stochastic independence of and for all , which results in
The rest of the argumentation is just calculating the expectations using that for all , and are uncorrelated with for all and stochastically independent of . Finally,
(E.9) |
and, in the same way, , which gives .
With similar arguments as above, we get
which can be further bounded by using
and
This results in
using the Hölder inequality and Lemma B.1.
For the summands in , we get
The first expectation is
while
Altogether, we have
The series can be bounded by . Using the Hölder inequality for , we have
Finally, relying again on Assumption 3 and Lemma B.1, also converges to 0 due to
Acknowledgments
The authors would like to thank Jan Johannes for helpfull discussions on the different estimation techniques in the functional linear regression model with and without endogeneity.
References
- [Alj et al. (2014)] Alj, Abdelkamel, Azrak, Rajae and Mélard, Guy. On Conditions in Central Limit Theorems for Martingale Difference Arrays Long Version. ECORE Discussion Paper (2014/12).
- [Cardot et al. (2006)] Cardot, Hervé, Mas, André and Sarda, Pascal. CLT in functional linear regression models. Probability Theory and Related Fields (2007) 138:325–361.
- [Cuesta-Albertos et al. (2019)] Cuesta-Albertos, Juan A., García-Portugués, Eduardo, Febrero-Bande, Manuel and González-Manteiga, Wenceslao. Goodness-of-fit tests for the functional linear model based on randomly projected empirical processes. The Annals of Statistics (2019) 47(1):439–467.
- [DasGupta (2008)] DasGupta, Anirban. Asymptotic Theory of Statistics and Probability. Springer Texts in Statistics. Springer, New York 2008.
- [Dorn (2021)] Dorn, Manuela. Tests auf Exogenität im funktionalen linearen Regressionsmodell unter schwacher Stationarität. Dissertation. University of Bayreuth 2021.
- [Florens et al. (2011)] Florens, Jean-Pierre, Johannes, Jan and Van Bellegem, Sébastien. Identification and estimation by penalization in nonparametric instrumental Regression. Econometric Theory (2011) 27:472–496.
- [Florens and Van Bellegem (2014)] Florens, Jean-Pierre and Van Bellegem, Sébastien. Instrumental variable estimation in functional linear models. ECORE Discussion Paper (2014/56).
- [García-Portugués et al. (2014)] García-Portugués, Eduardo, González-Manteiga, Wenceslao and Febrero-Bande, Manuel. A goodness-of-fit test for the functional linear model with scalar response. arXiv:1205.6167v6.
- [García-Portugués et al. (2020)] García-Portugués, Eduardo, Álvarez-Liébana, Javier, Álvarez-Pérez, Gonzalo and González-Manteiga, Wenceslao. Goodness-of-fit tests for functional linear models based on integrated projections. arXiv:2008.09885v1.
- [Gänssler et al. (1978)] Gänssler, Peter, Strobel, J. and Stute, Winfried. On central limit theorems for martingale triangular arrays. Acta Mathematica Academiae Scientiarum Hungaricae (1978) 31(3–4):205–216.
- [Hall und Heyde (1980)] Hall, Peter and Heyde, Christopher C. Martingale Limit Theory and Its Application. Probability and mathematical statistics. Academic Press, New York 1980.
- [Hausman (1978)] Hausman, Jerry A. Specification Tests in Econometrics, Econometrica (1978) 46(6):1251–1271.
- [Johannes (2013)] Johannes, Jan. Nonparametric estimation in functional linear models with second order stationary regressors. arXiv:0901.4266v1.
- [Johannes (2016)] Johannes, Jan. Functional linear instrumental regression under second order stationarity. arXiv:1603.01649v1.
- [Mammen (1993)] Mammen, Enno. Bootstrap and wild bootstrap for high dimensional linear models. The Annals of Statistics (1993) 21(1):255–285.
- [Müller and Stadtmüller (2005)] Müller, Hans-Georg and Stadtmüller, Ulrich. Generalized functional linear models. The Annals of Statistics (2005) 33(2):774–805.
- [Neubauer (1988a)] Neubauer, Andreas (1988). An a posteriori parameter choice for Tikhonov regularization in the presence of modeling error. Appl. Numer. Math. 4, 507–519
- [Ruymgaart et al. (2000)] Ruymgaart, Frits, Wang, Jing, Wei, Shih-Hsuan and Yu, Li. Some asymptotic theory for functional regression and classification. Texas Tech University, Lubbock 2000.
- [Tsybakov (2004)] Tsybakov, Alexandre B. Introduction à l’estimation non-paramétrique., Math. Appl. (Berl.) 41, x + 175
- [Serfling (1980)] Serfling, Robert J. Approximation theorems of mathematical statistics. Wiley Series in Probability and Mathematical Statistics. John Wiley and Sons, New York u. a. 1980.
- [Wong (1996)] Wong, Ka-fu. Bootstrapping Hausman’s exogeneity test. Economics Letters (1996) 53:139–143.
- [Wu (1973)] Wu, De-Min. Alternative tests of independence between stochastic regressors and disturbances. Econometrica (1973) 41(4):733–750.