This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Indirect inference for locally stationary ARMA processes with stable innovations

\nameShu Wei Chou-Chen and Pedro A. Morettin CONTACT Shu Wei Chou-Chen Email: [email protected] Institute of Mathematics and Statistics, University of São Paulo, Brazil

Indirect inference for locally stationary ARMA processes with stable innovations

\nameShu Wei Chou-Chen and Pedro A. Morettin CONTACT Shu Wei Chou-Chen Email: [email protected] Institute of Mathematics and Statistics, University of São Paulo, Brazil
Abstract

The class of locally stationary processes assumes that there is a time-varying spectral representation, that is, the existence of finite second moment. We propose the α\alpha-stable locally stationary process by modifying the innovations into stable distributions and the indirect inference to estimate this type of model. Due to the infinite variance, some of interesting properties such as time-varying autocorrelation cannot be defined. However, since the α\alpha-stable family of distributions is closed under linear combination which includes the possibility of handling asymmetry and thicker tails, the proposed model has the same tail behavior throughout the time. In this paper, we propose this new model, present theoretical properties of the process and carry out simulations related to the indirect inference in order to estimate the parametric form of the model. Finally, an empirical application is illustrated.

keywords:
Locally stationary process; stable distribution; indirect inference

1 Introduction

The class of locally stationary processes describes processes that are approximately stationary in a neighborhood of each time point but its structure, such as covariances and parameters, gradually changes throughout the time period [1, 2]. This type of processes has been proved to achieve meaningful asymptotic theory by applying infill asymptotics. The idea of this approach is that the time-varying parameters are rescaled to the unit interval, and thus, more available observations imply obtaining more contribution for each local structure. Consequently, statistical asymptotic results such as consistency, asymptotic normality, efficiency, locally asymptotically normal expansions, etc. are obtained. [3] provided a review of this type of process.

Most results of locally stationary processes assume innovations with finite second moment. However, different areas have observed phenomena with heavy tail distributions or infinite variance. In this work, we consider that the innovations of the locally stationary process follow α\alpha-stable distributions. The advantage of assuming α\alpha-stable distributions is its flexibility for asymmetry and thick tails. Also, it is closed under linear combinations and includes the Gaussian distribution as a special case. However, its estimation is difficult since the density function does not have a closed-form. Consequently, the usual estimation methods such as maximum likelihood and method of moments do not work.

Alternative estimation approaches such as methods based on quantiles [4] or on the empirical characteristic function [5] are proposed. However, those methods are only useful for the estimation of the α\alpha-stable distributions parameters and, therefore, they are difficult to apply for more complex models.

The strategy to estimate this kind of process is the indirect inference proposed by [6] and [7]. Since α\alpha-stable distributions can be easily simulated, the indirect approach, which is an intensive computationally simulation based method, can be a solution to overcome the estimation problem.

Models involving stable distribution were successfully implemented in indirect inference for independent samples from the α\alpha-stable distributions and α\alpha-stable ARMA processes [8]. Moreover, some time series models involving stable distributions are also successfully implemented using indirect inference [9, 10, 11, 12].

Our contribution in this work is twofold. First, we propose the locally stationary processes with stable innovations and present the theoretical properties of this model. We also justify the reason why we call them α\alpha-stable locally stationary processes. Second, we propose the indirect inference in order to estimate the models with linear time-varying coefficient.

The paper is organized as follows. In Section 2, we review the basic background on locally stationary processes, α\alpha-stable distribution and indirect inference. Then, properties of the α\alpha-stable locally stationary processes are presented in Section 3. Section 4 describes the indirect inference for this kind of processes. Simulations are performed to study the indirect inference approach in Section 5. A wind data application is illustrated in Section 6. Finally, conclusions are presented in Section 7.

2 Background

2.1 Locally stationary processes

Locally stationary processes were introduced by using a time-varying spectral representation [2]. However, we use the time-domain version as in [13] since stable distributions do not have finite second moment.

Definition 2.1.

The sequence of stochastic processes Xt,TX_{t,T} (t=1,,T)(t=1,...,T) is a linear locally stationary processes if Xt,TX_{t,T} has a representation

Xt,T=j=at,T(j)εtj,X_{t,T}=\sum_{j=-\infty}^{\infty}a_{t,T}(j)\varepsilon_{t-j}, (1)

where the following conditions are satisfied:

  1. (i)
    supt|at,T(j)|K(j),with K independent of T;\sup_{t}\left|a_{t,T}(j)\right|\leq\frac{K}{\ell(j)},~{}~{}~{}\text{with K independent of T;} (2)
  2. (ii)

    Let V(g)V(g) be the total variation of a function gg on [0,1]\left[0,1\right]; then, there exist funcions α(,j):(0,1]\alpha(\cdot,j):\left(0,1\right]\rightarrow\mathbb{R} with

    supu|a(u,j)|K(j),\sup_{u}\left|a(u,j)\right|\leq\frac{K}{\ell(j)}, (3)
    supjt=1T|at,T(j)a(tT,j)|K,\sup_{j}\sum_{t=1}^{T}\left|a_{t,T}(j)-a\left(\frac{t}{T},j\right)\right|\leq K, (4)
    V(a(,j))K(j);V(a(\cdot,j))\leq\frac{K}{\ell(j)}; (5)
  3. (iii)

    {εt}\left\{\varepsilon_{t}\right\} are i.i.d. with E[εt]=0E\left[\varepsilon_{t}\right]=0, E[εs,εt]=0E\left[\varepsilon_{s},\varepsilon_{t}\right]=0 for sts\neq t and E[εt2]=1E\left[\varepsilon_{t}^{2}\right]=1.

Note that the condition (iii) in the Definition 2.1 assumes that the innovations {εt}\left\{\varepsilon_{t}\right\} has zero mean and unit variance. In this case, there exists a spectral representation of the process XtX_{t} and time-varying spectral density. Note that classical stationary processes arise as a special case when all parameter curves are constant.

2.2 α\alpha-stable distribution

Stable distributions, as an extension of Gaussian distributions, can be defined by its characteristic function:

E{eiθX}={exp{σα|θ|α(1iβ(signθ)tanπα2)+iμθ},if α1,exp{σ|θ|(1+iβ2π(signθ)log|θ|)+iμθ},if α=1,E\left\{e^{i\theta X}\right\}=\left\{\begin{matrix}\exp\left\{-\sigma^{\alpha}|\theta|^{\alpha}(1-i\beta(\operatorname{sign}\theta)\tan\frac{\pi\alpha}{2})+i\mu\theta\right\},&\text{if $\alpha\neq 1$,}\\ \exp\left\{-\sigma|\theta|(1+i\beta\frac{2}{\pi}(\operatorname{sign}\theta)\log|\theta|)+i\mu\theta\right\},&\text{if $\alpha=1$,}\end{matrix}\right. (6)

where α(0,2]\alpha\in\left(0,2\right] is the index of stability (tail heaviness), σ>0\sigma>0 is the scale parameter, 1β1-1\leq\beta\leq 1 the asymmetry parameter and μ\mu\in\mathbb{R} the location parameter. It is denoted by Sα(σ,β,μ)S_{\alpha}(\sigma,\beta,\mu). Important properties can be found in detail in [14].

This class of distribution generalizes some important known distributions: normal distribution for α=2\alpha=2, Cauchy distribution (α=1,β=0\alpha=1,~{}\beta=0) and Lévy distribution (α=1/2,β=±1\alpha=1/2,~{}\beta=\pm 1). However, it does not have a closed-form density function in general. Moreover, the non-existence of moments greater than α\alpha makes it difficult to estimate parameters. Different estimation approaches have been proposed, such as methods based on quantiles [4] and methods based on the empirical characteristic function [5]. Nevertheless, they are only useful for independent samples from stable distributions and difficult to perform for more complex models.

When a random variable has the density function and distribution function, its simulation is an easy task. In stable distribution case, [15] proposed an algorithm to generate α\alpha-stable distribution. To simulate a random variable XSα(1,β,0)X\sim S_{\alpha}(1,\beta,0):

  1. 1.

    generate a random variable UU uniformly distributed on (π2,π2)\left(-\frac{\pi}{2},\frac{\pi}{2}\right), and a independent exponential random variable WW with mean 11, then

  2. 2.

    let Bα,β=arctan(βtanπα2)αB_{\alpha,\beta}=\frac{\arctan\left(\beta\tan\frac{\pi\alpha}{2}\right)}{\alpha} and Sα,β=[1+β2tan2πα2]1/(2α)S_{\alpha,\beta}=\left[1+\beta^{2}\tan^{2}\frac{\pi\alpha}{2}\right]^{1/(2\alpha)} , and compute

    X={Sα,βsin(α(U+Bα,β))(cosU)1/α[cos(Uα(U+Bα,β))W]1αα,if α1,2π[(π2+βU)tanUβlog(π2WcosUπ2+βU)],if α=1.X=\left\{\begin{matrix}S_{\alpha,\beta}\frac{\sin\left(\alpha(U+B_{\alpha,\beta})\right)}{(\cos U)^{1/\alpha}}\left[\frac{\cos(U-\alpha(U+B_{\alpha,\beta}))}{W}\right]^{\frac{1-\alpha}{\alpha}},&\text{if $\alpha\neq 1$,}\\ \frac{2}{\pi}\left[\left(\frac{\pi}{2}+\beta U\right)\tan U-\beta\log\left(\frac{\frac{\pi}{2}W\cos U}{\frac{\pi}{2}+\beta U}\right)\right],&\text{if $\alpha=1$.}\end{matrix}\right. (7)

Next, YSα(σ,β,μ)Y\sim S_{\alpha}(\sigma,\beta,\mu) is obtained by means of the standardization formula:

Y={σX+μ,if α1,σX+2πβσlogσ+μ,if α=1.Y=\left\{\begin{array}[]{ll}\sigma X+\mu,&\text{if $\alpha\neq 1$,}\\ \sigma X+\frac{2}{\pi}\beta\sigma\log\sigma+\mu,{}{}&\text{if $\alpha=1$.}\end{array}\right. (8)

Since stable distributions can be easily simulated, the indirect approaches proposed by [6] and [7] could be the solution to more complex models involving stable distributions.

2.3 Indirect inference

The indirect inference proposed by [6] is based on a very simple idea and it is suitable for situations where the direct estimation of the model is difficult. Let 𝒚\boldsymbol{y} be a sample of TT observations from a model of interest (IM) and θ^\hat{\theta} be the maximum likelihood estimator (MLE) of θΘ\theta\in\Theta in IM which is unavailable. Then, consider the auxiliary model (AM) depending on a parameter vector λΛ\lambda\in\Lambda whose likelihood function is easier to handle, but its MLE λ^\hat{\lambda} is not necessarily consistent. The indirect inference is carried out by the following steps:

Step 1

Compute λ^\hat{\lambda} based on 𝒚\boldsymbol{y}.

Step 2

Simulate a set of SS vectors of size TT from the IM on the basis of an arbitrary parameter vector θ^(0)\hat{\theta}^{(0)}. Let us denote each of those vectors as ys(θ^(0)),s=1,,Sy^{s}(\hat{\theta}^{(0)}),~{}s=1,...,S.

Step 3

Then, estimate parameters of the AM using simulated values from the IM,

λ^S(θ^(0))=argmaxλZs=1St=1Tln~[λ;ys(θ^(0))].\hat{\lambda}_{S}(\hat{\theta}^{(0)})=\operatorname*{argmax}_{\lambda\in Z}\sum_{s=1}^{S}\sum_{t=1}^{T}\ln\tilde{\mathcal{L}}\left[\lambda;y^{s}(\hat{\theta}^{(0)})\right]. (9)
Step 4

Numerically update the initial guess θ^(0)\hat{\theta}^{(0)} in order to minimize the distance

[λ^λ^S]Ω[λ^λ^S],\left[\hat{\lambda}-\hat{\lambda}_{S}\right]^{\prime}\Omega\left[\hat{\lambda}-\hat{\lambda}_{S}\right], (10)

where Ω\Omega is a symmetric nonnegative matrix defining the metric.

For choosing Ω\Omega, [6] proved that when the parameter vectors of both AM and IM have the same dimension and TT is sufficiently large, the estimator does not depend on the matrix Ω\Omega. In this paper, we consider Ω\Omega as identity matrix. Finally, the estimation step is performed with a numerical algorithm, such as Newton-Raphson. Then, for a given estimate θ^(p)\hat{\theta}^{(p)}, the procedure yields θ^(p+1)\hat{\theta}^{(p+1)} and the process will be repeated until the series of θ^(p)\hat{\theta}^{(p)} converges. The estimator is then given by

θ^=limpθ^(p).\hat{\theta}=\lim\limits_{p\rightarrow\infty}\hat{\theta}^{(p)}. (11)

3 tvARMA with stable innovations

An important example of the locally stationary process is the time-varying ARMA model, briefly tvARMA. In this section, we will consider this model with stable innovations.

Definition 3.1 (tvARMA with stable innovations).

Consider the system of difference equations

j=0pαj(tT)Xtj,T=k=0qβk(tT)γ(tkT)εtk,\sum_{j=0}^{p}\alpha_{j}\left(\frac{t}{T}\right)X_{t-j,T}=\sum_{k=0}^{q}\beta_{k}\left(\frac{t}{T}\right)\gamma\left(\frac{t-k}{T}\right)\varepsilon_{t-k}, (12)

where εt\varepsilon_{t} are i.i.d. and εtSα(1/2,β,0)\varepsilon_{t}\sim S_{\alpha}(\nicefrac{{1}}{{\sqrt{2}}},\beta,0) with α(0,2)\alpha\in(0,2). Assume α0(u)β0(u)1\alpha_{0}(u)\equiv\beta_{0}(u)\equiv 1 and αj(u)=αj(0)\alpha_{j}(u)=\alpha_{j}(0), βk(u)=βk(0)\beta_{k}(u)=\beta_{k}(0) for u<0u<0. Suppose also that all αj()\alpha_{j}(\cdot) and βk()\beta_{k}(\cdot), as well as γ2()\gamma^{2}(\cdot), are of bounded variation.

The reason that the scale parameter of the innovations is set to be σ=1/2\sigma=\nicefrac{{1}}{{\sqrt{2}}} is when α=2\alpha=2, the standardized Gaussian innovation is obtained. It is possible to define the equation (12) as:

Φt,T(B)Xt,T=Θt,T(B)zt,T,\Phi_{t,T}(B)X_{t,T}=\Theta_{t,T}(B)z_{t,T}, (13)

where zt,T=γ(tT)εtz_{t,T}=\gamma(\frac{t}{T})\varepsilon_{t}; Φt,T(B)=1+α1(tT)B++αp(tT)Bp\Phi_{t,T}(B)=1+\alpha_{1}(\frac{t}{T})B+\cdots+\alpha_{p}(\frac{t}{T})B^{p} and Θt,T(B)=1+β1(tT)B++βq(tT)Bq\Theta_{t,T}(B)=1+\beta_{1}(\frac{t}{T})B+\cdots+\beta_{q}(\frac{t}{T})B^{q} are the autoregressive (AR) and moving average (MA) operators, respectively.

There are several works related to stable linear processes. For instance, chapter 7 in [16] and Chapter 13 in [17] give a general review of stable linear processes. [18] study the infinite variance stable ARMA procseses and [19] study fractional ARIMA with stable innovations. [20] proposed a Whittle-type estimator to estimate the coefficients of the ARMA model. In the stable innovation and time-varying coefficient context, [21, 22] considered the univariate and multivariate case of the system (13) with symmetric stable innovations and assume γ()=1\gamma(\cdot)=1. However, they considered time-dependent coefficient without the local stationarity condition.

3.1 Existence and Uniqueness of a Solution

Before we study the local stationarity conditions on the time-varying coefficients, we present a set of regularity conditions of existence and uniqueness of solution of the system based on the concepts defined by [21, 22].

Definition 3.2.
  1. 1.

    The process (13) is AR regular (or causal) if there exist at,T(j)a_{t,T}(j) such that

    Xt,T=j=0at,T(j)εtj.X_{t,T}=\sum_{j=0}^{\infty}a_{t,T}(j)\varepsilon_{t-j}. (14)

    satisfying j=0|at,T(j)|δ<\sum\limits_{j=0}^{\infty}|a_{t,T}(j)|^{\delta}<\infty for all tt and δ=min{1,α}\delta=min\{1,\alpha\}.

  2. 2.

    The process (13) is MA regular (or invertible) if there exist bt,T(j)b_{t,T}(j) such that

    εt=j=0bt,T(j)Xtj,T.\varepsilon_{t}=\sum_{j=0}^{\infty}b_{t,T}(j)X_{t-j,T}. (15)

    satisfying j=0|bt,T(j)|δ<\sum\limits_{j=0}^{\infty}|b_{t,T}(j)|^{\delta}<\infty for all tt and δ=min{1,α}\delta=min\{1,\alpha\}.

The random series in (14) converges a.s. if and only if j=0|at,T(j)|α<{\sum\limits_{j=0}^{\infty}\left|a_{t,T}(j)\right|^{\alpha}<\infty}, and by applying the Proposition 13.3.1 in [17], it converges absolutely if and only if j=0|at,T(j)|δ<\sum\limits_{j=0}^{\infty}\left|a_{t,T}(j)\right|^{\delta}<\infty with δ=min{1,α}\delta=\min\{1,\alpha\}. Similar arguments are applied to the MA representation in (15).

To continue, we omit the subscript TT from above notation. Consider the homogeneous difference equation

Φt(B)ut=0.\Phi_{t}(B)u_{t}=0. (16)

If αp(tT)0\alpha_{p}(\frac{t}{T})\neq 0 for any tt, there exist pp linearly independent solution ψ1,t,ψ2,t,,ψp,t\psi_{1,t},\psi_{2,t},...,\psi_{p,t} such that

Ψ(t)=[ψ1,tψp,tψ1,t1ψp,t1ψ1,tp+1ψp,tp+1]\Psi(t)=\begin{bmatrix}\psi_{1,t}&\cdots&\cdots&\psi_{p,t}\\ \psi_{1,t-1}&\ddots&&\psi_{p,t-1}\\ \vdots&&\ddots&\vdots\\ \psi_{1,t-p+1}&\cdots&\cdots&\psi_{p,t-p+1}\end{bmatrix} (17)

is invertible for any tt [23]. Therefore, we can define

G(t,s)=Ψ(t)[Ψ(s)],G(t,s)=\Psi(t)\left[\Psi(s)\right]^{\prime}, (18)

the one-sided Green’s function matrix associated with the AR operator Φt(B)\Phi_{t}(B). It can be showed that G(t,s)G(t,s) is unique and invariant under different solutions Ψ(t)\Psi(t) obtained from the homogeneous difference equation (16). Furthermore, the one-sided Green’s function associated with the AR operator Φt(B)\Phi_{t}(B) is defined as the upper left-hand element in the matrix (18),

g(t,s)=[G(t,s)]11,g(t,s)=\left[G(t,s)\right]_{11}, (19)

which is also unique and invariant. Now, we are ready to establish the conditions for AR regularity and MA regularity.

Theorem 3.3.

Let {Xt,T}\left\{X_{t,T}\right\} be a sequence of stochastic process that satisfies (13). Suppose that αp(tT)0\alpha_{p}(\frac{t}{T})\neq 0 for all tt, and g(t,s)g(t,s), the one-sided Green’s functions associated with Φt(B)\Phi_{t}(B), is such that s=t|g(t,s)|δ<\sum\limits_{s=-\infty}^{t}|g(t,s)|^{\delta}<\infty, for all tt. Assume also that s=0q|βj()|2<\sum\limits_{s=-0}^{q}|\beta_{j}(\cdot)|^{2}<\infty for all tt, and Φt(z)\Phi_{t}(z) (Φt(z)0(\Phi_{t}(z)\neq 0 for |z|1)|z|\leq 1) and Θt(z)\Theta_{t}(z) have no common roots. Then, there is a valid solution, given by

Xt,T=j=0at,T(j)εtj,X_{t,T}=\sum_{j=0}^{\infty}a_{t,T}(j)\varepsilon_{t-j}, (20)

to (13) with coefficients uniquely determined by

at,T(j)={0,j<0,γ(tjT),j=0,γ(tjT)j=0kβk(tj+kT)g(t,tj+k),0jq,γ(tjT)j=0qβk(tj+kT)g(t,tj+k),j>q.a_{t,T}(j)=\left\{\begin{array}[]{l r}0,&j<0,\\ \gamma(\frac{t-j}{T}),&j=0,\\ \gamma(\frac{t-j}{T})\sum\limits_{j=0}^{k}\beta_{k}(\frac{t-j+k}{T})g(t,t-j+k),&0\leq j\leq q,\\ \gamma(\frac{t-j}{T})\sum\limits_{j=0}^{q}\beta_{k}(\frac{t-j+k}{T})g(t,t-j+k),&j>q.\end{array}\right. (21)
Proof.

By setting zt,T=γ(tT)εtz_{t,T}=\gamma(\frac{t}{T})\varepsilon_{t}, along with the absolute convergence conditions above, the proof is similar to [22]. ∎

Theorem 3.4.

Let {Xt,T}\left\{X_{t,T}\right\} be a sequence of stochastic process that satisfies (13). Suppose that βq(tT)0\beta_{q}(\frac{t}{T})\neq 0 for all tt, and h(t,s)h(t,s), the one-sided Green’s function associated with Θt(B)\Theta_{t}(B), is such that s=t|h(t,s)|δ<\sum\limits_{s=-\infty}^{t}|h(t,s)|^{\delta}<\infty, for all tt. Assume also that s=0p|αj()|2<\sum\limits_{s=-0}^{p}|\alpha_{j}(\cdot)|^{2}<\infty for all tt, and Φt(z)\Phi_{t}(z) and Θt(z)\Theta_{t}(z) (Θt(z)0(\Theta_{t}(z)\neq 0 for |z|1)|z|\leq 1) have no common roots. Then, the process (13) is invertible and its explicit inversion is given by

εt=j=0bt,T(j)Xtj,T.\varepsilon_{t}=\sum_{j=0}^{\infty}b_{t,T}(j)X_{t-j,T}. (22)

where Xt,TX_{t,T} denotes an arbitrary solution and the coefficients are uniquely determined by

bt,T(j)={0,j<0,1γ(tT),j=0,1γ(tT)l=0kαk(tj+kT)h(t,tj+k),0jp,1γ(tT)l=0qαk(tj+kT)h(t,tj+k),j>p.b_{t,T}(j)=\left\{\begin{array}[]{l r}0,&j<0,\\ \frac{1}{\gamma(\frac{t}{T})},&j=0,\\ \frac{1}{\gamma(\frac{t}{T})}\sum\limits_{l=0}^{k}\alpha_{k}(\frac{t-j+k}{T})h(t,t-j+k),&0\leq j\leq p,\\ \frac{1}{\gamma(\frac{t}{T})}\sum\limits_{l=0}^{q}\alpha_{k}(\frac{t-j+k}{T})h(t,t-j+k),&j>p.\end{array}\right. (23)
Theorem 3.5.

Let {Xt,T}\left\{X_{t,T}\right\} be a sequence of stochastic process that satisfies (12) that is AR regular. The solution Xt,TX_{t,T} of the form (14) is strictly stable and Xt,TSα(σ,β,0){X_{t,T}\sim S_{\alpha}(\sigma^{*},\beta^{*},0)}, with

σ=(12){j=0|at,T(j)|α}1/α,andβ=β{j=0sign[at,T(j)]|at,T(j)|αj=0|at,T(j)|α}.\sigma^{*}=\left(\frac{1}{\sqrt{2}}\right)\left\{\sum_{j=0}^{\infty}\left|a_{t,T}(j)\right|^{\alpha}\right\}^{1/\alpha},\text{and}~{}~{}\beta^{*}=\beta\left\{\frac{\sum\limits_{j=0}^{\infty}\operatorname{sign}\left[a_{t,T}(j)\right]\left|a_{t,T}(j)\right|^{\alpha}}{\sum\limits_{j=0}^{\infty}\left|a_{t,T}(j)\right|^{\alpha}}\right\}.
Proof.

The explicit form of the solution is straightforward since the linear combination of stable distributions is also stable. Moreover, the Property 1.2.6 from [14] implies that for each tt, the solution Xt,TX_{t,T} is strictly stable since each of them has location parameter equals 0. ∎

3.2 Local Stationarity

Similar to the Proposition 2.4 in [13], we can present the corresponding version for stable innovations. Since it is not a second-order process, the time-varying spectral density does not exist.

Theorem 3.6.

Consider the system of difference equations in (12) satisfying the AR regular conditions stated above. Suppose that all αj()\alpha_{j}(\cdot) and βk()\beta_{k}(\cdot), as well as γ2()\gamma^{2}(\cdot) are of bounded variation. Then, there exists a solution of the form

Xt,T=j=0at,T(j)εtj,X_{t,T}=\sum_{j=0}^{\infty}a_{t,T}(j)\varepsilon_{t-j},

which fullfills (3), (4) and (5).

Proof.

We give the proof for tvAR(p) process (i.e. q=0q=0) and then the extension to tvARMA(p,q) is straightforward. Since the process (12) is AR regular, there exists a solution of the form

Xt,T=j=0at,T(j)εtj,X_{t,T}=\sum_{j=0}^{\infty}a_{t,T}(j)\varepsilon_{t-j},

that is well defined and the coefficients are given by

at,T(j)=[=0j1𝜶(tT)]11γ(tjT)a_{t,T}(j)=\left[\prod_{\ell=0}^{j-1}\boldsymbol{\alpha}\left(\frac{t-\ell}{T}\right)\right]_{11}\gamma\left(\frac{t-j}{T}\right)

with

𝜶(u)=(α1(u)α2(u)αp(u)10000010)\boldsymbol{\alpha}(u)=\begin{pmatrix}-\alpha_{1}(u)&-\alpha_{2}(u)&\cdots&\cdots&-\alpha_{p}(u)\\ 1&0&\cdots&\cdots&0\\ 0&\ddots&\ddots&&\vdots\\ \vdots&\ddots&\ddots&\ddots&\vdots\\ 0&\cdots&0&1&0\end{pmatrix}

[for more detail see 23]. Then, the proof of the existence of the functions α(,j)\alpha(\cdot,j) satisfying (3), (4) and (5) follows the same proof to that of finite innovation case (see Appendix in [13]). ∎

Remark 1.
  1. 1.

    Note that since at,T(j)a(tT,j)a_{t,T}(j)\approx a\left(\frac{t}{T},j\right), Xt,TX_{t,T} can be approximated by

    X~t,T=j=0a(tT,j)εtj,\tilde{X}_{t,T}=\sum_{j=0}^{\infty}a\left(\frac{t}{T},j\right)\varepsilon_{t-j}, (24)

    which converges a.s. if and only if j=0|a(tT,j)|α<.\sum\limits_{j=0}^{\infty}\left|a\left(\frac{t}{T},j\right)\right|^{\alpha}<\infty.

    Moreover, X~t,TSα(σ+,β+,0)\tilde{X}_{t,T}\sim S_{\alpha}(\sigma^{+},\beta^{+},0), with

    σ+=12{j=0|a(tT,j)|α}1/α,andβ+=β{j=0sign[a(tT,j)]|a(tT,j)|αj=0|a(tT,j)|α}.\sigma^{+}=\frac{1}{\sqrt{2}}\left\{\sum_{j=0}^{\infty}\left|a\left(\frac{t}{T},j\right)\right|^{\alpha}\right\}^{1/\alpha},\text{and}~{}~{}\beta^{+}=\beta\left\{\frac{\sum\limits_{j=0}^{\infty}\operatorname{sign}\left[a\left(\frac{t}{T},j\right)\right]\left|a\left(\frac{t}{T},j\right)\right|^{\alpha}}{\sum\limits_{j=0}^{\infty}\left|a\left(\frac{t}{T},j\right)\right|^{\alpha}}\right\}.
  2. 2.

    Xt,TX_{t,T} in (14) can be expressed as a linear combination of α\alpha-stable random variables and Xt,TX_{t,T} is strictly stable with the same index of stability α\alpha.

  3. 3.

    Observe that Xt,TX_{t,T} is not strictly stationary, but it can be approximated by X~t,T\tilde{X}_{t,T} which is locally (strictly) stationary and strictly stable with the same index of stability.

  4. 4.

    Weak stationarity does not make sense since the second moment does not exist. Consequently, the time-varying spectral representation does not exist.

  5. 5.

    Let X1,T,,XT,TX_{1,T},...,X_{T,T} be the sequence of solutions defined in (14) and X~1,T,,X~T,T\tilde{X}_{1,T},...,\tilde{X}_{T,T} be the sequence of the stochastic process defined in (24). Both processes are strictly α\alpha-stable, since all linear combinations are strictly stable with the same index of stability. This means that the weak stationarity is lost but it is substituted by the same tail behavior throughout the time. This is the reason we call this process α\alpha-stable locally (strictly) stationary process.

If we consider the symmetric α\alpha-stable (SαSS\alpha S) innovations, i.e. β=0\beta=0, the simplest form is obtained.

Corollary 3.7 (tvARMA with symmetric stable innovations).

Let Xt,TX_{t,T} be a sequence of stochastic process that satisfies (12) with i.i.d. SαSS\alpha S innovations, that is, εtSα(1/2,0,0)\varepsilon_{t}\sim S_{\alpha}(\nicefrac{{1}}{{\sqrt{2}}},0,0) with α(0,2)\alpha\in(0,2). Then, there exists a solution of the form (14). This solution Xt,TX_{t,T} is symmetric, α\alpha-stable and Xt,TSα(σ,0,0)X_{t,T}\sim S_{\alpha}(\sigma^{*},0,0), with

σ=12{j=0|at,T(j)|α}1/α.\sigma^{*}=\frac{1}{\sqrt{2}}\left\{\sum_{j=0}^{\infty}\left|a_{t,T}(j)\right|^{\alpha}\right\}^{1/\alpha}.

Similarly to the general case, Xt,TX_{t,T} can be approximated by X~t,TSα(σ+,0,0)\tilde{X}_{t,T}\sim S_{\alpha}(\sigma^{+},0,0) as in (24), with

σ+={j=0|a(tT,j)|α}1/α.\sigma^{+}=\left\{\sum_{j=0}^{\infty}\left|a\left(\frac{t}{T},j\right)\right|^{\alpha}\right\}^{1/\alpha}.

In the symmetric case, in addition to the properties in Remark 1, note that the processes Xt,TX_{t,T} and X~t,T\tilde{X}_{t,T} are symmetric α\alpha-stable.

3.3 Some examples

Example 3.8.

The tvMA(q) model with stable innovations:

Xt,T=k=0qβk(tT)γ(tkT)εtk.X_{t,T}=\sum_{k=0}^{q}\beta_{k}\left(\frac{t}{T}\right)\gamma\left(\frac{t-k}{T}\right)\varepsilon_{t-k}. (25)
Example 3.9.

Consider the tvAR(p) model with stable innovations

j=0pαj(tT)Xtj,T=γ(tT)εt.\sum_{j=0}^{p}\alpha_{j}\left(\frac{t}{T}\right)X_{t-j,T}=\gamma\left(\frac{t}{T}\right)\varepsilon_{t}. (26)

Underregularity conditions, Xt,TX_{t,T} does not have a solution of the form

Xt,T=k=0ak(tT)εtk,X_{t,T}=\sum\limits_{k=0}^{\infty}a_{k}\left(\frac{t}{T}\right)\varepsilon_{t-k},

but only of the form (14) with

at,T(j)=[=0j1𝜶(tT)]11γ(tjT)a_{t,T}(j)=\left[\prod_{\ell=0}^{j-1}\boldsymbol{\alpha}\left(\frac{t-\ell}{T}\right)\right]_{11}\gamma\left(\frac{t-j}{T}\right) (27)

and

𝜶(u)=(α1(u)α2(u)αp(u)10000010)\boldsymbol{\alpha}(u)=\begin{pmatrix}-\alpha_{1}(u)&-\alpha_{2}(u)&\cdots&\cdots&-\alpha_{p}(u)\\ 1&0&\cdots&\cdots&0\\ 0&\ddots&\ddots&&\vdots\\ \vdots&\ddots&\ddots&\ddots&\vdots\\ 0&\cdots&0&1&0\end{pmatrix}

where 𝜶(u)=𝜶(0)\boldsymbol{\alpha}(u)=\boldsymbol{\alpha}(0) for u<0u<0 [13]. Moreover, at,T(j)a_{t,T}(j) can be approximated by a(u,j)=(𝜶(u)j)11γ(u)a(u,j)=\left(\boldsymbol{\alpha}(u)^{j}\right)_{11}\gamma(u) which satisfies (3), (4) and (5).

Figure 1 presents simulated tvAR(1) process of T=1000T=1000 observations with different innovation distribution (Gaussian, t(3)t_{(3)} and symmetric stable innovations, α=1.7,1.4,0.9,0.6\alpha=1.7,1.4,0.9,0.6) and a linear coefficient α1(u)=0.2+0.6u\alpha_{1}(u)=-0.2+0.6u and γ(u)=1\gamma(u)=1. We observe that for smaller α\alpha, the process seems to have more outliers.

Refer to caption
Figure 1: Simulated tvAR(1) assuming time-varying coefficient α1(u)=0.2+0.6u\alpha_{1}(u)=-0.2+0.6u and γ(u)=1\gamma(u)=1, with (a) Gaussian, (b) t(3)t_{(3)} innovations, and symmetric stable innovations with (c) α=1.7\alpha=1.7, (d) α=1.4\alpha=1.4, (e) α=0.9\alpha=0.9 and (f) α=0.6\alpha=0.6.

3.4 Prediction

Recalling that α\alpha-stable tvARMA has infinite variance, prediction results based on stable ARMA processes with dependent coefficients are presented by [21, 22]. Then, it is possible to predict future values along with the approach applied by [24], which considers the observed values X0,T,,XTh1,TX_{0,T},\cdots,X_{T-h-1,T} and rescaling the time interval to [0,1h+1T]\left[0,1-\frac{h+1}{T}\right], where hh is the forecasting horizon and the ratio h/Th/T tends to zero as TT tends to infinity. Here, we consider that the innovations are SαSS\alpha S random variables.

Suppose that we have the system of difference equation (12) that satisfies above regular conditions, and X0,T,,XT,TX_{0,T},\cdots,X_{T^{\prime},T} with T=Th1T^{\prime}=T-h-1 are observed. We are interested in predictions with horizon hh, i.e. XTh,T,,XT,TX_{T-h,T},...,X_{T,T}.

Since Xt,TX_{t,T} is AR regular, it can be expressed as

Xt,T=j=0at,T(j)εtj.X_{t,T}=\sum_{j=0}^{\infty}a_{t,T}(j)\varepsilon_{t-j}. (28)

Let X^T(l)\hat{X}_{T^{\prime}}(l) be the best linear predictor of XT+l,TX_{T^{\prime}+l,T} for l=1,,hl=1,...,h, namely

X^T(l)=j=0A(T,Tj)εTj,\hat{X}_{T^{\prime}}(l)=\sum\limits_{j=0}^{\infty}A(T^{\prime},T^{\prime}-j)\varepsilon_{T^{\prime}-j}, (29)

where A(T,Tj)A(T^{\prime},T^{\prime}-j) are some functions. Since the prediction error eT(l)=X^T+h,TXT(l)e_{T^{\prime}}(l)=\hat{X}_{T^{\prime}+h,T}-X_{T^{\prime}}(l) is also SαSS\alpha S random variable, it is possible to define its dispersion as d=σαd=\sigma^{\alpha} with σ\sigma its scale parameter. The idea is to minimize the dispersion dd. Note that

eT(l)=XT+l,TX^T(l)=j=0aT+l,T(j)εT+ljj=0A(T,Tj)εTj=j=0l1aT+l,T(j)εT+lj+j=laT+l,T(j)εT+ljj=0A(T,Tj)εTj=j=0l1aT+l,T(j)εT+lj+j=0(aT+l,T(j+l)A(T,Tj))εTj.\begin{array}[]{l l}e_{T^{\prime}}(l)&=X_{T^{\prime}+l,T}-\hat{X}_{T^{\prime}}(l)\\ &=\sum\limits_{j=0}^{\infty}a_{T^{\prime}+l,T}(j)\varepsilon_{T^{\prime}+l-j}-\sum\limits_{j=0}^{\infty}A(T^{\prime},T^{\prime}-j)\varepsilon_{T^{\prime}-j}\\ &=\sum\limits_{j=0}^{l-1}a_{T^{\prime}+l,T}(j)\varepsilon_{T^{\prime}+l-j}+\sum\limits_{j=l}^{\infty}a_{T^{\prime}+l,T}(j)\varepsilon_{T^{\prime}+l-j}-\sum\limits_{j=0}^{\infty}A(T^{\prime},T^{\prime}-j)\varepsilon_{T^{\prime}-j}\\ &=\sum\limits_{j=0}^{l-1}a_{T^{\prime}+l,T}(j)\varepsilon_{T^{\prime}+l-j}+\sum\limits_{j=0}^{\infty}\left(a_{T^{\prime}+l,T}(j+l)-A(T^{\prime},T^{\prime}-j)\right)\varepsilon_{T^{\prime}-j}.\end{array} (30)

Then, assuming εtSα(12,0,0)\varepsilon_{t}\sim S_{\alpha}(\frac{1}{\sqrt{2}},0,0) and using properties of SαSS\alpha S random variables, its dispersion is

disp[eT(l)]=(12)αj=0l1|aT+l,T(j)|α+(12)αj=0|(aT+l,T(j+l)A(T,Tj))|α.\text{disp}\left[e_{T^{\prime}}(l)\right]=\left(\frac{1}{\sqrt{2}}\right)^{\alpha}\sum\limits_{j=0}^{l-1}|a_{T^{\prime}+l,T}(j)|^{\alpha}+\left(\frac{1}{\sqrt{2}}\right)^{\alpha}\sum\limits_{j=0}^{\infty}|\left(a_{T^{\prime}+l,T}(j+l)-A(T^{\prime},T^{\prime}-j)\right)|^{\alpha}. (31)

Minimizing the expression (31), we obtain the following theorem.

Theorem 3.10.

The minimum dispersion predictor is given by

X^T(l)=j=0aT+l,T(j+l)εTj.\hat{X}_{T^{\prime}}(l)=\sum\limits_{j=0}^{\infty}a_{T^{\prime}+l,T}(j+l)\varepsilon_{T^{\prime}-j}. (32)
Proof.

From (31), it is straightforward to obtain

min disp[eT(l)]=(12)αj=0l1|aT+l,T(j)|α,\text{min disp}\left[e_{T^{\prime}}(l)\right]=\left(\frac{1}{\sqrt{2}}\right)^{\alpha}\sum\limits_{j=0}^{l-1}|a_{T^{\prime}+l,T}(j)|^{\alpha},

with aT+l,T(j+l)=A(T,Tj)a_{T^{\prime}+l,T}(j+l)=A(T^{\prime},T^{\prime}-j) for j=0,1,j=0,1,.... ∎

4 Indirect inference for α\alpha-stable tvARMA processes

The IM is the tvARMA with innovation εtSα(1/2,β,0)\varepsilon_{t}\sim S_{\alpha}(\nicefrac{{1}}{{\sqrt{2}}},\beta,0). We study the parameter estimation when the innovation parameters α\alpha and β\beta are known. Then, we study the case assuming unknown α\alpha.

Suppose that the parameter curves of model of interest can be parametrized by a finite-dimensional parameter θ\theta. The estimation strategy is to consider an auxiliary model with the same parametric time varying coefficient structure with Student’s t innovations. The conditional likelihood estimates, defined in the equation (30)(30) in [3], is,

𝝀^=argmin𝝀1Tt=1Tt,T(θλ(tT)),\boldsymbol{\hat{\lambda}}=\operatorname*{argmin}_{\boldsymbol{\lambda}}\frac{1}{T}\sum_{t=1}^{T}\ell_{t,T}\left(\theta_{\lambda}\left(\frac{t}{T}\right)\right), (33)

where t,T(𝜽)=logfθ(Xt,T|Xt1,T,,X1,T)\ell_{t,T}(\boldsymbol{\theta})=-\log f_{\theta}(X_{t,T}|X_{t-1,T},...,X_{1,T}). In practice, we will use t distribution with ν=3\nu=3 degrees of freedom for the case of known parameters since its tail is heavier than the Gaussian one. For unknown α\alpha case, we let ν\nu to be estimated in the AM.

5 Simulation study

This section presents a Monte Carlo (MC) simulation in order to investigate the properties of the indirect inference estimators. All the simulation programs and routines were implemented in R language. We present one scenario for each of the following models but still different values of α\alpha were selected. Some other scenarios were performed for each case and similar results were obtained and they are available upon request from the authors. For each scenario, simulations were done for T=500T=500, 10001000 and 15001500 observations based on R=1000R=1000 independent replications. The indirect inference was carried out using S=100S=100. For α\alpha known, we also performed the blocked Whittle estimation (BWE), proposed by [2], to compare the estimation of time structure of the model with indirect inference. The suggestion of block size N=T0.8N=\lfloor T^{0.8}\rfloor and shifting each block by S=0.2NS=\lfloor 0.2N\rfloor time units from [25] is used.

5.1 Known α\alpha case

5.1.1 α\alpha-stable tvAR(1)

Consider tvAR(p) in (26) with p=1p=1 and γ(tT)=γ\gamma\left(\frac{t}{T}\right)=\gamma:

Xt,T+α1(tT)Xt1,T=γεt,X_{t,T}+\alpha_{1}\left(\frac{t}{T}\right)X_{t-1,T}=\gamma\varepsilon_{t}, (34)

where εtSα(1/2,β,0)\varepsilon_{t}\sim S_{\alpha}\left(\nicefrac{{1}}{{\sqrt{2}}},\beta,0\right) with α\alpha and β\beta known.

We illustrate how the indirect inference can be employed to the tvAR(1) with the linear parametric form of the time varying coefficient α1(u)=θ0+θ1u\alpha_{1}\left(u\right)=\theta_{0}+\theta_{1}u, and we consider that εtSα(1/2,β,0)\varepsilon_{t}\sim S_{\alpha}(\nicefrac{{1}}{{\sqrt{2}}},\beta,0) for α\alpha known. Therefore, the parameters of the IM is θ=(θ0,θ1,γ)\theta=\left(\theta_{0},\theta_{1},\gamma\right).

The simulation was performed by assuming known parameters α=1.9\alpha=1.9 and β=0.9\beta=0.9 and unknown (θ0,θ1,γ)=(0.3,0.8,1)(\theta_{0},\theta_{1},\gamma)=(-0.3,0.8,1). It is important to report that since α\alpha is close to 22, all replications for the BWE converged. This outcome is expected since the innovation distributions approximate to the Gaussian distribution for α\alpha close to 22.

Table 1 reports the MC mean and standard error of both estimation methods. Notice that the MC mean from the indirect estimates seems to be consistent, that is, they approximate the real parameters and present lower standard errors as TT increases. On the other hand, the MC mean of the BWE are different from the real parameters and present higher standard errors compared to our estimation approach.

Table 2 presents the kurtosis and skewness of all estimates from both methods. In general, all indirect estimates present lower kurtosis and the skewness close to 0. Notice that since the second moment of the process does not exist, the parameter γ\gamma estimates from the BWE present highly positive asymmetry and they subestimate the true parameter.

Table 1: MC means and standard errors for different sample size TT, using indirect estimators and BWE assuming α\alpha and β\beta known, from α\alpha-stable tvAR(1) with (α,β,θ0,θ1,γ)=(1.9,0.9,0.3,0.8,1),(\alpha,\beta,\theta_{0},\theta_{1},\gamma)=(1.9,0.9,-0.3,0.8,1), based on R=1000R=1000 replications.
TT Indirect estimates BWE
θ0\theta_{0} θ1\theta_{1} γ\gamma θ0(W)\theta_{0}^{(W)} θ1(W)\theta_{1}^{(W)} γ(W)\gamma^{(W)}
500500 -0.2952 0.7897 0.9966 -0.2880 0.7825 1.2086
(0.0881) (0.1523) (0.0366) (0.1172) (0.2216) (0.6352)
10001000 -0.2975 0.7926 0.9996 -0.2917 0.7845 1.2197
(0.0585) (0.1028) (0.0260) (0.0811) (0.1545) (0.4734)
15001500 -0.2974 0.7958 0.9997 -0.2940 0.7926 1.2709
(0.0494) (0.0793) (0.0209) (0.0639) (0.1162) (0.8738)
Table 2: Kurtosis and skewness of indirect estimates and BWE for different sample size TT, assuming α\alpha and β\beta known, from α\alpha-stable tvAR(1) with (α,β,θ0,θ1,γ)=(1.9,0.9,0.3,0.8,1)(\alpha,\beta,\theta_{0},\theta_{1},\gamma)=(1.9,0.9,-0.3,0.8,1), based on R=1000R=1000 replications.
TT Indirect estimates BWE
θ0\theta_{0} θ1\theta_{1} γ\gamma θ0(W)\theta_{0}^{(W)} θ1(W)\theta_{1}^{(W)} γ(W)\gamma^{(W)}
500500 Kur 3.0330 2.8565 3.1076 3.3783 3.0944 375.8563
Skw 0.1388 -0.1354 0.1241 -0.0129 -0.0875 16.6778
10001000 Kur 3.1678 3.2835 2.7390 2.8722 2.9543 95.2261
Skw 0.0341 -0.0260 0.0437 0.0057 -0.1047 8.4487
15001500 Kur 3.1024 3.0645 2.9329 3.7487 6.1799 187.9560
Skw -0.0299 0.0248 0.0026 0.1707 -0.5935 12.4661

Figure 2 shows the density estimates of each parameter. They show that the standard error become smaller as TT increases. Along with the results from Tables 1 and 2, we can conclude that indirect estimates behave better than the BWE in terms of mean, standard error, skewness and kurtosis. Therefore, the simulation results show that the indirect inference performs well.

Refer to caption
Figure 2: Density estimates of θ0\theta_{0}, θ1\theta_{1} and γ\gamma for different sample sizes, based on R=1000R=1000 replications from α\alpha-stable tvAR(1) with (α,β,θ0,θ1,γ)=(1.9,0.9,0.3,0.8,1)(\alpha,\beta,\theta_{0},\theta_{1},\gamma)=(1.9,0.9,-0.3,0.8,1), using indirect inference.

5.1.2 α\alpha-stable tvMA(1)

In this section, we carried out simulations for a tvMA(q) in (25) with q=1q=1 and γ(tT)=γ\gamma\left(\frac{t}{T}\right)=\gamma:

Xt,T=γ{εt+β1(tT)εt1},X_{t,T}=\gamma\left\{\varepsilon_{t}+\beta_{1}\left(\frac{t}{T}\right)\varepsilon_{t-1}\right\}, (35)

where εtSα(1/2,β,0)\varepsilon_{t}\sim S_{\alpha}\left(\nicefrac{{1}}{{\sqrt{2}}},\beta,0\right) with α\alpha and β\beta known.

The indirect inference is employed for the linear parametric form of the time varying coefficient β1(u)=θ0+θ1u\beta_{1}\left(u\right)=\theta_{0}+\theta_{1}u, and we consider that εtSα(1/2,β,0)\varepsilon_{t}\sim S_{\alpha}(\nicefrac{{1}}{{\sqrt{2}}},\beta,0) for known α\alpha and β\beta. Hence, the vector of parameters of the model of interest is θ=(θ0,θ1,γ)\theta=\left(\theta_{0},\theta_{1},\gamma\right).

This scenario assumes known α=1.1\alpha=1.1 and β=0.2\beta=-0.2 and unknown (θ0,θ1,γ)=(0.35,0.6,1.2)(\theta_{0},\theta_{1},\gamma)=(0.35,-0.6,1.2). For the BWE case, we consider only R=939,978R=939,978 and 978978 replications with converged estimates for T=500,1000T=500,1000, and 15001500, respectively. This result is expected because BWE assumes finite second moment. The MC mean, standard error, kurtosis and skewness of estimates from the simulation are reported in the Table 3 and 4 and the density estimates in Figures 3.

Similarly to the previous case, the indirect estimates seem to be consistent and the standard error become smaller as TT increases. For this case, since α\alpha is smaller, the distribution of indirect estimates has heavier tails, and they have similar kurtosis and skewness than the BWE estimates, except for the parameter γ\gamma, when the indirect estimation behaves better. In addition, in term of standard error and MC mean, they still behave better than the BWE. We conclude that the indirect inference has a good performance.

Table 3: MC mean and standard error for different sample size TT, using indirect estimators and BWE assuming α\alpha and β\beta known, from α\alpha-stable tvMA(1) with (α,β,θ0,θ1,γ)=(1.1,0.2,0.35,0.6,1.2)(\alpha,\beta,\theta_{0},\theta_{1},\gamma)=(1.1,-0.2,0.35,-0.6,1.2), based on R=1000R=1000 replications.
TT Indirect estimates BWE111In tvMA(1) simulations, the BWE did not converge in some cases. Therefore, excluding those cases, R=939,978R=939,978 and 978978 replications are included for T=500,1000T=500,1000, and 15001500, respectively.
θ0\theta_{0} θ1\theta_{1} γ\gamma θ0(W)\theta_{0}^{(W)} θ1(W)\theta_{1}^{(W)} γ(W)\gamma^{(W)}
500500 0.3561 -0.5888 1.1989 0.3424 -0.5427 18.7932
(0.0298) (0.0577) (0.0600) (0.1418) (0.3084) (38.4343)
10001000 0.3545 -0.5953 1.1986 0.3386 -0.5532 47.5752
(0.0186) (0.0352) (0.0412) (0.0870) (0.1955) (232.0620)
15001500 0.3536 -0.5982 1.1986 0.3357 -0.5555 49.4572
(0.0131) (0.0244) (0.0331) (0.0747) (0.1690) (178.3984)
Table 4: Kurtosis and skewness of indirect estimates and BWE for different sample size TT assuming known α\alpha and β\beta from α\alpha-stable tvMA(1) with (α,β,θ0,θ1,γ)=(1.1,0.2,0.35,0.6,1.2)(\alpha,\beta,\theta_{0},\theta_{1},\gamma)=(1.1,-0.2,0.35,-0.6,1.2) based on R=1000R=1000 replications.
TT Indirect estimates BWE11footnotemark: 1
θ0\theta_{0} θ1\theta_{1} γ\gamma θ0(W)\theta_{0}^{(W)} θ1(W)\theta_{1}^{(W)} γ(W)\gamma^{(W)}
500500 Kur 7.9023 6.3460 2.9050 6.9952 7.3434 233.1652
Skw 1.2950 0.0800 0.2117 0.1961 1.1869 13.5324
10001000 Kur 9.8633 10.4926 2.8616 11.7454 8.6755 385.9194
Skw 1.6510 0.7121 0.0841 0.8633 1.5096 17.6374
15001500 Kur 8.1466 20.6873 2.8140 9.7011 10.5014 156.1843
Skw 1.5194 1.4762 0.1493 -0.1837 1.9926 11.4943
Refer to caption
Figure 3: Density estimates of θ0\theta_{0}, θ1\theta_{1} and γ\gamma for different sample sizes based on R=1000R=1000 replications from α\alpha-stable tvMA(1) with (α,β,θ0,θ1,γ)=(1.1,0.2,0.35,0.6,1.2)(\alpha,\beta,\theta_{0},\theta_{1},\gamma)=(1.1,-0.2,0.35,-0.6,1.2) using indirect inference.

5.1.3 α\alpha-stable tvARMA(1,1)

The third simulation was carried out with the case of tvARMA(p,q) with p=1p=1, q=1q=1 and γ(tT)=γ\gamma\left(\frac{t}{T}\right)=\gamma:

Xt,T+α1(tT)Xt1,T=γ{εt+β1(tT)εt1},X_{t,T}+\alpha_{1}\left(\frac{t}{T}\right)X_{t-1,T}=\gamma\left\{\varepsilon_{t}+\beta_{1}\left(\frac{t}{T}\right)\varepsilon_{t-1}\right\}, (36)

where εtSα(1/2,β,0)\varepsilon_{t}\sim S_{\alpha}\left(\nicefrac{{1}}{{\sqrt{2}}},\beta,0\right) with α\alpha and β\beta known.

We suppose a linear parametric form of the time varying coefficients α1(u)=θa0+θa1u\alpha_{1}\left(u\right)=\theta_{a0}+\theta_{a1}u and β1(u)=θb0+θb1u\beta_{1}\left(u\right)=\theta_{b0}+\theta_{b1}u. Therefore, the parameters of the IM is θ=(θa0,θa1,θb0,θb1,γ)\theta=\left(\theta_{a0},\theta_{a1},\theta_{b0},\theta_{b1},\gamma\right).

The simulation was done by assuming α=1.8\alpha=1.8, β=0.3\beta=0.3 and (θa0,θa1,θb0,θb1,γ)=(0.4,0.1,0.1,0.3,1.1)\left(\theta_{a0},\theta_{a1},\theta_{b0},\theta_{b1},\gamma\right)=(-0.4,0.1,0.1,0.3,1.1). For BWE, R=989,996R=989,996 and 994994 replications with converged estimates are included for T=500,1000T=500,1000, and 15001500, respectively.

The MC mean, standard error, kurtosis and skewness of estimates from the tvARMA(1,1) simulation are reported in the Table 5 and 6 and the density estimates in Figure 4. In general, the distribution of indirect estimates has heavier tails, and the kurtosis and skewness are similar to the BWE (except for the parameter γ\gamma, indirect estimates behave better). However, in terms of standard error and MC mean, they behave much better than the BWE. Therefore, the indirect inference works well for tvARMA(1,1).

Table 5: MC mean and standard error for different sample sizes TT, using indirect estimators and BWE assuming α\alpha and β\beta known, from α\alpha-stable tvARMA(1,1) with (α,β,θa0,θa1,θb0,θb1,γ)=(1.8,0.3,0.4,0.1,0.1,0.3,1)(\alpha,\beta,\theta_{a0},\theta_{a1},\theta_{b0},\theta_{b1},\gamma)=(1.8,0.3,-0.4,0.1,0.1,0.3,1), based on R=1000R=1000 replications.
TT Indirect estimates BWE222In tvARMA(1,1) simulations, the BWE did not converge in some cases. Therefore, excluding those cases, R=989,996R=989,996 and 994994 replications are included for T=500,1000T=500,1000, and 15001500, respectively.
θa0\theta_{a0} θa1\theta_{a1} θb0\theta_{b0} θb1\theta_{b1} γ\gamma θa0(W)\theta_{a0}^{(W)} θa1(W)\theta_{a1}^{(W)} θb0(W)\theta_{b0}^{(W)} θb1(W)\theta_{b1}^{(W)} γ(W)\gamma^{(W)}
500500 -0.4000 0.1061 0.0987 0.3097 0.9976 -0.3917 0.1021 0.1078 0.3049 1.4151
(0.1360) (0.2222) (0.1501) (0.2395) (0.0386) (0.1952) (0.3522) (0.2130) (0.3810) (0.7603)
10001000 -0.3921 0.0881 0.1064 0.2905 0.9982 -0.3850 0.0815 0.1105 0.2880 1.4919
(0.1001) (0.1617) (0.1053) (0.1652) (0.0290) (0.1409) (0.2535) (0.1470) (0.2599) (0.5806)
15001500 -0.3992 0.1021 0.0988 0.3060 0.9982 -0.3939 0.0926 0.1055 0.2964 1.5538
(0.0754) (0.1269) (0.0793) (0.1285) (0.0232) (0.1085) (0.1955) (0.1155) (0.2040) (0.8009)
Table 6: Kurtosis and skewness of indirect estimates and BWE for different sample size TT, assuming known α\alpha and β\beta from α\alpha-stable tvARMA(1,1) with (α,β,θa0,θa1,θb0,θb1,γ)=(1.8,0.3,0.4,0.1,0.1,0.3,1)(\alpha,\beta,\theta_{a0},\theta_{a1},\theta_{b0},\theta_{b1},\gamma)=(1.8,0.3,-0.4,0.1,0.1,0.3,1), based on R=1000R=1000 replications.
TT Indirect estimates BWE222In tvARMA(1,1) simulations, the BWE did not converge in some cases. Therefore, excluding those cases, R=989,996R=989,996 and 994994 replications are included for T=500,1000T=500,1000, and 15001500, respectively.
θa0\theta_{a0} θa1\theta_{a1} θb0\theta_{b0} θb1\theta_{b1} γ\gamma θa0(W)\theta_{a0}^{(W)} θa1(W)\theta_{a1}^{(W)} θb0(W)\theta_{b0}^{(W)} θb1(W)\theta_{b1}^{(W)} γ(W)\gamma^{(W)}
500500 Kur 3.3650 3.1791 3.3657 3.4297 2.9935 2.9112 3.0692 3.2168 3.2822 203.2823
Skw 0.2754 -0.2426 -0.0746 -0.1274 0.1839 0.1699 -0.1329 -0.2024 -0.0422 12.0737
15001500 Kur 3.3964 3.5054 3.4470 3.4690 3.0327 3.4242 3.2166 3.2601 3.0064 41.6803
Skw 0.2002 -0.1558 0.0100 -0.1184 0.2460 0.3149 -0.2253 0.0267 -0.1713 4.9608
15001500 Kur 3.5817 3.1935 3.7052 3.3930 2.9790 2.9176 2.8801 3.3685 3.3091 96.2273
Skw 0.2895 -0.1097 0.0137 -0.0730 0.0718 0.0809 0.0268 -0.1083 0.0372 7.7396
Refer to caption
Figure 4: Density estimates of θa0\theta_{a0}, θa1\theta_{a1}, θb0\theta_{b0}, θb1\theta_{b1} and γ\gamma for different sample sizes based on R=1000R=1000 replications from α\alpha-stable tvARMA(1,1) with (α,β,θa0,θa1,θb0,θa1,γ)=(1.8,0.3,0.4,0.1,0.1,0.3,1)(\alpha,\beta,\theta_{a0},\theta_{a1},\theta_{b0},\theta_{a1},\gamma)=(1.8,0.3,-0.4,0.1,0.1,0.3,1) using indirect inference.

5.2 Unknown α\alpha case

5.2.1 α\alpha-stable tvAR(1)

Consider the tvAR(1) model

Xt,T+α1(tT)Xt1,T=γ(tT)εt,X_{t,T}+\alpha_{1}\left(\frac{t}{T}\right)X_{t-1,T}=\gamma\left(\frac{t}{T}\right)\varepsilon_{t}, (37)

where εtSα(1/2,β,0)\varepsilon_{t}\sim S_{\alpha}\left(\nicefrac{{1}}{{\sqrt{2}}},\beta,0\right) with known β\beta. Here, the indirect inference is employed to the tvAR(1) in (34) with the linear parametric form of the time varying coefficient α1(u)=θ0+θ1u\alpha_{1}\left(u\right)=\theta_{0}+\theta_{1}u, and γ(u)=γ0+γ1u\gamma\left(u\right)=\gamma_{0}+\gamma_{1}u. The parameters of IM is θ=(θ0,θ1,α,γ0,γ1)\theta=\left(\theta_{0},\theta_{1},\alpha,\gamma_{0},\gamma_{1}\right). For AM, the same parametric form with the t-distribution assuming unknown ν\nu is used, that is, λ=(θ0(A),θ1(A),ν,γ0(A),γ1(A))\lambda=(\theta_{0}^{(A)},\theta_{1}^{(A)},\nu,\gamma_{0}^{(A)},\gamma_{1}^{(A)}).

The simulation was performed by assuming (α,β,θ0,θ1,γ0,γ1)=(1.4,0,0.35,0.6,0.5,0.1)(\alpha,\beta,\theta_{0},\theta_{1},\gamma_{0},\gamma_{1})=(1.4,0,\allowbreak 0.35,-0.6,0.5,0.1). Table 7 reports the MC mean and standard error of the estimates. Notice that the MC mean from the indirect estimates seems to be consistent. Table 8 presents the kurtosis and skewness of indirect estimates. All indirect estimates do not present kurtosis close to 3 and the skewness close to 0. Indeed, they are similar to the case when α\alpha is known.

Table 7: MC mean and standard error for different sample size TT using indirect estimators assuming (α,β,θ0,θ1,γ0,γ1)=(1.4,0,0.35,0.6,0.5,0.1)(\alpha,\beta,\theta_{0},\theta_{1},\gamma_{0},\gamma_{1})=(1.4,0,0.35,-0.6,0.5,0.1) with known β\beta from α\alpha-stable tvAR(1) based on R=1000R=1000 replications.
T Indirect estimates
Model of Interest Auxiliary model
θ0\theta_{0} θ1\theta_{1} α\alpha γ0\gamma_{0} γ1\gamma_{1} θ0(A)\theta_{0}^{(A)} θ1(A)\theta_{1}^{(A)} ν\nu γ0(A)\gamma_{0}^{(A)} γ1(A)\gamma_{1}^{(A)}
500500 0.3482 -0.5980 1.4083 0.4922 0.1111 0.3482 -0.5980 1.8853 0.3994 0.0897
(0.0406) (0.0715) (0.0737) (0.0527) (0.0960) (0.0407) (0.0716) (0.2351) (0.0446) (0.0778)
10001000 0.3492 -0.5986 1.4037 0.4974 0.1033 0.3492 -0.5986 1.8622 0.4033 0.0834
(0.0244) (0.0430) (0.0520) (0.0370) (0.0661) (0.0244) (0.0429) (0.1570) (0.0311) (0.0533)
15001500 0.3498 -0.5988 1.4000 0.4976 0.1011 0.3499 -0.5988 1.8478 0.4030 0.0818
(0.0187) (0.0323) (0.0417) (0.0305) (0.0546) (0.0187) (0.0323) (0.1244) (0.0255) (0.0441)
Table 8: Kurtosis and skewness of indirect estimates and BWE for different sample size TT assuming (α,β,θ0,θ1,γ0,γ1\alpha,\beta,\theta_{0},\theta_{1},\gamma_{0},\gamma_{1})=(1.4,0,0.35,0.6,0.5,0.11.4,0,0.35,-0.6,0.5,0.1) with known β\beta from α\alpha-stable tvAR(1) based on R=1000R=1000 replications.
T Indirect estimates
θ0\theta_{0} θ1\theta_{1} α\alpha γ0\gamma_{0} γ1\gamma_{1}
500500 kur 3.7767 3.6800 3.1078 2.8924 3.0343
skw -0.1369 0.1213 0.2730 0.1583 0.0156
10001000 kur 4.7209 3.8513 2.7680 3.1397 3.0710
skw -0.1548 0.1008 0.0889 0.0672 -0.0654
15001500 kur 4.2029 3.8664 2.7385 3.0881 3.0266
skw 0.1274 -0.0192 0.0967 0.0973 0.0108

Finally, Figure 5 shows the density estimates of each parameter. The density estimates show that the standard error become smaller as TT increases. We conclude that the distribution of indirect estimates seem to be consistent for these sample path length.

Refer to caption
Figure 5: Density estimates of θ0\theta_{0}, θ1\theta_{1}, α\alpha, γ0\gamma_{0} and γ1\gamma_{1} for different sample sizes based on R=1000R=1000 replications from α\alpha-stable tvAR(1) with (α,β,θ0,θ1,γ0,γ1\alpha,\beta,\theta_{0},\theta_{1},\gamma_{0},\gamma_{1})=(1.4,0,0.35,0.6,0.5,0.11.4,0,0.35,-0.6,0.5,0.1) using indirect inference.

5.2.2 α\alpha-stable tvMA(1)

The indirect inference for the model (35) with unknown α\alpha is illustrated. The parameter of IM is θ=(θ0,θ1,α,γ)\theta=(\theta_{0},\theta_{1},\alpha,\gamma) and the parameter of AM is λ=(θ0(A),θ1(A),ν,γ(A))\lambda=(\theta_{0}^{(A)},\theta_{1}^{(A)},\nu,\gamma^{(A)}). The simulation was performed by assuming (α,β,θ0,θ1,γ)=(1.75,0.2,0.35,0.4,0.7)(\alpha,\beta,\theta_{0},\theta_{1},\gamma)=(1.75,0.2,-0.35,0.4,0.7).

The MC mean and standard error of the estimates from both model (IM and AM) are reported in Table 9, and kurtosis and skewness are presented in Table 10. Along with the density estimates showed in Figures 6, the indirect estimates seem to be consistent with these sample path length. One interesting result is that while α<2\alpha<2 implies the IM has infinite variance, the AM was estimated with ν>2\nu>2, i.e. finite variance.

Table 9: MC mean and standard error for different sample size TT using indirect estimators assuming (α,β,θ0,θ1,γ)=(1.75,0.2,0.35,0.4,0.7)(\alpha,\beta,\theta_{0},\theta_{1},\gamma)=(1.75,0.2,-0.35,0.4,0.7) with known β\beta from α\alpha-stable tvMA(1) based on R=1000R=1000 replications.
T Indirect estimates
Model of Interest Auxiliary model
θ0\theta_{0} θ1\theta_{1} α\alpha γ\gamma θ0(A)\theta_{0}^{(A)} θ1(A)\theta_{1}^{(A)} ν\nu γ(A)\gamma^{(A)}
500500 -0.3518 0.4016 1.7566 0.7008 -0.3518 0.4016 3.9795 0.3810
(0.0699) (0.1245) (0.0739) (0.0296) (0.0694) (0.1237) (1.0183) (0.0390)
10001000 -0.3487 0.3987 1.7527 0.6999 -0.3486 0.3987 3.8307 0.3776
(0.0446) (0.0787) (0.0559) (0.0229) (0.0445) (0.0788) (0.6414) (0.0299)
15001500 -0.3504 0.4009 1.7525 0.7003 -0.3502 0.4007 3.7874 0.3785
(0.0375) (0.0663) (0.0457) (0.0187) (0.0373) (0.0661) (0.4852) (0.0242)
Table 10: Kurtosis and skewness of indirect estimates and BWE for different sample size TT assuming (α,β,θ0,θ1,γ)=(1.75,0.2,0.35,0.4,0.7)(\alpha,\beta,\theta_{0},\theta_{1},\gamma)=(1.75,0.2,-0.35,0.4,0.7) with known β\beta from α\alpha-stable tvMA(1) based on R=1000R=1000 replications.
T Indirect estimates
θ0\theta_{0} θ1\theta_{1} α\alpha γ0\gamma_{0}
500500 kur 3.8667 3.2718 2.8731 3.3445
skw -0.0413 -0.0030 -0.1140 0.0003
10001000 kur 3.7260 3.4565 2.8049 2.9412
skw 0.0436 0.0582 -0.0081 0.1446
15001500 kur 3.6876 3.4211 3.0489 3.0002
skw -0.0043 0.0187 -0.2133 0.0557
Refer to caption
Figure 6: Density estimates of θ0\theta_{0}, θ1\theta_{1}, α\alpha and γ\gamma for different sample sizes based on R=1000R=1000 replications from α\alpha-stable tvMA(1) with (α,β,θ0,θ1,γ)=(1.75,0.2,0.35,0.4,0.7)(\alpha,\beta,\theta_{0},\theta_{1},\gamma)=(1.75,0.2,-0.35,0.4,0.7) using indirect inference.

5.2.3 α\alpha-stable tvARMA(1,1)

Finally, the simulation was done for the case of tvARMA(1,1) in (36), but α\alpha is assumed to be unknown. The time varying coefficients are assumed to be linear, i.e. α1(u)=θa0+θa1u\alpha_{1}\left(u\right)=\theta_{a0}+\theta_{a1}u and β1(u)=θb0+θb1u\beta_{1}\left(u\right)=\theta_{b0}+\theta_{b1}u, and εtSα(1/2,β,0)\varepsilon_{t}\sim S_{\alpha}(\nicefrac{{1}}{{\sqrt{2}}},\beta,0) for known β\beta. Therefore, the parameters of IM is θ=(θa0,θa1,θb0,θb1,α,γ)\theta=\left(\theta_{a0},\theta_{a1},\theta_{b0},\theta_{b1},\alpha,\gamma\right), while AM has the parameter λ=(θa0(A),θa1(A),θb0(A),θb1(A),ν,γ(A))\lambda=\left(\theta_{a0}^{(A)},\theta_{a1}^{(A)},\theta_{b0}^{(A)},\theta_{b1}^{(A)},\nu,\gamma^{(A)}\right). The simulation was performed by assuming (α,β,θa0,θa1,θb0,θb1,α,γ)=(1.3,0,0.2,0.4,0.2,0.3,1.1)(\alpha,\beta,\theta_{a0},\theta_{a1},\theta_{b0},\theta_{b1},\alpha,\gamma)=(1.3,0,-0.2,-0.4,0.2,0.3,1.1).

The MC mean and standard error of the estimates from both IM and AM are reported in Table 11, and kurtosis and skewness are presented in Table 12. The density estimates are showed in Figures 6. Again, the indirect estimates seem to be consistent. Moreover, if we compare with simulation results from the known α\alpha, they present similar standard error, kurtosis and asymmetry.

Table 11: MC mean and standard error for different sample size TT using indirect estimators assuming (α,β,θa0,θa1,θb0,θb1,α,γ)=(1.3,0,0.2,0.4,0.2,0.3,1.1)(\alpha,\beta,\theta_{a0},\theta_{a1},\theta_{b0},\theta_{b1},\alpha,\gamma)=(1.3,0,-0.2,-0.4,0.2,0.3,1.1) with known β\beta from α\alpha-stable tvARMA(1,1) based on R=1000R=1000 replications.
T θa0\theta_{a0} θa1\theta_{a1} θb0\theta_{b0} θb1\theta_{b1} α\alpha γ\gamma
Model of Interest 500500 -0.2036 -0.3932 0.1971 0.3064 1.3018 1.0923
(0.0585) (0.0869) (0.0587) (0.0891) (0.0698) (0.0587)
10001000 -0.2005 -0.3986 0.2003 0.3004 1.3045 1.0976
(0.0319) (0.0489) (0.0329) (0.0504) (0.0471) (0.0433)
15001500 -0.1998 -0.3999 0.2012 0.2983 1.2998 1.0953
(0.0233) (0.0359) (0.0250) (0.0374) (0.0390) (0.0347)
T θa0(A)\theta_{a0}^{(A)} θa1(A)\theta_{a1}^{(A)} θb0(A)\theta_{b0}^{(A)} θb1(A)\theta_{b1}^{(A)} ν\nu γ(A)\gamma^{(A)}
Auxiliary model 500500 -0.2036 -0.3936 0.1971 0.3062 1.5904 0.7465
(0.0584) (0.0864) (0.0586) (0.0889) (0.1731) ( 0.0940)
10001000 -0.2006 -0.3986 0.2003 0.3006 1.5917 0.7542
(0.0319) (0.0483) (0.0329) (0.0505) (0.1160) (0.0668)
15001500 -0.1998 -0.4009 0.2012 0.2983 1.5772 0.7487
(0.0232) (0.0344) (0.0250) (0.0374) (0.0947) (0.0540)
Table 12: Kurtosis and skewness of indirect estimates and BWE for different sample sizes (T=500,1000,1500T=500,1000,1500) assuming (α,β,θa0,θa1,θb0,θb1,α,γ)=(1.3,0,0.2,0.4,0.2,0.3,1.1)(\alpha,\beta,\theta_{a0},\theta_{a1},\theta_{b0},\theta_{b1},\alpha,\gamma)=(1.3,0,-0.2,-0.4,0.2,0.3,1.1) with known β\beta from α\alpha-stable tvARMA(1,1) based on R=1000R=1000 replications.
T Indirect estimates
θa0\theta_{a0} θa1\theta_{a1} θb0\theta_{b0} θb1\theta_{b1} α\alpha γ\gamma
500500 kur 5.2893 4.5345 6.0807 5.5860 3.0705 3.3815
skw 0.2593 -0.1945 -0.1951 0.1297 0.1406 0.2709
10001000 kur 4.6288 4.1073 4.5984 4.1306 3.4796 2.9077
skw -0.1406 0.1144 0.0360 -0.0328 0.1167 -0.0713
15001500 kur 4.9301 4.0790 5.1471 4.5091 3.1301 3.0701
skw 0.0236 -0.1964 0.0964 -0.2378 0.1004 0.0833
Refer to caption
Figure 7: Density estimates of θa0\theta_{a0}, θa1\theta_{a1}, θb0\theta_{b0}, θb1\theta_{b1}, α\alpha and γ\gamma for different sample sizes based on R=1000R=1000 replications from α\alpha-stable tvARMA(1,1) with (α,β,θa0,θa1,θb0,θb1,γ)=(1.3,0,0.2,0.4,0.2,0.3,1.1)(\alpha,\beta,\theta_{a0},\theta_{a1},\theta_{b0},\theta_{b1},\gamma)=(1.3,0,-0.2,-0.4,0.2,0.3,1.1) using indirect inference.

6 Application

In this section, we illustrate an application for wind power generated in German offshore wind farms from 16/06/2015 at 00:00 to 27/07/2015 at 24:00 (T=1008T=1008 hours), obtained from the EMHIRES (European Meteorological High resolution RES time series) datasets [26]. For daily data, the Gaussian innovation assumption seems to be appropriate, but the hourly time series present heavy tails and Gaussian assumption is inadequate. Figure 8, panel (a) shows the original time series (yty_{t}) and its difference (Δyt\Delta y_{t}), while panel (b) shows the standardized histogram of the differenced time series, which shows heavy-tailed behavior. We select just a small segment of the data because the whole time series has more complex structure, such as seasonality, thus a non-parametric approach could be more appropriate.

Refer to caption
(a) Hourly wind power (yty_{t}) and its difference (Δyt\Delta y_{t})
Refer to caption
(b) Standardized histogram of Δyt\Delta y_{t}.
Figure 8: Hourly wind power from 16/06/2015 at 00:00 to 27/07/2015 at 24:00.

Figure 9 shows sample autocorrelation function (global), and partial autocorrelation function. Traditional models, like ARMA(1,1) and AR(4) seem to be appropriate, but the blocked smooth periodogram shows its slowly changed structure over the time.

Refer to caption
(a) ACF and partial ACF.
Refer to caption
(b) Blocked smooth periodogram
Figure 9: ACF, partial ACF and blocked smooth periodogram of wind power data.

To explore its local structure, we estimate ARMA(1,1) and AR(4) for 9 time blocks. Figures 10 and 11 present the smoothed estimated coefficients over time for both models. Both cases show that coefficients are approximately linear over time. Consequently, two models are proposed:

  • tvARMA(1,1) model with linear coefficients, α1(u)=θa0+θa1(u)\alpha_{1}(u)=\theta_{a0}+\theta_{a1}(u), β1(u)=θb0+θb1(u)\beta_{1}(u)=\theta_{b0}+\theta_{b1}(u) and γ(u)=γ0+γ1(u)\gamma(u)=\gamma_{0}+\gamma_{1}(u).

  • tvAR(4) model with linear coefficients, α1(u)=θa0+θa1(u)\alpha_{1}(u)=\theta_{a0}+\theta_{a1}(u), α2(u)=θb0+θb1(u)\alpha_{2}(u)=\theta_{b0}+\theta_{b1}(u), α3(u)=θc0+θc1(u)\alpha_{3}(u)=\theta_{c0}+\theta_{c1}(u), α4(u)=θd0+θd1(u)\alpha_{4}(u)=\theta_{d0}+\theta_{d1}(u) and γ(u)=γ0+γ1(u)\gamma(u)=\gamma_{0}+\gamma_{1}(u).

Refer to caption
Figure 10: (Smoothed) α(u)\alpha(u), β(u)\beta(u) and γ(u)\gamma(u) estimates of stationary ARMA(1,1) model for 99 block of size N=200N=200 with u=t/Tu=t/T center point of each block.
Refer to caption
Figure 11: (Smoothed) α1(u)\alpha_{1}(u), α2(u)\alpha_{2}(u), α3(u)\alpha_{3}(u), α4(u)\alpha_{4}(u) and γ(u)\gamma(u) estimates of stationary AR(4) model for 99 block of size N=200N=200 with u=t/Tu=t/T center point of each block.

After estimating both models, the residuals of tvARMA(1,1) are correlated, and we focus only on the tvAR(4). The parameter estimates are reported in Table 13. Figure 12 presents the residual analysis and the QQ-plot, box-plot and the histogram show that the distribution of error has heavy tail and the residuals are approximately white noise. Additionally, we estimated the skewness (0.350.35) and kurtosis (13.8413.84) and carried out Shapiro-Wilk and Jarque-Bera tests, which rejected the null hypothesis of normality. Moreover, Figure 13 presents the variogram of the first difference of the wind data and the residuals from the tvAR(4) model. It is clear to observe that both of the variograms diverge.

Table 13: BWE of tvAR(4) from wind power time series.
Parameter BWE
Estimate s.e. z-value p-value
θa0\theta_{a0} -1.5985 0.0768 -20.8171 0.0000
θa1\theta_{a1} 0.3305 0.1406 2.3508 0.0187
θb0\theta_{b0} 0.9135 0.1373 6.6536 0.0000
θb1\theta_{b1} 0.0207 0.2447 0.0847 0.9325
θc0\theta_{c0} -0.0585 0.1372 -0.4266 0.6697
θc1\theta_{c1} -0.7153 0.2445 -2.9254 0.0034
θd0\theta_{d0} -0.1316 0.0767 -1.7158 0.0862
θd1\theta_{d1} 0.5454 0.1405 3.8831 0.0001
γ0\gamma_{0} 0.0077 0.0003 24.5452 0.0000
γ1\gamma_{1} 0.0152 0.0007 21.6400 0.0000
Refer to caption
(a) QQ-plot
Refer to caption
(b) Autocorreltion function
Refer to caption
(c) Box-plot
Refer to caption
(d) Histogram with estimated stable curve (α=1.34\alpha=1.34, β=0\beta=0) and Gaussian curve.
Figure 12: Residual analysis using the BWE (standadized residual).
Refer to caption
(a) Δyt\Delta y_{t}
Refer to caption
(b) Residuals
Figure 13: Variogram of the first differenced wind data Δyt\Delta y_{t} and the residuals of the tvAR(4) model.

Since the residuals present heavy tail, we propose a more flexible model, α\alpha-stable tvAR(4). We performed indirect estimation assuming known and unknown α\alpha. In the first case, we assume α=1.34\alpha=1.34 and β=0\beta=0, which are obtained by MLE from the residuals of the BWE. Since estimation results are similar to the estimated model by assuming unknown α\alpha, we present only results of the second model here.

The vector of parameters of IM is (θa0,θa1,θb0,θb1,θc0,θc1,θd0,θd1,α,γ0,γ1)(\theta_{a0},\theta_{a1},\theta_{b0},\theta_{b1},\theta_{c0},\theta_{c1},\theta_{d0},\theta_{d1},\alpha,\gamma_{0},\gamma_{1}) and the indirect inference was done by assuming symmetric α\alpha-stable innovations. Table 14 reports indirect estimates assuming β=0\beta=0 with their MC standard error with R=1000R=1000 replications.

Table 14: Indirect estimates of α\alpha-stable tvAR(4) with S=40S=40 from wind data.
Parameter Indirect estimate Standard error
θa0\theta_{a0} -1.5434 0.0251
θa1\theta_{a1} -0.0316 0.0426
θb0\theta_{b0} 0.9036 0.0442
θb1\theta_{b1} 0.1083 0.0764
θc0\theta_{c0} -0.2818 0.0437
θc1\theta_{c1} -0.2235 0.0752
θd0\theta_{d0} 0.0639 0.0246
θd1\theta_{d1} 0.1496 0.0412
α\alpha 1.3875 0.0528
γ0\gamma_{0} 0.0065 0.0005
γ1\gamma_{1} 0.0033 0.0010

To evaluate the residual distribution with the stable distribution, [27] suggested using the stabilized probability plot (stabilized p-p plot), proposed by [28], instead of the QQ-plot because the last one is not suitable to evaluate heavy-tailed distribution. In QQ-plot, large fluctuation for the extreme values in case of the heavy-tailed distribution produce large standard errors in the tails. Let y1yny_{1}\leq\cdots\leq y_{n} be an ordered random sample of size nn from the distribution FF. The stabilized p-p plot is defined as the plot of si=(2π)arcsin(F12(yi))s_{i}=\left(\frac{2}{\pi}\right)\arcsin(F^{\frac{1}{2}}(y_{i})) versus ri=(2π)arcsin([(i12)/n]12)r_{i}=\left(\frac{2}{\pi}\right)\arcsin\left(\left[\left(i-\frac{1}{2}\right)/n\right]^{\frac{1}{2}}\right). In this way, the histogram and the stabilized p-p plot in figure 14 show that the stable distribution fits well the residuals.

Refer to caption
(a) Histogram with the stable curve
Refer to caption
(b) Stabilized p-p plot
Figure 14: Residual analysis from the α\alpha-stable tvAR(4) model assuming that the innovation distribution is stable with α=1.34\alpha=1.34 and β=0\beta=0.

Finally, we compare the Mean square error (MSE), Root mean square error (RMSE) and Mean absolute error (MAE) of tvARMA(1,1) and tvAR(4) using BWE and indirect estimates. Note that MSE and RMSE do not make sense theoretically if we assume α\alpha-stable tvAR(4). In Table 15, we observe that using BWE (assuming finite variance), MSE and RMSE are slightly lower, while using the indirect inference presents lower MAE.

Since the residual analysis indicates heavy-tails, α\alpha-stable tvAR(4) is a better model to describe the data. In this case, by assuming α=1.34\alpha=1.34, which is far from 22, the simulation done in the previous section shows that the BWE is not appropriate. Even though MSE and RMSE are lower for BWE, they are not appropriate for α\alpha-stable process since they cannot be theoretically handled. Based on MAE, the indirect inference performs slightly better. Moreover, the interpretation of estimated coefficients of the model also changed, i.e. α1(u)\alpha_{1}(u) and α2(u)\alpha_{2}(u) are constant, while α3(u)\alpha_{3}(u), α4(u)\alpha_{4}(u) and γ(u)\gamma(u) vary linearly.

Table 15: Goodness of fit of different models for the wind data.
Model MSE RMSE MAE
tvARMA(1,1) 0.000248 0.015739 0.009675
α\alpha-stable tvARMA(1,1) 0.000257 0.016028 0.009469
tvAR(4) 0.000242 0.015542 0.009468
α\alpha-stable tvAR(4) 0.000256 0.015993 0.009094

7 Conclusion

In this paper, we studied α\alpha-stable locally stationary ARMA processes and presented their properties. In contrast to the locally stationary processes with finite variance, this process involves the infinite variance observed in different fields. We also proposed an indirect inference method for the process with parametric time-varying coefficients. We performed simulations for basic models with linear parametric coefficients for known and unknown α\alpha. The results show that indirect inference appropriate. An application is also illustrated.

There are some limitations that still need to be solved in the future. Firstly, since the time-varying spectral representation does not exist, identifying the local structure using traditional methods (autocorrelation and partial autocorrelation) are an informal way to identify the time-varying structure. One possibility is the local version of the dependence measure called autocovariation [18]. Secondly, simulations should be done for more complex models and also consider the possibility of non-parametric models. Thirdly, the indirect inference is time-consuming but they are appropriate when heavy-tailed innovations are present. Simulations suggest that when α\alpha is close to 22, Gaussian innovations can be assumed. Model selection is still an open question. Also, there is few work related to prediction.

Finally, we are involved in research about the locally stationary process with tempered stable innovations, which are similar to stable distribution in its center but their tails are lighter and moments of all orders are finite.

Funding

The authors are grateful to the support of a CNPq grant (141607/2017-3) and the University of Costa Rica (SWC), and a Fapesp grant 2018/04654-9 (PAM).

References

  • [1] Dahlhaus R. Maximum likelihood estimation and model selection for locally stationary processes. Journal of Nonparametric Statistics. 1996;6(2-3):171–191.
  • [2] Dahlhaus R. Fitting time series models to nonstationary processes. The Annals of Statistics. 1997;25(1):1–37.
  • [3] Dahlhaus R. Locally stationary processes. In: Tata Subba Rao SSR, Rao C, editors. Time series analysis: Methods and applications. (Handbook of Statistics; Vol. 30). Elsevier; 2012. p. 351 – 413.
  • [4] McCulloch JH. Simple consistent estimators of stable distribution parameters. Communications in Statistics - Simulation and Computation. 1986;15(4):1109–1136.
  • [5] Koutrouvelis IA. An iterative procedure for the estimation of the parameters of stable laws. Communications in Statistics - Simulation and Computation. 1981;10(1):17–28.
  • [6] Gourieroux C, Monfort A, Renault E. Indirect inference. Journal of Applied Econometrics. 1993;8(S1):S85–S118.
  • [7] Gallant AR, Tauchen G. Which moments to match? Econometric Theory. 1996;12(4):657–681.
  • [8] Lombardi MJ, Calzolari G. Indirect estimation of α\alpha-stable distributions and processes. Econometrics Journal. 2008;11(1):193–208.
  • [9] Sampaio JM, Morettin PA. Indirect estimation of randomized generalized autoregressive conditional heteroskedastic models. Journal of Statistical Computation and Simulation. 2015;85(13):2702–2717.
  • [10] Sampaio JM, Morettin PA. Stable randomized generalized autoregressive conditional heteroskedastic models. Econometrics and Statistics. 2019;(to appear).
  • [11] Calzolari G, Halbleib R, Parrini A. Estimating garch-type models with symmetric stable innovations: Indirect inference versus maximum likelihood. Computational Statistics & Data Analysis. 2014;76:158 – 171.
  • [12] Calzolari G, Halbleib R. Estimating stable latent factor models by indirect inference. Journal of Econometrics. 2018;205(1):280 – 301.
  • [13] Dahlhaus R, Polonik W. Empirical spectral processes for locally stationary time series. Bernoulli. 2009;15(1):1–39.
  • [14] Samorodnitsky G, Taqqu M. Stable non-gaussian random processes: Stochastic models with infinite variance. Taylor & Francis; 1994. Stochastic Modeling Series.
  • [15] Weron A, Weron R. Computer simulation of lévy α\alpha-stable variables and processes. Berlin, Heidelberg: Springer Berlin Heidelberg; 1995. p. 379–392.
  • [16] Embrechts P, Klüppelberg C, Mikosch T. Modelling extremal events for insurance and finance. Springer-Verlag Berlin Heidelberg; 1997.
  • [17] Brockwell PJ, Davis RA. Time series: Theory and methods. Springer-Verlag New York; 1991.
  • [18] Kokoszka PS, Taqqu MS. Infinite variance stable ARMA processes. Journal of Time Series Analysis. 1994;15(2):203–220.
  • [19] Kokoszka PS, Taqqu MS. Fractional ARIMA with stable innovations. Stochastic Processes and their Applications. 1995;60(1):19 – 47.
  • [20] Mikosch T, Gadrich T, Kluppelberg C, et al. Parameter estimation for ARMA models with infinite variance innovations. The Annals of Statistics. 1995;23(1):305–326.
  • [21] Shelton Peiris M, Thavaneswaran A. On the properties of some nonstationary arma processes with infinite variance. International Journal of Modelling and Simulation. 2001;21(4):301–304.
  • [22] Shelton Peiris M, Thavaneswaran A. Multivariate stable arma processes with time dependent coefficients. Metrika. 2001 Nov;54(2):131–138.
  • [23] Miller K. Linear difference equations. W. A. Benjamin; 1968. Mathematics monograph series.
  • [24] Van Bellegem S, von Sachs R. Forecasting economic time series with unconditional time-varying variance. International Journal of Forecasting. 2004;20(4):611–627.
  • [25] Dahlhaus R, Giraitis L. On the optimal segment length for parameter estimates for locally stationary time series. Journal of Time Series Analysis. 1998;19(6):629–655.
  • [26] Gonzales-Aparicio I, Zucker A, Carerri F, et al. EMHIRES dataset. part I: Wind power generation European Meteorological derived high resolution res generation time series for present and future scenarios. EUR 28171 EN; 10.2790/831549; 2016.
  • [27] Nolan JP. Maximum likelihood estimation and diagnostics for stable distributions; 2002.
  • [28] Michael JR. The stabilized probability plot. Biometrika. 1983;70(1):11–17.