Chapter 14

ARMA Processes

Autoregressive-moving-average (ARMA) processes are linear processes that only depend on a finite number of parameters, which facilitates their use in statistics (see section 10.3).

14.1. Autoregressive processes

DEFINITION 14.1.– images is said to be a autoregressive process of order p (AR(p)) if:

[14.1] images

with πp ≠ 0, and where (εt) is a white noise such that:

images

Uniqueness of the decomposition.– If there exists a weakly stationary process (Xt) satisfying [14.1], then the decomposition is unique. Indeed, if:

images

then we have:

images1

Then, if image, we have images and, by stationarity, images, images. Hence, images, which contradicts images. Step by step, we deduce images, thus εt = εt.

Existence

LEMMA 14.1.– Equation [14.1] has a weakly stationary solution when the modulus of every root of the polynomial images is strictly greater than 1.

PROOF.– The rational fraction 1/P(z) does not have any poles within the disk with its center at 0 and radius 1 + η for small enough η > 0.

Consequently, we have the power series expansion:

images

with a0 = 1 and ∑ |aj| < +∞.

Now, we define a linear process by setting:

[14.2] images

To conclude, we use the identity:

images

More precisely, we set:

images

as well as

images

and it may be verified that:

images

B is called the “backward” operator.

EXAMPLE 14.1.–For |ρ| < 1, the equation Xt = ρXt−1 + εt, t images, has the unique solution:

images

since

images

THEOREM 14.1. Autocovariance.– The autocovariance (γt) of an autoregressive process of order p satisfies the Yule–Walker equations:

[14.3] images

where σ2 is the variance of εt.

PROOF.– For k ≥ 1

images

In addition

images

and

images

Asymptotic correlation: The autocorrelation of (Xt) is defined by setting:

images

From the first formula of [14.3], we have:

images

Yet

images

images is then a solution to (E). We deduce for (E) the general solution:

images

where the ci are constants. Therefore, (Pk) is in general a mix of decreasing exponentials and damped sinusoids. In any case

images

Partial autocorrelation

DEFINITION 14.2.–

1) Let X, Y, Z1,…,Zk ∈ L2 images be centered. The partial correlation coefficient between X and Y, relative to Z1,…, Zk, is defined by:

images

where X* and Y* are the orthogonal projections of X and Y onto the subspace of L2 images generated by Z1,…, Zk.

2) Given a weakly stationary, centered process images, its partial autocorrelation (rk, k ≥ 1) is defined as:

images

with the convention rk = 0 if σ (XtX*t) = σ (XtkX*tk) = 0.

THEOREM 14.2.– If (Xt) is anAR(p), then

images

PROOF.–

– For k = p, we deduce by projection from [14.1] that

images

Hence

images

Since εtXtpX*tp, we have

images

but by stationarity

images

therefore

images

– For k > p, we find:

images

It follows that

images

14.2. Moving average processes

DEFINITION 14.3.– images is said to be a moving average process of order q (MA(q)) if:

[14.4] images

where a0 = 1, aq ≠ 0; εt is a white noise such that images.

The expansion [14.4] is unique and, if the roots of images(z) = 1 + a1z + … + aqzq are of modulus > 1, we have:

images

with ∑ |πj| < ∞ and

images

Therefore,

images

EXAMPLE 14.2.– If Xt = εt + a1εt−1, |a1| < 1, we deduce that

images

Autocovariance: A direct calculation shows that

images

Partial autocorrelation: It is difficult to calculate. For an MA(1), we find:

images

This type of result is general: (rk) tends to 0 at an exponential rate for all MA(q).

14.3. General ARMA processes

An ARMA(p, q) process is defined by the equations:

[14.5] images

which may be symbolically written as:

images

with ϕpθq ≠ 0, supposing P(a) = 0 and images(a) = 0 have no common roots.

If the roots of P and images are outside of the unit disk, we have the representations

[14.6] images

and

[14.7] images

Therefore, (Xt) is a linear process with innovation (εt), and p, q, (ϕj), and (θj′) are unique.

Autocovariance: From [14.5], it follows that

images

For k > q, we obtain:

images

which is a Yule–Walker equation (see Theorem 14.1), therefore (γk) has the same asymptotic behavior as the autocovariance of an AR(p).

Partial autocovariance: Relation [14.6] shows that one may approach an ARMA (p, q) process by an MA(q′). Using this property, it may be established that the partial autocorrelation of an ARMA has the same asymptotic behavior as that of an MA.

Spectral density: Let us set:

images

Using Lemma 10.1 twice, we obtain:

images

where fY and fX are the spectral density of (Yt) and (Xt), respectively, and σ2/(2π) is the (constant) spectral density of (εt).

Consequently,

images

This rational form of the spectral density characterizes the ARMA process.

14.4. Non-stationary models

In practice, observed processes more often have a non-stationary part, which must be detected and eliminated to reduce the problem to the study of a stationary process. Some empirical methods were indicated in Chapter 9 (section 9.3). We now present some more elaborate methods.

14.4.1. The Box–Cox transformation

Let (Xt) be a process whose variance and mean are related by an equation of the form

images

where φ is strictly positive.

We may then stabilize the variance by transforming (Xt). Indeed, if T is a sufficiently regular function, we will have in the neighborhood of EXt:

images

that is

images

This (heuristic!) reasoning leads us to choose a transformation T such that

images

where k is a constant.

For example, if VarXt = c(EXt)2 and Xt > 0, we may choose T(Xt) = log Xt. If VarXt = cEXt and Xt > 0, we choose images.

More generally, we may use the Box–Cox transformation:

images

Then λ appears as an additional parameter that must be estimated.

14.4.2. Eliminating the trend by differentiation

When the trend of a process is deterministic, it may be estimated by the least-squares method (see section 9.3). If it is stochastic, we seek to eliminate it.

Consider, for example, a process (Xt) defined by:

images

where the εt are i.i.d.

E(Xt|Xt−1,…) = Xt−1 is then the trend and the process

images

is stationary.

This leads us to define an ARIMA(p, q, d)process as an (Xt) satisfying

[14.8] images

where P and images are polynomials of respective orders p and q, with roots that lie outside of the unit circle, and d is an integer.

(Xt) may then be interpreted as an ARMA process such that 1 appears among the roots of the autoregression polynomial.

Since we cannot invert P(B)(I−B)d to determine Xt as a function of the εtj, we require p + d initial values: Xt0−1, Xt0−2,…, Xt0pd that determine Xt0. When all the starting values are eliminated, the process reaches its cruising speed, and (I − B)d Xt coincides with an ARMA(p, q) process.

14.4.3. Eliminating the seasonality

If (Xt) has a trend, and period S, we may envisage a model of the form:

images

where

images

with dP2 = P and dimages2 = images.

(Xt) is then said to be a SARIMA (p, q, d; P, Q, D)S process.

The SARIMA (0,1,1;0,1,1)12 model is widely used in econometrics, and is written as:

images

14.4.4. Introducing exogenous variables

The previous models have the drawback of being closed: they only explain the present of Xt from its past values. It is more realistic to allow exterior variables to play a role: for example, the consumption of electricy is related to the temperature.

Then, letting (Zt) be the process associated with an “exogenous” variable, we may envisage the ARMAX model defined by:

images

where P, images, and R are polynomials.

More generally, we may consider the SARIMAX model obtained by introducing an exogenous variable into a SARIMA process. For details, we refer to the bibliography.

14.5. Statistics of ARMA processes

14.5.1. Identification

For simplicity, we suppose that the initially observed process is an ARIMA (p, q, d) model. To identify d, we may note that if d is strictly positive, the observed random variables are strongly correlated.

For example, if Xt = ε1 + …+εt,t ≥ 1, the correlation coefficient of (Xt,Xt+h) is written as:

images

thus it tends to 1 when t tends to infinity with h fixed, or faster than h.

The random variables X1,…,Xn being observed, the empirical correlation coefficients are given by:

images

If images vary slowly with h, and are not in the neighborhood of zero, then it is recognized that the model is not stationary, and we consider the differentiated series

images

We then consider the empirical correlation coefficients of (Yt) and we may continue to differentiate. It is advised to choose d ≤ 2, as each differentiation leads to a loss of information.

We are now in the situation where the observed process (Xt) is an ARMA(p, q): we identify (p, q), or more precisely, we construct an estimator images of (p, q).

Among the various methods that have been proposed, we choose two:

1) The Corner method (Beguin, Gouriéroux, Monfort)

This method is based on the following theorem.

THEOREM 14.3.–Let (Xt) be a stationary autocorrelationprocess (ρt). Consider the determinants

images

and the matrix M = (Δij)1≤i,jk (Xt) is then an ARMA(p, q)process (where p < k, q < k) if and only if M has a “corner” at the intersection of the qth line and the pth column:

images

PROOF.–See [GOU 83]. The method consists of forming the images that allow the construction of an estimator images, then seeking a “corner” in images. For details of the implementation of this method, we refer to [GOU 83].

2) The Akaike criterion

This is based on the interval between the true density, i.e. f0, of the observed vector (X1,…, Xn) and the family of densities associated with the ARMA(p, q) model. The chosen risk is the Kullback information:

images

The estimators of I that have been proposed are of the form:

images

where images is the maximum likelihood estimator of σ2 when (Xt) is a Gaussian ARMA(p,q) process, and (un) is a sequence which depends only on n.

Then, images = argmin images. If un = log n/n where un = c log log n/n with (c > 2), then images is an estimator that converges almost surely to (p, q) when n → ∞.

COMMENT 14.1.– Before using the methods that we have just outlined, it is useful to calculate the images, and to construct some estimators images of the partial autocorrelations. The results of sections 14.1 and 14.2 then provide the following empirical criteria:

– If images becomes small for h > q, the model is an MA(q).

– If images becomes small for k > p, it is an AR(p).

– If images and images decrease slowly enough, the model is mixed.

14.5.2. Estimation

The observed process is now assumed to be an ARMA(p, q), where p and q are known. It is necessary to estimate the unknown parameter:

images

where ϕj are the coefficients of the polynomial ρ, θj are those of images, and σ2 is the variance of εt.

When (Xt) is Gaussian, we may estimate η using the maximum likelihood method. This method has the advantage of providing estimators with minimal asymptotic variance, but its implementation is delicate, as the likelihood is complicated. In the context of an MA(q), we have:

images

therefore (X1,…, Xn) is the image of the Gaussian vector (ε1−q,…, εn) by linear mapping. This remark allows us to explicitly give the variance since the εt are i.i.d. with distribution images.

In the general case, one may obtain an approximation of the likelihood by approaching (Xt) with an MA(Q′).

If the process is an AR(p), it is preferable to use the conditional maximum likelihood method.

The process is of the form:

images

Denote by f the density of (X1−p,…, X0) and consider the vector (X1−p, …, X0, ε1,…, εn) with density:

images

The change of variables images, n, let us obtain the conditional density g of (X1,…, Xn) given (X1−p,…,X0 ):

images

Supposing the random variables (X1−p,…, X0, X1,…, Xn) to be observed, we obtain the conditional likelihood equations:

images

hence the estimator images.

Note that these equations are obtained from the Yule–Walker equations [14.3] by replacing the autocovariances with empirical autocovariances.

From this remark, it may be shown that even in the non-Gaussian case,

images

14.5.3. Verification

The previous operations allow the construction of images, and images, which completely determine the model.

To verify the suitability of the model to the observations, we define the residues images by:

images

where images and images are the estimators of the polynomials P and images, respectively.

To test the independence of the images, we consider the empirical autocorrelations images associated with the observed images, and we set:

images

Then, if K > p + q, it may be shown that Qn converges in distribution to a χ2 with Kpq degrees of freedom, whence the Box–Pierce test with critical region

images

where, if Z follows a χ2 distribution with Kpq degrees of freedom,

images

This test is of asymptotic size α.

If the model is revealed to be inadequate, the identification procedure must be re-examined.

If several models survive the verification procedure, we choose the model that has the best predictive power, i.e. the model for which the estimated prediction error images is the smallest.

14.6. Multidimensional processes

The study of multidimensional processes lies outside the scope of this book. We will only give some indications.

We will work in images, equipped with its Borelian σ-algebra images (the σ-algebra generated by open balls) and with its Euclidian structure (scalar product images, norm images.

Let images be a sequence of random variables with values in images. Supposing images, the expectation Xt = (Xt1 , . . . , Xtd ) is defined by setting

images

The cross-covariance operator of (Xs,Xt) is the linear map from images to images defined by images is called the covariance operator of Xt (written CXt).

The process (Xt) is then said to be stationary if EXt does not depend on t and

images

EXAMPLE 14.3: WHITE NOISE IN images.– Let images be a sequence of random vectors with values in images such that images, and

images

This is a stationary process.

EXAMPLE 14.4: MA(∞).– Letting (εt) be a white noise with values in images, we set:

[14.9] images

where the aj are linear operators from images to images such that images with images; series [14.9] is then convergentin mean square in images:

images

and the process (Xt) is stationary. Under certain conditions, (Xt) becomes a d-dimensional ARMA process (see [GOU 83]).

Extension to infinitely many dimensions is possible, notably in a Hilbert space (see [BOS 07]).

14.7. Exercises

EXERCISE 14.1.– Show that if (Xt) is a d-dimensional stationary process, its coordinates are stationary.

Explain why the converse is not necessarily true.

EXERCISE 14.2.– (AR(1)) Let (εt) be a white noise in images and ρ be a linear map from images to images. The process (Xt) is defined by setting:

[14.10] images

where images.

1) Show the equivalence of the following two conditions:

i) images
ii) images

2) Assuming i) to be satisfied, show that [14.10] has one unique stationary solution given by:

images

where the series converges in quadratic mean in images.

3) Determine EXt. Show that CXt−1, εt = 0 and deduce the relation:

images

where ρ′ is the transpose of ρ.

4) Establish the relation CXt−1 ,Xt = ρCX0.

EXERCISE 14.3.– (AR(1)) Consider the AR(1) defined in the previous exercise. We observe X1,…,Xn and seek to estimate the parameters of this process.

1) One estimator of m is defined by setting images. Show that the series images is convergent, and that

images

2) Supposing m = 0 and images is invertible, use the relation images to construct an empirical estimator of ρ. Study its convergence in probability.

EXERCISE 14.4.– (AR(1)) Consider the AR(1) model:

images

where (εt) is a Gaussian white noise with variance σ2.

We observe X1,…, Xn and wish to estimate θ = (ρ,σ2).

1) Calculate the covariance matrix of (X1,…, Xn) and deduce the expression of the density fn(x1,…,xn;θ).

2) Writing f(xt|xt−1; θ) for the density of Xt given Xt−1 = xt−1, show that

images

3) Determine the conditional maximum likelihood estimator images of θ by maximizing

images

Compare this estimator with the least-squares estimator.

4) Study the convergence of images.

EXERCISE 14.5.– Let images be a real, centered, regular, weakly stationary process. Supposing the autocorrelation (ρj,j ≥ 0) of (Xt) satisfies the following property:

images

show that (Xt) is a moving average of order q.

EXERCISE 14.6.– Let images be a white noise and ρ be a real number such that |ρ| > 1. We set:

[14.11] images

1) Show that this series converges in quadratic mean.

2) Show that (Xt) is the unique stationary solution to the equation

[14.12] images

3) Calculate Cov(Xt−1, εt). Is [14.11] the Wold decomposition of the process?

4) Determine Cov(Xs, Xt).

5) Setting

images

determine the spectral density of (ηt). Deduce the Wold decomposition of (Xt).

6) Now, supposing ρ = 1, calculate Var(Xt+hXt), h ≥ 1. Show that, if [14.12] has the stationary solution (Xt), we have:

images

Deduce that such a solution does not exist.

7) Treat the case where ρ = −1.

EXERCISE 14.7.– Let images be a white noise. Consider the moving average

images

1) Establish the relation:

images

2) Deduce

images

where the limit is in quadratic mean.

3) Show that (εt) is the innovation of (Xt), while the root of the associated polynomial has modulus 1.

EXERCISE 14.8.– Let images be a weak white noise, and θ ≠ 1. We set:

images

1) Compute the covariance function of images. Deduce that it is stationary, and calculate its spectral density.

2) Show that if |θ| < 1, then images. Deduce in this case the Wold representation of images.

3) Show that if |θ| > 1, then images. Is Xt = εtθεt−1 the Wold representation of the process?

EXERCISE 14.9.– Let images be a weak white noise with variance σ2. Supposing there exists a stationary process in the weak sense, which satisfies the equation:

images

determine its Wold representation and its spectral density.


1 For the definitions of images and images see Chapter 10.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset