Volume 2011, Issue 1 151823
Research Article
Open Access

Yule-Walker Estimation for the Moving-Average Model

Chrysoula Dimitriou-Fakalou

Corresponding Author

Chrysoula Dimitriou-Fakalou

Department of Statistical Science, University College London, London WC1E 6BT, UK ucl.ac.uk

Search for more papers by this author
First published: 14 August 2011
Citations: 4
Academic Editor: Qing Zhang

Abstract

The standard Yule-Walker equations, as they are known for an autoregression, are generalized to involve the moments of a moving-average process indexed on any number of dimensions. Once observations become available, new moments estimators are set to imitate the theoretical equations. These estimators are not only consistent but also asymptotically normal for any number of indexes. Their variance matrix resembles a standard result from maximum Gaussian likelihood estimation. A simulation study is added to conclude on their efficiency.

1. Introduction

At first, processes taking place on more than one dimension were considered on surfaces or in the three-dimensional space for the sake of spatial dynamics. Next, the time series Autoregressive Moving-Average model could be as well used to clothe the covariance dependence of spatial point processes. Nevertheless, the assumptions of causality and invertibility were based on conventional orderings of the vector indexes, as these were introduced by Whittle [1] and generalized by Guyon [2], and they failed to look natural on the spatial axes. The more general bilateral equations considered by Whittle [1] also related to serious problems during the estimation of the parameters. Besag’s [3] automodels as well as Besag’s [4] equations for observations on irregular sites moved in a new direction but the weaknesses during the estimation remained.

The inclusion of the time axis has provided more convenience for the spatial analysis that may be summarized in two ways. On the one hand when the observations are collected on irregular sites, the asymptotic properties of the estimators for the spatial dependence parameters can now be established as the number of timings only increases to infinity; for example, Dimitriou-Fakalou [5] has proposed a methodology on how to proceed. On the other hand when the regular recordings are increasing towards all the spatial as well as the time axes, the unilateral ordering of indexes can now be arranged in a meaningful way and the causal and invertible ARMA model may be preferred as a natural spatiotemporal equation.

Special cases of the ARMA model are the pure autoregressive and moving-average equations and they both relate to useful properties as in the spectral density of interest. The autoregressive process uses a spectral density with “finite denominator”, which for Gaussian processes [3] translates into a finite conditional expectation of any value based on all other values. A moving-average process uses a “finite numerator” in the spectral density, resulting in the autocovariance function with nonzero values on a finite set of “lags” only.

For a causal and finite autoregressive equation, the method of moments estimators according to the Yule-Walker equations almost coincides with the least squares or maximum Gaussian likelihood estimators and their consistency as well as their asymptotic normality may be established. Nevertheless, if the estimation for a pure moving-average or ARMA multi-indexed process takes place using its infinite autoregressive representation, a problem known as the “edge-effect” [2] resurrects and the asymptotic normality cannot be guaranteed. A modification of the Gaussian likelihood, which has been written from the autoregressive representation of each observation based on its “previous” ones, has been proposed by Yao and Brockwell [6] to confine the edge-effect with a weak condition of a finite second moment, but this works for the case of processes indexed on two dimensions only.

In this paper, for the multidimensional processes that may be clothed via an invertible and finite moving-average equation, a new method of estimation is proposed imitating the Yule-Walker equations. The finite number of nonzero autocovariances of the moving-average process is used most advantageously to eliminate the edge-effect. As a result, the proposed estimators are not only consistent but also asymptotically normal using relaxed conditions of finite second moments.

2. Theoretical Yule-Walker Equations

For = {0, ±1, ±2, …} and d any positive integer, real-valued processes indexed on d will be considered, such as time series or spatial processes on one line transect (d = 1), spatial processes in the plane (d = 2) or in the three-dimensional space, as well as the combination of spatiotemporal processes on 2, 3, or 4, respectively. For any vd, an invertible moving-average process {Y(v)} will be defined by the following equation:
()
where {ε(v)} are uncorrelated, zero-mean random variables with positive and finite variance σ2. In (2.1) a unilateral ordering of the fixed lags 0 < i1 < ⋯<iq has been considered. For d = 1, this is the standard ordering of integer numbers and, for d = 2, the conventional unilateral ordering has been introduced by Whittle [1]. Guyon [2] has generalized the ordering of two vector indexes on d, d ≥ 3. As for the invertibility condition, it is only imposed to make sure that it can be written ε(v) = Y(v) + ∑i>0ΘiY(vi), ∑i>0i| < also using the ordering i > 0. Anderson and Jury [7] have provided a sufficient condition on the invertibility of such filters.
For the complex numbers z1, …, zd and the vector z = (z1, …, zd), the polynomial is considered, where for i1, …, id. Similarly, the backwards operators B1, …, Bd and the vector backwards operator B = (B1, …, Bd) are considered, such that Biε(v) ≡ ε(vi). So (2.1) may now be rewritten as
()
Writing , the autoregressive sequence {X(v)} may be defined from {ε(v)} according to the following:
()
The auto-regression is unilateral but not causal, in the sense that it can be written that X(v) = ε(v) + ∑i>0Θiε(v + i), ∑i>0i| < as a function of the “future” terms ε(v + i), i ≥ 0. Moreover, writing the two polynomials
()
it holds that
()
The set ≡ {in, −in, inim,   n, m = 1, …, q} in (2.4) has members the vector “lags”, where {Y(v)} have nonzero autocorrelations. The set {in, inim,   n, m = 1, …, q, n > m} that uses the “positive” lags of will have cardinality q*.
The relationship in (2.5) provides the way to write Yule-Walker equations for a moving-average process. The processes {Y(v)} and {X(v)} are not only “caused” by the same sequence {ε(v)}, one from its “past” and the other from its “future” values, but also have “reverse” second-order dependence; that is, the spectral density of the auto-regression is
()
where according to (2.4), E(X(v)X(vi)) = σ2ci and the spectral density of the moving-average process is
()
which also implies that E(Y(v)Y(vi)) = σ2γi. While the original Yule-Walker equations for {X(v)} dictate
()
()
from (2.9), the following (q + 1) equations may be written for i > 0, that is,
()
and their sum ∑jγjcij = ∑jγjcji = ∑jγjcji = ∑jγjcj+i = 0, which may be rewritten as
()
Further to (2.11), it holds that
()
from (2.8) and (2.9). It is true that (2.11) and (2.12) are equivalent to γ(z)c(z) = 1. Like the standard Yule-Walker equations, the general versions (2.11) and (2.12) are constructed from the coefficients . As the finite number of nonzero autocovariances of {Y(v)} is involved now, they will be used as theoretical prototypes to be imitated by their sample analogues for the sake of estimation of the moving-average parameters.
Finally, as the assumption of invertibility of the moving-average equation used for parameterization will be essential, it should be noted that Dimitriou-Fakalou [8] has presented in detail the merits of a causal and invertible spatiotemporal ARMA model. As a special case, the moving-average process {Yt(s)}, where t is the time index and s are spatial indexes, may satisfy the equation , where {εt(s)} are zero mean, uncorrelated variables with identical, finite variance, qt is a finite positive integer, have a finite cardinality, and it holds that
()
that is, the model is invertible. Note that, in either equation, no unnatural ordering is forced over space as the unilateral conventions “run” over the past timings only. Hence the model both uses a space-time unilateral ordering of indexes and is meaningful. As a result, for the special case of a stationary process with autocovariance function identified to take zero values after a reasonably small number of lags on all spatial and temporal dimensions, the invertible moving-average equation should be preferred and the new results presented next will be useful.

3. Method of Moments Estimators

For 𝒮d being a set of finite cardinality N, there are available {Y(v),   v𝒮} from (2.2) with true parameters that are to be estimated. The following condition will be assumed to be true.

Condition C. The parameter space Θ ⊂ q is a compact set containing the true value θ0 as an inner point. Further, for any the moving-average model (2.2) is invertible.

Based on 𝒮, a set v will be defined for any vd. More specifically, it will hold that iv only if vi𝒮. Next, for the corrected set 𝒮* with N* elements, it will hold v𝒮* if v + i𝒮 for every i. Thus for any v𝒮*, it holds that v. The reduction from 𝒮 to 𝒮* is essential as it will guarantee later that the edge-effect will fade away; this is because the set 𝒮*𝒮 includes the locations v that have all their “neighbors” vi with E(Y(v)Y(vi)) ≠ 0 being available. As the source of the edge-effect is the speed of the absolute bias to zero, 𝒮* will guarantee that whatever “is missing” from the finite sample will not add at all to the bias of the estimators.

By imitating (2.11) the estimators are set to be the solutions of equations:
()
where and . Using Proposition 3.1, the equations
()
will be used additionally, in order to make sure that the estimators are consistent. The proposition guarantees that the use of q equations used as prototypes for estimation can provide a unique way to compute the q parameters of the invertible equation.

Proposition 3.1. The polynomials and θ0(z) −1 = 1 + ∑i>0Θi,0zi with ∑i>0i,0| < as well as γ0(z) = θ0(z)θ0(z−1) ≡ ∑jγj,0zj and are considered. For any , such that it holds that and θ(z)−1 = 1 + ∑i>0Θizi with ∑i>0i| < , as well as γ(z) = θ(z)θ(z−1) ≡ ∑jγjzj and , the unique solution that satisfies the q equations , n = 1, …, q, and , n, m = 1, …, q, n > m, is , n = 1, …, q.

For mathematical convenience, new variables depending on the sampling set are defined HY(v) ≡ Y(v), v𝒮 and HY(v) ≡ 0, v𝒮, and (3.1) may be rewritten as
()
or
()
where the zero subindex relates to θ0. In (3.3) and (3.4), it is Jn = (Jn,1, …, Jn,q) with elements
()
The last term in (3.5) will become more obvious when the properties of the estimators will be established but for now it might be justified by a second-order Taylor expansion, where it holds
()
due to the fact that will be shown to be consistent estimators in Theorem 4.1. Further, the sums over 𝒮* involving the second derivatives of the Taylor expansion, when divided by N* or N (N*/N → 1), will also tend in probability to the relevant finite expectations when {ε(v)} will be assumed to be independent and identically distributed random variables with a finite variance.
Finally by imitating (2.12), the estimator of the error variance may be defined as
()

Example 3.2. For the simplest one-dimensional case when q* = q = 1, it is demonstrated in detail how to compute the new estimator; some other cases will briefly be considered later. Suppose that are available {Y(1), …, Y(N)} from Y(v) = e(v) + θe(v − 1), where |θ| < 1, θ ≠ 0, and {e(v)} are uncorrelated random variables with variance unity. The original set 𝒮 ≡ {1,2, …, N} will be corrected to 𝒮* ≡ {2,3, …, N − 1} that it is the maximal set, such that for all v𝒮* it holds that E(Y(v)Y(v*)) = 0, v*𝒮. Note that the “lags” on which the autocorrelations of Y are not equal to zero are as in the set = {1,0, −1}; further, for any v it holds that v = {v − 1, …, vN} and, thus, for v𝒮* it is true that v.

The autocovariance of the relevant auto-regression at the “lag” j is (−θ)|j|/(1 − θ2)  . As a result, the estimator will be the solution of the following equation:

()
Furthermore, when {e(v)} are independent and identically distributed random variables, it will hold in probability that ∑vY(v)Y(vi)/N → 0 for |i| = 2,3, … and the estimator may be approximated as a solution of the quadratic equation:
()
Later, it will be investigated whether the approximation is worthwhile. The estimator reduces to
()
or
()
with
()
For the actual nonzero auto-correlation ρ = θ/(1 + θ2), it holds that |ρ| < 1/2 and D = 1 − 4ρρ > 0. As a result, if 0 < θ < 1 and 0 < ρ < 1/2, then 1/2ρ > 1, and the value
()
is bigger than 1. If, on the other hand, −1 < θ < 0, and −1/2 < ρ < 0 then 1/2ρ < −1 and the same value is smaller than −1. Thus, the estimator is approximately
()
Although the distribution of is not known for small sample sizes and risk to be outside the (−0.5,0.5) range with some positive probability, both the consistency and the asymptotic normality will be established next. As opposed to what is known for the maximum Gaussian likelihood estimators, the proposed estimators will be asymptotically normal even for processes on more than one dimension. Thus, the new methods might be more useful for higher number of dimensions as well as large sample sizes.

4. Properties of Estimators

The asymptotic normality of standard estimators for the parameters of stationary processes indexed on d, d ≥ 2, has not been established yet. More specifically as Guyon [2] demonstrated, the bias of the genuine Gaussian likelihood estimators, computed from N regular recordings increasing to infinity on all dimensions and at equal speeds, is of order N−1/d. Nevertheless in order to secure the asymptotic normality, the absolute bias when multiplied by N1/2 should tend to zero, which is only true for d = 1 as in Brockwell and Davis [9].

Regarding the proposed estimators of this paper, their bias will stem from (3.3) as in the expected value of , n = 1, …, q, which expresses what is “missing” from the sample. Even when the bias is multiplied by N*1/2, the random variable , n = 1, …, q has a zero mean due to the selection 𝒮* and the fact that the “finite” autocovariance function of a moving-average process is being used. Thus, the edge-effect will not be an obstacle to establish the asymptotic normality of the estimators.

A series of ARMA model-based arguments have been used before to deal with the edge-effect by Yao and Brockwell [6] for two-dimensional processes as well as a weak condition of finite second moments. Guyon [2], on the other hand, used the periodogram as in the Gaussian likelihood of observations proposed by Whittle [1] and required that the fourth moments would be finite. Dahlhaus and Künsch [10] improved certain weaknesses of Guyon′s [2] modification on the likelihood but paid the price of having to reduce the dimensionality to secure their results. The proposed estimators of this paper will also use the weak condition of finite second moments, which is a great advantage, since they will refer to moving-average processes on any number of dimensions.

Condition 4. (i) For a set 𝒮𝒮Nd of cardinality N, it is written N if the length M of the minimal hypercube including 𝒮, say 𝒮𝒞M, and the length m of the maximal hypercube included in 𝒮, say 𝒞m𝒮, are such that M, m. (ii) Further, as M, m it holds that M/m is bounded away from .

Theorem 4.1. If {ε(v)} ~ IID(0, σ2), then under Conditions C1 and C2(i) and as N, it holds that

()
From (3.3) and (3.4), all the q equations together may be written as
()
where and
()
which depends on the sampling set 𝒮. The next proposition reveals what happens to the part (J/N) in (4.2). Then Theorem 4.3 establishes the asymptotic normality of the estimators.

Proposition 4.2. Let the polynomial . If {ε(v)} ~ IID(0, σ2), then under Conditions C1 and C2(i) and as N, it holds that

()

Theorem 4.3. For {W(v)} ~ IID(0,1), the auto-regression {η(v)} is defined by θ0(B)η(v) ≡ W(v). Also the vector and the variance matrix

()
are considered. If {ε(v)} ~ IID(0, σ2), then under Conditions C1 and C2 and as N, it holds that
()

5. Empirical Results

In this section, an empirical comparison of the proposed estimators to maximum Gaussian likelihood estimators for moving-average processes is presented in detail. The theoretical foundations in favour of the new estimators have been provided already when their asymptotic normality on any number of dimensions has been established based on finite second moments only. As a result, the speed of the bias to zero is not expected to cause them to perform worse than maximum likelihood estimators, especially when the dimensionality is large.

On the other hand, Theorem 4.3 attributes to the new estimators the variance matrix when, according to Hannan [11], efficient estimation results in with Wq ≡ Var (ξ) and the same notation as in Theorem 4.3. Wq and are defined similarly but they are not, in general, equal. It seems that as the number of moving-average parameters of the process increases, the two types elements get closer. Nevertheless, as the pure moving-average model will be preferred when it is parsimonious, a decision is made here whether the new estimators are efficient then.

The investigation has started with the one-dimensional case by generating {Y(1), …, Y(N)} from the model with one parameter only. The moments estimator has been approximated as in the example earlier, while the minimizer of with respect to the parameter θ has been considered to be the Gaussian likelihood estimator. The true values θ = 0.5, −0.8 have been considered and the sample size has been set equal to N = 30,   100,   200,   300. Very encouraging results for the efficiency of the proposed estimator are presented in Table 1 as even when the sample size is still small, extreme differences in the variances of the two types estimators cannot be detected. It is the bias of the moments estimator that seems to be the only reason why it might be outperformed by the likelihood estimator in terms of the Mean Squared Error. Nevertheless, the speed with which the bias tends to zero is much faster as one would expect and, eventually, the new estimator for θ = 0.5 performs better altogether.

Table 1. Approximate bias, variance, and Mean Squared Error of the Method of Moments and the Maximum Likelihood Estimator for the MA(1) parameter from 1000 replications.
θ = 0.5 θ = −0.8
N = 30 Bias Var MSE Bias Var MSE
MME −0.403345 0.00020972 0.162897 0.0931052 0.00097912 0.00964771
MLE 0.01795 0.00015984 0.00048204 0.020949 0.00062347 0.00104331
N = 100 Bias Var MSE Bias Var MSE
MME −0.123812 0.00021444 0.0155439 0.0217928 0.00097912 0.00145404
MLE 0.00521 0.00031329 0.00034043 0.00599 0.00063936 0.00067524
N = 200 Bias Var MSE Bias Var MSE
MME −0.0422653 0.00021173 0.00199808 0.0123275 0.00097912 0.00113109
MLE −0.00116 0.00022068 0.00022202 0.00275 0.00062348 0.00063104
N = 300 Bias Var MSE Bias Var MSE
MME −0.00708814 0.00023768 0.00028792 −0.00166597 0.00097912 0.00098189
MLE 0.00313 0.00029131 0.00036115 0.00386 0.00065544 0.00067034
Next for the two-dimensional case, {Y(u, v),   u, v = 1, …, n} have been generated from
()
with {e(u, v),   u, v} being independent, Gaussian random variables with zero mean and variance unity. Using similar arguments as in the example, the moments estimators of β and γ might be derived by solving simultaneously
()
as well as the third equation
()
which is needed to secure the consistency (Proposition 3.1). All of the sums above extend over 𝒮* ≡ {2, …, n − 1} × {2, …, n − 1}. The estimators introduced here are , and in the three equations, it has been considered that , where
()
and {e*(u, v)} are zero mean, uncorrelated random variables all with variance unity. Under the condition , a special form of “causality” is implied and the standard Yule-Walker equations may be used to easily write the needed autocorrelations , as functions of and . Again, the maximum likelihood estimators have been approximated by the minimizers of ∑u,v(Y(u,v)−βe(u−1,v)−γe(u,v−1))2 with respect to β and γ.

Table 2 verifies once more the conclusions drawn from the one-dimensional case. It is safe to consider that the new estimators are efficient and this is very apparent in the case that the parameters are in absolute value further away from zero (β = 0.5 and γ = 0.45). The striking case with the small sample size of around 50 points observed in the plane reveals that the variances of the moments estimators may even happen to be smaller. On the other hand, the bias heavily affects the results for the MSE for small sample sizes. Nevertheless as the sample size increases, the absolute bias of the likelihood estimators does not seem to decrease at all versus the bias of the proposed estimators that speedily reaches the zero value. As a result, the new estimators eventually equalize the MSE performance of the standard estimators.

Table 2. Approximate bias, variance, and mean squared error of the method of moments and the maximum likelihood estimators (based on N = n2 observations) for the two-dimensional MA parameters β and γ from 1000 replications.
β = 0.5 γ = 0.45
n = 7 Bias Var MSE Bias Var MSE
MME 0.02842 0.00010879 0.00091649 −0.07858 0.00001209 0.0061869
MLE 0.0006 0.0002023 0.00020266 −0.00604 0.00006244 0.00009892
n = 30 Bias Var MSE Bias Var MSE
MME 0.0149 0.00037173 0.00059374 −0.00746 0.0002023 0.00025795
MLE −0.00184 0.0003022 0.00030558 0.00066 0.00016793 0.00016837
n = 100 Bias Var MSE Bias Var MSE
MME 0.00576 0.00025984 0.00029302 0.0065 0.00023986 0.00028211
MLE −0.00044 0.00022068 0.00022087 −0.00024 0.0002023 0.00020236
β = 0.15 γ = −0.3
n = 30 Bias Var MSE Bias Var MSE
MME 0.02638 0.00004406 0.00073996 0.03274 0.00001688 0.00108879
MLE 0.00012 0.00002248 0.00002249 0.00004 0.00010879 0.00010879
n = 100 Bias Var MSE Bias Var MSE
MME −0.0422653 0.00021173 0.00199808 0.0123275 0.00097912 0.00113109
MLE −0.00058 0.00002248 0.00002281 0.00026 0.00000402 0.00008408

Acknowledgments

Some results of this paper are part of the Ph.D. thesis titled “Statistical inference for spatial and spatiotemporal processes,” which has been submitted to the University of London. The author would like to thank the Leverhulme Trust, which funded her studies and Professor Q. Yao.

    Appendix

    A. Outline Proofs

    A.1. Proof of Proposition 3.1

    Under the invertibility condition for the polynomials θ(z), there is a one-to-one correspondence between the q coefficients and the q autocorrelations , where ρj = γj/γ0, jd. It is true that the coefficients , n = 1, …, q, generate the numbers cj,0, jd, and they are a solution to the q equations of interest. If there is another solution, say ,n = 1, …, q generating cj,1, jd, then it must hold that
    ()
    On the other hand, the general Yule-Walker equations for this solution imply that
    ()
    Thus, the q linear equations with q* unknowns are derived as
    ()
    with a unique solution as it is explained next; the autocovariances cj,1, jd refer to a (weakly) stationary process, say {X1(v)}, with spectral density bounded away from 0 and under Condition C1. Consequently, the variance matrix
    ()
    where , is nonsingular and there is a unique solution to the equations above, that is, ρj,1 = ρj,0,  j.

    A.2. Proof of Theorem 4.1

    For n = 1, …, q, (3.1) may be rewritten as
    ()
    Under the assumption that {ɛ(v)} are independent and identically distributed, it holds as N that , according to Proposition 6.3.10 or the Weak Law of Large Numbers and Proposition 7.3.5 of Brockwell and Davis [9], extended to include the cases d ≥ 2. Then for n = 1, …, q and the first of two terms in (A.5), it may be written as
    ()
    or
    ()
    For the next term, it may be written as
    ()
    due to the Cauchy-Schwartz inequality and the independence of the random variables Y(v), Y(vj),j. Now for any θ ∈ Θ, it holds that cjcj(θ) is the corresponding autocovariance function of a causal auto-regression. This guarantees that the autocovariance function decays at an exponential rate and there are constants C(θ) > 0 and α(θ)∈(0,1), such that , j = (j1, …, jd). Similarly for the estimator , it holds that
    ()
    with probability one and . For the case of observations on a hyperrectangle and under Condition C2(ii), it can be easily verified that , for example, the arguments of Yao and Brockwell [6] when d = 2. In general, it can be written as Condition C2(i) that
    ()
    After combining the results for the two terms in (A.5), it can be written that
    ()
    where the first term has been set equal to zero. Thus, exactly like the theoretical analogue implies. Since q instead of q equations have been used, there is a unique solution θ0, according to Proposition 3.1 and as N and C2(i) holds. Finally, the consistency for implies, according to (3.7), that
    ()

    A.3. Proof of Proposition 4.2

    According to (3.5), for the (n, m)th element of J/N,   n, m = 1, …, q, it suffices to look at
    ()
    where the last term tends to zero in probability, thanks to the consistency of the estimators from the use of all the q equations. For the polynomial
    ()
    the second term in (A.1) can be written as
    ()
    This comes straight from the fact that
    ()
    as N and Condition C2(i) holds. The limit comes from the same argument, as for the proof of the consistency for the estimators. For example, if Condition C2(ii) is true, it may be written , since for any i = (i1, …, id) ∈ d, it holds that for constants C > 0 and α ∈ (0,1). Similar action might be taken for the first term in (A.1).
    For the auto-regression {X(v)} as it was defined in (2.3), it can be seen immediately that Y(v) is uncorrelated to X(v + i), i > 0, since these are linear functions of “future” error terms only. In general, it can be written according to (2.5) that
    ()
    and back to the general Yule-Walker equations. Thus,
    ()
    and Y(v) is uncorrelated to X(vi),   i ≠ 0. As a result, it holds for n, m = 1, …, q, that
    ()
    and, for nm, that
    ()
    The proof is completed when it is seen that both Y(v) and X(v) are linear functions of members from the sequence {ε(v)} and, thus,
    ()

    A.4. Proof of Theorem 4.3

    From (3.3), (4.2), and (4.3) and for n = 1, …, q, it can be written that
    ()
    The convergence in probability to zero of the remainder might be justified by the fact that its expected value is equal to zero due to the selection 𝒮*, and that its variance is
    ()
    where
    ()
    given the sampling set 𝒮. First, {ε(v)} are independent and identically distributed and then
    ()
    without the assumption of a finite third or fourth moment. Under Condition C2(ii), it can be written that
    ()
    and a similar argument may be used for the cross-terms due to the Cauchy-Schwartz inequality. For the case d = 2 and recordings on a rectangle, there is a justification for that by Yao and Brockwell [6]. Thus, as N and Condition C2 holds, which guarantees the convergence in probability to zero.
    Equation (4.2) may now be rewritten as
    ()
    where U(v) τ ≡ (X(v + i1), …, X(v + iq))Y(v), vd. It holds that Y(v) is a linear function of ε(vi),   i = 0,   i1, …, iq, and X(v) is a function of ε(v + i),   i ≥ 0. Then for n, m = 1, …, q, it can be written for j ≥ 0 that
    ()
    Thus, for , Cj,0E(X(v)X(v+j)τ)/σ2, it can be written that E(U(v)) = 0 and
    ()
    Now, for any positive integer K, the set
    ()
    is defined. According to the infinite moving-average representation of X(v), for fixed K the new process {X(K)(v)} may be defined from
    ()
    Similarly, U(K)(v) τ ≡ (X(K)(v + i1), …, X(K)(v + iq))Y(v),   vd, and
    ()
    are defined here. For the same reasons as before, it holds that E(U(K)(v)) = 0 and that
    ()
    For any vector λq, it holds that {λτU(K)(v)} is a strictly stationary and K*-dependent process, for a positive and finite integer K*. The definition of K-dependent processes as well as a theorem for the asymptotic normality for strictly stationary and K-dependent processes on d might be given similarly to the one-dimensional case by Brockwell and Davis [9]. Then
    ()
    as N and under Condition C2, where . Similarly, for
    ()
    it holds as K that λτMKλλτMλ. Using Chebychev′s inequality, it may be verified that
    ()
    as K and, thus, it holds that
    ()
    as N and under Condition C2. According to (A.14), the (n, m)th element of M is equal to
    ()
    and back to the general Yule-Walker equations with MIq, that is, the identity matrix of order q. Equation (A.16) may be rewritten
    ()
    as N and under Condition C2. Combining (A.6) and (A.18) with Proposition 4.2, it can be written that
    ()
    as N and under Condition C2. The proof will be completed when it is shown that . Indeed, for the vector and for with R being a (q × 1) random vector that is independent of W, since it is a linear function of W(−i1i),   i > 0,   ii2i1, …, iqi1, the required result might be obtained then, as
    ()
    Note that the decomposition , as it is justified in the end of proof of Theorem 4.3, guarantees that has a nonzero determinant.

      The full text of this article hosted at iucr.org is unavailable due to technical difficulties.