Beruflich Dokumente
Kultur Dokumente
(s, t)
(s, t) = p ,
(s, s)(t, t)
where
(s, t) = E[(Ys E(Ys ))(Yt E(Yt ))]
is the autocovariance function, provided the expectations exist.
A strictly stationary time series is one for which the probabilistic behavior of every collection
of values {Yt1 , Yt2 , . . . , Ytk } is identical to that of the time shifted set {Yt1 +h , Yt2 +h , . . . , Ytk +h }.
That is,
P{Yt1 c1 , . . . , Ytk ck } = P rob{Yt1 +h c1 , . . . , Ytk +h ck }
for all k = 1, 2, . . ., all time points t1 , t2 , . . . , tk , all numbers c1 , c2 , . . . , ck , and all time shifts
h = 0, 1, 2, . . ..
A second-order stationary time series, Yt , is a finite variance process such that
(a) the mean value function, t = E(Yt ) is constant and does not depend on time t, and
(b) the covariance function, (s, t) = E[(Ys s )(Yt t )] depends on s and t only through their
difference |s t|.
A common tool to remove trends is differencing. It allows to remove linear or polynomial trends,
but it usually complicates the dependence structure of the process.
(a) We have
Yt = + Yt1 + t
= 2 + Yt2 + t1 + t
= ...
Xt
= t + Y0 + k .
k=1
(b) We have
t
!
X
t = E(Yt ) = E t + Y0 + k = t
k=1
and
( s, t) = cov(Ys , Yt )
s t
!
X X
= cov s + Y0 + k , t + Y0 + k
k=1 k=1
min(s,t) min(s,t)
X X
= cov k , k
k=1 k=1
= min(s, t) 2 .
1
(c) We have
(t 1, t)
(t 1, t) = p
(t 1, t 1)(t, t)
(t 1) 2
= p
(t 1) 2 t 2
t1
= 1, as t .
t
This means two observations far apart are strongly correlated.
(d) We showed in (b) that t depends on t and (s, t), on min(s, t).
(e) Let us consider the differenced series Yt = Yt Yt1 = t + . We have then t = , which
does not depend on t and cov(Ys , Yt ) = cov(s + , t + ) = 2 (s t), where (s t) = 1 if
s = t and 0 otherwise.
The function F is called the spectral distribution function of h , and its derivative f , if it exists,
is called the spectral density function. If h |h | < , then f exists. A function f () defined on
P
R[1/2, 1/2] is the spectrum of a stationary process if and only if f () = f (), f () 0, and
f ()d < .
We have
2
hence,
X
f () = (h)e2h
h
XX
= 2 j jh e2ij e2i(jh)
h j
X X
= 2 j e2ij k e2ik
j k
2
= 2 |()| .
We also have
2
|(| = (1 1 e2i )(1 1 e2i )
and
2
|()| = (1 + 1 e2i )(1 + 1 e2i )
hence
2
|()|
fY () = 2 2
|()|
(1 + 1 e2i )(1 + 1 e2i )
= 2
(1 1 e2i )(1 1 e2i )
1 + 22 cos(2i) + 12
= .
1 21 cos(2i) + 21
4. A time series {Yt } is an autoregressive-moving average process of order p, q, ARMA(p, q), model,
if it is stationary and of the form
X
t = j Ytj = (B)Yt ,
j=0
An ARMA(p, q) process (B)Yt = (B)t is causal iff (z) 6= 0 within the unit disk D. If so, then
the coefficients of (z) satisfy (z) = (z)/(z) for z D. The process is invertible iff (z) 6= 0
for for z D. If so, then the coefficients of (z) satisfy (z) = (z)/(z) for for z D.
Yt = t 0.5t1
and so this is a causal and invertible MA(1) process ((z) = 1 and (z) = 1 0.3z has root
1/0.3). We have
(h) = 1.25 2 (h) 0.5 2 ((h + 1) + (h 1))
so except 0 = 1, well only have 1 = 0.4.
3
(b) Since the roots of 1 x + 0 5x2 are (1 i)/2, this is an ARMA(2,1) process. Moreover,
it is neither causal nor invertible since the two roots of (z) lie inside the unit disk and the
root of (z) is 1.
(c) Here, we have
(h) = 5 2 (h) 2 2 ((h + 1) + (h 1))
and so we have the same ACF as we had in (a). The two processes are the same (provided
the variance of the white noise is such that the two autocovariance functions are identical),
but the model in (a) is invertible, whereas the model in (c) is not.
where the t and t are mutually independent. We suppose that data y1 , . . . , yn are available.
Let Ht denote the information available at time t. Filtering is the estimation of t using Ht ,
smoothing is the estimation of t using Hn and prediction is the forecasting t+h fot h > 0 using
Ht .
(a) We have
iid
State equation: Xt = 0.9Xt2 + t , t N (0, 2 ),
iid
Observation equation: Yt = Xt + t , t N (0, 2 ).
2
02 = 12 = .
1 0.81
(c) The left time plot (Xt ) shows clearly the AR(2) structure, whereas on the right time plot, it
is more difficult to see, because of the noise t . The range is also more important (from -10
to 5 instead of -8 to 2). The left ACF is typical from an AR(2) model with such parameters.
On the right one, the added noise reduces the proportion of information in the observation
and the correlation, so the values are diminished on the plot. Finally, on the left PACF, we
clearly find the model structure, whereas on the right one, we also have the consequences of
the added noise.