Beruflich Dokumente
Kultur Dokumente
Christian Brownlees
Abstract
These notes contain the derivations of some of the results presented in the course.
1
1 Intro to Time Series
{yt : t ∈ T}
{yt (ω) : t ∈ T}
Consider the vector t = (t1 , t2 , ..., tk )0 with ti ∈ T and k ∈ N. Note that the k tuple
yt1 , yt2 , ..., ytk is a multivariate random variable in Rk . The distribution of the random
vector
is called the nite dimensional distribution of the time-series {yt } for t = (t1 , t2 , . . . , tk ).
Theorem 1.1. Kolmogorov Extension Theorem The set of all nite dimensional
distributions F is the set of all nite dimensional distributions of some stochastic process
lim Fyt1 ,yt2 ,...,yti ,...,ytk (y1 , . . . , yi , . . . , yk ) = Fyt1 ,yt2 ,...,yti−1 ,yti+1 ,...,ytk (y1 , . . . , yi−1 , yi+1 , . . . , yk ) .
yti →∞
2
1.2 Stationarity
Denition 1.4. Strict Stationarity. Consider the vector t = (t1 , t2 , ..., tk )0 with
ti ∈ T and k ∈ N.
The time series {yt } is strictly stationary if for each t the nite dimensional distribution
of
and
Denition 1.5. Autocovariance Function. Let {yt } be a time series such that
Var(yt ) < ∞ for each t. The autocovariance function γ(s, t) of {yt } is dened as
stationary if
3. For any t, s,
that is, the autocovariance of the process only depends on the lag k = t − s.
Denition 1.7. White Noise. A covariance stationary time series {yt } is white noise
yt ∼ W N (σ 2 )
3
Exercises
Exercise 1.1. Consider the simple Gaussian random walk process for t = 1, 2, ...
t
X
xt = zi
i=1
where zi are iid Gaussian with zero mean and unit variance. Is the process strictly station-
ary? Explain. Is the process covariance stationary? Explain. Is the process Gaussian?
Explain.
distribution F . Can the iid sample be considered a stochastic process? Explain. If yes,
is the process strictly stationary? Explain. If yes, is the process covariance stationary?
Exercise 1.3. Show that if {yt } is covariance stationary, then the process {xt } dened
Exercise 1.4. Let {yt } be covariance stationary. Consider the process {xt } dened as
xt = at + byt where a, b ∈ R and the process {zt } dened as zt = xt − xt−1 . Note that the
{xt } process can be thought of a process made up of a deterministic linear trend (at) and
a stochastic component (byt ). Is {xt } covariance stationary? Explain. Is {zt } covariance
stationary? Explain.
Exercise 1.5. Let {yt } and {xt } be covariance stationary processes. Also, the {yt } and
{xt } processes are uncorrelated, meaning that is Cor(yt , xs ) = 0 for any t, s. Is the {zt }
process dened as zt = yt + xt covariance stationary? Explain.
4
1. Expectation of the Sample Mean
E(y T ) = µ
∞
X
2
lim T Var(y T ) = γk = σLR
T →∞
k=−∞
where
∞
X
2
σLR = γk .
k=−∞
T
! T
1X 1X
E yt = E (yt ) .
T t=1 T t=1
Note that this property holds irrespective of the dependence properties of the the process
{yt }.
Part 2. We begin by noting that
!2
T
1 X
E(y T − µ)2 = E (yt − µ)
T2 t=1
T T T
!
1 X X X
= E (y1 − µ) (yt − µ) + (y2 − µ) (yt − µ) + . . . (yT − µ) (yt − µ)
T2 t=1 t=1 t=1
1
= E ((γ0 + γ1 + . . . + γT −1 ) + (γ1 + γ0 + . . . + γT −2 ) + . . . + (γT −1 + γT −2 + . . . + γ0 ))
T2
1
= (T γ0 + 2(T − 1)γ1 + 2(T − 2)γ2 + . . . + 2γT −1 )
T 2
1 T −1 T −2 1
= γ0 + 2γ1 + 2γ2 + . . . + 2γT −1 . (1)
T T T T
5
First we show that the variance of the mean goes to zero as T increases.
2
T − 1 T − 2 1
T E(y T − µ) = γ0 + 2γ1 + 2γ2 + . . . + 2γT −1
T T T
T −1 T −2 1
≤ |γ0 | + 2|γ1 | + 2|γ2 | + . . . + 2|γT −1 |
T T T
≤ |γ0 | + 2|γ1 | + 2|γ2 | + . . . + 2|γT −1 | .
Note that the rhs of the last expression is nite since we are assuming the the covariances
are absolutely summable, thus E(y t − µ)2 → 0. Next we have to show that
T −1 T −2 1
lim T E(y T − µ)2 = lim γ0 + 2γ1 + 2γ2 + . . . + 2γT −1
T →∞ T →∞ T T T
+∞
X
= γj .
j=−∞
This results is established in Hamilton (1994) in Section 7.2. The intuition of the result is
the following. Consider the sum of the autocovariances in (1) and let j denote the index
of the autocovariances. For large j the autocovariances are small and do not aect the
sum. For small j the autocovariances have a weight that approches 1 as T grows. Thus,
Part 3. The last claim of the theorem is a straightforard implication of Parts 1 and 2
P+∞
Var(y) j=−∞ γj
P(|y T − µ| ≥ ε) ≤ = .
ε2 T ε2
Thus P+∞
j=−∞ γj
lim P(|y T − µ| ≥ ε) ≤ lim =0,
T →∞ T →∞ T ε2
6
Theorem 1.3. Central Limit Theorem. Let yt be
∞
X
yt = µ + φi t−i
i=0
P∞
where t is a sequence of iid random variables with nite variance and i=0 |φi | < ∞.
Then,
a 1 2
yT ∼ N µ, σLR
T
where
∞
X
2
σLR = γk
k=−∞
7
2 Linear Time Series Models
Denition 2.1. Linear Time Series Process. The time series yt is said to be linear
if it can be written as
∞
X
yt = µ + ψi t−i
i=0
Proposition 2.1. Covariance Stationarity. Let {yt } be a linear time series process
∞
X
yt = µ + ψi t−i
i=0
P∞
where t is a white noise process with variance σ2 and ψ0 = 1. If i=0 ψi2 < ∞, then yt
is weakly stationary and it has the following properties
E(yt ) = µ
∞
X
Var(yt ) = σ2 ψi2
i=0
∞
X
Cov(yt , yt−k ) = σ2 ψi ψi−k
i=0
yt = c + t + θ1 t−1 t ∼ W N (σ2 ).
8
MA(1) process. Then the unconditional mean and unconditional variance are respectively
and
Var(c + t + θ1 t−1 ) = Var(t + θ1 t−1 ) = Var(t ) + θ12 Var(t−1 ) = σ2 + θ12 σ2 .
MA(1) process. Then the conditional mean and conditional variance given the information
and
9
the autocovariance and autocorrelation functions of the process for k ≥ 1 are
θσ 2 θ
k=1
1+θ2
k=1
γk = and ρk = .
otherwise otherwise
0
0
= θE(2t−1 ) = θσ2 ,
= 0.
where φi = −θi with |θ| < 1. Then, yt can equivalently be represented as the MA(1)
process
yt = φ0 (1 − θ1 ) + t − θ1 t−1 .
10
Proof of Proposition 2.5. Note that
yt−1 + θ1 yt−2 + θ12 yt−3 + θ13 yt−4 + θ14 yt−5 + ... = φ0 + t−1 .
If you multiply the second equation by θ1 and subtract it from the rst one you get
yt = φ0 + t − θ1 c − θ1 t−1 = φ0 (1 − θ1 ) + t − θ1 t−1 .
yt = c + φ1 yt−1 + t t ∼ W N (σ2 ).
denition 2.3 such that |φ1 | < 1. Then {yt } is a stationary linear time series process with
representation
∞
c X
yt = + φj t−j .
1 − φ1 j=0
11
Proof of proposition 2.6. We begin
yt = c + φ1 yt−1 + t
= c + φ1 (c + φ1 yt−2 + t−1 ) + t
∞
c X j
yt = + φ t−j
1 − φ1 j=0 1
AR(1) process. Then the unconditional mean and unconditional variance are respectively
c σ2
E(yt ) = and Var(yt ) = .
1 − φ1 1 − φ21
∞
!
c X j
E(yt ) = E + φ t−j
1 − φ1 j=0 1
∞
!
c X
= +E φj1 t−j
1 − φ1 j=0
∞
c X j
= + φ E (t−j )
1 − φ1 j=0 1
c
= ,
1 − φ1
12
and
∞
!
c X j
Var(yt ) = Var + φ t−j
1 − φ1 j=0 1
∞
!
X
= Var φj1 t−j
j=0
∞
!
X
= φ2j
1 σ2
j=0
σ2
= .
1 − φ21
process. Then the autocovariance and autocorrelation functions of the process for k > 0
are
σ 2 φk1
γk = and ρk = φk1 .
1 − φ21
γ1 = Cov(yt , yt−1 )
Next we nd an expression for the autocovariance of order k > 1 as a function of the
autocovariance of order k − 1
γk = Cov(yt , yt−k )
13
By applying recursively the last expression we get the autocovariance function formula
process as in denition 2.4 such that |φ1 | < 1. Then {yt } is a stationary linear time
∞
c X
yt = + (φ1 + θ1 ) φj t−j .
1 − φ1 j=0
yt = c + φ1 yt−1 + θ1 t−1 + t
14
As the stochastic process stretches back in time, k → ∞, one gets
∞
c X
yt = + t + (φ1 + θ1 ) φj−1
1 t−j
1 − φ1 j=1
stationary ARMA(1,1) process. Then the unconditional mean and unconditional variance
are respectively
c θ2 + 2θ1 φ1 + 1
E(yt ) = and Var(yt ) = σ2 1 .
1 − φ1 1 − φ21
= c + φ1 E(yt−1 )
= c + φ1 E(yt ) .
c
E(yt ) = .
1 − φ1
15
Solving with respect to Var(yt ) gives is
θ12 + 2θ1 φ1 + 1
Var(yt ) = σ2 .
1 − φ21
ARMA(1,1) process. Then the autocovariance function of the process for k > 0 is
(φ1 +θ1 )(1+φ1 θ1 )
1−φ21
k=1
γk = .
φ1k−1 γ1 otherwise
γ1 = Cov(yt , yt−1 )
Next we nd an expression for the autocovariance of order k > 1 as a function of the
autocovariance of order k − 1
γk = Cov(yt , yt−k )
By applying recursively the last expression we get the autocovariance function formula
16
2.3 Forecasting Under Square Loss
Proposition 2.12. Let yt be a stationary time series. Let ŷT +h|T denote the forecast of
Then, under the square loss, the optimal forecast of yT +h is its conditional mean given
The last equation follows from the fact that ET (yT +h − µT +h|T )(µT +h|T − ŷT +h|T ) = 0.
Notice that the ŷT +h|T only enters in the second term of the last expression. Thus,
arg min ET (yT +h − ŷT +h|T )2 = arg min ET (µT +h|T − ŷT +h|T )2 = µT +h|T .
ŷT +h|T ŷT +h|T
Proposition 2.13. Let yt be a MA(1) process. Then, the optimal h-step ahead forecast
and h-step ahead forecast error variance conditional on the information available at time
T are
σ 2
c + θ1 T
h=1
h=1
ŷT +h|T = and Var(eT +h|T ) = .
otherwise σ2 (1 + θ12 ) otherwise
c
and
VarT (eT +1|T ) = VarT (yT +1 − ŷT +1|T ) = VarT (T +1 ) = σ2 .
17
For h > 1 we have
and
VarT (eT +1|T ) = VarT (yT +h − ŷT +h|T ) = VarT (θ1 T +h−1 + T +h ) = (1 + θ12 )σ2 .
Proposition 2.14. Let yt be a stationary AR(1) process. Then, the optimal h-step ahead
forecast and h-step ahead forecast error variance conditional on the information available
at time T are
h
X
ŷT +h|T = µ + φh1 (yT − µ) and Var(eT +h|T ) = φ2i 2
1 σ
i=0
and
VarT (eT +1|T ) = VarT (yT +1 − ŷT +1|T ) = VarT (T +1 ) = σ2 .
For h = 2 we have
= µ + φ1 ET (yT +1 − µ)
= µ + φ21 (yT − µ)
18
and
VarT (eT +2|T ) = VarT (yT +2 − ŷT +2|T ) = VarT (φ1 (yT +1 − µ) + T +2 − φ21 (yT − µ))
= VarT (φ1 T +1 + T +2 )
= (1 + φ21 )σ2 .
19
3 Modelling the Conditional Variance
p
yt = σt2 zt zt ∼ iidD(0, 1)
2
σt2 = ω + αyt−1
process. Then the unconditional mean and unconditional variance are respectively
ω (1 − α2 )
E(yt ) = 0, Var(yt ) = and kurt(yt ) = µ4
1−α (1 − µ4 α2 )
where µ4 = E(zt4 ).
p p p
E(yt ) = E( σt2 zt ) = E(Et−1 ( σt2 zt )) = E( σt2 Et−1 (zt )) = 0
and
Var(yt ) = ω + αVar(yt−1 )
σ 2 = ω + ασ 2 .
Solving the last equation with respect to σ 2 gives the claim of the proposition.
20
Last note that
Note that
[E(σt2 )]2 = (σ 2 )2
and
E((σt2 )2 ) = E( (σ 2 + α(yt−1
2
− σ 2 ))2 )
= (σ 2 )2 + α2 E( (yt−1
2
− σ 2 )2 )
= (σ 2 )2 + α2 E(yt−1
4
) + α2 (σ 2 )2 − 2α2 E(yt−1
2
)σ 2
= (σ 2 )2 + α2 E(yt−1
4
) − α2 (σ 2 )2
= (σ 2 )2 + α2 µ4 E((σt−1
2
)2 ) − α2 (σ 2 )2
1 − α2
E((σt2 )2 ) = (σ 2 )2
1 − µ4 α 2
p p
Cov(yt , yt−k ) = E(yt yt−k ) = E(Et−1 ( σt2 zt yt−k )) = E( σt2 yt−k Et−1 (zt )) = 0 .
Proposition 3.3. ACF of yt2 of an ARCH(1). Let {yt } be an ARCH(1) process. Then
21
Proof of proposition 3.3. Note that
Cov(yt2 , yt−1
2
) = E((yt2 − σ 2 )(yt−1
2
− σ 2 ))
= E((σ 2 + α(yt−1
2
− σ 2 ))zt2 − σ 2 )(yt−1
2
− σ 2 ))
= E(σ 2 (yt−1
2
− σ 2 )(zt2 − 1)) + αE((yt−1
2
− σ 2 )(yt−1
2
− σ 2 )zt2 )
2
= αE((yt−1 − σ 2 )2 )
Next
Cov(yt2 , yt−k
2
) = E((yt2 − σ 2 )(yt−k
2
− σ 2 ))
= E((σ 2 + α(yt−1
2
− σ 2 ))zt2 − σ 2 )(yt−k
2
− σ 2 ))
= E(σ 2 (yt−1
2
− σ 2 )(zt2 − 1)) + αE((yt−1
2
− σ 2 )(yt−k
2
− σ 2 )zt2 )
2
= αE((yt−1 − σ 2 )(yt−k
2
− σ 2 ))
= αCov(yt2 , yt−k+1
2
)
= αγk−1
ARCH(1) process. Then the h-step ahead forecast of the conditional variance is
k−1
X
σT2 +h|T = ω αi + αk rT2 .
i=1
22
σT2 +2|T = ET (σT2 +2 ) = ET (ω + αrT2 +1 ) = ω + αET (rT2 +1 ) = ω + αET (σT2 +1 )
p
yt = σt2 zt zt ∼ iidD(0, 1)
2
σt2 = ω + αyt−1 2
+ βσt−1
ω
Var(yt ) = .
1−α−β
σ 2 = ω + ασ 2 + βσ 2 .
Solving the last equation with respect to σ 2 gives the claim of the proposition.
23
GARCH(1.1) process as in denition 3.2. Then
∞
ω X
σt2 = +α β j−1 rt−j
2
.
1−β j=1
2
σt2 = ω + αrt−1 2
+ βσt−1
2 2 2
= ω + αrt−1 + β(ω + αrt−2 + βσt−1 )
2 2 2
= ω(1 + β) + αrt−1 + αβrt−2 + αβσt−2
..
.
k−1
X k
X
j
= ω β +α β j−1 rt−j
2
+ β k σt−k
2
j=0 j=1
rt2 = ω + (α + β)rt−1
2
− βηt−1 + ηt
σt2 = ω + αrt−1
2 2
+ βσt−1
2
rt2 − rt2 + σt2 = ω + αrt−1 2
+ βσt−1 2
− βrt−1 2
+ βrt−1
rt2 = ω + (α + β)rt−1
2 2
− β(rt−1 2
− σt−1 ) + (rt2 − σt2 )
24
GARCH(1,1) process. Then the h-step ahead forecast of the conditional variance is
ω + αrT2 + βσT2
h=1
σT2 +h|T =
σ 2 + (α + β)h−1 (σT2 +1|T − σ 2 ) otherwise
25
4 Factor ARCH
Denition 4.1. Factor ARCH Process. Let (r1 t , ..., rn−1 t , rm t )0 be generated as
q
ri t = λi rm t + σi2t zi t zm t ∼ iidD(0, 1)
p
2
rm t = σm t zm t zm t ∼ iidD(0, 1)
given t − 1.
Proposition 4.1. Covariance Matrix of the Factor ARCH Process. Let (r1 t , ..., rn−1 t , rm t )0
q
Vart−1 (ri t ) = E(ri2t )
= Et−1 (λi rm t + σi t zi t )2 2
q
= Et−1 (λ2i rm
2
t + σ 2 2
z
it it + 2λ r
i mt σi2t zi t )
q p
2 2 2 2 2
= Et−1 (λi σm t zm t + σi t zi t + 2λi σi2t σm 2
t zm t zi t )
= λ2i σm
2 2
t + σi t .
q q
2 2
Covt−1 (ri t , rj t ) = E(ri t rj t ) = Et−1 (λi rm t + σi t zi t )(λj rm t + σj t zj t )
q q q q
2 2 2 2
= Et−1 (λi λj rm t + λ r
j mt σ z
it it + λ r
i mt σ z
jt jt + σ z
it it σj2 t zj t )
2
= λi λj σm t .
26