Beruflich Dokumente
Kultur Dokumente
Jian Li
Department of Electrical and Computer Engineering
University of Florida
Gainesville, FL 32611, USA
ck ej2kFo t
k=
ck
Fo
=
=
1
Tp
x(t)ej2kFo t dt,
Tp
1
.
Tp
x(t)
Ex.
/2
Tp
FT
ejo t 2( o )
x()
2F0
2 c0
2c1
0
0
Ex.
s(t) =
k=
(t kT )
s(t)
t
-2T
Ck =
1
T
-T
for all k
2T
S ()
2 /T
2 /T
Remark:
Discrete signals
x ()
x(t)
Ex:
x(t) s(t)
Remark:
2/
Aliasing Problem:
Ex.
10
Y ()
=
=
n=
Z
y(t)ejt dt
=
=
DT F T :
Y ()
x(nT )(t nT )
n=
x(nT )ejnT
n=
X
n=
11
x(nT )ejnT
Discrete-Time
Signal
y()
x(nT)
DTFT
nT
0
2
T
x(n)
DTFT
-2
OR
-1
DT F T
-1/2
1/2
X() = P
jn
n= x(n)e
P air :
R
x(n) = 1 X()ejn d
2
13
x(n)
DTFT
Ex.
2
x(n) = 1 PN 1 X(k)ej2k Nn ,
k=0
N
DF T P air :
X(k) = PN 1 x(n)ej2k Nn
n=0
14
X(k)
DFT
-10
10
-10
10
k
0
Effects of Zero-Padding:
X ()
x(n)
DTFT
x(n)
X(k)
DFT
5 points
1
n
x(n)
X(k)
DFT
10 points
3
0
8
9
Z-Transform
X(z) = P
n
n= x(n)z
R
x(n) = 1
X(z)z n1 dz
2j
X(z)
N
1
X
x(n)z n
n=0
j 2 k
N
z=e
Im
Re
17
ak y(n k) =
M
X
k=0
bk x(n k)
Impulse Response:
(n)
h(n) = y(n)|x(n)=(n)
PM
bk z k
H(z) = PNk=0 k .
k=0
18
ak z
N=0.
IIR Filter:
N>0.
Minimum Phase: All poles and zeroes of H(z) are inside the unit
circle.
19
n=
Let X() =
n=
|x(n)| <
x(n)ejn
R
2
1
P
n= |x(n)| = 2 S()d,
2
S() = |X()|
2
Let (k) =
(k)ejk
n=
x(n)x (n k).
k= n=
k=
"
x(n)ejn
n=
2
= |X()| = S().
#"
x(s)ejs
s=
21
22
r(0)
r(k)
r (k) r(0)
h
i
x(n)
x (n) x (n k)
= E
x(n k)
r2 (0) |r(k)| 0.
23
Covariance matrix:
r(0)
r(1)
r(m 2) r(m 1)
..
..
.
.
r (1)
r(0)
r(m 2)
..
..
..
..
..
R=
.
.
.
.
.
..
..
..
..
..
.
.
.
.
.
r (1)
r(0)
r (m 1) r (m 2)
It is easy to show that R is positive semidefinite.
R is also Toeplitz.
Since R = RH , R is Hermitian.
24
Eigendecomposition of R
R = UUH ,
where UH U = UUH = I
(U is unitary matrix whose columns are eigenvectors of R)
= diag(1 , ..., m ),
(i are the eigenvalues of R, real, and 0).
25
X
P () =
r(k)ejk
k=
Z
1
r(k) =
2
Or
P (f ) =
P ()ejk d
r(k)ej2f k
k=
r(k) =
1
2
P (f )ej2f k df
12
r(0) =
1
2
PSD
1 2
27
2
1
1 NX
.
P () = lim E
x(n)ejn
N
N
n=0
N
1
X
k=N +1
|k| |r(k)| = 0
x(n)
H ()
y(n)
Py () = Px ()|H()| .
Complex (DE) Modulation.
y(n) = x(n)ej0 n .
It is easy to show that
ry (k) = rx (k)ej0 k .
Py () = Px ( 0 ).
29
P
2
N
1
Periodogram = Pp () = N1 n=0 x(n)ejn .
Correlogram = Pc () =
r(k)ejk .
k=(N 1)
k 0,
k < 0,
r(k) =
1
N k
PN 1
r(k) = r (k)
31
i=k
x(i)x (i k)
x(i)
Ex.
i
0
r(0) =
1
3
P2
0 (1)(1)
= 1, (average of 3 points)
P2
1
r(1) = r(1) = 2 1 (1)(1) = 1, (average of 2 points)
P2
1
r(2) = r(2) = 1 2 (1)(1) = 1, (average of 1 point)
r(3) = r(3) = 0.
pc ()
r(k)
k
2 1 0
32
Remark:
r(k) is a bad estimate of r(k) for large k.
E[
r(k)] = r(k) (unbiased )
Proof:
"
1
E[
r(k)] = E
N k
N
1
X
i=k
x(i)x (i k)
N
1
X
1
=
r(k) = r(k)
N k
i=k
33
Remark:
N 1
1 X
E[x(i)x (i k)]
E[
r(k)] =
N
i=k
N 1
1 X
N k
r(k)
=
r(k) =
N
N
i=k
r(k), as N
(Asymptotically unbiased)
34
x(i)
Ex.
i
0
r(0) =
1
3
P2
0 (1)(1)
= 1.
P2
r(1) = r(1) = 13 1 (1)(1) = 23 .
P2
1
r(2) = r(2) = 3 2 (1)(1) = 13 .
r(k)
Pc ( )
DTFT
-2 -1 0 1 2
35
Remark:
With biased r(k), Pc () = Pp () 0,for all
E[
r(k)] 6= r(k)
E[
r(k)] r(k), as N Asymptotically unbiased.
r(0)
r(1)
r(N 1)
(1)
r
(0)
(N
2)
R=
..
..
..
..
.
.
.
.
r (N 1) r (N 2)
r(0)
is positive semidefinite.
with r(k) biased estimate. Then R
36
N
1
X
x(i)ej N ki
i=0
1
2
Pp (k) = |X(k)| .
N
37
Let
W
j 2
N
= e
X(k) =
N
1
X
, N = 2m
x(n)W kn
n=0
N
2
1
X
x(n)W kn +
n=0
N
1
X
x(n)W kn
n= N
2
N
2
1
X
Nk
N
W 2 W kn
x(n) + x n +
2
n=0
Note:
W
Nk
2
j 2
N
Nk
2
=e
= ejk
1,
even k
=
1, odd k
38
X(2p) = PN 1 x(n) + x(n + N ) W kn , k = 2p = 0, 2, ...
n=0
2
N
kn
P
N
2 1
X(2p + 1) =
x(n) x(n + ) W , k = 2p + 1,
n=0
which requires 2
N 2
2
complex multiplication
39
n
o
2
Variance = E (
a E[
a])
a
a as Number of measurements .
40
Ex. Measurement
y = a + e,
Where a is an unknown constant and e is N (0, 2 ).
Find a
from y ?
f(y|a)
Pdf of y:
y
a
41
5 = a ML
a
M L = y
V ar[
aM L ] = V ar[y] = 2
E[
aM L ] = E[y] = E[a + n] = a
42
Ex. y = a + e
Three independent measurements y1 , y2 , y3 are taken.
a
M L = ? Bias = ? Variance = ?
f (yi |a) =
1 e
2
(yi a)2
2 2
1
e
f (y1 , y2 , y3 |a) = 3i=1 2
f (y1 ,y2 ,y3 |a)
|a=aM L
a
(yi a)2
2 2
=0
1
a
M L = (y1 + y2 + y3 ).
3
1
E[
aM L ] = E (y1 + y2 + y3 ) = a.
3
V ar[
aM L ] =
=
1
V ar(y1 + y2 + y3 )
9
2
1 2
( + 2 + 2 ) =
.
9
3
43
M L = x
x
ML = x
M L =max (x1 , x2 ).
x1
44
x2
Cram
er - Rao Bound.
Let B(a) = E [
a(r)|a] a denote the bias of a
(r), where r is the
measurement.
Then
h
i
2
M SE = E (
a(r) a) |a
1+ a
B(a)]
[
.
2
E [ a ln f (r|a)] |a
45
Proof: B(a)
= E [
a(r) a|a]
Z
[
a(r) a] f (r|a)dr
=
Z
Z
[
a(r) a]
f (r|a)dr
B(a) =
f (r|a)dr
a
a
|
{z
}
=1
Z
1
1+
[
a(r) a] f (r|a) f (r|a)
B(a) =
dr
a
a
f (r|a)
a f (r|a)
ln f (r|a) =
But
a
f (r|a)
Z
B(a) =
ln f (r|a)dr
1+
[
a(r) a] f (r|a)
a
a
nR
h
i o2
p
p
[
a(r) a] f (r|a) a
ln f (r|a)
f (r|a) dr
2
= 1 + a B(a) .
46
Schwarz Inequality:
Z
12 Z
g1 (x)g2 (x)dx
g1 2 (x)dx
g2 2 (x)dx
21
2
1+
B(a)
[
a(r) a] f (r|a)dr
a
)
(Z
2
ln f (r|a) f (r|a)dr
.
a
|
{z
}
I(a)
a
(r) a = c a
ln f (r|a).
Efficient Estimate:
An estimate is efficient if
(a.) It is unbiased
o
n
2
(b.) It achieves the CR - bound, i.e, E [
a(r) a] |a = CRB.
Ex. r = a + e
1
ln f (r|a) = 2 2(r a)
a
2
1
= 2 (a r).
f (r|a) =
48
=0
ln f (r|a)
a
a=
aM L
a
M L = r
1
(a a
M L )
ln f (r|a) =
a
2
2
ln f (r|a) = a
M L a
a
h
i
2
E (
aM L a) a = CRB
a
M L efficient
E [
a ] = E [r] = a, unbiased
ML
1 2
1
1
ln f (r|a) a = E
I(a) = E
(a
r)
=
=
a
2
4
2
1
CRB =
aM L ].
= 2 = V ar[
I(a)
49
Remarks:
(1) If a
(r) is unbiased, V ar[
a(r) ] CRB.
(2) If an efficient estimate a
(r) exists, i.e,
ln f (r|a) = c[
a(r) a]. (c is independent of r.)
a
then
0=
M L (r) = a
(r).
ln f (r|a)|a=aM L (r) results in a
a
If an efficient estimate exists, it is a
M L .
N
1 X
=
ri .
N i=1
a
M L has Gaussian distribution as N .
51
Asymptotic Properties of a
M L (r1 , ..., rN )
(a) a
M L (r1 , ..., rN ) a (
aM L is a consistent estimate.)
N
(b) a
M L is asymptotically efficient.
(c) a
M L is aymptotically Gaussian.
Ex. r = g 1 (a) + e,
Let
e N (0, 2 ). a
M L =?
efficient ?
dg 1 (a)
1
1
ln f (r|a) = 2 r g (a)
|a=aM L = 0
a
da
a
M L = g(r) = g(bM L ).
Invariance property of ML estimator
If a = g(b) then a
M L = g(bM L ).
a
M L may not be efficient. a
M L is not efficient if g() is a
nonlinear function.
52
PROPERTIES OF PERIODOGRAM
Bias Analysis
When r(k) is a biased estimate,
i
h
i
E Pp () = E Pc () = E
h
k < 0,
N
1
X
k=(N 1)
r(k)ejk
N k
k 0, E [
r(k)] =
r(k),
N
N +k
N |k|
E [
r(k)] = E [r (k)] =
r (k) =
r(k),
N
N
N
1
i
h
X
|k|
1
r(k)ejk .
E Pp () =
N
k=(N 1)
53
k
-(N-1)
N-1
i
X
E Pp () =
[wB (k)r(k)] ejk
k=
Let
h
DT F T
wB (k) WB ()
i
E Pp () =
1
2
54
P ()WB ( )d.
i
E Pp () =
1
2
P ()WR ( )d .
DT F T
wR (k) WR ()
w R (k)
1
k
-(N-1)
P( )
N-1
E [ P()]
W ()
B,R
55
W B ()
Main lobe
Side lobes
2
N
(or
1
N
in Hz) .
2
N
cannot be
Remark:
The side lobes of WB () transfer power from high power
frequency bins to low power frequency bins leakage.
Smearing and leakage cause more problems to peaky P () than
to flat P ().
If P () = 2 , for all , E[Pp ()] = P ().
Bias of Pp () decreases as N . (asymptotically unbiased.)
57
Variance Analysis
We shall consider the case x(n) is zero-mean circularly symmetric
complex Gaussian white noise.
is equivalent to:
E [ Re(x(n))Re(x(k))] = 2 (n k).
2
E [Im(x(n))Im(x(k))] = 2 (n k).
E [Re(x(n))Im(x(k))] = 0.
Remark: The real and imaginary parts of x(n) are N (0, 2 ) and
independent of each other.
58
k=(N 1)
Pp ()
|k|
1
N
r(k)ejk = 2
r(k)ejk = 2
k=
i
= E Pp () .
59
i
E Pp (1 )Pp (2 )
N 1 N 1 N 1 N 1
1 X X X X
E
[x(k)x
(l)x(m)x
(n)]
N2
m=0 n=0
k=0 l=0
= 4 + 2
ej(1 2 )(kl)
N
k=0 l=0
2
N
1
4 X
= 4 + 2
ej(1 2 )k
N
k=0
(
)2
N
4
sin[(1 2 ) 2 ]
4
= + 2
N
sin (1 2 )
2
60
i
lim E Pp (1 )Pp (2 ) = P (1 )P (2 ) + P 2 (1 )(1 2 ).
N
nh
ih
io
lim E Pp (1 ) P (1 ) Pp (2 ) P (2 )
N
P 2 ( ), =
1
1
2
=
0,
1 6= 2 ( uncorrelated if 1 6= 2 )
X
h(k)x(n k),
y(n) =
k=0
h (n)
y(n)
61
REFINED METHODS
Decrease variance of P () by increasing bias or
decreasing resolution .
Blackman - Tukey (BT) Method
Remark: The r(k) used in Pc () is poor estimate for large lags k.
M <N :
M
1
X
PBT () =
w(k)
r(k)ejk ,
k=(M 1)
M
N
1
M.
M
fixed
0 .
63
DT F T
(N 1) n N 1
Then Y () 0 iff
y(0)
y(1) y(N 1)
0
y(1)
y(0) y(N 2) y(N 1)
..
..
.
.
y[(N 1)]
y(0)
y(1)
..
..
..
..
.
.
.
.
..
.
is positive semidefinite.
BT =
R
w(0)
r(0)
w(M 1)
r(M 1)
..
..
.
.
w[(M 1)]
r[(M 1)]
..
..
.
w(0)
r(0)
..
BT 0.
is positive semidefinite, i.e, R
65
w(0)
BT =
R
w[(M 1)]
..
.
r(0)
J
r[(N 1)]
..
.
w(M 1)
..
.
..
.
w(0)
..
r(N 1)
..
.
..
.
r(0)
..
Theorem:
If A 0 (positive semidefinite) B 0 then A
B 0.
67
Time-Bandwidth Product
Equivalent Time Width Ne :
PM 1
w(n)
n=(M 1)
Ne =
w(0)
Ex.
PM 1
k=(M 1) (1)
Ne =
= 2M 1.
1
w (n)
R
n
-(M-1)
M-1
68
Ex.
1 |n| , (M 1) n (M 1)
M
wB (n) =
0,
else
w (n)
B
Ne = M
n
0
-(M-1)
69
M-1
Equivalent Bandwidth e :
2e =
W ()d
W (0)
DT F T
M
1
X
w(n)ejn .
n=(M 1)
W (0) =
M
1
X
n=(M 1)
70
w(n)
Ne e =
PM 1
Ne e = 1
n=(M 1) w(n)
R
1
2 W ()d
W ()d
PM 1
2 n=(M 1)
w(n)
=1
Remark:
If a signal decays slowly in one domain, it is more concentrated in
the other domain.
Window shape determines the side lobe level relative to W (0).
71
Ex:
1
.
x(2n) X
2
2
DT F T
X()
x(n)
1
1/2
n
x(2n)
1/2
1/4
n
0
Variance of PBT ()
1
m .
N
10 .
Bartlett Method
} | {z
}
x(n):
}
| {z
| {z
x1 (n)
x2 (n)
xL (n)
1X
PB () =
Pl ().
L
l=1
Remark:
PB () 0, .
Welch Method:
xl (n) may overlap in the Welch method.
xl (n) may be windowed before computing Periodogram.
x1 (n)
x 2 (n)
x s (n)
75
2
M
1
X
1
w(n)xl (n)ejn
Pl () =
MP
n=0
S
X
1
PW () =
Pl ()
S
l=1
76
Daniell Method:
PD () =
P ()
1
2
R +
Pp ()d.
1 , [, ]
DT
F
T
w(n) in PBT () W () =
0, else .
The larger the , the lower the variance, but the poorer the
resolution.
77
Implementation of PD ()
Zero pad x(n) so that x(n) has N points, N >> N.
2
k,
k = 0, , N 1.
k+J
X
1
PD (k ) =
Pp (j ).
2J + 1
j=kJ
Pp ()
| {z }
PD (k )
78
PARAMETRIC METHODS
Parametric Modeling
Ex.
r(0) 12
e
P (f ) =
2f
f
f
2
1
, |f |
2
P(f)
Parsimony Principle:
Better estimates may be obtained by using an
appropriate data model with fewer unknowns.
Appropriate Data Model.
If data model wrong, P (f ) will always be biased.
Estimate
True PSD
f
80
Rational Spectra:
2
P () = 2 B()
A()
A() = 1 + a1 ej + + ap ejp
B() = 1 + b1 ej + + bq ejq .
Remark: We mostly consider real valued signals here.
a1 , , ap , b1 , , bq are real coefficients.
Any continuous PSD can be approximated arbitrarily close by a
rational PSD.
81
u(n)
H () = B ()
A ()
x(n)
Remark:
2
2 B()
Pxx () =
.
A()
82
B()
.
A()
p
X
k=1
ak x(n k) +
In Z-transform Form, z = ej
H(z) =
q
X
k=0
bk u(n k).
B(z)
,
A(z)
A(z) = 1 + a1 z 1 + + ap z p
B(z) = 1 + b1 z 1 + + bq z q
x(n)
z -1
x(n-1)
z 1 x(n) = x(n 1)
B(z)
u(n)
Then: x(n) =
A(z)
83
AR Model: AR(p)
MA Model: MA(q)
2
2 B()
.
P () =
A()
2
2 1
P () =
.
A()
2
P () = 2 |B()| .
84
Spectral Factorization:
H() =
B()
A()
2
2 B()B ()
2 B()
=
.
P () =
A()
A()A ()
A() = 1 + a1 ej + + ap ejp
= 1 + a1 ej + + ap ejp
1
1
1
= 1 + a1 + + ap p = A( )
z
z
z
P (z) =
B(z)B( )
2 A(z)A( 1z ) .
z
B(z)B ( z1 )
P (z) =
1
A(z)A z
2
85
Consider
1
B(z)B(
2
z)
P (z) =
1 .
A(z)A( z )
1
.
1
.
Re
86
Remark:
If poles of
1
A(z)
H(z) =
B(z)
A(z)
B(z)
A(z)
is BIBO stable.
is minimum
We chose H(z) so that both its zeroes and poles are inside unit
circle.
u(n)
H (z) =
B (z)
A (z)
x(n)
Stable and
Minimum Phase system
87
1 + 0.9z 1
H(z) =
=
1 + 0.8z 1
H(z)
=
=
=
ARMA(1,1)
1
1
(1 + 0.8z 1 ) (1+0.9z
1 )
1
(1 + 0.8z 1 )(1 0.9z 1 + 0.81z 2 + )
AR().
Remark:Let ARMA(p,q) =
B(z)
A(z)
1
C(z)
= AR().
B(z)
1
Since
=
B(z)C(z) = A(z)
A(z)
C(z)
1
q
1
1 + b1 z + + bq z
1 + c1 z +
1
p
= 1 + a1 z + + ap z
1
0 0
c1
1
0 0
..
.
.. ..
..
1
.
.
.
.. .. ..
b1 .
.
.
.
..
cp
2 =
.. .. ..
.
.
. 1
cp+1
a
.. p
.
..
0
.. .. ..
.
.
.
.
c1
bq
.
..
..
.
cp
..
.
89
()
cp+1
.
.
.
cp+q
cp
..
.
cp+q1
cp
..
.
cpq+1
..
..
.
.
..
.
cp
1
b1
..
.
bq
0
0
..
.
0
b
c
cpq+1
1
p+1
..
.
..
. = ..
.
cp
bq
cp+q
90
.()
a1 , , ap can be
1
A(z)
x(n) =
= 1 + 1 z 1 + 2 z 2 +
1
A(z) u(n)
= u(n) + 1 u(n 1) +
E [x(n)u(n)] = 2
E [x(n k)u(n)] = 0, k 1
h
i a
1
= u(n)
x(n) x(n 1) x(n p)
ap
91
k = 0,
h
i a1
=
E x(n) x(n) x(n 1) x(n p)
.
ap
h
i
92
a1
..
.
ap
2 . ()
2 .
k 1,
h
i
a1
=
E x(n k) x(n) x(n 1) x(n p)
.
..
ap
h
i
93
a1
..
.
ap
0. ()
0.
r(0)
r(1)
r(p)
r(0)
r(p + 1)
r(1)
..
..
..
.
.
.
r(p) r(p 1)
r(0)
Ra = r
1
a1
..
.
ap
2
0
..
.
0
r(p + 1)
a
r(1)
1
..
..
..
. = . .
.
ap
r(p 1)
r(0)
r(p)
r(0)
..
.
94
Remarks:
When we only have N samples, {r(k)} is not available. {
r(k)}
may be used to replace {r(k)} to obtain a
1 , , a
p .
This is the Yule - Walker Method.
R is a positive semidefinite matrix. R is positive definite unless
x(n) is a sum of less than p2 sinusoids.
R is Toeplitz.
Levinson - Durbin algorithm is used to solve for a efficiently
AR models are most frequently used in practice.
Estimation of AR parameters is a well-established topic.
95
Remarks:
If {
r(k)} is a positive definite sequence and if a1 , , ap are found
by solving Ra = r, then the roots of polynomial
1 + a1 z 1 + + ap z p are inside the unit circle.
The AR system thus obtained is BIBO stable
Biased estimate {
r(k)} should be used in YW-equation to obtain
a stable AR system:
96
a = r
or R
1
Gohberg - Semencul Formula for R1 or R
97
Rn+1
r(0)
r(1)
r(n)
r(0)
r(1)
,
= .
.
..
..
r(n) r(n 1)
r(0)
n = 1, 2, , p
a
n,1
.
Let n = ..
an,n
98
( real signal )
LDA solves
Rn+1 =
n
0
99
bn
bn1
=
, with b =
Let b
..
b1
Then if c = Ab
c = Ab
100
b1
..
.
bn
Proof:
Aij
ci
A=
= a|ij|
= cni+1 =
a0
a1
..
.
an1
n
X
a1
..
.
..
.
an1
..
..
.
.
..
.
a1
a1
a0
Ani+1,k bk
k=1
=
=
n
X
k=1
n
X
a|ni+1k| bk
m=1
a|mi| bnm+1 =
(Ab)
i
101
n
X
m=1
Am,ibm
(m = n k + 1)
Consider:
Rn+2
n =
..
Rn+1
. r(n + 1)
..
.
r(n)
..
..
.
.
..
.
r(1)
..
r(0)
r(n + 1) .
Let
rn =
Then n = r(n + 1) + n T rn .
102
r(1)
..
.
r(n)
n 0
0
n
Result:
n
Let kn+1 = . Then
n
n
+ kn+1
.
n+1 =
0
1
2
n+1 = n (1 kn+1
)
103
Proof:
Rn+2
1
n+1
= Rn+2
+
k
n+1
n
= 0 + kn+1 0
n
n
n + kn+1 n
n+1
= 0
=
0
n + kn+1 n
0
104
LDA: Initialization:
n=1:
R2 =
r(0) r(1)
r(1) r(0)
1 = r(1)
r(0)
1 = r(0)
r 2 (1)
r(0)
1
1
1
0
O(1) flops
For n = 1, 2, , p 1, do:
T
r(n+1)+ n
rn
=
n flops
n
2
n+1 = n (1 kn+1
)
O(1) flops
n
+ kn+1
. n flops
n+1 =
0
1
105
O(1) flops
k1 = 1
kn+1
Ex:
a1 = 0
.
0
1
a2
Straightforward Solution:
a
1
1 =
a2
=
2
(1 ) 1
2
2 = 1 2 .
=
0
106
LDA: Initialization:
r(1)
r(0)
1 =
k2
1 = r(0)
r 2 (1)
r(0)
k1 = 1 = .
= 1 2 .
r1 = ,
r(2) + 1T r1
=
1
2 + ()
=
=0
1 2
= 1 (1 k22 ) = (1 2 )(1 02 )
=
1 2 = 2
107
1
0
+ k2
+ 0
1
1
a1
a2
Properties of LDA:
|kn | < 1,
n = 1, 2, , p,
n = 1, 2, , p,
108
iff
Rn+1 > 0
1 an,1 an,n
..
..
.
.
0
1 a1,1
{z
UT
n+1
1
r(0) r(n)
an,1
..
..
.
r(n)
r(0)
{z
} an,n
|
|
}
Rn+1
..
0
1
r(0)
{z
Dn+1
109
()
1
..
a1,1
{z
Un+1
n=1:
r(0) r(1)
r(1) r(0)
1 a1,1
1
a1,1
r(0) r(1)
r(0)
110
r(1) r(0)
r(1)
1
0 r(0)
1 a1,1
.
1
a1,1
=
Since
Tk
UTk
r(0)
rTk
rk
Rk
r(0) + Tk rk
rTk + Tk Rk
UTk rk
Uk Rk
Rk+1
1
k
111
k
0
0
Uk
Uk
r(0)
rTk
rk
Rk
r(0) + rTk k = k
k
0
r(0) + Tk rk
rTk + Tk RTk
rk + Rk k = 0
rTk + Tk Rk
UTk rk
UTk Rk
= k
=0
1
Uk
0
k 0
= Dk+1
=
0 Dk
112
1
R
Since U1
n+1 n+1
proven !
1
T
= D1
Un+1
n+1 ,
1
T
R1
n+1 = Un+1 Dn+1 Un+1 .
12
Un+1 Dn+1
det(Rn+1 ) = n det(Rn )
Rn+1 > 0,
n = 1, 2, , p,
and k > 0,
k = 1, 2, , p.
113
iff
r(0) > 0
Recall
2
).
n+1 = n (1 kn+1
If Rn+1 > 0,
n > 0,
r(0) > 0,
n = 1, 2, , p,
n n+1
=
n
Since n n+1 < n ,
2
kn+1
2
<1
kn+1
If
|kn | < 1,
|kn+1 | < 1.
r(0) > 0,
= r(0) > 0,
0
n+1 = n (1 k 2
2
< 1.
kn+1
n+1 )
> 0,
114
n = 1, 2, , p 1
MA Signals:
x(n)
= B(z)u(n)
= u(n) + b1 u(n 1) + + bq u(n q)
r(k)
= E [x(n)x(n k)]
= E {[u(n) + + bq u(n q)]
[u(n k) + + bq u(n q k)]}
|k| > q :
|k| < q :
r(k)
qk
X
bl bl+k ,
l=0
r(k)
= 2
r(k) = 0
= r(k).
q>k0
q <k <0
real.
b0 = 1, b1 , , bq =
Pq
P () = k=q r(k)ejk .
115
A simple estimator is
q
X
r(k)ejk .
k=q
ARMA Signals:
(1 + a1 z 1 + + ap z p )x(n) = (1 + b1 z 1 + + bq z q )u(n).
Let us write x(n) as MA():
x(n) = u(n) + h1 u(n 1) + h2 u(n 2) +
E [x(n)u(n)] = 2 .
E [u(n)x(n k)] = 0, k 1
ARMA model can be written as
x(n)
h
i x(n 1)
h
= 1 b1
1 a1 ap
..
x(n p)
u(n)
i
u(n 1)
bq
..
u(n q)
k= 0:
1 a1
k = 1:
i r(1)
h
ap
.. = 1 b1
.
r(p)
1 a1
r(0)
ap
r(1)
r(0)
..
.
r(p 1)
= 1 b1
..
.
118
i
2 h1
bq
..
.
2 hq
bq
2 h1
..
.
2 hq1
k q+1
1 a1
r(k)
i
r(k + 1)
ap
..
r(k + p)
= 1 b1
bq
r((q + 1))
r(q)
r((q + 1) + p)
..
..
.
.
119
0
0
..
.
0
1
a1
..
.
ap
= 0.
= 0.
r(q + 1)
r(q)
r(q p + 1)
1
r(q + 1)
r(q p + 2) a1
r(q + 2)
..
.. = 0.
..
.
.
.
r(q + p) r(q + p 1)
r(q p + 1)
a1
..
.
.
r(q + 1)
..
..
.
a
p
r(q + p 1)
r(q)
r(q)
120
ap
r(q)
r(q + 1)
r(q + 2)
=
..
r(q + p)
Remarks:
(1) Replacing r(k) for r(k) above, we can solve for a
1 , , a
p .
(2) The matrix on the left side
is nonsingular under mild conditions.
is Toeplitz.
is NOT symmetric.
Levinson - type fast algorithms exist.
121
y(n) = (1 + b1 z 1 + + bq z q )u(n).
1
A(z)
x(n)
x(n)
A(z)
y(n)
1 2
Py ().
Px () =
A()
122
= E [y(n)y(n k)]
= E [A(z)x(n)A(z)x(n k)]
p
p
X
X
= E
ai x(n i)
aj x(n j k)
i=0
p X
p
X
i=0 j=0
j=0
ai aj r(k + j i).
= Pp Pp r(k + j i)
a
a
,
a
=
1, k = 0, 1, , q
k
i j
0
i=0
j=0
k = k .
123
Remarks:
Pq
jk
e
k
k=q
2
A()
The AR estimates a
1 , , a
p have reasonable accuracy if the
ARMA poles and zeroes are well inside the unit circle.
Very poor estimates a
1 , , a
p occur when ARMA poles and
zeroes are closely-spaced and nearby unit circle. (This is
narrowband signal case).
124
Ex: Consider
x(n) = cos(1 n + 1 ) + cos(2 n + 2 ),
where 1 and 2 are independent and uniformly distributed on
[0,2].
1
1
r(k) = cos(1 k) + cos(2 k).
2
2
cos 2k
cos 1k
r(q)
r(q p + 1)
..
..
.
.
..
r(q + p 1)
r(q)
.
..
..
p
.
.
r(q + M 1) r(q + M p)
126
(M > p)
r(q + 1)
..
.
r(q + p)
..
r(q + M )
Remarks:
The overdetermined linear equations may be solved with
Least Squares or Total Least Squares Methods.
M should be chosen based on the trade-off between information
contained in the large lags of r(k) and the accuracy of r(k).
Overdetermined YW -equation may also be obtained for AR
signals.
127
A=
If b =
If
b=
1
1
3
0
1
0
, x = 3.
, x =?
128
e = Ax b
so that eH e
is minimized.
e
1
e2
Let e =
..
.
em
e2
e1
Ex:
130
eH e
= (Ax b) (Ax b)
= xH AH Ax xH AH b bH Ax + bH b
h
iH
h
i
1
1
= x (AH A) AH b (AH A) x (AH A) AH b
i
h
1
+ bH b bH A(AH A) AH b
x = (AH A)
AH b
131
LS Solution
Illustration of LS solution:
Let
..
A = [a1 . a2 ].
x1
xLS =
x2
a1
b
x1 a 1
A x LS
a2
x2 a 2
132
Ex:
A=
xLS
1
0
,b =
1
1
xLS =?
= (AH A)1 AH b
1
h
i 1
h
= 1 0
1
0
AxLS
eLS = AxLS
1
0
b=
(1) =
1
0
133
1
1
1
1
0
1
Computational Aspects of LS
Solving Normal Equations
H
A A xLS = AH b.
This equation is called Normal equation.
Let
AH A = C,
CxLS = g,
AH b = g.
134
(1)
Cholesky Decomposition:
where
L=
C = LDLH ,
1
0 0
l21 1 0
(Lower Triangular Matrix )
..
..
.
.
ln1 ln2 1
d
0
1
..
D=
, di > 0.
.
0
dn
135
l21
..
.
ln1
ln2
y = DLH xLS .
y1
0
0 y2
=
..
..
.
.
y1 = g1
y2 = g2 l21 y1
..
yk = gk Pk1 lkj yj ,
j=1
136
yn
g1
..
.
gn
k = 3, , n.
1 l21
Remarks:
ln1
ln2
..
.
x =
n
xk =
yn
dn
yk
dk
x1
..
.
xn
y1
d1
= LH xLS = D1 y = ..
yn
dn
Pn
l
j=k+1 jk xj ,
k = n 1,
137
Ex.
3 3
4 4+
x1
x2
x1
x2
1
1
1
1
Ax = b
25 +
1
. 25
H
H
.
A A=
,A b =
25 + 25 + 2
1 + 2
138
49
1
+2
H
H
A b=
.
x= A A
49
139
T
z1
.... x = QAx = Qb = ,
z2
0
Ex.
3
4
3
4
x1
x2
1
1
1 3 4
Q=
.
5 4 3
1
5 5 + 5
x1
= 5 .
QAx = Qb gives
7
0 7
x
2
5
5
x =1
2
AxLS = b + eLS .
min.
bT LS ]||F =
minimum,
gij = (ij)th
element of G.
142
B].
VVH = VH V = I.
..
n+1
0
0
1 2 n+1 0, i are real
144
Let
1
V11
V=
V21
Remarks:
..
. V12 n
..
1
. V22
xT LS = V12 V22 1
ARMA Signals:
Two Stage Least Squares Method
Step 1: Approximate ARM A(p, q) with AR(L) for a large L.
2 , , a
L .
YW Equation may be used to estimate a
1 , a
u
(n)
= x(n) + a
1 x(n 1) + + a
L x(n L).
=
N
X
1
u
2 (n).
N L
n=L+1
146
x(n)
B(z)
A(z)
Let
x=
x(0)
x(1)
..
.
x(N 1)
147
u
=
u
(0)
u
(1)
..
.
u
(N 1)
H =
x(1)
x(0)
..
a1
a2
..
.
= ap .
b1
.
.
.
bq
u
(1)
x(p)
x(p + 1)
u
(0)
u
(q)
u
(q + 1)
x(N 2) x(N p 1) u
(N 2) u
(N q 1)
148
LS Solution
Remarks:
(real signals) .
x = H + u
1 T
T
)
= H H
H (x u
Remark: The difficult case for this method is when ARMA zeroes
are near unit circle.
149
n-2
n-3
n
n-1
x(n)
m
i=1
f
x (n)
f
a x(n-i)
i
f
e (n)
-
150
ef (n) = x(n) x
f (n).
h
i
2
f = E ef (n)
Goal: Minimize f
= E
2 i
e (n)
f
= E x(n) +
= rxx (0) +
m
X
m
X
i=1
!2
afi x(n i)
i=1
m
X
j=1
f
afi
=0
m X
m
X
i=1 j=1
rxx (i) +
m
X
j=1
151
afj rxx (j i) = 0.
rxx (0)
rxx (1)
rxx (m)
rxx (0)
rxx (m 1)
rxx (1)
..
f
a1
..
.
afm
m
p
152
f
0
..
.
0
n+3
n+1
n
n+2
n+4
x
b (n) =
m
X
i=1
eb (n) = x(n m) x
b (n m)
h
2 i
b
b
= E e (n)
.
153
To minimize b , we obtain
rxx (0)
..
.
rxx (1)
rxx (m)
1
.
..
rxx (0)
abm
for all i
154
b
0
..
.
0
m
X
i=1
afm,i x(n i)
m
X
i=1
= [x(n m) x(n m + 1)
= [x(n)
1
m
abm,i x(n m + i)
x(n)]
1
m
m
Recall LDA:
efm (n) =
m =
[x(n) x(n 1)
m1
0
+ km
m1
1
x(n m)]
m1
= [x(n) x(n 1)
+ km
m1
x(n m + 1)]
156
m1
m1
x(n m)]
1
Similarly,
157
k1
x(n)
f
e (n)
2
km
k2
k1
f
e (n)
m
k2
-1
km
-1
b
e (n)
1
b
e (n)
2
-1
b
e (n)
m
-1
1+ a z +
1
+ a
u(n)
-p
z
Whitening Filter
158
x
(n) =
p
X
k=1
ak x(n k).
r(1)
r(0)
r(p 1)
a
1
.
.
..
.
..
.
.
.
=
.
.
.
.
.
r(p)
a
p
r(p 1)
r(0)
159
p
X
k=1
ak x(n k) + u(n),
In matrix form,
x(p 1)
x(p)
..
.
x(N 2)
x(p)
x(p + 1)
..
x(N 1)
x(p 2) ..
x(p + 1) ..
..
n = 0, 1, , N 1
x(0)
x(1)
x(N p 1)
160
a1
a2
..
.
ap
u(p)
u(p + 1)
+
..
.
u(N 1)
x(p 1)
x(0)
..
x(N 2) x(N p 1)
to the overdetermined
a1
..
.
ap
x(p)
..
.
x(N 1)
Remarks:
The Covariance or Prony Method minimizes
"
#2
p
N
1
N
1
X
1 X 2
1 X
2
u
(n) =
x(n) +
a
k x(n k)
=
N p n=p
N p n=p
k=1
161
X
1 X
2
a
k x(n k)
x(n) +
=
N n=
k=1
where those x(n) that are NOT available are set to zero.
For large N , the YW and Prony methods yield similar results.
For small N , YW method gives poor performance. The Prony
p for small N . The Prony
method can give good estimates a
1 , , a
method gives exact estimates for x(n) =sum of sinusoids.
Since biased r(k) are used in YW method, the estimated poles
are inside unit circle. Prony method does not guarantee stability.
162
p
X
k=1
x(0)
x(1)
..
.
x(N p 1)
afk = abk = ak ,
x(1)
x(2)
..
.
k = 1, , p
x(2)
x(p)
x(3) x(p + 1)
x(N p)
x(N 1)
a1
a2
..
.
ap
x(p 1)
..
x(N 2)
x(1)
..
x(N p)
x(0)
..
.
x(N p 1)
x(p)
..
x(N 1)
a1
a2
..
.
ap
x(p)
..
.
x(N 1)
x(0)
..
.
x(N p 1)
Remarks: The F/B method does not guarantee poles inside the
unit circle. In Practice, the poles are usually inside the unit circle.
164
x(p 1)
x(0)
..
..
.
.
x(N 2) x(N p 1)
x (p)
x (1)
..
..
.
.
x (N p)
x (N 1)
165
a1
a2
..
.
ap
x(p)
..
.
x(N 1)
x (0)
..
.
x (N p 1)
Remarks on
2:
In YW method,
2 = r(0) +
p
X
a
k r(k).
k=1
In Prony Method,
Let
eLS
e(p)
..
.
e(N 1)
N
1
X
1
2
2 =
|e(n)|
N p n=p
166
In F/B Method,
Let
eLS
=
2(N p)
2
ef (p)
..
.
ef (N 1)
eb (0)
..
.
eb (N p 1)
)
(N 1
N
p1
X
X
ef (n)2 +
eb (n)2
n=p
n=0
167
Burg Method
Consider real data and real model. Recall LDA:
n
n
+ kn+1
n+1 =
0
1
Thus, if we know n and kn+1 , we can find n+1 .
Recall
()
ef (n) = ef (n) + k eb (n 1)
m m1
m
m1
eb (n) = eb (n 1) + km ef (n),
m
where
m1
m1
m1
X
k=1
m1
X
k=1
168
a
m1,k x(n k).
a
m1,k x(n m + 1 + k)
km
PN 1 f
em1 (n)
ebm1 (n1)
.
= PN 1 n=m
2
2
f
b
[em1 (n)] +[em1 (n1)]
n=m
Initialization
1
N
r(0) =
0 = r(0)
PN 1
n=0
169
x2 (n)
n = 1, 2, , N 1
n = 0, 1, , N 2.
()
For m = 1, 2, , p,
2
m = m1 (1 km
)
m1
m1
m =
, ( 1 = k1 ).
+ km
0
1
2.
Remarks: p =
Since a2 + b2 2ab,
km < 1,
170
* Calculate a1 , , ap , 2 .
* Obtain P ()
DT F T
* r(k) P ().
171
N
X
P (xi ) ln P (xi ),
i=1
P (xi ) = prob(x = xi )
For continuous random variable,
Z
H(x) =
f (x) ln f (x)dx.
f (x) = pdf of x.
172
x(0)
..
x=
.
x(N 1)
N (0, RN )
1
ln(detRN ).
2
N , we consider Entropy Rate:
HN =
Since HN
as
HN
h = lim
N N + 1
h is maximized with respect to r(p + 1), r(p + 2), .
Remark: For Gaussian case, we obtain Yule-Walker equations .... !
173
A(z)
real outputs
E[u(n)] = 0,
V ar[u(n)] = 2
E[u(i)u(j)] = 0, i 6= j,
2
f = f x(p), , x(N 1)| x(0), , x(p 1), a1 , , ap ,
2
f x(0), , x(p 1)|a1 , , ap ,
2
* Consider first f1 = f x(0), , x(p 1)|a1 , , ap ,
1
1 T 1
f1 =
exp x0 Rp x0 .
p
1
2
2
(2) det 2 (Rp )
x(0)
r(0)
r(p 1)
..
..
.
.
.
.
,
R
=
x0 =
.
.
p
.
.
.
x(p 1)
r(p 1)
r(0)
175
* Consider next
2
f2 = f x(p), , x(N 1)|x(0), , x(p 1), a1 , , ap ,
x(n) +
p
X
k=1
ak x(n k) = u(n)
176
u(p)
..
=
.
u(N 1)
x(p)
a1 1
0 0
x(p + 1)
a2 a1 1 0
..
.
..
..
.
.
x(N 1)
0 ap 1
a1 x(p 1) + + ap x(0)
a2 x(p 1) + + ap x(1)
..
ap x(p 1)
..
0
177
Let
u=
u(p)
..
.
u(N 1)
,x =
x(p)
..
.
x(N 1)
1
0
a1 1
J=
.
..
..
.
ap
det(J) = 1
178
0
..
f (u) =
1
(2 2 )
f2
N p
2
1 T
exp 2 u u
2
= f [u(x)] |det(J)|
= f [u(x)].
Let
x(p)
x(p 1)
x(0)
x(p)
x(1)
x(p + 1)
X=
..
179
a=
f2 =
1
a1
..
.
ap
u = Xa
1 T T
1
X X
a .
N p exp 2 2 a
(2 2 )
180
An Approximate ML Estimator
p ,
2 are found by maximizing f2 .
a
1 , , a
a
1 , , a
p are found by minimizing
aT XT X
a = uT u
u(p)
1
x(p)
x(0)
x(1)
a1 u(p + 1)
x(p + 1)
..
..
.. =
.
.
.
x(N 1) x(N p 1)
ap
u(N 1)
p
N
1
X
X
1
x(n) +
2 =
a
j x(n j) .
N p n=p
j=1
182
k1 , , kp ,
2 , P ()
183
FPE(k) =
N +k
N k k
AIC(k) = N ln k + 2k .
Remarks:
As N , AICs probability of error in choosing correct order
does NOT 0.
As N , AIC tends to overestimate model order.
184
MDL(k) = N ln k + k ln N .
minimizes
k
1 X 1
1
CAT(k) =
,
N i=1 i
k
i =
N
i
N i
Noisy AR Processes:
y(n) = x(n) + w(n)
x(n) = AR(p) process.
2
w(n) = White Gaussian noise with zero-mean and variance w
= Pxx () + Pww ()
=
=
2
2
|A()|
2
+ w
2
2
2 + w
|A()|
2
|A()|
186
2
2 , w
may be estimated by
* ARMA methods.
* A large order AR approximation.
* Compensating the effect of w(n).
* Bootstrap or adaptive filtering and AR methods.
187
e(n)
x(n)
188
2.) m = 0,
filtering problem
3.) m < 0,
smoothing problem.
189
e(n)
hk z k .
k=
hk z k
k=0
p
X
k=0
190
hk z k
=E
("
d(n)
k=
= rdd (0)
+
o
n
2
E = E |e(n)|
#"
hk y(n k)
l=
d(n)
hl rdy (l)
k= l=
l=
# )
hl y(n l)
hk rdy
(k)
k=
ryy (l k)hk hl
Remark: For Causal and FIR filters, only limits of sums differ.
Let
E
= 0,
i
hi = i + ji
rdy (i) =
k=
191
E
= 0.
i
i
In Z - domain
Pdy (z) = H o (z)Pyy (z)
which is the optimum Non-causal Wiener Filter.
Ex : d(n)
Pxx (z)
Pww (z)
x(n)
= x(n),
and w(n)
are uncorrelated.
Optimal filter ?
Pyy (z)
= rxx (k).
Pdy (z)
= Pxx (z)
Pyy (z)
o
H (z) =
Pdy (z)
=
ho (k) =
h
0.36
1.6 (1 0.5z 1 ) (1 0.5z)
|k|
1
0.3
2
(k)
0.3
k
0
193
hk z k
k=0
Split H(z) as
194
B(z)B
1
z
195
1
Pyy (z)
1
z
=1
rd (i) = gi , i = 0, 1, 2,
196
hi = gi bi .
Note that
rd (i) =
=
E {d(n + i) (n)}
(
"
# )
X
E d(n + i)
bk y(n k)
k=0
bk rdy (i + k).
k=0
bk rdy (i + k).
1
z
rd (i) = gi , for i = 0, 1, , ON LY .
197
Let
[X(z)]+ =
"
xk z k
k=
G (z) =
=
+
xk z k .
k=
gk z k
k=
G (z) = Pdy (z)B
1
z
H (z) = B(z) Pdy (z)B
198
1
z
x(n) + w(n)
H(z)
e(n)
Causal
Pdy (z)
1 1 0.8z 1
199
Pdy (z)B
1
z
=
=
=
1
Pdy (z)B
z +
0.36
1 1 0.8z
.
1
1.6 (1 0.8z ) (1 0.5z)
5
5
0.36
3
6z
+
1 0.5z
1.6 1 0.8z 1
5
0.36
o
3
=
G
(z)
1
1
0.8z
1.6
200
5
0.36
1
1 1 0.8z 1
o
3
H (z) =
= 0.375
.
1
1
1
1
0.8z
1
0.5z
1
0.5z
1.6
1.6
h (k) =
201
3 1 k
U (k),
8 2
k = 0, 1, 2, .
p
X
hk z k
k=0
p
X
k=0
rdy (0)
rdy (1)
..
rdy (p)
ryy (0)
ryy (1)
ryy (p)
ryy (0)
ryy (1)
=
..
..
.
.
ryy (p) ryy (p 1)
ryy (0)
ho0
ho1
..
.
hop
= x(n) + w(n)
=
K
X
k ej(k n+k )
k=1
uniform distribution on
k
= angular frequencies
w(n)
[, ]
203
Remarks:
Applications: Radar, Communications, .
We are mostly interested in estimating 1 , , K .
Once 1 , , K are estimated,
1, ,
K ,
found readily from
1, ,
K
Let
k ejk = k
1
y(0)
ej 1
y(1)
..
..
.
.
y(N 1)
ej(N 1)1
1
ej 2
1 , , K can be
1
ej K
..
.
ej(N 1)K
204
1
2
..
.
k
Remarks:
ryy (k)
= E {y (n)y(n + k)}
K
X
i2 eji k + 2 (k)
i=1
Pyy ()
2
i=1
K
X
2
2
i2 ( i ) + 2 .
2
3
1
N
|w(n)|
1
exp
2
2
Remark: The real and imaginary parts of w(n) are real Gaussian
2
random variables with zero-mean and variance 2 .
The two parts are independent of each other.
206
1
N
( 2 )
( P
)
N 1
2
n=0 |w(n)|
exp
2
,
1
N
( 2 )
y(N 1) is
( P
)
N 1
2
|y(n) x(n)|
exp n=0
2
k=1
207
y(0)
1
1
..
..
..
Let y =
,
=
,
.
.
.
y(N 1)
K
K
B=
ej1
..
.
ej2
ej(N 1)1
208
1
ejK
..
.
ej(N 1)K
(y B) (y B) .
h
h
iH
i
1
1
= BH B
BH B BH B
BH y
BH y
1 H
H
H
H
B y.
+ y yy B B B
=
1
argmax yH B BH B
BH y .
1
= BH B
.
BH y
is a consistent estimate of
Remarks:
209
For large N ,
h
) (
)
E (
6 2
N3
1
21
..
.
1
2K
CRB
However,
is difficult to implement.
The maximization to obtain
* The search may not find global maximum.
* Computationally expensive.
210
Special Cases:
1.) K = 1
h
= argmax yH B BH B
|
{z
B=
g1
ej
..
.
ej(N 1)
211
1
BH y ,
}
H
, B B = N.
BH y
1 ej
N
1
X
ej(N 1)
y(0)
..
.
y(N 1)
y(n)ejn
n0
2
N
1
1 X
argmax
y(n)ejn
N
n=0
212
2.)
2
.
N
1
Since V ar (
k k ) 3
N
1
k k 3 .
N2
2
.
i
k | >
infi6=k |
N
We can resolve all K sine waves by evaluating g1 at FFT points:
= infi6=k |i k | >
2
i,
i =
N
Any K
of these
i = 0, , N 1
BH B = N I, I = Identity matrix.
2
K
N
1
X 1 X
j
k n
g1 =
y(n)e
.
N
gives
k=1
n=0
213
2
N
The K
i that maximizes g1 correspond to the
largest K peaks of the Periodogram.
Remarks:
k estimates obtained by using the K largest peaks of
Periodogram have accuracy
k k 2
N
The periodogram is a good frequency estimator.
introduced by Schuster a century ago !)
214
(This was
215
jk 1
is an Annihilating filter for xk (n).
1e z
K
jk 1
Let A(z) = k=1 1 e z
A(z)x(n) = 0
y(n) = x(n) + w(n)
A(z)y(n) = A(z)w(n)
216
()
Remark:
It is tempting to cancel A(z) from both sides above, but this is
wrong since y(n) 6= w(n) !
1+a
1 z 1 + + a
L z L = A(z)A(z)
1
a
a
L
where
217
we get
ryy (L)
..
.
ryy (L + 1)
..
.
ryy (L + M )
ryy (1)
ryy (L + M 1) ryy (M )
y (n L M )
ryy (1)
= 0.
.
..
ryy (M )
a
L
a
1
ryy (L + 1)
..
..
. =
.
a
L
ryy (L + M )
a =
218
y (n L 1)
..
.
Remarks:
When y(0), , y(N 1) are the only data available, we first
estimate ryy (i) and replace ryy (i) in above equation with estimate
ryy (i)
{
K } are the angular positions of the K roots nearest the unit
circle
Increasing L and M will
* give better performance due to using the information in
higher lags of r(i)
Increasing L and M too much will
* give worse performance due to increased variance in r(i) for
large i
219
y(n)
y(n
1)
i (n) =
Proof: Let y
..
y(n i + 1)
(n) =
x
x1 (n)
..
.
xK (n)
i (n) =
w
w(n)
w(n 1)
..
.
w(n i + 1)
220
i (n) =
y
ej1
..
.
ej2
ej(i1)1
1
ejK
ej(i1)K
{z
Ai
Ai = i K Vandermonde matrix.
rank(Ai ) = K if i K and k 6= l for k 6= l.
i (n) = Ai x
(n) + w
i (n)
y
221
x
i (n)
(n) + w
y(n L 1)
..
Thus = E
.
y(n L M )
y (n 1) y (n L)
(n L 1)
x (n
= E AM x
where
PL+1
= AM PL+1 AH
L,
H
(n L)
x (n)
= E x
222
1)AH
L
E {xi (n)}
o
n
= E i ej(i n+i )
Z
1
=
i eji n eji di = 0
2
n
o
j[i (nk)+i ]
j(i n+i )
i e
= E i e
= i2 eji k
E xi (n k)xj (n) = 0, i 6= j
223
PL+1
x1 (n L 1)
x2 (n L 1)
h
x (n 1) x (n 1)
= E
.
1
xK (n L 1)
0
12 ej1 L
..
.
.
.
.
=
.
.
.
2 jK L
0
K
e
224
Consider
ryy (L)
ryy (1)
..
ryy (L + M 1) ryy (M )
=
Remarks: rank ()
r (L + 1)
1
yy
..
..
.
.
a
L
ryy (L + M )
min(M, L)
225
[U1
U2 ]
V1H
V2H
K
LK
(Diagonal
denote the singular
value
decomposition
(SVD)
of
.
1 0
= V1 1
U
a
1
1
226
()
Remark:
K to replace gives better frequency estimation.
Using
K is closer to
This result may be explained by the fact that
than .
The rank approximation step is referred as noise cleaning .
227
with (**)
Step 2: Compute the SVD of
Step 3: Compute the roots of
1+a
1 z 1 + + a
L z L = 0
Pick the K roots that are nearest the unit circle and obtain the
frequency estimates as the angular positions (phases) of these roots.
Remarks: Rule of Thumb for selecting L and M :
LM
N
L+M
3
Although one cannot guarantee that the K roots nearest the unit
circle give the best frequency estimates, empirical evidence shows
that this is true most often .
228
||x||22 = xH x.
Proof:
||Ub||22 = bH UH Ub = bH b = ||b||2 .
Consider Ax b,
where
is
M L,
x is
L 1,
b is
M 1,
229
is of rank K
SVD of A:
A = UV
U1
U2
V1H
V2H
= ||UH Ax UH b||22
= ||y UH b||22
2
H
1 0
y1
U1 b
y2
b
0 0
UH
2
2
2
H
2
= ||1 y1 UH
1 b||2 + ||U2 b||2
230
H
y1 = 1
1 U1 b .
y1
H
V x=y=
0
i y
h
1
= V1 y1
x = Vy = V1 V2
0
H
x = V1 1
U
1 b.
1
||x||22 = ||y||22 =
231
minimum
Recall:
= (1 + a
1 z 1 + + a
L z L )w(n). (L K)
At not too low SNR,
h
i
y(L)
y(L 1)
y(0)
y(L)
y(1)
y(L + 1)
..
a
1
.. 0
.
a
L
1
a
.. 0
.
a
L
()
x(L 1)
x(0)
..
X=
.
x(N 2) x(N L 1)
Note
x(0)
..
.
x(N L 1)
ej1
..
.
ej(N L1)1
233
1
ejK
..
.
ej(N L1)K
.
..
x(1)
..
.
x(N L)
X=
..
ej1
..
.
..
..
.
ejK
ej(N L1)1
..
j1
e
1
..
ejK
K
ej(N L1)K
1
1
ej1
ejK
..
..
.
.
0
j(N L1)1
j(N L1)K
e
e
ej(L1)1 ej1 1
j(L1)2
ej2 1
e
..
..
..
.
.
.
ej(L1)K
ejK
234
0
..
.
K
y(L 1)
y(0)
..
..
.
.
y(N 2) y(N L 1)
{z
|
Y
y(L)
a
1
..
..
.
.
y(N 1)
a
L
}
{z
|
y
h
i
V1H
K
0
1
Let Y = U1 U2
V2H
LK
0 2
235
a
1
..
.
L
a
H
U
= V1 1
1
1
y(L + 1)
..
.
y(N 1)
()
a with ()
Step 3. Compute the roots from
M (n) =
y
AM
y(n)
y(n 1)
..
.
y(n M + 1)
ej1
..
.
ej(M 1)1
237
ejK
..
.
ej(M 1)K
(n) =
x
x1 (n)
..
.
xK (n)
M (n) =
w
w(n)
..
.
w(n M + 1)
M (n) = AM x
(n) + w
M (n)
y
Let
R = E
H
M (n)
y
yM
(n)
(n)AH
M
(n)
x
= E AM x
H
M (n)w
M (n)
+ E w
238
2
R = AM PAH
M + I,
0
12
..
P=
.
.
2
0
K
If M K, AM PAH
M has K positive eigenvalues and M K zero
eigenvalues. We shall consider M K below.
Let the positive eigenvalues of AM PAH
M be denoted
1
2
K
=
k + 2 , k = 1, , K.
k
Two groups
k = 2 , k = K + 1, , M
239
2
0
..
RG = G
= 2 G
.
0 2
H
2
RG =
AM PAM + I G
=
2
AM PAH
MG + G
AM PAH
MG = 0
240
AH
MG = 0
Remark:
Let the linearly independent K columns of AM define
K -dimensional signal subspace
* Then the eigenvectors of R that correspond to the M K
smallest eigenvalues are orthogonal to the signal subspace.
* The eigenvectors of R that correspond to the K largest
eigenvalues of R span the same signal subspace as AM .
AM = SC for a K K
non-singular C.
241
MUSIC:
K
ej
aM () =
..
ej(M 1)
Steps in MUSIC:
PN
1
H
M (n)
Step 1: Compute R = N n=M y
yM
(n), and its
eigendecomposition.
whose columns are the eigenvectors of R
that correspond
Form G
242
aM ()GG aM ()
[, ]
243
aM (z)|z=ej = aM ()
244
ESPRIT Method
(Estimation of Signal Parameters by Rotational Invariance
Techniques )
ej1
ejK
AM =
..
ej(M 1)1
ej(M 1)K
B2 = last M 1 rows of AM .
B2 D = B1 ,
D=
ej1
0
..
.
ejK
0
245
S = B C = B DC.
1
1
2
S2 = B2 C
S2 C1 = B2
S1 = S2 C1 DC = S2 .
1
H
S2 S2
=
SH
2 S1 .
The diagonal elements of D are the eigenvalues of .
1
H S
H S
= S
S
Steps of ESPRIT: Step 1:
2 2
2 1
eigenvalues of .
246
Remarks:
1
2 S
S
can also be solved with Total Least Squares Method
Since is K K matrix, we do not need to pick K roots nearest
the unit circle, which could be wrong roots.
ESPRIT does not require the search over parameter space, as
required by Spectral MUSIC.
All of these remarks make ESPRIT a recommended method !
247
K
X
k ejk n + e(n)
k=1
Unknown frequencies.
e(n)
2
K
X
jk n
k e
y(n)
N
1
X
n=0
k=1
248
Remarks:
k ,
k and
When e(n) is zero mean Gaussian white noise, this NLS method
is the ML method.
When e(n) is non-white noise, NLS method gives asymptotically
(N ) statistically efficient estimates of
k and k despite the
fact that NLS is not an ML method for this case.
The non-linear minimization is a difficult problem.
249
Remarks:
Concentrating out {k } gives
h
i
1
= argmax yH B BH B
BH y
= B B
1
B y
H
250
Preparation:
Let
yk (n) = y(n)
K
X
i ej i n
i=1,i6=k
* i and
i , i 6= k, are assumed given, known, or estimated.
Let
gk =
N
1
X
n=0
* Minimizing gk gives:
yk (n) k ejk n 2 .
2
1
NX
k = argmaxk
yk (n)ejk n .
n=0
N
1
X
1
j
n
k =
.
yk (n)e k
N
n=0
k =
k
251
Remarks:
N
1
X
yk (n)ejk n
n=0
252
2 from y2 (n)
and
Obtain
and reestimate
1 and 1 from y1 (n)
Step 3: Assume K = 3.
1 ,1 ,
2 , 2 . Obtain
3 and 3 from y3 (n) .
Obtain y3 (n) from
2 , 2 ,
3 , 3 . Reestimate
1 and 1 from y1 (n).
Obtain y1 (n) from
Obtain y2 (n) from
1 , 1 ,
3 , 3 . Reestimate
2 and 2 from y2 (n).
Iterate until g does not decrease significantly anymore !
253
Step 4: Assume K = 4,
..
.
Continue until K is large enough!
Remark:
RELAX is found to perform better than existing
254