Beruflich Dokumente
Kultur Dokumente
L 2p
= 21 22
M
M
O
M
L
pp
p1 p2
Note that is a square p p matrix. Also let be a symmetrix matrix.
Variance-covariance matrices and correlation matrices are always symmetric.
Definition 4.1
p
Definition 4.2
p
Example
The characteristics equation of a second order (2 2)
matrix A can be written as:
Eigenvalues
Remark
Eigenvectors
Let be an eigenvalue of A. Then there exists
a vector v such that
Av v
The vector x is called an eigenvector of A associated
with the eigenvalue . Ordinarily we normalise v so that
it has length one, that is,
v v 1
T
Property of Eigenvectors
When two eigenvalues of (convariance or correlation matrix)
are not equal, their corresponding eigenvectors(vi , v j ) will be
orthogonal to one another, i.e.,
viT v j 0 when i j .
Thus,
2/ 13
3/ 13
Eigen/diagonal Decomposition
Let
be a square matrix
are eigenvalues of
Diagonal decomposition:
why/how
Thus SU=U .
And S=U U1.
... v p
...
Problem
2
What is the eigen decomposition of S=
1
1
2 ?
Example continued
1 / 2 1 / 2 1 0 1 / 2
Then, S=
1 / 2 1 / 2 0 3 1 / 2
U
1/ 2
1/ 2
(U-1= UT )
is a symmetric matrix:
S QQ
where Q is orthogonal:
Q-1= QT
Columns of Q are normalized eigenvectors
Columns are orthogonal.
definite matrix.
semidefinite matrix.
If a matrix is either positive definite or positive semidenfinite, the matrix is
defined to be a nonnegative
matrix.
Vector Subtraction
Expectation
The expected value of a random
variable X, denoted by E(X), is the longrun average value of the random variable
over many repeated trials or occurrences.
E ( X ) xf ( x)
x
E ( X ) xf ( x)dx
.5
Dawson Supply, Inc.
.4
RelativeFrequency
R
elativeFrequency
.4
.3
.2
.2
.1
.1
0
9
10
11
10
11
12
Mean=10.3
Which one would you choose?
13
14
15
Variance
The variance is the average of the squared
deviations (the difference between data value and the
mean). Formally, suppose X has mean E ( X ) . The
variance of X, is defined as:
Var ( X ) E[( X ) 2 ]
It follows that Var ( X ) 0
The variance provides a measure of the spread or
dispersion of the distribution around its mean.
The variance can be affected greatly by extreme
values.
Y
X
Var(X)=
, Var(Y)=
. The covariance of X
and Y, is defined
Cov ( X , Y )as
E[( X X )(Y Y )]
X2 Y2
If
,
, then Cov(X,Y) will be
finite.
Cov(X,Y) can be positive, negative, or zero.
Correlation
If 0 X2 0 , Y2
, the correlation
of X and Y, is defined as
( X ,Y )
Cov ( X , Y )
XY
Relationship between
independent and uncorrelated
r = -1
r = -.6
X
Y
r = +1
r=0
r = +.3
r=0
Linear Correlation
Linear relationships
Y
Curvilinear relationships
Y
X
Y
X
Y
Linear Correlation
Strong relationships
Y
Weak relationships
Y
X
Y
X
Y
Linear Correlation
No relationship
Y
X
Y
f(x)
1
e
2
where = mean
= standard deviation
2 2
f(x)
%
1
p 2
where - xi , i =
12
'
x x -1
% %
% %
2
squared generalized
distance from x to
1,,p.
~
2
21 22 L 2p
=
, =
M M O M
%
% M
p
p1 p2 L pp
f(x) =
%
12
'
x x -1
% %
% %
2
squared generalized
distance from x to
where - xi , i = 1, 2.
%
% 2
21
22
Example
Consider a bivariate normal distribution with
8
=
6
6 2
and
=
,
2 3
how is the plot given by the equation
1
12
'
x -1 x
%%
%%
2
or
( x - ) ' 1 ( x - ) c2 ?
c1
TheBivariateNormalDistribution
f(x,y)
f(x,y)
f(x,y)
1
2
1
2
y
1
2
ContourPlotsoftheBivariateNormalDistribution
y
1
2
1
2
1
2
ScatterPlotsofdatafromtheBivariateNormalDistribution
y
1
2
1
2
1
2
Properties of the
Multivariate Normal
Distribution
For any multivariate
normal random vector X
~
1.
The density
f(x)
%
1
p 2
=
%
12
'
x x -1
% %
% %
2
at
1
2.
The density
f(x)
%
1
p 2
12
'
x x -1
% %
% %
2
a'X =
%%
i=1
'
ai X i ~ N p a,
a 'a
%% %%%
a1i X i
i=1
a 2i X i
'
'
'
A X = i=1
~
N
A,A
A
q
%%
%% %%%
M
p
a
X
qi i
i=1
X1
qx1
___ , = ___
= ,
px1
2
X2
p-q x1
p-q x1
qx1
X =
px1
12
11
qx p-q
qxq
pxp
21
p-q xq
22
p-q x p-q
~
~
8.
1 11
% , %
2 0 22
%
%
sample size
n %
%
X (CLT)
,
2
n (
)
2
(x )
x ~ N , or
~ N 0 , 1
n
SE
2
SE
58