Beruflich Dokumente
Kultur Dokumente
INTRODUCTION
TO
MACHİNE
LEARNİNG
3RD EDİTİON
ETHEM ALPAYDIN
© The MIT Press, 2014
alpaydin@boun.edu.tr
http://www.cmpe.boun.edu.tr/~ethem/i2ml3e
CHAPTER 5:
MULTİVARİATE METHODS
Multivariate Data
3
X11 X 12 X d1
2 2 2
X1 X 2 Xd
X
N N
X1 X 2 X d
N
Multivariate Parameters
4
t 1 i
Nt
x
Sample mean m : mi , i 1,..., d
N
x N t
mi x tj m j
t 1 i
Covariance matrix S : sij
N
sij
Correlatio n matrix R : rij
si s j
5
Estimation of Missing Values
6
x ~ N d μ, Σ
1 1
px exp x μ Σ x μ
T 1
2 Σ
d /2 1/ 2
2
Multivariate Normal Distribution
8
21 2 1 21
2
2
zi x i i / i
Bivariate Normal
9
10
Independent Inputs: Naive Bayes
11
If p (x | Ci ) ~ N ( μi , ∑i )
1 1
px|C i exp x μ T
Σ 1
x μ
i
2 Σ i 1/ 2 i i
d/2
2
Discriminant functions
gi x log px|C i log P C i
d 1 1
log 2 log Σ i x μ i T Σ i 1 x μ i log P C i
2 2 2
12
Estimation of Parameters
13
PˆC i
t i
r t
mi
t i x
r t t
t i
r t
r x m i x m i
t t t T
Si t i
t i
r t
1 1
gi x log S i x m i T S i 1 x m i log PˆC i
2 2
Different Si
Quadratic discriminant
1
2
1 T 1
gi x log S i x S i x 2xT S i 1m i m i T S i 1m i log PˆC i
2
xT Wi x w i T x wi 0
where
1
Wi S i 1
2
w i S i 1m i
1 T 1 1
wi 0 m i S i m i log S i log PˆC i
2 2
14
likelihoods
discriminant:
P (C1|x ) = 0.5
posterior for C1
15
Common Covariance Matrix S
16
1
gi x x mi T S1 x mi log P̂ Ci
2
which is a linear discriminant
gi x w i T x wi 0
where
1 T 1
w i S m i wi 0 m i S m i log PˆC i
1
2
Common Covariance Matrix S
17
Diagonal S
18
gi x j
log PˆC i
2 j 1 s j
variances may be
different
Diagonal S, equal variances
20
gi x log PˆC i
2s 2
2
*?
Model Selection
22
Estimated parameters
p̂ij
t j ri
x t t
t i
r t
Discrete Features
25
px |C i pijkjk
z
j 1 k 1
pˆijk
t jk ri
z t t
t i
r t
Multivariate Regression
26
r t
g x |w
t
0 , w1 ,...,wd
Multivariate linear model
w0 w1x1t w2 x 2t wd xdt
2
Multivariate polynomial model:
Define new higher-order variables
z1=x1, z2=x2, z3=x12, z4=x22, z5=x1x2
and use the linear model in this new z space
(basis functions, kernel trick: Chapter 13)