Sie sind auf Seite 1von 16

Soft Computing: Neural Networks

1
Neural Networks
(Review)
Piero P. Bonissone
GE Corporate Research & Development
Bonissone@crd.ge.com
(Adapted from Roger Jang)
Soft Computing: Neural Networks
2
Outline
Introduction
Classifications
Amount of supervision
Architectures
Output Types
Node Types
Connection Weights
Reflections
Soft Computing: Neural Networks
3
Neural Nets: Classification
Supervised Learning
Multilayer perceptrons
Radial basis function networks
Modular neural networks
LVQ (learning vector quantization)
Unsupervised Learning
Competitive learning networks
Kohonen self-organizing networks
ART (adaptive resonant theory)
Others
Hopfield networks
Soft Computing: Neural Networks
4
Supervised Neural Networks
Requirement:
known input-output relations
input pattern output
Soft Computing: Neural Networks
5
Perceptrons
-Rosenblatt: 1950s
-Input patterns represented is binary
-single layer network can be trained easily
-output o is computed by
where
w
i
is a (modifiable) weight
x
i
is the input signal
is some threshold
f(
.
) is the activation function
o f w x
i i
i
n

_
,


1
f x x ( ) sgn( )

'

1
0
if x > 0
otherwise
Soft Computing: Neural Networks
6
Single-Layer Perceptrons
Network architecture
x1
x2
x3
w1
w2
w3
w0
y = signum(wi xi + w0)
Soft Computing: Neural Networks
7
Single-Layer Perceptron
Example: Gender classification
h
v
w1
w2
w0
Network Arch.
y = signum(hw1+vw2+w0)
-1 if female
1 if male
=
y
Training data
h (hair length)
v

(
v
o
i
c
e

f
r
e
q
.
)
Soft Computing: Neural Networks
8
Perceptron
Learning:
select an input vector
if the response is incorrect, modify all weights
where
t
i
is a target output
is the learning rate
If there exists a set of weights, then the method is
guaranteed to converge
w t x
i i i

Soft Computing: Neural Networks
9
XOR
Minsky and Papert reported a severe shortcoming
of single layer perceptrons, the XOR problem
not linearly separable
which (together with a lack or proper training
techniques for multi-layer perceptrons) all but killed
interest in neural nets in the 70s and early 80s.
x1 x2 output
0 0 0
0 1 1
1 0 1
1 1 0
x
x O
O 1
1 0
0
Soft Computing: Neural Networks
10
ADALINE
Single layer network (proposed by Widrow and Hoff)
Output is a weighted linear combination of weights
The error is described as
(for pattern p)
where
t
p
is the target output
o
p
is the actual output
o w x w
i i
i
n

0
1
( )
E t o
p p p

2
Soft Computing: Neural Networks
11
ADALINE
To decrease the error, the derivative with
respect to the weights is taken
The delta rule is:
( )

p i p p i
w t o x
( )

E
w
t o x
p
i
p p i
2
Soft Computing: Neural Networks
12
Multi-Layer Perceptrons
-Recall the output
-and the squared error measure
which is approximated by
-and the activation function
then the learning rule for each node can be derived
using the chain rule...
o f w x
i i
i
n

_
,


1
( )
E t o
p p p

2
( )
p
k
p
k
p
k
k
n
E t o

2
1
( )
f x
e
x

+

1
1
Soft Computing: Neural Networks
13
Multilayer Perceptrons (MLPs)
Learning rule:
Steepest descent (Backprop)
Conjugate gradient method
All optim. methods using first derivative
Derivative-free optim.
Network architecture
x1
x2
y1
y2
hyperbolic tangent
or logistic function
Soft Computing: Neural Networks
14
Backpropagation
to propagate the error back through a multi-
layer perceptron.
w
E
w
ki
p
ki
p

Soft Computing: Neural Networks


15
Improvements
Momentum term
smoothes weight updating
can speed learning up
w E w
w prev
+
Soft Computing: Neural Networks
16
MLP Decision Boundaries
A B
B A
A
B
XOR Interwined General
1-layer: Half planes
A B
B A
A
B
2-layer: Convex
A B
B A
A
B
3-layer: Arbitrary

Das könnte Ihnen auch gefallen