Beruflich Dokumente
Kultur Dokumente
Outline
Objective Fuzzy Control
Background, Technology & Typology
ANFIS:
as a Type III Fuzzy Control as a fuzzification of CART Characteristics Pros and Cons Opportunities Applications References
Copyright 2002 by Piero P. Bonissone 2
ANFIS Objective
To integrate the best features of Fuzzy Systems and Neural Networks:
From FS: Representation of prior knowledge into a set of constraints (network topology) to reduce the optimization search space From NN: Adaptation of backpropagation to structured network to automate FC parametric tuning
FC Types
Type I: RHS is a monotonic function Type II: RHS is a fuzzy set Type III: RHS is a (linear) function of state
4
FC Technology (Background)
Fuzzy KB representation
Scaling factors, Termsets, Rules
ANFIS Network
Inputs
IF-part
Rules + Norm
THEN-part
Output
1
& N
Input 1
x1
& N
& N
Output
Input 2
x2
& N
Layers:
Rules [ Differentiable T-norm - usually product ] Normalization [ Sum and arithmetic division ] Functions [ Linear regressions and multiplication with , i.e., normalized weights , ] Output [ Algebraic Sum ]
pj is the probability of class j in node c Gini(c) measure the amount of impurity (incorrect classification) in node c
Copyright 2002 by Piero P. Bonissone 9
CART Problems
Discontinuity Lack of locality (sign of coefficients)
10
x1
x1 a1 > x1
x2
x2 a2 > a2 a2 > a2
a1 x1
x2
a2 x2
a2 > x2 a2 x2
f1(x1,x2)
f2(x1,x2)
f3 (x1,x2)
Partition Tree
Rule Table
11
y3= f3(x11, . )
x1
x1 a1
x2
a1 > x1
x2
y1= f1(x11, . )
x2 a2
f1(x1,x2 )
f2 (x1,x2)
f3 (x1,x2)
(a )(x1)
1
X1
Let's assume two inputs: I1=(x11,x12), and I2=(x21,x22) such that: x11 = ( a1- ) x21 = ( a1+ ) x12 = x22 < a2 Then I1 is assigned f1(x11,x12) while I2 is assigned f3(x1,x2) 0
X1
x11 a1 x21
Copyright 2002 by Piero P. Bonissone 12
!Y w
j
Y =
Copyright 2002 by Piero P. Bonissone
j =1 n
!w
j =1
j
13
ANFIS Characteristics
Adaptive Neural Fuzzy Inference System (ANFIS)
Algorithm defined by J.-S. Roger Jang in 1992 Creates a fuzzy decision tree to classify the data into one of 2n (or pn) linear regression models to minimize the sum of squared errors (SSE):
SSE = ! e2 j
j
where:
ej is the error between the desired and the actual output p is the number of fuzzy partitions of each variable n is the number of input variables
Copyright 2002 by Piero P. Bonissone 14
ANFIS Architecture
Rule Set:
IF ( x 1 is A 1 ) AND (x 2 is B 1 ) THEN IF ( x 1 is A 2 ) AND (x 2 is B 2 ) THEN ...
x1 x2 x1 x2 A1 A2 B1 B2 1 1 2 2 N 1 N 2 x1 x2 4 1 f1 2 f2 5 y
f 1 = p 1 x 1 + q 1 x 2 + r1 f 2 = p 2 x 1 + q 2 x 2 + r2
Layers: 0
16
Fuzzy reasoning
A1 B1
w1
z=
A2 x y
B2
w1*y1+w2*y2 w1+w2
w2
x y
A2 B1 B2
w1
w1*y1 w2*y2
wi*yi
wi
Y
17
w2
O1,i = Ai (x1 )
O1,i = Bi 2 ( x 2 )
A ( x1 ) =
1 x1 c i 1+ ai
2b
18
A (x) =
1+
1 x ci ai
2b
0 -10
-5
10
19
O2 ,i = wi = Ai (x1 ) Bi (x2 )
(for i=1,2)
O 3 ,i
wi = wi = w1 + w 2
(for i=1,2)
O4,i = wi fi = wi ( pi x1 + qi x2 + ri )
Consequent parameters {pi, qi, ri} Node output: Evaluation of Right Hand Side Polynomials
Copyright 2002 by Piero P. Bonissone 22
!
i
wi fi =
!w f !w
i i i i
Note:
Output is linear in consequent parameters p,q,r:
= w1 w2 f1 + f2 w1 + w 2 w1 + w 2
= w 1 ( p 1 x 1 + q 1 x 2 + r1 ) + w
= w 1 x 1 p 1 + w 1 x 2 q 1 + w 1 r1 + w 2 x 1 p 2 + w 2 x 2 q 2 + w
( )
(p 2 x1
+ q 2 x 2 + r2
( )r
2
ANFIS Network
Inputs
IF-part
Rules + Norm
THEN-part
Output
1
& N
Input 1
x1
& N
& N
Output
Input 2
x2
& N
Layers:
24
L-Type # Nodes
Inputs Values Rules Normalize Lin. Funct. Sum n (pn) pn pn pn 1
# Param
0 3(pn)=|S1| 0 0 (n+1)pn=|S2| 0
25
{c {
c p 0 , c p 1 ,..., c p n } },..., {
n n n
}
26
Backward pass:
S2 is fixed and S1 is computed using a gradient descent algorithm (usually Back-propagation)
27
linear parameters
w1*y1
w1
x1 x2
A2 B1 B2
wi*yi
w4 w4*y4
wi
/
A1 B1 A2 B2
Y
x2
B2 B1
A1
A2 x1
Forward stroke Backward stroke MF param. (nonlinear) Coef. param. (linear) fixed least-squares steepest descent fixed
28
ANFIS Least Squares (LSE) Batch Algorithm LSE used in Forward Stroke:
Parameter Set: S = {S1U S 2}, and {S1 I S 2 = }
Output = F I , S , where I is the input vector
( )
( )
For given values of S1, using K training data, we can transform the above equation into B=AX, where X contains the elements of S2 This is solved by: (ATA)-1AT B=X* where (ATA)-1AT is the pseudo-inverse of A (if ATA is nonsingular) The LSE minimizes the error ||AX-B||2 by approximating X with X*
Copyright 2002 by Piero P. Bonissone 29
X * = Xk
Copyright 2002 by Piero P. Bonissone
30
ANFIS Back-propagation
Error measure Ek (for the kth (1kK) entry of the training data)
Ek =
where : N(L) = number
N ( L) i =1 2 ( d x ) ! i L ,i
d i = i th component x L , i = i th component
Ek
31
+E = i
where :
33
ANFIS Pros
Characteristics
Translates prior knowledge into network topology & initial fuzzy partitions Network's first three layers not fully connected (inputs-valuesrules)
Advantages
Faster convergency than typical feedforward NN
++
+++
34
ANFIS Cons
Characteristics
Translates prior knowledge into network topology & initial fuzzy partitions
Disadvantages
Surface oscillations around points (caused by high partition number)
--
P"
--
35
Advantages
+++
++
36
Disadvantages
Not possible to represent known monotonic relations
to
Error gradient calculation requires derivability of fuzzy partitions and T-norms used by FLC
--
37
ANFIS Opportunities
Changes to decrease ANFIS complexity
Use dont care values in rules (no connection between any node of value layer and rule layer) Use reduced subset of state vector in partition tree while evaluating linear functions on complete state
X = (X r U X ( nr ) )
Use heterogeneous partition granularity (different partitions pi for each state variable, instead of p)
# RULES = pi
i =1
Copyright 2002 by Piero P. Bonissone 38
39
ANFIS Applications at GE
Margoil Oil Thickness Estimator Voltage Instability Predictor (Smart Relay) Collateral Evaluation for Mortgage Approval Prediction of Time-to-Break for Paper Web
40
ANFIS References
ANFIS: Adaptive-Network-Based Fuzzy Inference System, J.S.R. Jang, IEEE Trans. Systems, Man, Cybernetics, 23(5/6):665-685, 1993. Neuro-Fuzzy Modeling and Control, J.S.R. Jang and C.-T. Sun, Proceedings of the IEEE, 83(3):378-406 Industrial Applications of Fuzzy Logic at General Electric, Bonissone, Badami, Chiang, Khedkar, Marcelle, Schutten, Proceedings of the IEEE, 83(3):450-465 The Fuzzy Logic Toolbox for use with MATLAB, J.S.R. Jang and N. Gulley, Natick, MA: The MathWorks Inc., 1995 Machine Learning, Neural and Statistical Classification Michie, Spiegelhart & Taylor (Eds.), NY: Ellis Horwood 1994 Classification and Regression Trees, Breiman, Friedman, Olshen & Stone, Monterey, CA: Wadsworth and Brooks, 1985
Copyright 2002 by Piero P. Bonissone 41