Beruflich Dokumente
Kultur Dokumente
ABSTRACT: In this paper, neural network Introduction Two simple examples are given to illus-
architecture is presented as one approach to trate control and optimization with neural
The purpose of this paper is to serve as an
the design and implementation of intelligent introduction to neural networks and to sug- network architecture. The first example de-
control systems. Neural networks can be velops a two-layer discrete-time neural net-
gest that neural network architecture can be
considered as massively parallel distributed used for design and application of intelligent work called a bidirectional associative mem-
processing systems with the potential for control systems [ 11. The neural network ap- ory, which can be used to increase the
ever-improving performance through dy- reliability of a control system through fault
proach is introduced by defining the problem
namical learning. The nomenclature and isolation. The second example uses a single-
of control as the mapping of measured sig-
characteristics of neural networks are out- nals for “change” into calculated controls layer continuous-time network to optimize
lined, and two simple examples are pre- for “action.” Neural networks represent analog-to-digital conversion.
sented to illustrate applications to control massively parallel distributed processing ca-
systems. pability with the potential for ever-improv- Motivation for Intelligent Control
ing performance through dynamical learn- One of the primary purposes of classical
ing. The development of fast architectures feedback control is to increase robustness for
An early version of this paper was presented at
makes implementation in real time feasible. a control system, i.e., increase the degree to
the 1987 IEEE International Conference on Neural
Networks, San Diego, California, June 21-24, Furthermore, neural networks represent the which the system performs when there is un-
1987. Behnam Bavarian is with the Electrical En- nonalgorithmic class of information process- certainty. Classical linear control provides
gineering Department, University of California, ing as opposed to current conventional al- robustness over a relatively small range of
Imine, CA 92717. gorithmic signal processing. uncertainty. Adaptive control techniques
cW
+ 21,k=I
yutput
d ( x ) dk(xk) (3)
y(k + 1) = f[WT(k)X(k)] Ik
t(x) = i:
- 1 = I a,(.&)d : ( x , ) b , ( x , )
Yl Y2
- i:
k= I w,kdk(xk)]2 (4)
Fig. 2 . Example of processing element Fig. 3. Topology of a two-layer network
characteristics with a hard limiter. for a BAM. Their paper then derives conditions on the
April 1988 5
functions and establishes convergence of tra- as Eq. (7) (where superscript T indicates A W T = (1 5 5 1)
jectories. Among the most notable condi- transpose), it follows readily that, for any
+(I 1 1 l)=L
tions required are symmetry in the weight update defined by Eq. (5) or (6), the change
matrix [Wv] and the requirement for d i ( x i ) in energy A E is not positive, and, hence, the L W = (4 -4 4 -4 4)
to be differentiable and monotonic nonde- system is stable for every W , i.e., every ma-
trix is bidirectionally stable [15]. -+(I 0 I O l)=A
creasing.
There are still many questions that remain Hence, (A, L ) , ( B , M), and (C, N) are fixed
unanswered. What is the capacity of a neural E = -yWX’ (7)
points for the BAM defined by Eqs. (5) and
network? That is, how many patterns can be The Hebbian outer-product rule can be (6) and the W matrix defined in Eq. (8).
stored in a network with n processing ele- used to encode the q association (Xi, inx) Furthermore, with vector ( A 6) (a slight +
ments? What should be done with the spu- a BAM by changing the binary representa- variation to the A feature vector) introduced
rious equilibrium points? These questions tion of the vectors to bipolar representation, to the BAM, there is convergence to the
are, in general, still open questions in re- i.e., changing 0’s to -1’s. Let ( A i , B,) be closest failure label L. For example,
search. In what follows, two specific ex- the desired bipolar representation, then the
amples of neural networks are presented to resulting weighting matrix W is given by 6=(0 1 0 0 0)
help clarify concepts. These are simple ex- A+6=(1 1 1 0 1)
amples with more elaborate techniques given W = BTA, + B:A2 + . . . + B i A , (8)
in the references. The problem of sensor or actuator failure
(A + 6 ) WT = ( 2 2 2 2)
isolation is of interest in improving the con- +(I 1 1 I)=L
Fault Isolation Mapping Example trol system reliability. Here a BAM is used
as an associative mapping from system fail- More on the error-correcting properties and
A bidirectional associative memory is used the capacity of the BAM is discussed in [ 151.
to illustrate fault isolation and control. The ure feature space to system failure label
space. Define three classes of failure by the It should be noted that learning in the BAM
BAM is a two-layer feedback network of in- is fixed, hence it lacks robustness in the sense
terconnected processing elements that recalls following feature vectors and the corre-
sponding label vectors: that a one-bit shift in one of the patterns may
stored associations ( X I , Y,) for i = I , result in convergence to a wrong association.
2 , . . . , q. Thus, input of the n-dimensional The BAM is an example of a mapping net-
vector X , should give output of the m-di- Feature Vectors Label Vectors work. The single-layer network of Hopfield
mensional vector Y,, and vice versa. Figure is an example of an optimizing network.
3 shows the topology of the BAM. The net- A=(10101) L =(1111)
work is connected by a fixed-weight matrix B=(10100) M = ( O l 1 0 )
C = ( 0 1 0 1 1 ) N =(1001) Hopfield Network Optimizing
W of dimension m X n. The network dy-
namics for the ith processing element in layer Example
Y is given by Eq. (5). To encode the association using the preced-
The Hopfield network [IO] is a single large
ing Hebbian law, these vectors are converted
autoassociative neural network where the
y,(k + I) = I if y:(k + I) > O to bipolar forms.
outputs of the processing elements are fed
= y(k) if y:(k + 1) = 0
back to all the elements via a crossbar mesh
Feature Vectors with resistive weights connecting them to the
=O if y:(k + I) <O A’ = (1 - 1 1 -1 1)
input. Figure 5 shows a schematic diagram
B’ = of the network. The excitation dynamics are
(1 - 1 1 - 1 -1)
(5) defined by the differential equation in (9),
C’ = ( - 1 1 -1 1 1)
n where the nonlinearity f is a sigmoid curve
Y:@) = c WllS
]=I
Label Vectors
given by Eq. (10).
=
The update can be done synchronously
when all the processing element updates oc-
cur at a clock cycle, or asynchronously when
only some subset is updated each time. De- Using the BAM equations (5) and (6), the
fining an energy function E for this network following example shows associative recall. Fig. 5. Hopfield crossbar network.