Sie sind auf Seite 1von 5

&VSPQFBO$POUSPM$POGFSFODF

&$$

"VHVTUo4FQUFNCFS ,BSMTSVIF (FSNBOZ

$1HXUDO3UHGLFWLRQ0RGHOIRU0RQLWRULQJDQG)DXOW'LDJQRVLVRI
D3ODVWLF,QMHFWLRQ0RXOGLQJ3URFHVV
1&RVWD%5LEHLUR

CISUC – Centro de Informática e Sistemas da Universidade de Coimbra


Department of Engineering Informatics, Coimbra University, Portugal
Fax: +351 39 701266 Telephone: +351 39 790000
e-mail: nfilipe@eden.dei.uc.pt, bribeiro@eden.dei.uc.pt

.H\ZRUGV Industrial Intelligent Control, Principal Artificial neural networks (ANN), because they are cost-
Component Analysis, Neural Networks, Prediction. effective, easy to understand and because of their ability to
learn from examples, have found many applications in
$EVWUDFW process modeling and control, as intelligent sensors, to
estimate variables that usually cannot be measured on-line, in
In engineering systems, early detection of the occurrence of dynamic system identification, in fault detection and
faults is critical in avoiding product defects. This problematic diagnosis and, finally, in process control.
is here discussed in the framework of an industrial process, This paper addresses the use of artificial neural networks
namely, an injection moulding plastic machine. The in process modeling. Namely, in a plastics injection moulding
relationships between the process state and the product industrial process. In section 2, a description of the process is
quality are achieved through Principal Component Analysis. given. In section 3, neural data prediction is discussed.
After having identified the main variables, two neural Following the need to understand the process behaviour and
network architectures were investigated, TDNN and Elman how it responds to changes in the environment, the data
networks, with respect to one-step ahead prediction. The processing analysis is performed in section 4. Finally, in
results show that TDNN exhibited lower training times with section 5, an industrial intelligent system based on neural
respect to a desired performance criteria. However, for time networks is proposed and the obtained results displayed.
series in which temporal dependency is large, the recurrent
networks with time delayed inputs could lead to better results. ,QMHFWLRQPRXOGLQJSURFHVVGHVFULSWLRQ
In principle, injection moulding is a simple process. A
,QWURGXFWLRQ thermoplastic, in the form of granules, passes from a feed
hopper into the barrel where it is heated and melted. It is then
Neural networks have achieved, in recent years, a high degree injected into a cold mould, which is clamped tightly closed.
of importance. The availability of process control computers When the plastic has had enough time to return to a solid state
as well as data historians made it easy to develop neural the mould opens and the produced part is extracted. All
network solutions for process modeling and control. Their details of the mould are reproduced in the finished part.
usefullness is no longer questioned. From medical to Although the injection moulding process can be applied to
industrial applications, neural networks have been applied in a large variety of polymers, the manufacture of only one
a countless number of situations. Lately, mainly due to the particular grade is considered here. The part, from the
increasing pressure from the consumption markets, a lot of automotive industry, was moulded on a DN502 mould and
research has been done under the field of fault detection and using the grade Hostacom D M2 T06.
diagnosis. Such systems, shortly refered to as FDD systems,
have become extremely important among industrial
applications. The needs to produce more and better, imposed
by the more demanding and better informed markets, has led
to the investigation of new methodologies in the quality
control area. Neural computing, as one of such techniques,
became an attractive approach in this area, since neural
)LJ Injection moulding machine.
networks are adaptable to an evolving environment, and are
able to take a quick decision once they have learned the
The process variables for which, in the current mould,
proper control function.
data is available are cycle time, dosage time, injection time
and cushion. The last two variables can be seen in figure 2.

*4#/ 
InjectionTime 3URFHVV
,QSXW
 3URGXFW
3URFHVV /DE7HVW
4XDOLW\

)LJ Conventional quality control procedure.



             
3ULQFLSDO&RPSRQHQW$QDO\VLV
Cushion
In order to identify the relationships between variables and to
 detect the possible causes of faulty situations, some statistics
 have been used. By examining the way the system reacts to

variations in the process variables, it is possible to identify the
most significant parameters related with product quality, and

to detect the causes for process anomalies. However, these
 relationships are generally unknown and not easily identified.
             
The relationships between the process state and the product
)LJ Industrial plastic injection process data for a set of 782 quality can be statistically described either by conventional
samples. approaches or Principal Component Analysis.
Common methods are useful to determine the
relationships between process variables, by examining the
covariance matrices or the correlation coefficients. High
1HXUDOGDWDSUHGLFWLRQ coefficients clearly show strong correlation between the
parameters. However, in the presence of time dependencies or
One of the most promising current applications of neural non-linear relationships, the analysis will fail, since it only
networks is the development of intelligent sensors. Often, the works for linear time-independent processes. Nevertheless,
important variables in an industrial process are not available although the injection moulding process is known to be
during the production run, but with some nonlinear statistics extremely non-linear, some variables seem to be related, as
it is possible to associate them with the available indicated by the relatively high calculated correlation
measurements. coefficients. Alternatively, Principal Component Analysis
In the injection moulding process, 10 to 20 variables are allows clusters of linked variables to be derived. One may
usually available. However, even though only a few number think on PCA as a technique to find the directions in which a
of them are truly significant, the existence of strong cloud of data points is most stretched. In this way, it is
nonlinearities and interdependencies between them, increases possible to extract the main relations in data as well as to
the complexity of the process. The prediction of such key detect its abnormal behaviour. When the process is moving
process variables is therefore a crucial task. out-of-control variables indicative of the cause of the
Neural networks are known for being able to efficiently deviation must be identified. Examining their contribution to
approximate and interpolate multivariate data. Their ability to the calculated scores, it is possible to identify the set of
deal with nonlinearities and their capability to generalize from original variables whose contribution may be reflective of the
the acquired knowledge during the training phase, makes non-conforming behaviour.
them extremely suitable for prediction. Figures 4 and 5 show the scores plots for principal
Until recently, quality control was performed off-line in components one and two and principal components three and
some kind of quality control laboratory, as shown in Figure 3. four, respectively ( '+' define the nominal data, whilst 'o'
Obviously, a time delay is incurred in analysing the test define abnormal data). As can be seen, the evidence of a
samples, which can be in the range of some minutes to some problem is most clearly highlighted by principal component
hours. This could be too late to make the necessary four. The identification of the combination of variables,
adjustments. which contribute to the particular principal component of
One can go even further in terms of prediction and look interest (pc4), allows the identification of the most probable
into an important aspect such as the one concerned with causes of the faults. In our case, cushion and injection time
process control. Poor quality parts are usually associated with were the ones that contributed the most to the pc4.
some deviation in the variable values. If the system could
predict in advance the process behaviour, it would be able to
identify future deviations and proceed attemptly with the
necessary corrections.


6&25(63/27 A neural network is trained on historical data, obtained

from the injection moulding machine during running

operation, to predict process variables one step ahead. We can
thus construct a neural model of the real process. The main
objective is the detection of defects on the produced parts as
3ULQFLSDO&RPSRQHQW



well as the identification of its causes. Parts with burning


 marks, spots, streaks or anomalous dimensions should be
detected and isolated. Once the faults are known and its

causes identified, one can easily train a neural network to act

on the system's responses in order to detect anomalies and
identify its causes. Figure 6 shows the overall proposed
 scheme.
Due to the dynamic behaviour of the process, if we

        assume that the input variables xt are samples of a time
3ULQFLSDO&RPSRQHQW
series, and a future value is to be predicted, then the neural
)LJ Scores plot for principal components 1 and 2 with network becomes dynamic. This leads to another problem.
process data. The one of finding structure in a time varying data sequence.
There are basically two different approaches. The simpliest
6&25(63/27
 one, known as time lagging or temporal windowing, is to
convert a time varying signal into a static vector. The

samples x1 , K , x n can be stored in a delay line, which
serves as the input layer to a feedforward multi-layer
3ULQFLSDO&RPSRQHQW

perceptron (MLP) network. As each new data point becomes


 available, the data is shifted along to accommodate it [Swi1].
This can be regarded as a moving window which covers the

last n values in the sequence. Networks that use data in this

way are known as TimeDelayNeuralNetworks, TDNN.
Another method, for representing time, is to feed a
 recurrent network with single values in a sequence and allow
it to learn its own short time memory. Elman networks have

      
been used in this work. These networks have a multi-layered
3ULQFLSDO&RPSRQHQW structure similar to the one of MLPs. The main difference is
)LJ Scores plot for principal components 3 and 4 with the addition of a special hidden layer, often called the context
process data. or state layer [Pham2], which acts as a short-term dynamic
memory responsible for the ability to learn to recognise and
generate temporal patterns, as well as spatial patterns [Elm3].
1HXUDOQHWZRUNV This memory, which is simply a copy of the hidden layer, acts
as an extension to the input layer, holding the feedback
The foundation of building intelligent sensors is that some signals from the hidden units from the previous time step. The
variables do not have on-line sensors or the available sensors outputs of these context neurons are then fed forward to the
are nor cost-effective or reliable due to operating conditions. hidden layer for the current time step. Since only the forward
However, they often have a functional relationship with other connections are to be adapted and the feedback connections
process variables which can be measured on-line. It is, are preset to constant values, then these networks can be seen
therefore, clear that large benefits could be obtained by as feedforward networks and the above BP algorithm can be
applying neural networks to data prediction. used [Pham2].
Two learning algorithms have been used for neural
< network approximation of a non-linear activation function f.
; 3URFHVV
The standard and well known backpropagation algorithm
(BP), and the Levenberg-Marquardt method. The former is
$11 1HXUDO)'' basically a gradient descent method, in which the change in
0RGHO V\VWHP the weight of a connection between two neurons is given by
º
wk +1 = wk − α k g k
)LJ Fault Detection and Diagnosis general proposed
scheme.


where w k is a vector of current weights and biases, g k is the
current gradient, and α k is the learning rate. In order to The purpose was to train a network that could predict the
process behaviour in terms of the injection time and cushion
improve the convergence speed and possible avoidance of parameters, as these were found to be related to the detected
local minima, an additional term, called momentum, is used faults. All data points were previously scaled to zero mean
in the generalised delta rule. This way, the gradient is also and variance one.
influenced by the previous weight change.
The Levenberg-Marquardt algorithm, like the
Obtained Epochs
quasi-Newton methods, approaches second-order training
SSE
speed without having to compute the Hessian matrix (second
derivatives) of the performance index at the current values of (OPDQ 0.6054 500
the weights and biases. Instead, the Hessian matrix is 7'11 0.2544 500
approximated as
7DEOH – Network convergence evaluation after 500 epochs.
T
H =J J
To select the best injection moulding fitting model,
several neural networks were built and trained on 781 data
and the gradient computed as
samples using the BP algorithm. TDNN and Elman network
architectures were used and tested on a validation set
g = JTe constituted by 784 independent samples. Although both
network types have similar performances, the TDNN, with 4
where J is the Jacobian matrix, which contains first input units, 15 hidden units, with tan-sigmoid activation
derivatives of the network errors with respect to the weights functions, and 2 output units, with linear activation functions,
and biases, and e is a vector of network errors. The Jacobian exhibits lower training times, converging more rapidly to the
matrix can be computed through a standard backpropagation desired previously set error value. Table 2 shows the
technique that is much less complex than computing the networks performance in terms of sum-squared error (SSE)
Hessian matrix. after 500 epochs. Figures 7, 8 and 9 show the obtained results
Whichever the method chosen, before training a neural for the TDNN. Particular attention should be given to figure
network, data preprocessing is required. The first step is to 8, where it can be seen that the network outputs are quite
identify and remove outliers as there is no guarantee that close to the real data except for a number of points close to
network training will ignore outliers and fit well on good sample 100, which can be explained by a slight deviation of
samples. Further, since neural networks are nonlinear the measures (see figure 2). The same analysis can be
approximators, it is experienced that they are more sensitive performed on the error histograms. While for the first output
to outliers than linear regression methods because they can the network exhibits a healthy behaviour with a peak at
bend to reach the outliers [Qin4]. Table 1 shows the around zero, falling off as the number of errors of greater
minimum, maximum and the average values for each one of magnitude diminishes [Swi1], for the second output that does
the process variables. It can be easily seen that each variable not happen, which can be explained by the variability of the
covers a different range of values. Therefore, the errors due to process variable cushion.
the higher valued variables will have a greater effect during
training than the errors due to the lower ones, as their
magnitude will be greater. If we are able to ensure that each
variable covers the same range than we also ensure that the
errors on each variable contribute with the same proportion to
the change in the network weights.

0LQLPXP 0D[LPXP $YHUDJH


YDOXH YDOXH YDOXH
&\FOH7LPH V 34.000 245.000 37.975

'RVDJH7LPH V 8.940 9.9000 9.341

,QMHFWLRQWLPH V 6.010 6.790 6.654

&XVKLRQ PP 7.400 15.400 14.622 )LJ Error evolution during the training process.


7DEOH – Data measurement ranges.


0HDVXUHPHQWGDWD [1HWZRUNRXWSXW R RXWSXW $XWRFRUUHODWLRQIXQFWLRQRISUHGLFWLRQHUURU RXWSXW
 

 

 

 
              
3UHGLFWLRQHUURU ODJ
+LVWRJUDPRYHUSUHGLFWLRQHUURUV
 

 

 

 
               

0HDVXUHPHQWGDWD [1HWZRUNRXWSXW R RXWSXW $XWRFRUUHODWLRQIXQFWLRQRISUHGLFWLRQHUURU RXWSXW


 







 
              
3UHGLFWLRQHUURU ODJ
+LVWRJUDPRYHUSUHGLFWLRQHUURUV
 

 

 

 
               

)LJ Network outputs and error prediction results. )LJ Error correlation plots and histograms.

&RQFOXVLRQV $FNQRZOHGJHPHQWV
Linear modelling of the existing real-world problems in The authors would like to thank PLASFIL - Plásticos da
industry is inadequate, since they are non-linear, time-varying Figueira, Lda. for the availability of the mould and for
and uncertain. making the moulding data available. The authors would also
This paper addresses a system architecture which like to express their gratitude to PRAXIS XXI, Sub-Programa
combines PCA with the so-called intelligent system Ciência e Tecnologia do 2º Quadro Comunitário de Apoio,
methodologies such as neural networks. which financially supports this project.
Neural networks can account for temporal dependencies
directly via a delay window or indirectly using a recurrent 5HIHUHQFHV
copy of the hidden layer. For time series in which temporal
dependency is large, it can be appropriate to use recurrent
[Swi1] Kevin Swingler, “Applying Neural Networks: A
networks with time delayed inputs.
practical guide”, Academic Press, (1996).
The data transformation by PCA accounts for obtaining
[Pham2] D. T. Pham, “An introduction to artificial neural
more precise and reasonable numerical results. Further
networks”. In A. B. Bulsari, Neural Networks for Chemical
analysis for diagnosis requires the development of a neural
Engineers, Elsevier, (1995).
network system able to respond in real-time to the evolving
[Elm3] J. L. Elman, “Finding structure in time”, Cognitive
industrial environment.
Science, , pp 179-211, (1990).
[Qin4] S. Joe Qin, “Neural Networks for Intelligent Sensors
and Control – Practical Issues and Some Solutions”, In Omid
Omidvar and David L. Elliot, Neural Systems for Control,
Academic Press, (1997).



Das könnte Ihnen auch gefallen