Beruflich Dokumente
Kultur Dokumente
1. INTRODUCTION
Function approximation is one of the fundamental issues
in machine learning [SI and has been exploited by
computer scientists in a number of applications,
including data classification and data mining [3]. One
widely-used form of function approximators is a linear
combination of spherical Gaussian functions as shown
in the following:
Then, we have
and
g($1
+)...,
+)
Let z = exp --
;2)
j=l
j=O
1 + 2 C r ( 2 j ) *= 2 x z ( 2 j + 1 ? .
=.f@i)
(3)
vk-1
(8)
1038
Furthermore, we have
sz
approximator uniformly to -.
81nk
With the variances of the Gaussian basis functions
determined, the next issue to solve is to determine the
weights of these Gaussian functions.
Based on
equation (3), we can derive one equation with each
training sample. As a result, if we have n training
samples, then we have n equations with n variables,
which are the weights of the Gaussian functions.
Accordingly, we can solve these equations and find the
appropriate weights. However, the time complexity for
solving n equations with n variables is very high,
exceeding O(n2). Therefore, the empirical approach
elaborated in the following is employed in this paper.
Based on equation (6), the empirical approach
begins with setting
f (Si)
w .=
< 22; = 6 . 4 6 6 4 ~ 1 0 - ~ .
In summary, we have
)[
0 < 1+ 2 2 zp)2- 2 z
j=l
j=o
where
p=?=\J&.
Accordingly, the
RBF
[ I1 (.. i; 11
where 2 = 1+ 2 C e x p - 7
space o f f , and vl,
samples of x.
VZ,
, x is a vector in the
Number of terms
taken from each side
of equation (8)
6
7
8
9
1
root
Error bound
0.882248
0.897739
0.909615
0.919010
2.8013 x lo-
1.2580 x IO+
5.6212 x lo-
2.5010 x lo-*
[i-;)
1039
>'
08-
06-\,
04-
-\\m\',
*,I, /-\
i,
02-
46-
\\
\
08-
,L.,",
-2
-15
-1
4 5
\\
/
iJ,i
\h
44-
'k$
41
\\\>
0 2O -
y;\
I/
$\'
\'\<?
\
05
15
vur[ E ( s ~ ) + E ( v ~ ) + . . . + ~ ( v ~ )
q+l
With the low-pass filter incorporated, the function
approximator based on equation (11)becomes:
-2
-1.5
-1
0.5
0.5
1.5
i\
o.2
r+l
h i )z
h)
4.8
1
-2
-1.5
?-;.".
-1
i/
6.5
0.5
1.5
3(v,>
=f(vJ
h v 2 ) 2 f(v2)
- * e ?
3. EXPERIMENTAL RESULTS
This section reports the experiments conducted to
evaluate the proposed learning algorithm.
The
following three functions are used in the experiments:
(i)
f , ( x ) = 1+ (1-x+ 2x2)e-", x
[-5,51;
where x is a vector in the space off, VI. v2, ..., yk are the
k nearest training samples of x, and ,8 and A are the same
constants defined in equation (11).
As far as the time complexity of the learning
algorithm is concerned, there are four major tasks
carried out by the learning algorithm in the construction
of a RBF network. The first task is to construct a
hypercube data structure to store the training samples.
This task can be done in O(lSl), where S is the set of
training samples. The second task is to coinduct
low-pass filtering at the input of the function
approximator. The time complexity of this task is
O(qlS)), where 4 is number of neighboring samples
involved in equation (12). The third task :is to
construct a function approximator based on equation
(13). The only work involved in this task is to
determine the values of ,8 and A , which can be done in
constant time. The last task that the learning algoiithm
carries out is executing the compensation procedure.
This task can be done in O(dSl), where r is the number
of neighboring samples in equation (14) and z i s the
number of iterations of the outer loop of the
compensation procedure. As a result, the overall time
complexity of the learning algorithm in the construction
of a FU3F network is O(ISI+ qlSl+n$SI). If q, z, and r
are treated as constants, then the time complexity is; in a
linear order.
With equation (15), the time complexity of' the
function approximator for predicting the function values
of input vectors is O(k)+O(kIq), where T is the set of
input vectors and k is the number of nearest training
samples included in equation (15). The term O(k) is
due to the time taken to figure out the k nearest training
samples of an incoming vector. Because the training
samples are evenly spaced, this task can be done in O(k).
If k is treated as a constant, then the time complexity is
1041
I Setting
I Gaussian
Bayesian information
criterion
Bias
I true
Scales
1 [I, 0.5,0.2,0.1]
Table 3. The parameterloption settings in rbf-fs-2,
rbf-rr-2, and rbf-rt-2.
Model selection criterion
fi
rbf-fs-2
rbf-rr-2
rbf-rt-2
proposed
f2
0.036+0.008 0.025k0.002
0.09W.022
0.04OH.008 0.021M.002 0.122f0.024
0.045f0.012 0.016f0.003 0.097M.007
0.04W.007 0.016k0.002 0.065f0.005
4. CONCLUSIONS
In this paper, an efficient learning algorithm for
constructing function approximators with RBF networks
is proposed. The proposed learning algorithm features
a linear time complexity for constructing the RBF
network. On the other hand, the most well-known
existing learning algorithms suffer high time complexity
due to the need to compute the inverse of a matrix.
Experimental results reveal that the RBF network
constructed with the proposed learning algorithm is able
to approximate the target function at the same level of
accuracy as the RBF networks constructed with the
existing learning algorithms. As far as the time
complexity for predicting the function values of input
vectors is concerned, the RBF network constructed with.
the proposed learning algorithm can complete the task in
@)U), where T is the set of input vectors. Another
important feature of the proposed learning algorithm is
that the space complexity of the RBF network
5. REFERENCES
[ l ] E. J. Hartman, J. D. Keeler, and J. M. Kowalski,
"Layered neural networks with Gaussian hidden
units as universal approximations," Neural
Computation, vol. 2, no. 2, pp. 210-215, 1990.
[2] S. Haykin, "Neural networks: A comprehensive
foundation," New York McMillan, 1994.
[3] V. Kecman, Learning and Soft Computing: Support
Vector Machines, Neural Networks, and Fuuy
Logic Models, The MIT Press, Cambridge,
Massachusetts, London, England, 2001.
[4] D. J. C. Mackay, "Bayesian interpolation," Neural
Computation, vol. 4, no. 3, pp. 415-447, 1992.
[SI T. Mitchell, Machine Learning, McGraw-Hill,
1997.
[6] M. J. L. Orr, "Regularisation in the selection of
radial basis function centres," Neural Computation,
vol. 7, no. 3, pp. 606-623, 1995.
[7] M. J. L. OK, "Introduction to radial basis function
networks," Technical report, Center for Cognitive
Science, University of Edinburgh, 1996.
[8] M. J. L. Orr, "Optimising the widths of radial basis
functions," Proceedings of the Fifth Brazilian
Symposium on Neural Networks, 1998.
[9] M. J. L. Orr, "Recent advances in radial basis
function networks," Technical report, Center for
Cognitive Science, University of Edinburgh, 1999.
[lo] M. J. L. Orr, "Matlab functions for radial basis
function networks," http://www.anc.ed.ac.uk/-mjol
software1rbf.zip.
[ l l ] T. Poggio and F. Girosi, "Networks for
approximation and learning,'' IEEE Proceedings,
vol. 78, no. 9,pp. 1481-1497,1990.
[12] L. R. Rabiner and B. Gold, Theory and Application
of Digital Signal Processing, Prentice-Hall,
Englewood Cliffs, New Jersey, 1974.
1042