Introduction
In this project, we introduce a novel learning rule by applying Rousseeuw's
least trimmed squares (LTS) method to increase the robustness of
feedforward neural networks. The particular paradigm of networks under
consideration is the radial basis function network (RBFN); it has
significant advantages over multilayer perceptrons (MLP), namely faster
convergence, smaller extrapolation errors, higher reliability, and a more
well-developed theoretical analysis.
Similar robust learning rules that have been applied to different
paradigms, for example, MLP and GMDP networks. In order to compensate for
the loss of efficiency, an adaptive version of the proposed learning rule
is developed in an attempt to maximize the robustness and efficiency.
In contrast to most present robust learning rules, our method neither is
dependent upon any assumptions about error distribution nor is it necessary
to estimate the error distribution. More importantly, an upper bound on
the number of outliers that the resulting robust RBFN can withstand is
established.
Robust Radial Basis Function Networks
Radial basis function networks (RBFNs) have increasingly attracted interest
for engineering applications due to their advantages over traditional
multilayer perceptrons, namely faster convergence, smaller extrapolation
errors, and higher reliability. RBFN is a class of single hidden layer
feedforward networks where the activation functions for hidden units are
defined as radially symmetric basis functions phi such as
the Gaussian function.
Each hidden unit forms a localized receptive field in the input
space X, whose centroid is located on c with width
sigma. The fraction of overlap between each hidden unit and its
neighbors is decided by the width sigma such that a smooth
interpolation over the input space is allowed. Therefore, unit i gives
a maximum response for input stimuli close to c.
In general, training RBFNs involves two phases: one is unsupervised
learning (clustering) on the hidden layer to determine N receptive
field centroids in the training data set and the associated widths; the
other is supervised learning on the output layer to estimate the connection
weights w. The output layer simply implements a multiple linear
regression and hence can be trained by the iterative gradient descent
method based on the least squares approach. Much work has been devoted to
improving the learning performance for clustering and determining the
optimal complexity for networks since it plays a critical role in achieving
good approximation precision although the choice of the radial basis
functions is not crucial for performance; for example, see (CHENS-91,
MOOD-89). In contrast to this work, we focus on the issue of enhancing
reliability of RBFNs in the presence of gross errors.
Experimental Results and Analysis
A number of simulations have been conducted to evaluate the efficiency
and the stability against the presence of outliers for the proposed
AR2BF learning rule. Simulation results using the
aforementioned learning rules, namely RBF, R2BF, and
AR2BF rules on the problems of 1-D and 2-D function
approximation will be discussed.
Bibliography
[BURR-91] P. Burrascano, "A Norm Selection Criterion for the Generalized
Delta Rule", IEEE Trans. on Neural Networks, vol. 2, pp. 125-130, 1991.
[BURR-94] P. Burrascano, "Robust Learning in the presence of Outliers",
1994 International Symposium on Artificial Neural Networks, Taiwan, 1994.
[CHEN-94] D. S. Chen and R. C. Jain, "A Robust Back Propagation Learning
Algorithm for Function Approximation", IEEE Trans. on Neural
Networks, vol. 5, pp. 467-479, 1994.
[CHENS-91] Chen, S, Cowan, C. F. N., and Grant, P. M., "Orthogonal
Least Squares Learning Algorithm for Radial Basis Function Networks",
IEEE Trans. on Neural Networks, vol. 2, No. 2, pp. 302-309, 1991.
[CICH-93] A. Cichocki and R. Unbehauen, "Robust Estimation of Principal
Components by using Neural Network Learning Algorithms",
IEE Electronics Letters, vol. 29, pp 1869-1870.
[COVE-65] T. M. Cover, "Geometrical and Statistical Properties pf Systems
of Linear Inequalities with Applications in Pattern Recognition", IEEE
Trans. on Electronic Computers, vol. EC-14, pp. 326-334, 1965.
[GEMA-92] S. Geman and E. Bienenstock, "Neural Networks and the
Bias/Variance Dilemma", Neural Computation, vol. 4, pp. 1-58, 1992.
[GIRO-92] Federico Girosi, "Some Extensions of Radial Basis Functions and
Their Applications in Artificial Intelligence",
Computers Math. Applic., vol. 24, no. 61-80, 1992.
[HAMP-85] F. R. Hampel, E. M. Ronchetti, P. J. Rousseeuw, and W. A. Stahel,
Robust Statistics: The Approach Based on Influence Functions,
John Wiley & Sons, Inc., 1985.
[HART-90] E. J. Hartman, J. D. Keeler and J. M. Kowalski,
"Layered Neural Networks with Gaussian Hidden Units as Universal
Approximators", Neural Computation, vol. 2, pp. 210-215, 1990.
[HORN-89] K. Hornik, M. Stinchcombe, and H. White,
"Multilayer Feedforward Networks are Universal Approximators",
Neural Networks, vol. 2, pp. 359-366, 1989.
[HWAN-92] J. N. Hwang, H. Li, M. Maechler, D. Martin, and J.
Schimert, "Projection Pursuit Learning Networks for Regression",
Engr. Applic. Artificial Intelligence, vol. 5, pp.193-204, 1992.
[JOIN-92] J. A. Joines and Mark W. White,
"Improved Generalization Using Robust Cost Functions",
Proc. IEEE ICNN, vol. 3, pp. 911-918, 1992.
[KARA-92] N. B. Karayiannis and A. N. Venetsanopoulos,
"Fast Learning Algorithms for Neural Networks",
IEEE Trans. on Circuits and Systems, vol. 39, pp. 453-474, 1992.
[KIM-94] J. Kim and J. H. Seo, "Implementation of a Robust Feedforward
Neural Network Using the Classifier Structure", IEEE ICNN, vol.
3, pp. 1427-1432.
[LI-94-0] S. Li, Y. Chen, and E. L. Leiss, "GMDP: a Novel Unified Neuron
Model for Multilayer Feedforward Neural Networks",
Proc. IEEE ICNN, Orlando, pp. 107-112, 1994.
[LI-94-1] S. Li and E. L. Leiss, "Robust Learning in Feedforward Neural
Networks with GMDP Units", 1994 International Symposium on Artificial
Neural Networks, Taiwan, 1994.
[LI-94-2] S. Li and E. L. Leiss,
"Enhancing Robust Learning in Multilayer Perceptrons",
The Sixth International Conference on Artificial
Intelligence and Expert Systems Applications, Houston, 1994.
[LIAN-94] K. Liano,
"A Robust Approach to Supervised Learning in Neural Network",
Proc. IEEE ICNN, Orlando, vol. 1, pp. 513-516, 1994.
[MEER-91] P. Meer, D. Mintz, A. Rosenfeld, and D. Y. Kim,
"Robust Regression Methods for Computer Vision: A Review",
Internation Journal of Computer Vision, vol. 6, pp. 59-70, 1991.
[MOOD-89] J. E. Moody and C. J. Darken, "Fast Learning in
Networks of Locally-tuned Processing Units", Neural
Computation, vol. 1, pp. 281-294, 1989.
[NILS-90] N. J. Nilsson, The Mathematical Foundations of
Learning Machines, Morgan Kaufmann Publishers, 1990.
[PARK-91] J. Park and I. W. Sandberg, "Universal
Approximation using Radial-Basis Function Networks", Neural
Computation, vol. 3, pp. 246-257, 1991.
[PARL-94] A. G. Parlos, B. Fernandez, A. F. Atiya, J. Muthusami, and W.
K. Tsai, "An Accelerated Learning Algorithm for Multilayer Perceptron
Networks", IEEE Transactions on Neural Networks, vol. 5, pp.
493-497, 1994.
[POGG-90] Poggio, T. and Girosi, F., "Networks for Approximation and
Learning", Proc. of the IEEE, vol. 78, No. 9, pp. 1481-1497, 1990.
[ROUS-84] P. J. Rousseeuw, "Least Median of Squares
Regression", Journal of the American Statistical Association,
vol. 79, pp. 871-880, 1984.
[ROUS-87] P. J. Rousseeuw and A. M. Leroy, Robust
Regression and Outlier Detection, John Wiley \& Sons, Inc., 1987.
[WHIT-89] White, H., "Learning in Artificial Neural Networks: A
Statistical Perspective", Neural Computation, vol. 1, pp.
425-464, 1989.