Abstract
Incremental Net Pro (IncNet Pro) with local learning feature and statistically controlled growing and pruning of the network is introduced. The architecture of the net is based on RBF networks. Extended Kalman Filter algorithm and its new fast version is proposed and used as learning algorithm. IncNet Pro is similar to the Resource Allocation Network described by Platt in the main idea of the expanding the network. The statistical novel criterion is used to determine the growing point. The Bi-radial functions are used instead of radial basis functions to obtain more flexible network.
Preview
Unable to display preview. Download preview PDF.
References
J. V. Candy. Signal processing: The model based approach. McGraw-Hill, New York, 1986.
W. Duch and G. H. F. Diercksen. Feature space mapping as a universal adaptive system. Computer Physics Communications, 87:341–371, 1994.
W. Duch and N. Jankowski. New neural transfer functions. Jour. of Applied Math. and Computer Science. submitted.
S. E. Fahlman and C. Lebiere. The cascade-correlation learning architecture. In D. S. Touretzky, editor, NIPS. Morgan Kamufmann, 1990.
E. Fiesler. Comparative bibliography of ontogenic neural networks. In Proceedings of the International Conference on Artificial Neural Networks, 1994.
F. Girosi and T. Poggio. Networks and the best approximation property. AI Lab. Memo, MIT, 1989.
B. Hassibi and D. G. Stork. Second order derivatives for network pruning: Optimal brain surgeon. In NIPS, 1993.
V. Kadirkamanathan. A statistical inference based growth criterion for the RBF network. In Proc. IEEE. Workshop on Neural Networks for Signal Processing, 1994.
V. Kadirkanianathan and M. Niraujan. A function estimation approach to sequential learning with neural networks. Neural Computation, 5(6):954–975, 1993.
Y. LeCan, J. Denker, S. Solla., R. E. Howard, and L. D. Ja.ckel. Optimal brain damage. In D. S. Touretzky, editor, Advances in Neural Information Processing Systems II. Morgan Kauffman, 1990.
J. Platt. A resource-allocating network for function interpolation. Neural Computation, 3:213–225, 1991.
T. Poggio and F. Girosi. Network for approximation and learning. Proc. IEEE, 78:1481–1497, 1990.
M. J. D. Powell. Radial basis functions for multivariable interpolation: A review. In J. C. Mason and M. G. Cox, editors, Algorithms for Approximation of Functions and Data, pages 143–167. Oxford University Press, 1987.
Y. Shang and W. Wah. Global optimization for neural network training. IEEE ] Computer, 29, 1996.
B. Šter and A. Dobnikar. Neural networks in medical diagnosis: Comparison with other methods. In A. B. B. et al., editor, Proceedings of the, International Conference EANN '96, pages 427–430, 1996.
A. S. Weigend, D. E. Rumelhart, and B. A. Huberman. Back-propagation, weight elimination and time series prediction. In Proceedings of the 1990 Connectionist Models Summer School, pages 65–80. Morgan Kaufmann, 1990.
L. Yingwei, N. Sundararajan, and P. Saratchandran. A sequential learning scheine for function approximation using minimal radial basis function neural networks. Neural Computations, 9:461–478, 1997.
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1997 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Jankowski, N., Kadirkamanathan, V. (1997). Statistical control of RBF-like networks for classification. In: Gerstner, W., Germond, A., Hasler, M., Nicoud, JD. (eds) Artificial Neural Networks — ICANN'97. ICANN 1997. Lecture Notes in Computer Science, vol 1327. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0020185
Download citation
DOI: https://doi.org/10.1007/BFb0020185
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-63631-1
Online ISBN: 978-3-540-69620-9
eBook Packages: Springer Book Archive