Abstract
This paper presents an incremental learning algorithm primarily intended for adjusting and optimizing high dimensional backpropagation ANN-based supervised classification systems. The algorithm avoids the highly time-consuming pre-processing stage used to reduce dimensionality through the deletion or averaging of redundant information and the establishment of an appropriate processing window. The proposed algorithm acts during the training process of the ANN by automatically obtaining the optimal window size and transformation parameter values needed for a given set of classification requirements. During this process, it changes the network topology on line by adaptively appending input units and their corresponding connections to the existing network. The proposed process allows the ANN to learn incrementally by adapting to the new topology without forgetting what had been learnt earlier. This process could also be used as an incremental learning system in, for instance, robotic systems; when new instances are fed, as it would not need to perform a whole training stage. Instead, the knowledge encoded in the new instances could be learnt through the proposed adjustment of network topology.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Benediktsson, J.A., Swain, P.H., Ersoy, O.K., Hong, D.: Classification of High Dimensional Data Using Neural Networks. In: Geoscience and Remote Sensing Symp. (1990)
Donoho, D.: High dimensional data analysis - the curses and blessings of dimensionality. In: AMS conference Math Challenges of the 21st Century, Los Angeles, August 6-11 (2000)
Verleysen, M.: Learning high-dimensional data. In: NATO Advanced Research Workshop on Limitations and Future Trends in Neural Computing, Siena, Italy, pp. 22–24 (2001)
Jimenez, L.O., Landgrebe, D.A.: Supervised classification in high-dimensional space: geometrical,statistical, and asymptotical properties of multivariate data. IEEE Transactions on Systems, Man and Cybernetics 28(1), 39–54 (1998)
Ravi Kanth, K.V., Agrawal, D., Singh, A.: Dimensionality Reduction for Similarity Searching in Dynamic Databases. In: SIGMOD Conference, pp. 166–176 (1998)
Duda, R., Hart, P., Stork, D.: Pattern Classification, 2nd edn. Wiley Interscience, Hoboken (2000)
Grollman, D.H., Jenkins, O.C., Wood, F.: Discovering natural kinds of robot sensory experiences in unstructured environments. J. Field Robotics 23(11-12), 1077–1089 (2007)
Yairi, T.: Map Building without Localization by Dimensionality Reduction Techniques. In: Proceedings of the 24th International Conf. on Machine Learning, Corvallis, OR (2007)
Willoughby, C.T., Folkman, M.A., Figueroa, M.A.: Application of hyperspectral-imaging spectrometer systems to industrial inspection. In: Proc. SPIE, vol. 2599, pp. 264–272 (1996)
Pan, Z., Healey, G.E., Prasad, M., Tromberg, B.J.: Hyperspectral face recognition for homeland security. In: Proc. Infrared Tech. and Applications XXIX, vol. 5074, pp. 767–776 (2003)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Prieto, A., Bellas, F., Duro, R.J., Lopez-Peña, F. (2009). An Incremental Learning Algorithm for Optimizing High-Dimensional ANN-Based Classification Systems. In: Köppen, M., Kasabov, N., Coghill, G. (eds) Advances in Neuro-Information Processing. ICONIP 2008. Lecture Notes in Computer Science, vol 5506. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-02490-0_126
Download citation
DOI: https://doi.org/10.1007/978-3-642-02490-0_126
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-02489-4
Online ISBN: 978-3-642-02490-0
eBook Packages: Computer ScienceComputer Science (R0)