Abstract
Artificial Neural Networks (ANN) are currently exploited in many scientific domains. They had shown to act as doable, practical, and fault tolerant computational methodologies. They are equipped with solid theoretical background and proved to be effective in many demanding tasks such as approximating complex functions, optimizing search procedures, detecting changes in behaviors, recognizing familiar patterns, identifying data structures. ANNs computational limitations, essentially related to the presence of strong nonlinearities in the data and their poor generalization ability when provided of fully connected architectures, have been hammered by more sophisticated models, such as Modular Neural Networks (MNNs), and more complex learning procedures, such as deep learning. Given the multidisciplinary nature of their use and the interdisciplinary characterization of the problems they approach, ranging from medicine to psychology, industrial and social robotics, computer vision, and signal processing (among many others) ANNs may provide the bases for a redefinition of the concept of information processing. These reflections are supported by theoretical models and applications presented in the chapters of this book.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Caelli, T., Guan, L., Wilson Wen, W.: Modularity in neural computing. Proc. IEEE 87(9), 1497–1518 (1999)
Cattaneo, M.: A continuous updating rule for imprecise probabilities. In: Laurent, A., et al. (eds.) Information Processing and Management of Uncertainty in Knowledge Based Systems, pp. 426–435. Springer (2014)
Comon, P.: Independent component analysis, a new concept? Sig. Process. 36, 287–314 (1994)
Deco, G., Brauer, W.: Nonlinear higher-order statistical decorrelation by volume-conserving neural architectures. Neural Netw. 8(4), 525–535 (1996)
Esposito, A.: The amount of information on emotional states conveyed by the verbal and nonverbal channels: some perceptual data. In: Stylianou, Y., et al. (eds.) LNCS 4391, 249–268. Springer-Verlag, Berlin (2007)
Klir, G.J.: Generalized information theory: aims, results, and open problems. Reliab. Eng. Syst. Saf. 85(1–3), 21–38 (2004)
Shannon, C. E.: A mathematical theory of communication. First published: July 1948 (1948). doi:10.1002/j.1538-7305.1948.tb01338.x
Shannon, C., Weaver, W.: The mathematical theory of communication. University of Illinois Press, Urbana, IL (1963)
Yang, H.H., Amari, S.: Adaptive on-line learning algorithms for blind separation—maximum entropy and minimum mutual information. Neural Comput. 9, 1457–1482 (1997)
Yang, M., Chen, T., Wu, K.: Generalized belief function, plausibility function, and Dempster’s combination rule to fuzzy sets. Int. J. Intel. Syst. 18(8), 925–937 (2003)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG
About this chapter
Cite this chapter
Esposito, A., Faundez-Zanuy, M., Morabito, F.C., Pasero, E. (2018). Redefining Information Processing Through Neural Computing Models. In: Esposito, A., Faudez-Zanuy, M., Morabito, F., Pasero, E. (eds) Multidisciplinary Approaches to Neural Computing. Smart Innovation, Systems and Technologies, vol 69. Springer, Cham. https://doi.org/10.1007/978-3-319-56904-8_1
Download citation
DOI: https://doi.org/10.1007/978-3-319-56904-8_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-56903-1
Online ISBN: 978-3-319-56904-8
eBook Packages: EngineeringEngineering (R0)