Abstract
In this paper, we have first considered a number of popular recurrent neural network architectures. Then, two subclasses of general recurrent neural network architectures are introduced. It is shown that all these popular recurrent neural network architectures can be grouped under either of these two subclasses of general recurrent neural network architectures. It is also inferred that these two subclasses of recurrent neural network architectures are distinct, in that it is not possible to transform from one form to the other. Two recently introduced recurrent neural network architectures specifically designed for special purposes, viz., for overcoming long term temporal dependency, and for data structure classifications are also considered.
Once the architectural aspects of the class of networks are settled, then one could consider the training aspects. This will be considered in a companion paper [31].
Preview
Unable to display preview. Download preview PDF.
References
Albertini, F., Sontag, E. “For neural networks, function determines form”. Neural Networks. Vol 6, pp 975–990, 1993.
Back, A.D., Tsoi, A.C. “FIR and IIR synapses, a new neural network architecture for time series modelling”. Neural Computation. Vol. 3, No. 3, pp 375–385, 1991.
Baldi, P., Chauvin, Y. “Hybrid modelling, HMM/NN architectures, and protein modelling”. Neural Computation. Vol 8, No. 7, pp 1541–1565, 1996.
Bengio, Y., Simard, P., Frasconi, P. “Learning Long term dependencies with gradient descent is difficult”. IEEE Trans Neural Networks. Vol. 5, pp 157–166, 1994.
Box, G. E. P., Jenkins, G. Time Series Analysis. Holden Day, 1967.
Calder, B., Grunwald, D., Jones, M., Lindsay, D., Martin, J., Mozer, M., Zorn, B. “Evidence-based static branch prediction using machine learning”. ACM Transaction on on Programming Languages and Systems, Vol. 19, pp 188–222, 1997.
Chen, S., Billings, S., Grant, P. “Nonlinear system identification using neural networks”. International Journal of Control. Vol. 51, No. 6, pp. 1191–1214, 1990.
Elman, J. “Finding structure in time”. Cognitive Science. Vol. 14, pp 179–211, 1990.
Frasconi, P., Gori, M., Soda, G. “Local feedback multilayered networks”. Neural Computation. Vol. 4, pp 120–130, 1992.
Haykin, S. Neural Networks, A comprehensive foundation. MacMillan College Pub Co. 1994.
Hornik, K. “Approximation capabilities of multilayer feedforward neural networks”. Neural Networks. Vol. 4, pp 251–257, 1990.
Hochreiter, S., Schmidhuber, J. “Long short-term memory”. Neural Computation. Vol 9, pp 1735–1780, 1997.
Jordan, M. “Supervised learning and systems with excess degree of freedom”. Massachusetts Institute of Technology, COINS Technical Report 88-27, May, 1988.
Kailath, T. Linear Systems. Prentice Hall, Englewood Cliffs, N.J., 1980.
Lawrence, S., Giles, L., Back, A., Tsoi, A. C. “The gamma MLP — multiple temporal resolutions, the curse of dimensionality, and gradient descent learning”. Neural Computation To appear.
Lapedes, A., Farber, R. “Nonlinear signal processing using neural networks prediction and system modelling”. Los Alamos Natioanl Laboratory, Los Alamos, LA-UR-262, 1987.
Lin, T., Horne, B.G., Giles, L. “How embedding memory in recurrent neural network architecture helps learning long term temporal dependencies”. Technical Report, University of Maryland. Report Number UMIACS-TR-96-76, and CS-TR-3706, Institute for Advanced Computer Studies, University of Maryland, College Park, Maryland, 1996.
Marple, S.L. Digital spectral analysis and applications. Englewood, N.J.: Prentice Hall, 1987.
Narendra, K.P., Parthasarathy, K. “Identification and Control of Dynamical Systems using Neural Networks”. IEEE Trans Neural Networks. Vol 1, pp 4–27, 1990.
Nerrand, O., Roussel-Ragot, P., Personnaz, L., Dreyfus, G., Marcos, S. “Neural Networks and nonlinear adaptive filtering: Unifying concepts and new algorithms”. Neural Computation. Vol 5, pp 165–197, 1993.
Pindea, F. “Dynamics and architecture for neural computation in recurrent neural networks”. Journal of Complexity. Vol 4., pp 216–245, 1988.
Principe, J., de Vries, B., Oliveira, P. “The gamma filter — a new class of adaptive IIR filters with restricted feedback”. IEEE Trans Signal Processing. Vol. 41, pp 649–656, 1993.
Robinson, A., J. Dynamic error propagation networks. PhD thesis, University of Cambridge, Cambridge, U.K., 1989.
Scarselli, F. Tsoi, A.C. “Universal approximation using feedforward neural networks: A survey of some existing methods, and some results”. Neural Networks. To appear.
Siegelmann, H., Home, B., Giles, L. “Computational capabilites of recurrent NARX neural networks”. IEEE Trans System, Man and Cybernetics. Part B, Vol 27, pp 208–218, 1997.
Sontag, E. “Neural networks for control”. In Essay on Control: Perspectives in the Theory and its applications. H. L. Trentelman, J. C. Willems, Ed. Boston: Birkhauser, pp. 339–380, 1993.
Sperduti, A. “Labelling RAAM”. Connection Science. Vol. 6, No. 4, pp 429–459, 1994.
Sperduti, A., Starita, A. “Supervised neural networks for the classification of structures”. IEEE Trans Neural Networks. Vol 8, pp 714–735, 1997.
Tsoi, A.C., Back, A.D. “Locally recurrent globally feedforward networks: a critical review of architectures”. IEEE Trans on Neural Networks. Vol. 5, No. 2, pp 229–239, 1994.
A C Tsoi, “Application of neural network methodology to the modelling of the yield strength in a steel rolling plate mill”, Advances in Neural Information Processing Systems, Vol 4. Ed. Moody, J, Hansen, S, Lippmann, R, Morgan Kaufmann Publishers, 1992.
Tsoi, A.C. “Gradient based learning methods”. This volume.
Tsoi, A.D., Back, A.D. “Discrete time recurrent neural network architectures: a unifying review”. Neurocomputing. Vol. 15, pp 183–224, 1997.
Waibel, A., Hanazawa, T., Hinton, G., Shikano, K., Lang, L. “Phonemic recognition using time delay neural networks” IEEE Trans Acoustic Speech and Signal Processing. Vol. 37, No. 3, pp 328–339, 1989.
Wan, E. “Temporal backpropagation for FIR neural networks”. Proc Int Joint Conf Neural Networks. San Diego, June, 1990, pp 575–580, 1990.
Williams, R., Zipser, D. “A learning algorithm for continually running fully recurrent neural networks”. Neural Computation. Vol. 1, pp 270–280, 1989.
Zomaya, A., Mills, P.M., Tade, M.O. Neuron-adaptive process control, a practical approach. Wiley, 1996.
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1998 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Tsoi, A.C. (1998). Recurrent neural network architectures: An overview. In: Giles, C.L., Gori, M. (eds) Adaptive Processing of Sequences and Data Structures. NN 1997. Lecture Notes in Computer Science, vol 1387. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0053993
Download citation
DOI: https://doi.org/10.1007/BFb0053993
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-64341-8
Online ISBN: 978-3-540-69752-7
eBook Packages: Springer Book Archive