Abstract
A simple result is presented that links the learning of hidden Markov models to results in complexity theory about nonlearnability of finite automata under certain cryptographic assumptions. Rather than considering all probability distributions, or even just certain specific ones, the learning of a hidden Markov model takes place under a distribution induced by the model itself.
Supported by a Marie Curie fellowship of the European Union under grant no. ERB-FMBI-CT98-3248. Most of this research was done while the author was working at the University of Munich.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
M. Anthony and N. Biggs, Computational learning theory: an introduction, Cambridge University Press, 1992.
P. Baldi and Y. Chauvin, Smooth on-line learning algorithms for hidden Markov models, Neural Computation 6 (1994) 307–318.
L. E. Baum, An inequality and associated maximization technique in statistical estimation for probabilistic functions of a Markov process, Inequalities 3, 1972, 1–8.
A. Blumer, A. Ehrenfeucht, D. Haussler, and M. K. Warmuth, Learnability and the Vapnik-Chervonenkis dimension, J. of the ACM 36(4), 929–965, 1989.
E. Charniak, Statistical Language Learning, MIT Press, 1993.
P. Clote and R. Backofen, Computational molecular biology: an introduction, Wiley, 2000.
O. Goldreich, S. Goldwasser, and S. Micali, How to construct random functions, J. of the ACM 33(4) (1986) 792–807.
M. J. Kearns, The computational complexity of machine learning, MIT Press, 1990.
M. J. Kearns and L. G. Valiant, Cryptographic limitations on learning boolean formulae and finite automata, J. of the ACM 41(1) (1994) 67–95.
M. J. Kearns, U. V. Vazirani, An introduction to computational learning theory, MIT Press, 1994.
M. Kharitonov, Cryptographic hardness of distribution-specific learning, Proc. 25th ACM Symp. on the Theory of Computing, 372–381, ACM Press, N.Y., 1993.
L. Pitt and M. K. Warmuth, Prediction-preserving reducibility, J. Computer and System Sci. 41(3) (1990) 430–467.
L. G. Valiant, A theory of the learnable, Communications of the ACM 27(11) (1984) 1134–1142.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2002 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Terwijn, S.A. (2002). On the Learnability of Hidden Markov Models. In: Adriaans, P., Fernau, H., van Zaanen, M. (eds) Grammatical Inference: Algorithms and Applications. ICGI 2002. Lecture Notes in Computer Science(), vol 2484. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45790-9_21
Download citation
DOI: https://doi.org/10.1007/3-540-45790-9_21
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-44239-4
Online ISBN: 978-3-540-45790-9
eBook Packages: Springer Book Archive