Abstract
Hidden Markov Model (Hmm) is one of the most popular language models. To improve its predictive power, one of Hmm hypotheses, named limited history hypothesis, is usually relaxed. Then Higher-order Hmm is built up. But there are several severe problems hampering the applications of high-order Hmm, such as the problem of parameter space explosion, data sparseness problem and system resource exhaustion problem. From another point of view, this paper relaxes the other Hmm hypothesis, named stationary (time invariant) hypothesis, makes use of time information and proposes a non-stationary Hmm (NSHmm). This paper describes NSHmm in detail, including its definition, the representation of time information, the algorithms and the parameter space and so on. Moreover, to further reduce the parameter space for mobile applications, this paper proposes a variant form of NSHmm (VNSHmm). Then NSHmm and VNSHmm are applied to two sequence labeling tasks: pos tagging and pinyin-to-character conversion. Experiment results show that compared with Hmm, NSHmm and VNSHmm can greatly reduce the error rate in both of the two tasks, which proves that they have much more predictive power than Hmm does.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Jelinek, F.: Self-Organized Language Modeling for Speech Recognition. In: IEEE ICASSP (1989)
Nagy, G.: At the Frontier of OCR. Processing of IEEE 80(7) (1992)
Xu, Z., Wang, X., Zhang, K., Guan, Y.: A Post Processing Method for Online Handwritten Chinese Character recognition. Journal of Computer Research and Development 36(5) (May 1999)
Brown, P.F., Della Pietra, S.A., Della Pietra, V.J., Mercer, R.L.: The Mathematics of Statistical Machine Translation: Parameter Estimation. Computational Linguistics 19(2) (1992)
Bingquan, L., Xiaolong, W., Yuying, W.: Incorporating Linguistic Rules in Statistical Chinese Language Model for Pinyin-to-Character Conversion. High Technology Letters 7(2), 8–13 (2001)
Manning, C.D., Schutze, H.: Foundation of Statistic Natural Language Processing. The MIT Press, Cambridge (1999)
Brown, P.F., Della Pietra, V.J., deSouza, P.V., Lai, J.C., Mercer, R.L.: Class-based n-gram models of natural language. Computational Linguistics 18(4), 467–479 (1992)
Ghahramani, Z., Jordan, M.: Factorial hidden Markov models. Machine Learning 29 (1997)
Fritsch, J.: ACID/HNN: A framework for hierarchical connectionist acoustic modeling. In: Proc. IEEE ASRU, Santa Barbara (December 1997)
Goodman, J.: A bit of progress in language modeling. Computer Speech and Language, 403–434 (2001)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
JingHui, X., BingQuan, L., XiaoLong, W. (2005). Principles of Non-stationary Hidden Markov Model and Its Applications to Sequence Labeling Task. In: Dale, R., Wong, KF., Su, J., Kwong, O.Y. (eds) Natural Language Processing – IJCNLP 2005. IJCNLP 2005. Lecture Notes in Computer Science(), vol 3651. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11562214_72
Download citation
DOI: https://doi.org/10.1007/11562214_72
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-29172-5
Online ISBN: 978-3-540-31724-1
eBook Packages: Computer ScienceComputer Science (R0)