Abstract
We have proposed previously a boosted maximum entropy model to overcome three major problems in applying the maximum entropy models to text chunking [1]: (i) feature selection, (ii) high computational complexity, and (iii) highly-imbalanced data. To cope with the first problem, the boosted ME models adopt a decision tree as a constructor of the high-order features. Because decision trees can be represented as a set of if-then rules, the features for ME models are automatically constructed by transforming a decision tree into if-then rules. Active learning is adopted to solve the high computational complexity, and the AdaBoost is used to overcome the highly imbalance in natural language resources. In this paper, we apply the boosted maximum entropy model to two major tasks in natural language processing: POS tagging and PP attachment.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Park, S.-B., Zhang, B.-T.: A Boosted Maximum Entropy Model for Learning Text Chunking. In: Proceedings of ICML 2002, pp. 482–489 (2002)
Ratnaparkhi, A.: A Maximum Entropy Model for Part-of-speech Tagging. In: Proceedings of EMNLP 1996, pp. 133–142 (1996)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2004 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Park, SB., O, J., Lee, SJ. (2004). Part-of-Speech Tagging and PP Attachment Disambiguation Using a Boosted Maximum Entropy Model. In: Zhang, C., W. Guesgen, H., Yeap, WK. (eds) PRICAI 2004: Trends in Artificial Intelligence. PRICAI 2004. Lecture Notes in Computer Science(), vol 3157. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-28633-2_99
Download citation
DOI: https://doi.org/10.1007/978-3-540-28633-2_99
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-22817-2
Online ISBN: 978-3-540-28633-2
eBook Packages: Springer Book Archive