Abstract
Most of previous methods on question retrieval treat all words as equally important. This paper employs a bidirectional long short-term memory network to predict word salience weight in the question, which is hinted by the word’s matching status in the answer. Our method is trained on a large corpus of natural question-answer pairs, and so it requires no human annotation. We conduct experiments on question retrieval in a cQA dataset. The results show that our model outperforms traditional methods by a wide margin.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Chris, D., Miguel, B., Wang, L., Austin, M., Noah, A.: Transition based dependency parsing with stack long short-term memory. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (2015)
Di, W., Eric, N.: A long short-term memory model for answer sentence selection in question answering. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (2015)
Hua-Ping, Z., Hong-Kui, Y., De-Yi, X., Qun, L.: Hhmm-based chinese lexical analyzer ictclas. In: SIGHAN 2003 Proceedings of the Second SIGHAN Workshop on Chinese Language Processing (2003)
John, C., Elad, H., Yoram, S.: Adaptive sub gradient methods for online learning and stochastic optimization. J. Mach. Learn. Res. 12(7), 257–269 (2011)
Michael, B., Bruce Croft, W.: Discovering key concepts in verbose queries. In: Proceedings of the 31st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (2008)
Michael, B., Metzler, D., Croft, B.: Learning concept importance using a weighted dependence model. In: Proceedings of the Third ACM International Conference on Web Search and Data Mining (2010)
Nal, K., Edward, G., Phil, B.: A convolutional neural network for modelling sentences. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (2014)
Richard, S., et al.: Recursive deep models for semantic compositionality over a sentiment treebank. In: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing (2013)
Sepp, H., Jurgen, S.: Long-short term memory. Neural Compute 9(8), 1735–1780 (1997)
Tomas, M., Kai, G.C., Jeffrey, D.: Efficient estimation of word representations in vector space. In: Workshop at ICLR (2013)
Tomas, M., Martin, K., Lukas, B., Jan, C., Sanjeev, K.: Recurrent neural network based language model. In: INTERSPEECH (2010)
Xin, W., Yuanchao, L., Chengjie, S., Baoxun, W., Xiaolong, W.: Predicting polarities of tweets by composing word embeddings with long short-term memory. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (2015)
Ming, Z.-Y., Chua, T., Cong, G.: Discovering key concepts in verbose queries exploring domain-specific term weight in archived question search. In: Proceedings of the 19th ACM International Conference on Information and Knowledge Management (2010)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Huang, X., Dai, X. (2018). Learning Term Weight with Long Short-Term Memory for Question Retrieval. In: Hong, JF., Su, Q., Wu, JS. (eds) Chinese Lexical Semantics. CLSW 2018. Lecture Notes in Computer Science(), vol 11173. Springer, Cham. https://doi.org/10.1007/978-3-030-04015-4_53
Download citation
DOI: https://doi.org/10.1007/978-3-030-04015-4_53
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-04014-7
Online ISBN: 978-3-030-04015-4
eBook Packages: Computer ScienceComputer Science (R0)