Abstract
Answer selection in community question answering (cQA) is a common task in natural language processing. Recent progress focuses on not only pure question-answer (QA) match but also support answers [4]. In this paper, we argue that the performance can drop dramatically if noisy support answers are selected. To tackle the above issue, we propose a novel way to leverage the contributions of support answers: the match scores which are firstly normalized by the correlations between the question and the corresponding similar questions, such that the negative effect from the noisy answers can be reduced. The model applies word-to-word attention to improve QA match and employs cosine similarity as the normalization factor for support answers. Compared with previous work, experiments on the Yahoo! Answers L4 dataset show that our model achieves superior P@1 and MRR results.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
References
Qiu, X., Huang, X.: Convolutional neural tensor network architecture for community-based question answering. In: IJCAI, pp. 1305–1311 (2015)
Lu, Z., Li, H.: A deep architecture for match short texts. In: Advances in Neural Information Processing Systems, pp. 1367–1375 (2013)
Tan, M., dos Santos, C., Xiang, B., Zhou, B.: Improved representation learning for question answer match. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (2016)
Pang, L., Lan, Y., Guo, J., Xu, J., Cheng, X.: SPAN: understanding a question with its support answers. In: AAAI, pp. 4250–4251 (2016)
Yih, W.T., Chang, M.W., Meek, C., Pastusiak, A.: Question Answering Using Enhanced Lexical Semantic Models (2013)
Hou, Y., Tan, C., Wang, X., Zhang, Y., Xu, J., Chen, Q.: HITSZICRC: exploiting classification approach for answer selection in community question answering. In: Proceedings of the 9th International Workshop on Semantic Evaluation, SemEval, vol. 15, pp. 196–202 (2015)
Wang, M., Manning, C.D.: Probabilistic tree-edit models with structured latent variables for textual entailment and question answering. In: Proceedings of the 23rd International Conference on Computational Linguistics, pp. 1164–1172. Association for Computational Linguistics (2010)
Yao, X., Van Durme, B., Callison-Burch, C., Clark, P.: Answer extraction as sequence tagging with tree edit distance. In: HLT-NAACL, pp. 858–867. Citeseer (2013)
Severyn, A., Moschitti, A.: Automatic feature engineering for answer selection and extraction. In: EMNLP, vol. 13, pp. 458–467 (2013)
Kalchbrenner, N., Grefenstette, E., Blunsom, P.: A convolutional neural network for modelling sentences. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (2014)
Kim, Y.: Convolutional neural networks for sentence classification. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1746–1751. Association for Computational Linguistics (2014)
Donahue, J., Anne Hendricks, L., Guadarrama, S., Rohrbach, M., Venugopalan, S., Saenko, K., Darrell, T.: Long-term recurrent convolutional networks for visual recognition and description. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2625–2634 (2015)
Sainath, T.N., Vinyals, O., Senior, A., Sak, H.: Convolutional, long short-term memory, fully connected deep neural networks. In: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4580–4584. IEEE (2015)
Xiang, Y., Zhou, X., Chen, Q., Zheng, Z., Wang, X., Qin, Y.: Incorporating label dependency for answer quality tagging in community question answering via CNN-LSTM-CRF. In: COLING (2016)
Hu, B., Lu, Z., Li, H., Chen, Q.: Convolutional neural network architectures for match natural language sentences. In: Advances in Neural Information Processing Systems, pp. 2042–2050 (2014)
Wang, D., Nyberg, E.: A long short-term memory model for answer sentence selection in question answering. In: Meeting of the Association for Computational Linguistics and the International Joint Conference on Natural Language Processing, pp. 707–712 (2015)
Feng, M., Xiang, B., Glass, M.R., Wang, L., Zhou, B.: Applying deep learning to answer selection: a study and an open task. In: 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU), pp. 813–820. IEEE (2015)
Yu, L., Hermann, K.M., Blunsom, P., Pulman, S.: Deep learning for answer sentence selection. arXiv preprint arXiv:1412.1632 (2014)
Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014)
Hermann, K.M., Kocisky, T., Grefenstette, E., Espeholt, L., Kay, W., Suleyman, M., Blunsom, P.: Teaching machines to read and comprehend. In: Advances in Neural Information Processing Systems, pp. 1693–1701 (2015)
Rocktäschel, T., Grefenstette, E., Hermann, K.M., Kociskỳ, T., Blunsom, P.: Reasoning about entailment with neural attention. arXiv preprint arXiv:1509.06664 (2015)
Cui, Y., Chen, Z., Wei, S., Wang, S., Liu, T., Hu, G.: Attention-over-attention neural networks for reading comprehension. arXiv preprint arXiv:1607.04423 (2016)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Graves, A., Jaitly, N., Mohamed, A.R.: Hybrid speech recognition with deep bidirectional LSTM. In: 2013 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU), pp. 273–278. IEEE (2013)
Robertson, S., Zaragoza, H.: The probabilistic relevance framework: BM25 and beyond. Found. Trends Inf. Retrieval 3(4), 333–389 (2009)
Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., Corrado, G.S., Davis, A., Dean, J., Devin, M., et al.: TensorFlow: large-scale machine learning on heterogeneous distributed systems. arXiv preprint arXiv:1603.04467 (2016)
Pennington, J., Socher, R., Manning, C.D.: Glove: global vectors for word representation. In: EMNLP, vol. 14, pp. 1532–1543 (2014)
Kingma, D., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
Acknowledgments
This paper is supported by the National High Technology Development 863 Program of China (No. 2015AA015405) and the Maker Special Fund of Shenzhen (No. GRCK20160 82611002620). We thank the reviewers for their constructive suggestions on this paper.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG
About this paper
Cite this paper
Zheng, Z. et al. (2018). Answer Selection in Community Question Answering by Normalizing Support Answers. In: Huang, X., Jiang, J., Zhao, D., Feng, Y., Hong, Y. (eds) Natural Language Processing and Chinese Computing. NLPCC 2017. Lecture Notes in Computer Science(), vol 10619. Springer, Cham. https://doi.org/10.1007/978-3-319-73618-1_57
Download citation
DOI: https://doi.org/10.1007/978-3-319-73618-1_57
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-73617-4
Online ISBN: 978-3-319-73618-1
eBook Packages: Computer ScienceComputer Science (R0)