Abstract
In open-domain question answering (QA), the system needs to answer questions from various fields and forms according to given passages. Machine reading comprehension (MRC) can assist the system in comprehending passages and questions, hence often used in improving the performance of the QA system. More passages will bring more features to the result of comprehension, making the reading result more accurate. However, there is still a lack of effective ways when facing multiple passages because it is difficult to use the effective information brought by multiple passages while processing noisy information. This study introduces a new MRC model called RRQA (Reconfirmed Reader for Open-Domain Question Answering) to integrate the QA scenario into traditional MRC, which focuses on the interaction between questions and passages, considering the noisy information. The proposed model consists of two parts: answer extraction and answer verification. In the part of answer extraction, a multi-layer neural network model extracts candidate answers from each passage according to the question, taking into account that the question may have no answer. In the answer verification part, the final answer is verified by combining the features of all candidate answers, which can reduce the incompleteness of the independent answer extraction. Experiments show that the model RRQA outperforms the state-of-the-art models on the WebQA dataset.
Similar content being viewed by others
References
Cui W et al (2017) KBQA: Learning Question answering over QA corpora and knowledge bases. Proc VLDB Endowment 10(5):565–576
Liu S et al (2019) Neural machine reading comprehension: methods and trends. Appl Sci 9 (18):3698
Reddy S et al (2019) CoQA: a conversational question answering challenge. Trans Association Comput Linguist 7:249–266
Xu Y et al (2020) Topic-aware multi-turn dialogue modeling. AAAI:14176–14184
Seo M J et al (2016) Bidirectional attention flow for machine comprehension ICLR (poster)
Huang H-Y et al (2017) Fusionnet: fusing via fully-aware attention with application to machine comprehension international conference on learning representations
Zhang Z et al (2020) Retrospective reader for machine reading comprehension. AAAI:14506–14514
Rajpurkar P et al (2018) Know what you don’t know: unanswerable questions for SQuAD. Proc 56th Annual Meet Association Comput Linguist (Volume 2: Short Papers) 2:784– 789
He W et al (2018) Dureader: a chinese machine reading comprehension dataset from real-world applications. Proc Workshop Mach Read Question Answer:37–46
Dhingra B et al (2017) Gated-Attention Readers for text comprehension
Wang W et al (2017) Gated self-matching networks for reading comprehension and question answering. Proc 55th Annual Meeting of the Association Comput Linguist (Volume 1: Long Papers) 1:189–198
Cui Y et al (2019) Pre-training with whole word masking for chinese BERT. arXiv:1906.08101
Vaswani A et al (2017) Attention is all you need. Proc 31st Int Conf Neural Inf Process Syst 30:5998–6008
Anusha N et al (2018) Neural network dynamics for model-based deep reinforcement learning with model-free fine-tuning. In: 2018 IEEE international conference on robotics and automation (ICRA), pp 7559–7566
Li P et al (2016) Dataset and neural recurrent sequence labeling model for Open-Domain factoid question answering. arXiv:1607.06275
Vinyals O et al (2015) Pointer networks. In: NIPS’15 proceedings of the 28th international conference on neural information processing systems - Volume 2, vol 28, pp 2692–2700
Weissenborn D et al (2017) Making neural QA as simple as possible but not simpler. In: Proceedings of the 21st conference on computational natural language learning (CoNLL 2017), pp 271–280
Devlin J et al (2018) BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 conference of the north American chapter of the association for computational linguistics: human language technologies, (Long and Short Papers), vol 1, pp 4171–4186
Liu Y et al (2019) RoBERTa: a robustly optimized BERT pretraining approach. arXiv:1907.11692
Lan Z et al (2020) ALBERT: a lite BERT for self-Supervised learning of language representations ICLR 2020: eighth international conference on learning representations
Yang Z et al (2019) XLNEt: generalized autoregressive pretraining for language understanding. Adv Neural Inf Process Syst 32:5753–5763
Bordes A et al (2015) Large-scale simple question answering with memory networks. arXiv:1506.02075
Weissenborn D et al (2017) Making neural QA as simple as possible but not simpler. In: Proceedings of the 21st conference on computational natural language learning (CoNLL 2017), pp 271–280
Chen D et al (2017) Reading wikipedia to answer Open-Domain questions. Proceedings of the 55th annual meeting of the association for computational linguistics (Volume 1: Long Papers), vol 1, pp 1870–1879
Tan C et al (2017) S-Net: from answer extraction to answer generation for machine reading comprehension. AAAI:5940–5947
Zhuang Y, Wang H (2019) Token-level dynamic self-attention network for multi-passage reading comprehension. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 2252–2262
Cong Y, Wu Y, Liang X et al (2021) PH-Model: enhancing multi-passage machine reading comprehension with passage reranking and hierarchical information. Appl Intell 51:5440–5452
Wang Y et al (2018) Multi-passage machine reading comprehension with cross-passage answer verification
Wang S et al (2017) Evidence aggregation for answer Re-Ranking in Open-Domain question answering. In: International conference on learning representations, p 1
Hu M et al (2019) Retrieve, read, rerank: towards end-to-end multi-document reading comprehension. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 2285–2295
Choi E et al (2017) Coarse-to-fine question answering for long documents. In: Proceedings of the 55th annual meeting of the association for computational linguistics (Volume 1: Long Papers), vol 1, pp 209–220
Wang Z et al (2018) Joint training of candidate extraction and answer selection for reading comprehension. In: Proceedings of the 56th annual meeting of the association for computational linguistics (Volume 1: Long Papers), vol 1, pp 1715–1724
Pennington J et al (2014) Glove: global vectors for word representation. In: Proceedings of the 2014 Conference on empirical methods in natural language processing (EMNLP), pp 1532–1543
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Wang J et al (2020) SRQA: synthetic reader for factoid question answering. Knowl Based Syst 193:105415
N.L.C. Group (2017) R-NET: machine reading comprehension with self-matching networks. In: Proceedings of the 55th annual meeting of the association for computational linguistics (Volume 1: Long Papers), vol 1, pp 189–198
Lin CY (2004) Rouge: a package for automatic evaluation of summaries
Acknowledgements
This work is supported by the Fundamental Research Funds for the Central Universities (Grant No. 2572019BH03).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Li, S., Zhang, W. RRQA: reconfirmed reader for open-domain question answering. Appl Intell 53, 18420–18430 (2023). https://doi.org/10.1007/s10489-023-04461-3
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-023-04461-3