Skip to main content
Log in

RRQA: reconfirmed reader for open-domain question answering

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

In open-domain question answering (QA), the system needs to answer questions from various fields and forms according to given passages. Machine reading comprehension (MRC) can assist the system in comprehending passages and questions, hence often used in improving the performance of the QA system. More passages will bring more features to the result of comprehension, making the reading result more accurate. However, there is still a lack of effective ways when facing multiple passages because it is difficult to use the effective information brought by multiple passages while processing noisy information. This study introduces a new MRC model called RRQA (Reconfirmed Reader for Open-Domain Question Answering) to integrate the QA scenario into traditional MRC, which focuses on the interaction between questions and passages, considering the noisy information. The proposed model consists of two parts: answer extraction and answer verification. In the part of answer extraction, a multi-layer neural network model extracts candidate answers from each passage according to the question, taking into account that the question may have no answer. In the answer verification part, the final answer is verified by combining the features of all candidate answers, which can reduce the incompleteness of the independent answer extraction. Experiments show that the model RRQA outperforms the state-of-the-art models on the WebQA dataset.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

References

  1. Cui W et al (2017) KBQA: Learning Question answering over QA corpora and knowledge bases. Proc VLDB Endowment 10(5):565–576

    Article  Google Scholar 

  2. Liu S et al (2019) Neural machine reading comprehension: methods and trends. Appl Sci 9 (18):3698

    Article  Google Scholar 

  3. Reddy S et al (2019) CoQA: a conversational question answering challenge. Trans Association Comput Linguist 7:249–266

    Article  MathSciNet  Google Scholar 

  4. Xu Y et al (2020) Topic-aware multi-turn dialogue modeling. AAAI:14176–14184

  5. Seo M J et al (2016) Bidirectional attention flow for machine comprehension ICLR (poster)

  6. Huang H-Y et al (2017) Fusionnet: fusing via fully-aware attention with application to machine comprehension international conference on learning representations

  7. Zhang Z et al (2020) Retrospective reader for machine reading comprehension. AAAI:14506–14514

  8. Rajpurkar P et al (2018) Know what you don’t know: unanswerable questions for SQuAD. Proc 56th Annual Meet Association Comput Linguist (Volume 2: Short Papers) 2:784– 789

    Article  Google Scholar 

  9. He W et al (2018) Dureader: a chinese machine reading comprehension dataset from real-world applications. Proc Workshop Mach Read Question Answer:37–46

  10. Dhingra B et al (2017) Gated-Attention Readers for text comprehension

  11. Wang W et al (2017) Gated self-matching networks for reading comprehension and question answering. Proc 55th Annual Meeting of the Association Comput Linguist (Volume 1: Long Papers) 1:189–198

    Article  Google Scholar 

  12. Cui Y et al (2019) Pre-training with whole word masking for chinese BERT. arXiv:1906.08101

  13. Vaswani A et al (2017) Attention is all you need. Proc 31st Int Conf Neural Inf Process Syst 30:5998–6008

    Google Scholar 

  14. Anusha N et al (2018) Neural network dynamics for model-based deep reinforcement learning with model-free fine-tuning. In: 2018 IEEE international conference on robotics and automation (ICRA), pp 7559–7566

  15. Li P et al (2016) Dataset and neural recurrent sequence labeling model for Open-Domain factoid question answering. arXiv:1607.06275

  16. Vinyals O et al (2015) Pointer networks. In: NIPS’15 proceedings of the 28th international conference on neural information processing systems - Volume 2, vol 28, pp 2692–2700

  17. Weissenborn D et al (2017) Making neural QA as simple as possible but not simpler. In: Proceedings of the 21st conference on computational natural language learning (CoNLL 2017), pp 271–280

  18. Devlin J et al (2018) BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 conference of the north American chapter of the association for computational linguistics: human language technologies, (Long and Short Papers), vol 1, pp 4171–4186

  19. Liu Y et al (2019) RoBERTa: a robustly optimized BERT pretraining approach. arXiv:1907.11692

  20. Lan Z et al (2020) ALBERT: a lite BERT for self-Supervised learning of language representations ICLR 2020: eighth international conference on learning representations

  21. Yang Z et al (2019) XLNEt: generalized autoregressive pretraining for language understanding. Adv Neural Inf Process Syst 32:5753–5763

    Google Scholar 

  22. Bordes A et al (2015) Large-scale simple question answering with memory networks. arXiv:1506.02075

  23. Weissenborn D et al (2017) Making neural QA as simple as possible but not simpler. In: Proceedings of the 21st conference on computational natural language learning (CoNLL 2017), pp 271–280

  24. Chen D et al (2017) Reading wikipedia to answer Open-Domain questions. Proceedings of the 55th annual meeting of the association for computational linguistics (Volume 1: Long Papers), vol 1, pp 1870–1879

  25. Tan C et al (2017) S-Net: from answer extraction to answer generation for machine reading comprehension. AAAI:5940–5947

  26. Zhuang Y, Wang H (2019) Token-level dynamic self-attention network for multi-passage reading comprehension. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 2252–2262

  27. Cong Y, Wu Y, Liang X et al (2021) PH-Model: enhancing multi-passage machine reading comprehension with passage reranking and hierarchical information. Appl Intell 51:5440–5452

    Article  Google Scholar 

  28. Wang Y et al (2018) Multi-passage machine reading comprehension with cross-passage answer verification

  29. Wang S et al (2017) Evidence aggregation for answer Re-Ranking in Open-Domain question answering. In: International conference on learning representations, p 1

  30. Hu M et al (2019) Retrieve, read, rerank: towards end-to-end multi-document reading comprehension. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 2285–2295

  31. Choi E et al (2017) Coarse-to-fine question answering for long documents. In: Proceedings of the 55th annual meeting of the association for computational linguistics (Volume 1: Long Papers), vol 1, pp 209–220

  32. Wang Z et al (2018) Joint training of candidate extraction and answer selection for reading comprehension. In: Proceedings of the 56th annual meeting of the association for computational linguistics (Volume 1: Long Papers), vol 1, pp 1715–1724

  33. Pennington J et al (2014) Glove: global vectors for word representation. In: Proceedings of the 2014 Conference on empirical methods in natural language processing (EMNLP), pp 1532–1543

  34. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780

    Article  Google Scholar 

  35. Wang J et al (2020) SRQA: synthetic reader for factoid question answering. Knowl Based Syst 193:105415

    Article  Google Scholar 

  36. N.L.C. Group (2017) R-NET: machine reading comprehension with self-matching networks. In: Proceedings of the 55th annual meeting of the association for computational linguistics (Volume 1: Long Papers), vol 1, pp 189–198

  37. Lin CY (2004) Rouge: a package for automatic evaluation of summaries

Download references

Acknowledgements

This work is supported by the Fundamental Research Funds for the Central Universities (Grant No. 2572019BH03).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Wenqian Zhang.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Li, S., Zhang, W. RRQA: reconfirmed reader for open-domain question answering. Appl Intell 53, 18420–18430 (2023). https://doi.org/10.1007/s10489-023-04461-3

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-023-04461-3

Keywords

Navigation