Skip to main content
Log in

Deep learning-based question answering: a survey

  • Survey Paper
  • Published:
Knowledge and Information Systems Aims and scope Submit manuscript

Abstract

Question Answering is a crucial natural language processing task. This field of research has attracted a sudden amount of interest lately due mainly to the integration of the deep learning models in the Question Answering Systems which consequently power up many advancements and improvements. This survey aims to explore and shed light upon the recent and most powerful deep learning-based Question Answering Systems and classify them based on the deep learning model used, stating the details of the used word representation, datasets, and evaluation metrics. It aims to highlight and discuss the currently used models and give insights that direct future research to enhance this increasingly growing field.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Notes

  1. https://github.com/tmikolov/word2vec, Date of Access: 5th Jun, 2022.

  2. https://github.com/stanfordnlp/GloVe, Date of Access: 5th Jun, 2022.

  3. Download link: https://trec.nist.gov/data/qa.html, Date of Access: 5th Jun, 2022.

  4. The Text REtrieval Conference (TREC) is a series of workshops that provides the needed infrastructure for text retrieval methodologies with large-scale evaluation.

  5. Download link: https://download.microsoft.com/download/E/5/F/E5FCFCEE-7005-4814-853D-DAA7C66507E0/WikiQACorpus.zip, Date of Access: 5th Jun, 2022.

  6. Download link: https://github.com/shuzi/insuranceQA, Date of Access: 5th Jun, 2022.

  7. Download link: https://ciir.cs.umass.edu/downloads/wikipassageqa/WikiPassageQA.zip, Date of Access: 5th Jun, 2022.

  8. Download link: https://deepai.org/dataset/squad, Date of Access: 5th Jun, 2022.

  9. Download link: https://github.com/abisee/cnn-dailymail, Date of Access: 5th Jun, 2022.

  10. Download link: https://research.fb.com/downloads/babi/, Date of Access: 5th Jun, 2022.

  11. Download link: https://www.cs.cmu.edu/~glai1/data/race/, Date of Access: 5th Jun, 2022.

  12. Download link: https://sheng-z.github.io/ReCoRD-explorer/, Date of Access: 5th Jun, 2022.

  13. Download link: http://nlp.cs.washington.edu/triviaqa/, Date of Access: 5th Jun, 2022.

  14. Download link: https://github.com/deepmind/narrativeqa, Date of Access: 5th Jun, 2022.

  15. Download link: https://github.com/deepmind/narrativeqa, Date of Access: 5th Jun, 2022.

  16. Download link: https://hotpotqa.github.io/, Date of Access: 5th Jun, 2022.

  17. Download link: https://microsoft.github.io/msmarco/, Date of Access: 5th Jun, 2022.

  18. Download link: https://quac.ai/, Date of Access: 5th Jun, 2022.

References

  1. Toshevska M, Mirceva G, Jovanov M (2020) Question answering with deep learning: a survey. Faculty of Computer Science and Engineering Ss Cyril and Methodius University Skopje, Macedonia

  2. Srba I, Bielikova M (2016) A comprehensive survey and classification of approaches for community question answering. ACM Trans Web 10:1–63. https://doi.org/10.1145/2934687

    Article  Google Scholar 

  3. (2019) A survey on machine reading comprehension. J Beijing Univ Posts Telecommun 42:1

  4. Huang Z, Xu S, Hu M et al (2020) Recent trends in deep learning based open-domain textual question answering systems. IEEE Access 8:94341–94356. https://doi.org/10.1109/ACCESS.2020.2988903

    Article  Google Scholar 

  5. Palasundram K, Mohd Sharef N, Kasmiran KA, Azman A (2020) Enhancements to the sequence-to-sequence-based natural answer generation models. IEEE Access 8:45738–45752. https://doi.org/10.1109/ACCESS.2020.2978551

    Article  Google Scholar 

  6. Abbasiantaeb Z, Momtazi S (2020) Text-based question answering from information retrieval and deep neural network perspectives: a survey. Wiley Interdiscip Rev: Data Min Knowl Discov 11:e1412. https://doi.org/10.1002/widm.1412

    Article  Google Scholar 

  7. Li H (2018) Deep learning for natural language processing: advantages and challenges. Natl Sci Rev 5:24–26. https://doi.org/10.1093/nsr/nwx110

    Article  Google Scholar 

  8. Xiang Y, Chen Q, Wang X, Qin Y (2017) Answer selection in community question answering via attentive neural networks. IEEE Signal Process Lett 24:505–509. https://doi.org/10.1109/LSP.2017.2673123

    Article  Google Scholar 

  9. Otter DW, Medina JR, Kalita JK (2020) A survey of the usages of deep learning in natural language processing. IEEE Trans Neural Netw Learn Syst 32:604–624

    Article  Google Scholar 

  10. Vanitha G, Sanampudi SK, Guda V (2011) Approaches for question answering systems. Int J Eng Sci Technol (IJEST) 3:990–995

    Google Scholar 

  11. Riloff E, Wiebe J (2003) Learning extraction patterns for subjective expressions. In: Proceedings of the 2003 conference on empirical methods in natural language processing (EMNLP), pp 105–112

  12. Riloff E, Thelen M (2020) A rule-based question answering system for reading comprehension tests. In: ANLP-NAACL 2000 workshop: reading comprehension tests as evaluation for computer-based language understanding systems

  13. Echihabi A, Marcu D (2003) A noisy-channel approach to question answering. In: Association for computational linguistics (ACL)

  14. Heie MH, Whittaker EWD, Furui S (2012) Question answering using statistical language modelling. Comput Speech Lang 26:193–209. https://doi.org/10.1016/j.csl.2011.11.001

    Article  Google Scholar 

  15. Wang M, Smith NA, Mitamura T (2007) What is the jeopardy model? A quasi-synchronous grammar for QA. In: Proceedings of the 2007 joint conference on empirical methods in natural language processing and computational natural language learning (EMNLP-CoNLL), pp 22–32

  16. Choi E, Hewlett D, Uszkoreit J et al (2017) Coarse-to-fine question answering for long documents. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 209–220

  17. Reddy S, Chen D, Manning CD (2019) CoQA: a conversational question answering challenge. Trans Assoc Comput Linguist 7:249–266. https://doi.org/10.1162/tacl_a_00266

    Article  Google Scholar 

  18. Hinton GE, Osindero S, Teh Y-W (2006) A fast learning algorithm for deep belief nets. Neural Comput 18:1527–1554. https://doi.org/10.1162/neco.2006.18.7.1527

    Article  MathSciNet  MATH  Google Scholar 

  19. Severyn A, Moschittiy A (2015) Learning to rank short text pairs with convolutional deep neural networks. In: SIGIR 2015—proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval. Association for Computing Machinery, Inc, pp 373–382

  20. Miller GA (1995) WordNet. Commun ACM 38:39–41. https://doi.org/10.1145/219717.219748

    Article  Google Scholar 

  21. Akbik A, Blythe D, Vollgraf R (2018) Contextual string embeddings for sequence labeling. In: COLING 2018, 27th international conference on computational linguistics, pp 1638–1649

  22. Adhikari A, Ram A, Tang R, Lin J (2019) DocBERT: BERT for document classification. CoRR. arXiv:1904.08398

  23. Zhang H, Xu J, Wang J (2019) Pretraining-based natural language generation for text summarization. In: Proceedings of the 23rd conference on computational natural language learning (CoNLL). Association for Computational Linguistics, pp 789–797

  24. Zhou C, Neubig G, Gu J (2019) Understanding knowledge distillation in non-autoregressive machine translation. In: Proceedings of the 2019 international conference on learning representations

  25. Mikolov T, Chen K, Corrado G, Dean J (2013) Distributed representations of words and phrases and their compositionality. In: Advances in neural information processing systems (NeurIPS), pp 3111–3119

  26. Pennington J, Socher R, Manning CD (2014) GloVe: global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 1532–1543

  27. Devlin J, Chang M-W, Lee K et al (2019) BERT: pre-training of deep bidirectional transformers for language understanding. In: Annual conference of the North American chapter of the association for computational linguistics (NAACL)

  28. Peters M, Neumann M, Iyyer M et al (2018) Deep contextualized word representations. In: Proceedings of the 2018 conference of the North American Chapter of the association for computational linguistics: human language technologies, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 2227–2237

  29. Yamada I, Asai A, Shindo H et al (2020) LUKE: deep contextualized entity representations with entity-aware self-attention. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), pp 6442–6454

  30. Clark K, Luong M-T, Le QV, Manning CD (2020) ELECTRA: Pre-training text encoders as discriminators rather than generators. In: International conference on learning representations (ICLR)

  31. Yang Z, Dai Z, Yang Y et al (2019) XLNet: generalized autoregressive pretraining for language understanding. In: Wallach H, Larochelle H, Beygelzimer A et al (eds) Advances in neural information processing systems. Curran Associates Inc, Red Hook, NY

    Google Scholar 

  32. Zaheer M, Guruganesh G, Dubey KA et al (2020) Big bird: transformers for longer sequences. In: Larochelle H, Ranzato M, Hadsell R et al (eds) Advances in neural information processing systems. Curran Associates Inc, Red Hook, NY, pp 17283–17297

    Google Scholar 

  33. Jun C, Jang H, Sim M et al (2022) ANNA: enhanced language representation for question answering. In: Proceedings of the 7th workshop on representation learning for NLP. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 121–132

  34. Vaswani A, Shazeer N, Parmar N et al (2017) Attention is all you need. In: Advances in neural information processing systems, pp 5998–6008

  35. Goldberg Y (2019) Assessing BERT’s syntactic abilities. CoRR. arXiv:1901.05287

  36. Lan Z, Chen M, Goodman S et al (2019) ALBERT: a lite BERT for self-supervised learning of language representations. In: International conference on learning representations (ICLR)

  37. Sanh V, Debut L, Chaumond J, Wolf T (2019) DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. CoRR. arXiv:1910.01108

  38. Wang W, Bi B, Yan M et al (2020) StructBERT: incorporating language structures into pre-training for deep language understanding. In: 8th international conference on learning representations (ICLR)

  39. Liu Y, Ott M, Goyal N et al (2019) RoBERTa: a robustly optimized BERT pretraining approach. CoRR. arXiv:1907.11692

  40. He P, Liu X, Gao J, Chen W (2021) DeBERTa: decoding-enhanced BERT with Disentangled Attention. In: Proceedings of the 9th international conference on learning representations (ICLR)

  41. Jiang Z-H, Yu W, Zhou D et al (2020) ConvBERT: improving BERT with span-based dynamic convolution. In: Larochelle H, Ranzato M, Hadsell R et al (eds) Advances in neural information processing systems. Curran Associates Inc, Red Hook, NY, pp 12837–12848

    Google Scholar 

  42. Papineni K, Roukos S, Ward T, Zhu W-J (2002) BLEU: a method for automatic evaluation of machine translation. In: Proceedings of the 40th annual meeting of the association for computational linguistics

  43. Banerjee S, Lavie A (2005) Meteor: an automatic metric for MT evaluation with improved correlation with human judgments. In: Proceedings of the ACL workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization

  44. Lin C-Y (2004) ROUGE: a package for automatic evaluation of summaries. In: Text summarization branches out: proceedings of the 2004 association for computational linguistics (ACL-04) Workshop, Barcelona, Spain, pp 74–81

  45. Zhang T, Kishore V, Wu F et al (2020) BERTScore: evaluating text generation with bert. In: Proceedings of the international conference on learning representations (ICLR)

  46. Lee H, Yoon S, Dernoncourt F et al (2021) KPQA: a metric for generative question answering using keyphrase weights. In: Proceedings of annual conference of the North American chapter of the association for computational linguistics (NAACL), pp 2105–2115

  47. Feng M, Xiang B, Glass MR et al (2015) Applying deep learning to answer selection: a study and an open task. In: 2015 IEEE workshop on automatic speech recognition and understanding (ASRU), pp 813–820

  48. Rao J, He H, Lin J (2016) Noise-contrastive estimation for answer selection with deep neural networks. In: International conference on information and knowledge management, proceedings. Association for Computing Machinery, pp 1913–1916

  49. Wang Z, Mi H, Ittycheriah A (2016) Sentence similarity learning by lexical decomposition and composition. COLING, Association for Computational Linguistics (ACL), pp 1340–1349

  50. Madabushi HT, Lee M, Barnden J (2018) Integrating question classification and deep learning for improved answer selection. In: Proceedings of the 27th international conference on computational linguistics, pp 3283–3294

  51. Wang Z, Hamza W, Florian R (2017) Bilateral multi-perspective matching for natural language sentences. In: Proceedings of the twenty-sixth international joint conference on artificial intelligence. International Joint Conferences on Artificial Intelligence Organization, California, pp 4144–4150

  52. Tay Y, Phan MC, Tuan LA, Hui SC (2017) Learning to rank question answer pairs with holographic dual LSTM architecture. In: SIGIR 2017—proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval. Association for Computing Machinery, Inc, pp 695–704

  53. Mihaylov T, Kozareva Z, Frank A (2017) Neural skill transfer from supervised language tasks to reading comprehension. Workshop on learning with limited labeled data: weak supervision and beyond at NIPS

  54. di Gennaro G, Buonanno A, di Girolamo A et al (2020) Intent classification in question-answering using LSTM architectures. Progr Artif Intell Neural Syst. https://doi.org/10.1007/978-981-15-5093-5_11

    Article  Google Scholar 

  55. Zhang L, Lin C, Zhou D et al (2021) A Bayesian end-to-end model with estimated uncertainties for simple question answering over knowledge bases. Comput Speech Lang 66:101167. https://doi.org/10.1016/j.csl.2020.101167

    Article  Google Scholar 

  56. Hu M, Peng Y, Wei F et al (2018) Attention-guided answer distillation for machine reading comprehension. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 2077–2086

  57. Ran Q, Li P, Hu W and Zhou J (2019) Option comparison network for multiple-choice reading comprehension. CoRR. arXiv arXiv:1903.03033

  58. Yang A, Wang Q, Liu J et al (2019) Enhancing pre-trained language representations with rich knowledge for machine reading comprehension. Association for Computational Linguistics, Stroudsburg, PA

    Book  Google Scholar 

  59. Shoeybi M, Patwary M, Puri R et al (2019) Megatron-LM: training multi-billion parameter language models using model parallelism. CoRR. arXiv:1909.08053

  60. Garg S, Vu T, Moschitti A (2020) TANDA: transfer and adapt pre-trained transformer models for answer sentence selection. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 7780–7788.https://doi.org/10.1609/aaai.v34i05.6282

  61. Zhu P, Zhang Z, Zhao H, Li X (2022) DUMA: reading comprehension with transposition thinking. IEEE/ACM Trans Audio Speech Lang Process 30:269–279. https://doi.org/10.1109/TASLP.2021.3138683

    Article  Google Scholar 

  62. Guu K, Lee K, Tung Z et al (2020) Retrieval augmented language model pre-training. In: III HD, Singh A (eds) Proceedings of the 37th international conference on machine learning. PMLR, pp 3929–3938

  63. Wu B, Zhang Z, Zhao H (2021) Graph-free multi-hop reading comprehension: a select-to-guide strategy. CoRR. arXiv:2107.11823

  64. Li X-Y, Lei W-J, Yang Y-B (2022) From easy to hard: two-stage selector and reader for multi-hop question answering. CoRR. arXiv:2205.11729

  65. Guan Y, Li Z, Leng J et al (2021) Block-skim: efficient question answering for transformer. CoRR. arXiv:2112.08560

  66. Zhou X, Hu B, Chen Q, Wang X (2018) Recurrent convolutional neural network for answer selection in community question answering. Neurocomputing 274:8–18. https://doi.org/10.1016/j.neucom.2016.07.082

    Article  Google Scholar 

  67. Cohen D, Yang L, Croft WB (2018) WikiPassageQA: a benchmark collection for research on non-factoid answer passage retrieval. In: 41st international ACM SIGIR conference on research and development in information retrieval, SIGIR 2018. Association for Computing Machinery, Inc, pp 1165–1168

  68. Zhang X, Li S, Sha L, Wang H (2017) Attentive interactive neural networks for answer selection in community question answering. In: Proceedings of the AAAI conference on artificial intelligence, vol 31, no 1

  69. Bian W, Li S, Yang Z et al (2017) A compare-aggregate model with dynamic-clip attention for answer selection. In: International conference on information and knowledge management, proceedings. Association for Computing Machinery, pp 1987–1990

  70. Yoon S, Dernoncourt F, Kim DS et al (2019) A compare-aggregate model with latent clustering for answer selection. In: Proceedings of the 28th ACM international conference on information and knowledge management, pp 2093–2096

  71. Peng Y, Liu B (2018) Attention-based neural network for short-text question answering. In: ACM International conference proceeding series. Association for Computing Machinery, pp 21–26

  72. Yu AW, Dohan D, Luong M-T et al (2018) QANet: combining local convolution with global self-attention for reading comprehension. CoRR. arXiv:1804.09541

  73. Miller A, Fisch A, Dodge J et al (2016) Key-value memory networks for directly reading documents. In: Proceedings of the 2016 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 1400–1409

  74. Yang L, Ai Q, Guo J, Croft WB (2016) aNMM: ranking short answer texts with attention-based neural matching model. In: International conference on information and knowledge management, proceedings. Association for Computing Machinery, pp 287–296

  75. Shao T, Guo Y, Chen H, Hao Z (2019) Transformer-based neural network for answer selection in question answering. IEEE Access 7:26146–26156. https://doi.org/10.1109/ACCESS.2019.2900753

    Article  Google Scholar 

  76. Sukhbaatar S, Szlam A, Weston J, Fergus R (2015) End-to-end memory networks. In: Advances in neural information processing systems, pp 2440–2448

  77. Kumar A, Irsoy O, Ondruska P et al (2016) Ask me anything: dynamic memory networks for natural language processing. In: International conference on machine learning, pp 1378–1387

  78. Pan B, Li H, Zhao Z et al (2017) MEMEN: multi-layer embedding with memory networks for machine comprehension. In: AAAI conference on artificial intelligence, (AAAI-18), the 30th innovative applications of artificial intelligence (IAAI-18), and the 8th AAAI symposium on educational advances in artificial intelligence (EAAI-18)

  79. Back S, Yu S, Indurthi SR et al (2018) MemoReader: large-scale reading comprehension through neural memory controller. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 2131–2140

  80. Xiong C, Zhong V, Socher R (2016) Dynamic coattention networks for question answering. In: International conference on learning representations (ICLR)

  81. Wang S, Yu M, Chang S, Jiang J (2018) A co-matching model for multi-choice reading comprehension. In: Association for computational linguistics (ACL), pp 746–751

  82. Xiong C, Zhong V, Socher R (2017) DCN+: mixed objective and deep residual coattention for question answering. CoRR. arXiv:1711.00106

  83. McCann B, Keskar NS, Xiong C, Socher R (2018) The Natural language decathlon: multitask learning as question answering. CoRR. arXiv:1806.08730

  84. Wang W, Yan M, Wu C (2018) Multi-granularity hierarchical attention fusion networks for reading comprehension and question answering. In: Proceedings of the 56th annual meeting of the association for computational linguistics, vol 1: long papers, pp 1705–1714

  85. Tay Y, Tuan LA, Hui SC (2018) Multi-cast attention networks. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery and data mining. ACM, New York, NY, USA, pp 2299–2308

  86. Min S, Seo M, Hajishirzi H (2017) Question answering through transfer learning from large fine-grained supervision Data. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 2: short papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 510–517

  87. Golub D, Huang P-S, He X, Deng L (2017) Two-stage synthesis networks for transfer learning in machine comprehension. In: Proceedings of the 2017 conference on empirical methods in natural language processing. association for computational linguistics, Stroudsburg, PA, USA, pp 835–844

  88. Seo M, Kembhavi A, Farhadi A, Hajishirzi H (2016) Bidirectional attention flow for machine comprehension. In: International conference on learning representations (ICLR)

  89. Liu X, Shen Y, Duh K, Gao J (2018) Stochastic answer networks for machine reading comprehension. In: Proceedings of the 56th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 1694–1704

  90. Xiong W, Yu M, Guo X et al (2019) Simple yet effective bridge reasoning for open-domain multi-hop question answering. In: Proceedings of the 2nd workshop on machine reading for question answering. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 48–52

  91. Hermann KM, Kočiský T, Grefenstette E et al (2015) Teaching machines to read and comprehend. In: Cortes C, Lawrence N, Lee D, Sugiyama M, Garnett R (eds) Advances in neural information processing systems, vol 28. Curran Associates Inc, Red Hook, NY

    Google Scholar 

  92. Kadlec R, Schmid M, Bajgar O, Kleindienst J (2016) Text understanding with the attention sum reader network. In: Proceedings of the 54th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 908–918

  93. Trischler A, Ye Z, Yuan X et al (2016) Natural language comprehension with the EpiReader. In: Proceedings of the 2016 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 128–137

  94. Wang S, Jiang J (2017) Machine comprehension using match-LSTM and answer pointer. In: International conference on learning representations (ICLR), pp 1–15

  95. Cui Y, Chen Z, Wei S et al (2017) Attention-over-attention neural networks for reading comprehension. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 593–602

  96. Dhingra B, Liu H, Yang Z et al (2017) Gated-attention readers for text comprehension. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 1832–1846

  97. Wang W, Yang N, Wei F et al (2017) Gated self-matching networks for reading comprehension and question answering. In: ACL 2017—55th annual meeting of the association for computational linguistics, proceedings of the conference (long papers). Association for Computational Linguistics (ACL), pp 189–198

  98. Liu R, Wei W, Mao W, Chikina M (2017) Phase conductor on multi-layered attentions for machine comprehension. CoRR. arXiv:1710.10504

  99. Huang H-Y, Zhu C, Shen Y, Chen W (2017) FusionNet: fusing via fully-aware attention with application to machine comprehension. CoRR. arXiv:1711.07341

  100. Zhu H, Wei F, Qin B, Liu T (2018) Hierarchical attention flow for multiple-choice reading comprehension. In: Hierarchical Attention flow for multiple-choice reading comprehension, vol 32, no 1

  101. Kundu S, Ng HT (2018) A question-focused multi-factor attention network for question answering. In: Proceedings of the AAAI conference on artificial intelligence, vol 32, no 1

  102. Tan C, Wei F, Yang N et al (2018) S-Net: from answer extraction to answer synthesis for machine reading comprehension. In: Proceedings of the AAAI conference on artificial intelligence, vol 32, no 1

  103. Zhu C, Zeng M, Huang X (2018) SDNet: contextualized attention-based deep network for conversational question answering. CoRR. arXiv:1812.03593

  104. LeeKim HH (2020) GF-Net: improving machine reading comprehension with feature gates. Pattern Recognit Lett 129:8–15. https://doi.org/10.1016/j.patrec.2019.10.030

    Article  Google Scholar 

  105. Huang X, Zhang J, Li D, Li P (2019) Knowledge graph embedding based question answering. In: WSDM 2019—proceedings of the 12th ACM international conference on web search and data mining. Association for Computing Machinery, Inc, pp 105–113

  106. Chen Y, Wu L, Zaki MJ (2020) GraphFlow: exploiting conversation flow with graph neural networks for conversational machine comprehension. In: Proceedings of the twenty-ninth international joint conference on artificial intelligence. International Joint Conferences on Artificial Intelligence Organization, California, pp 1230–1236

  107. Cao X, Liu Y (2022) Coarse-grained decomposition and fine-grained interaction for multi-hop question answering. J Intell Inf Syst 58:21–41. https://doi.org/10.1007/s10844-021-00645-w

    Article  Google Scholar 

  108. Nishida K, Nishida K, Nagata M et al (2019) Answering while summarizing: multi-task learning for multi-hop QA with evidence extraction. In: Proceedings of the 57th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Florence, Italy, pp 2335–2345

  109. Xiao Y, Qu Y, Qiu L et al (2019) dynamically fused graph network for multi-hop reasoning. In: Proceedings of the 57th annual meeting of the association for computational linguistics, Florence, Italy, pp 6140–6150

  110. Cao Y, Fang M, Tao D (2019) BAG: bi-directional attention entity graph convolutional network for multi-hop reasoning question answering. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL): human language technologies, pp 357–362

  111. Fang Y, Sun S, Gan Z et al (2020) Hierarchical graph network for multi-hop question answering. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), pp 8823–8838

  112. Zhang M, Li F, Wang Y et al (2020) Coarse and fine granularity graph reasoning for interpretable multi-hop question answering. IEEE Access 8:56755–56765. https://doi.org/10.1109/ACCESS.2020.2981134

    Article  Google Scholar 

  113. Tu M, Huang K, Wang G et al (2020) Select, answer and explain: interpretable multi-hop reading comprehension over multiple documents. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, no 05, pp 9073–9080

  114. Zhao C, Xiong C, Rosset C et al (2020) Transformer-XH: multi-evidence reasoning with extra hop attention. In: International conference on learning representations (ICLR)

  115. Zhang X, Bosselut A, Yasunaga M et al (2022) GreaseLM: Graph REASoning Enhanced Language Models for question answering. CoRR. arXiv:2201.08860

  116. Shi J, Cao S, Hou L et al (2021) TransferNet: an effective and transparent framework for multi-hop question answering over relation graph. In: Proceedings of the 2021 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 4149–4158

  117. Zhang Y, Nie P, Ramamurthy A, Song L (2021) Answering any-hop open-domain questions with iterative document reranking. In: SIGIR 2021—proceedings of the 44th international ACM SIGIR conference on research and development in information retrieval. Association for Computing Machinery, Inc, pp 481–490

  118. Ren H, Dai H, Dai B et al (2021) LEGO: latent execution-guided reasoning for multi-hop question answering on knowledge graphs. In: International conference on machine learning, pp 8959–8970

  119. Xiong W, Li XL, Iyer S et al (2020) Answering complex open-domain questions with multi-hop dense retrieval. In: Proceedings of the conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing, EMNLP-IJCNLP 2019, pp 2590–2602

  120. Wu J, Mu T, Thiyagalingam J, Goulermas JY (2020) Building interactive sentence-aware representation based on generative language model for community question answering. Neurocomputing 389:93–107. https://doi.org/10.1016/j.neucom.2019.12.107

    Article  Google Scholar 

  121. Bi B, Wu C, Yan M et al (2019) Incorporating external knowledge into machine reading for generative question answering. In: Conference on empirical methods in natural language processing and international joint conference on natural language processing (EMNLP-IJCNLP)

  122. Bauer L, Wang Y, Bansal M (2018) Commonsense for generative multi-hop question answering tasks. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 4220–4230

  123. Izacard G, Grave E (2021) Leveraging passage retrieval with generative models for open domain question answering. In: Proceedings of the 16th conference of the European chapter of the association for computational linguistics: main volume. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 874–880

  124. Yavuz S, Hashimoto K, Zhou Y et al (2022) Modeling multi-hop question answering as single sequence prediction. In: Proceedings of the 60th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 974–990

  125. Shen Y, Huang P sen, Gao J, Chen W (2017) ReasoNet: learning to stop reading in machine comprehension. In: Proceedings of the ACM SIGKDD international conference on knowledge discovery and data mining. Association for Computing Machinery, pp 1047–1055

  126. Buck C, Bulian J, Ciaramita M et al (2018) Ask the right questions: active question reformulation with reinforcement learning. In: International conference on learning representations (ICLR)

  127. Xu Y, Liu J, Gao J et al (2017) Dynamic fusion networks for machine reading comprehension. CoRR. arXiv:1711.04964

  128. Hu M, Peng Y, Huang Z et al (2018) Reinforced mnemonic reader for machine reading comprehension. In: Proceedings of the twenty-seventh international joint conference on artificial intelligence. International Joint Conferences on Artificial Intelligence Organization, California, pp 4099–4106

  129. Santoro A, Raposo D, Barrett DGT et al (2017) A simple neural network module for relational reasoning. In: Guyon I, Von Luxburg U, Bengio S, Wallach H, Fergus R, Vishwanathan S, Garnett R (eds) Advances in neural information processing systems, vol 30. Curran Associates Inc, Red Hook, NY

    Google Scholar 

  130. Swayamdipta S, Parikh AP, Kwiatkowski T (2017) Multi-mention learning for reading comprehension with neural cascades. In: International conference on learning representations (ICLR)

  131. Tay Y, Tuan LA, Hui SC (2018) Hyperbolic representation learning for fast and efficient neural question answering. In: WSDM 2018—proceedings of the 11th ACM international conference on web search and data mining. Association for Computing Machinery, Inc, pp 583–591

  132. Seonwoo Y, Kim J-H, Ha J-W, Oh A (2020) Context-aware answer extraction in question answering. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), pp 2418–2428

  133. Wu Y, Zhao S (2021) Community answer generation based on knowledge graph. Inf Sci 545:132–152. https://doi.org/10.1016/j.ins.2020.07.077

    Article  Google Scholar 

  134. Zhou G, Xie Z, Yu Z, Huang JX (2021) DFM: a parameter-shared deep fused model for knowledge base question answering. Inf Sci 547:103–118. https://doi.org/10.1016/j.ins.2020.08.037

    Article  Google Scholar 

  135. He H, Gimpel K, Lin J (2015) Multi-perspective sentence similarity modeling with convolutional neural networks. In: Proceedings of the 2015 conference on empirical methods in natural language processing. Association for Computational Linguistics, pp 1576–1586

  136. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9:1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735

    Article  Google Scholar 

  137. Schuster M, Paliwal KK (1997) Bidirectional recurrent neural networks. IEEE Trans Signal Process 45:2673–2681. https://doi.org/10.1109/78.650093

    Article  Google Scholar 

  138. Wang S, Jiang J (2017) A compare-aggregate model for matching text sequences. In: Proceedings of the 5th international conference on learning representations (ICLR 2017)

  139. Vinyals O, Fortunato M, Jaitly N (2015) Pointer networks. In: Cortes C, Lawrence N, Lee D, Sugiyama M, Garnett R (eds) Advances in neural information processing systems, vol 28. Curran Associates Inc, Red Hook, NY

    Google Scholar 

  140. Gong Y, Bowman SR (2018) Ruminating reader: reasoning with gated multi-hop attention. In: Proceedings of the workshop on machine reading for question answering. Association for Computational Linguistics, pp 1–11

  141. Liu H, Singh P (2004) ConceptNet—a practical commonsense reasoning tool-kit. BT Technol J 22:211–226. https://doi.org/10.1023/B:BTTJ.0000047600.45421.6d

    Article  Google Scholar 

  142. Yang Y, Yih W-T, Meek C (2015) WIKIQA: a challenge dataset for open-domain question answering. In: Proceedings of the 2015 conference on empirical methods in natural language processing, pp 2013–2018

  143. Filice S, Croce D, Moschitti A, Basili R (2016) KeLP at SemEval-2016 task 3: learning semantic relations between questions and answers. In: Proceedings of the 10th international workshop on semantic evaluation (SemEval-2016), pp 1116–1123

  144. Rajpurkar P, Zhang J, Lopyrev K, Liang P (2016) SQuAD: 100,000+ questions for machine comprehension of text. In: Proceedings of the 2016 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 2383–2392

  145. Weston J, Bordes A, Chopra S et al (2015) Towards AI-complete question answering: a set of prerequisite toy tasks. CoRR. arXiv:1502.05698

  146. Lai G, Xie Q, Liu H et al (2017) RACE: large-scale ReAding Comprehension Dataset From Examinations. In: Proceedings of the 2017 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 785–794

  147. Zhang S, Liu X, Liu J et al (2018) ReCoRD: bridging the gap between human and machine commonsense reading comprehension. CoRR. arXiv:1810.12885

  148. Joshi M, Choi E, Weld DS, Zettlemoyer L (2017) TriviaQA: a large scale distantly supervised challenge dataset for reading comprehension. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 1: long papers

  149. Kočiský T, Schwarz J, Blunsom P et al (2018) The NarrativeQA reading comprehension challenge. Trans Assoc Comput Linguist 6:317–328. https://doi.org/10.1162/tacl_a_00023

    Article  Google Scholar 

  150. Nguyen T, Rosenberg M, Song X et al (2016) MS MARCO: a human generated machine reading comprehension dataset. In: CoCo@ NIPS

  151. Yang Z, Qi P, Zhang S et al (2018) HotpotQA: a dataset for diverse, explainable multi-hop question answering. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics

  152. Wang S, Yu M, Chang S, Jiang J (2018) A co-matching model for multi-choice reading comprehension. arXiv preprint. arXiv:1806.04068

  153. Choi E, He H, Iyyer M et al (2018) QuAC: question answering in context. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics Brussels, Belgium, pp 2174–2184

  154. Brown T, Mann B, Ryder N et al (2020) Language models are few-shot learners. In: Larochelle H, Ranzato M, Hadsell R et al (eds) Advances in neural information processing systems. Curran Associates Inc, Red Hook, NY, pp 1877–1901

    Google Scholar 

  155. Smith S, Patwary M, Norick B et al (2022) Using DeepSpeed and megatron to train megatron-turing NLG 530B, a large-scale generative language model. CoRR. arXiv:2201.11990

  156. Ahmed W, Anto BP (2017) Question answering system based on neural networks. Int J Eng Res 6:142–144

    Google Scholar 

  157. Romeo S, da San MG, Belinkov Y et al (2019) Language processing and learning models for community question answering in Arabic. Inf Process Manage 56:274–290. https://doi.org/10.1016/j.ipm.2017.07.003

    Article  Google Scholar 

  158. Aouichat A, Hadj Ameur MS, Geussoum A (2018) Arabic question classification using support vector machines and convolutional neural networks. In: International conference on applications of natural language to information systems, pp 113–125

  159. Park C, Lee C, Hong L et al (2019) S2-Net: machine reading comprehension with SRU-based self-matching networks. ETRI J 41:371–382. https://doi.org/10.4218/etrij.2017-0279

    Article  Google Scholar 

  160. Liu J, Yang Y, Lv S et al (2019) Attention-based BiGRU-CNN for Chinese question classification. J Ambient Intell Humaniz Comput. https://doi.org/10.1007/s12652-019-01344-9

    Article  Google Scholar 

  161. Lai Y, Feng Y, Yu X et al (2019) Lattice CNNs for matching based Chinese question answering. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, no 01, pp 6634–6641

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Heba Abdel-Nabi.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Abdel-Nabi, H., Awajan, A. & Ali, M.Z. Deep learning-based question answering: a survey. Knowl Inf Syst 65, 1399–1485 (2023). https://doi.org/10.1007/s10115-022-01783-5

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10115-022-01783-5

Keywords

Navigation