Abstract
Question Answering is a crucial natural language processing task. This field of research has attracted a sudden amount of interest lately due mainly to the integration of the deep learning models in the Question Answering Systems which consequently power up many advancements and improvements. This survey aims to explore and shed light upon the recent and most powerful deep learning-based Question Answering Systems and classify them based on the deep learning model used, stating the details of the used word representation, datasets, and evaluation metrics. It aims to highlight and discuss the currently used models and give insights that direct future research to enhance this increasingly growing field.










Similar content being viewed by others
Explore related subjects
Discover the latest articles and news from researchers in related subjects, suggested using machine learning.Notes
https://github.com/tmikolov/word2vec, Date of Access: 5th Jun, 2022.
https://github.com/stanfordnlp/GloVe, Date of Access: 5th Jun, 2022.
Download link: https://trec.nist.gov/data/qa.html, Date of Access: 5th Jun, 2022.
The Text REtrieval Conference (TREC) is a series of workshops that provides the needed infrastructure for text retrieval methodologies with large-scale evaluation.
Download link: https://download.microsoft.com/download/E/5/F/E5FCFCEE-7005-4814-853D-DAA7C66507E0/WikiQACorpus.zip, Date of Access: 5th Jun, 2022.
Download link: https://github.com/shuzi/insuranceQA, Date of Access: 5th Jun, 2022.
Download link: https://ciir.cs.umass.edu/downloads/wikipassageqa/WikiPassageQA.zip, Date of Access: 5th Jun, 2022.
Download link: https://deepai.org/dataset/squad, Date of Access: 5th Jun, 2022.
Download link: https://github.com/abisee/cnn-dailymail, Date of Access: 5th Jun, 2022.
Download link: https://research.fb.com/downloads/babi/, Date of Access: 5th Jun, 2022.
Download link: https://www.cs.cmu.edu/~glai1/data/race/, Date of Access: 5th Jun, 2022.
Download link: https://sheng-z.github.io/ReCoRD-explorer/, Date of Access: 5th Jun, 2022.
Download link: http://nlp.cs.washington.edu/triviaqa/, Date of Access: 5th Jun, 2022.
Download link: https://github.com/deepmind/narrativeqa, Date of Access: 5th Jun, 2022.
Download link: https://github.com/deepmind/narrativeqa, Date of Access: 5th Jun, 2022.
Download link: https://hotpotqa.github.io/, Date of Access: 5th Jun, 2022.
Download link: https://microsoft.github.io/msmarco/, Date of Access: 5th Jun, 2022.
Download link: https://quac.ai/, Date of Access: 5th Jun, 2022.
References
Toshevska M, Mirceva G, Jovanov M (2020) Question answering with deep learning: a survey. Faculty of Computer Science and Engineering Ss Cyril and Methodius University Skopje, Macedonia
Srba I, Bielikova M (2016) A comprehensive survey and classification of approaches for community question answering. ACM Trans Web 10:1–63. https://doi.org/10.1145/2934687
(2019) A survey on machine reading comprehension. J Beijing Univ Posts Telecommun 42:1
Huang Z, Xu S, Hu M et al (2020) Recent trends in deep learning based open-domain textual question answering systems. IEEE Access 8:94341–94356. https://doi.org/10.1109/ACCESS.2020.2988903
Palasundram K, Mohd Sharef N, Kasmiran KA, Azman A (2020) Enhancements to the sequence-to-sequence-based natural answer generation models. IEEE Access 8:45738–45752. https://doi.org/10.1109/ACCESS.2020.2978551
Abbasiantaeb Z, Momtazi S (2020) Text-based question answering from information retrieval and deep neural network perspectives: a survey. Wiley Interdiscip Rev: Data Min Knowl Discov 11:e1412. https://doi.org/10.1002/widm.1412
Li H (2018) Deep learning for natural language processing: advantages and challenges. Natl Sci Rev 5:24–26. https://doi.org/10.1093/nsr/nwx110
Xiang Y, Chen Q, Wang X, Qin Y (2017) Answer selection in community question answering via attentive neural networks. IEEE Signal Process Lett 24:505–509. https://doi.org/10.1109/LSP.2017.2673123
Otter DW, Medina JR, Kalita JK (2020) A survey of the usages of deep learning in natural language processing. IEEE Trans Neural Netw Learn Syst 32:604–624
Vanitha G, Sanampudi SK, Guda V (2011) Approaches for question answering systems. Int J Eng Sci Technol (IJEST) 3:990–995
Riloff E, Wiebe J (2003) Learning extraction patterns for subjective expressions. In: Proceedings of the 2003 conference on empirical methods in natural language processing (EMNLP), pp 105–112
Riloff E, Thelen M (2020) A rule-based question answering system for reading comprehension tests. In: ANLP-NAACL 2000 workshop: reading comprehension tests as evaluation for computer-based language understanding systems
Echihabi A, Marcu D (2003) A noisy-channel approach to question answering. In: Association for computational linguistics (ACL)
Heie MH, Whittaker EWD, Furui S (2012) Question answering using statistical language modelling. Comput Speech Lang 26:193–209. https://doi.org/10.1016/j.csl.2011.11.001
Wang M, Smith NA, Mitamura T (2007) What is the jeopardy model? A quasi-synchronous grammar for QA. In: Proceedings of the 2007 joint conference on empirical methods in natural language processing and computational natural language learning (EMNLP-CoNLL), pp 22–32
Choi E, Hewlett D, Uszkoreit J et al (2017) Coarse-to-fine question answering for long documents. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 209–220
Reddy S, Chen D, Manning CD (2019) CoQA: a conversational question answering challenge. Trans Assoc Comput Linguist 7:249–266. https://doi.org/10.1162/tacl_a_00266
Hinton GE, Osindero S, Teh Y-W (2006) A fast learning algorithm for deep belief nets. Neural Comput 18:1527–1554. https://doi.org/10.1162/neco.2006.18.7.1527
Severyn A, Moschittiy A (2015) Learning to rank short text pairs with convolutional deep neural networks. In: SIGIR 2015—proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval. Association for Computing Machinery, Inc, pp 373–382
Miller GA (1995) WordNet. Commun ACM 38:39–41. https://doi.org/10.1145/219717.219748
Akbik A, Blythe D, Vollgraf R (2018) Contextual string embeddings for sequence labeling. In: COLING 2018, 27th international conference on computational linguistics, pp 1638–1649
Adhikari A, Ram A, Tang R, Lin J (2019) DocBERT: BERT for document classification. CoRR. arXiv:1904.08398
Zhang H, Xu J, Wang J (2019) Pretraining-based natural language generation for text summarization. In: Proceedings of the 23rd conference on computational natural language learning (CoNLL). Association for Computational Linguistics, pp 789–797
Zhou C, Neubig G, Gu J (2019) Understanding knowledge distillation in non-autoregressive machine translation. In: Proceedings of the 2019 international conference on learning representations
Mikolov T, Chen K, Corrado G, Dean J (2013) Distributed representations of words and phrases and their compositionality. In: Advances in neural information processing systems (NeurIPS), pp 3111–3119
Pennington J, Socher R, Manning CD (2014) GloVe: global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 1532–1543
Devlin J, Chang M-W, Lee K et al (2019) BERT: pre-training of deep bidirectional transformers for language understanding. In: Annual conference of the North American chapter of the association for computational linguistics (NAACL)
Peters M, Neumann M, Iyyer M et al (2018) Deep contextualized word representations. In: Proceedings of the 2018 conference of the North American Chapter of the association for computational linguistics: human language technologies, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 2227–2237
Yamada I, Asai A, Shindo H et al (2020) LUKE: deep contextualized entity representations with entity-aware self-attention. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), pp 6442–6454
Clark K, Luong M-T, Le QV, Manning CD (2020) ELECTRA: Pre-training text encoders as discriminators rather than generators. In: International conference on learning representations (ICLR)
Yang Z, Dai Z, Yang Y et al (2019) XLNet: generalized autoregressive pretraining for language understanding. In: Wallach H, Larochelle H, Beygelzimer A et al (eds) Advances in neural information processing systems. Curran Associates Inc, Red Hook, NY
Zaheer M, Guruganesh G, Dubey KA et al (2020) Big bird: transformers for longer sequences. In: Larochelle H, Ranzato M, Hadsell R et al (eds) Advances in neural information processing systems. Curran Associates Inc, Red Hook, NY, pp 17283–17297
Jun C, Jang H, Sim M et al (2022) ANNA: enhanced language representation for question answering. In: Proceedings of the 7th workshop on representation learning for NLP. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 121–132
Vaswani A, Shazeer N, Parmar N et al (2017) Attention is all you need. In: Advances in neural information processing systems, pp 5998–6008
Goldberg Y (2019) Assessing BERT’s syntactic abilities. CoRR. arXiv:1901.05287
Lan Z, Chen M, Goodman S et al (2019) ALBERT: a lite BERT for self-supervised learning of language representations. In: International conference on learning representations (ICLR)
Sanh V, Debut L, Chaumond J, Wolf T (2019) DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. CoRR. arXiv:1910.01108
Wang W, Bi B, Yan M et al (2020) StructBERT: incorporating language structures into pre-training for deep language understanding. In: 8th international conference on learning representations (ICLR)
Liu Y, Ott M, Goyal N et al (2019) RoBERTa: a robustly optimized BERT pretraining approach. CoRR. arXiv:1907.11692
He P, Liu X, Gao J, Chen W (2021) DeBERTa: decoding-enhanced BERT with Disentangled Attention. In: Proceedings of the 9th international conference on learning representations (ICLR)
Jiang Z-H, Yu W, Zhou D et al (2020) ConvBERT: improving BERT with span-based dynamic convolution. In: Larochelle H, Ranzato M, Hadsell R et al (eds) Advances in neural information processing systems. Curran Associates Inc, Red Hook, NY, pp 12837–12848
Papineni K, Roukos S, Ward T, Zhu W-J (2002) BLEU: a method for automatic evaluation of machine translation. In: Proceedings of the 40th annual meeting of the association for computational linguistics
Banerjee S, Lavie A (2005) Meteor: an automatic metric for MT evaluation with improved correlation with human judgments. In: Proceedings of the ACL workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization
Lin C-Y (2004) ROUGE: a package for automatic evaluation of summaries. In: Text summarization branches out: proceedings of the 2004 association for computational linguistics (ACL-04) Workshop, Barcelona, Spain, pp 74–81
Zhang T, Kishore V, Wu F et al (2020) BERTScore: evaluating text generation with bert. In: Proceedings of the international conference on learning representations (ICLR)
Lee H, Yoon S, Dernoncourt F et al (2021) KPQA: a metric for generative question answering using keyphrase weights. In: Proceedings of annual conference of the North American chapter of the association for computational linguistics (NAACL), pp 2105–2115
Feng M, Xiang B, Glass MR et al (2015) Applying deep learning to answer selection: a study and an open task. In: 2015 IEEE workshop on automatic speech recognition and understanding (ASRU), pp 813–820
Rao J, He H, Lin J (2016) Noise-contrastive estimation for answer selection with deep neural networks. In: International conference on information and knowledge management, proceedings. Association for Computing Machinery, pp 1913–1916
Wang Z, Mi H, Ittycheriah A (2016) Sentence similarity learning by lexical decomposition and composition. COLING, Association for Computational Linguistics (ACL), pp 1340–1349
Madabushi HT, Lee M, Barnden J (2018) Integrating question classification and deep learning for improved answer selection. In: Proceedings of the 27th international conference on computational linguistics, pp 3283–3294
Wang Z, Hamza W, Florian R (2017) Bilateral multi-perspective matching for natural language sentences. In: Proceedings of the twenty-sixth international joint conference on artificial intelligence. International Joint Conferences on Artificial Intelligence Organization, California, pp 4144–4150
Tay Y, Phan MC, Tuan LA, Hui SC (2017) Learning to rank question answer pairs with holographic dual LSTM architecture. In: SIGIR 2017—proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval. Association for Computing Machinery, Inc, pp 695–704
Mihaylov T, Kozareva Z, Frank A (2017) Neural skill transfer from supervised language tasks to reading comprehension. Workshop on learning with limited labeled data: weak supervision and beyond at NIPS
di Gennaro G, Buonanno A, di Girolamo A et al (2020) Intent classification in question-answering using LSTM architectures. Progr Artif Intell Neural Syst. https://doi.org/10.1007/978-981-15-5093-5_11
Zhang L, Lin C, Zhou D et al (2021) A Bayesian end-to-end model with estimated uncertainties for simple question answering over knowledge bases. Comput Speech Lang 66:101167. https://doi.org/10.1016/j.csl.2020.101167
Hu M, Peng Y, Wei F et al (2018) Attention-guided answer distillation for machine reading comprehension. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 2077–2086
Ran Q, Li P, Hu W and Zhou J (2019) Option comparison network for multiple-choice reading comprehension. CoRR. arXiv arXiv:1903.03033
Yang A, Wang Q, Liu J et al (2019) Enhancing pre-trained language representations with rich knowledge for machine reading comprehension. Association for Computational Linguistics, Stroudsburg, PA
Shoeybi M, Patwary M, Puri R et al (2019) Megatron-LM: training multi-billion parameter language models using model parallelism. CoRR. arXiv:1909.08053
Garg S, Vu T, Moschitti A (2020) TANDA: transfer and adapt pre-trained transformer models for answer sentence selection. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 7780–7788.https://doi.org/10.1609/aaai.v34i05.6282
Zhu P, Zhang Z, Zhao H, Li X (2022) DUMA: reading comprehension with transposition thinking. IEEE/ACM Trans Audio Speech Lang Process 30:269–279. https://doi.org/10.1109/TASLP.2021.3138683
Guu K, Lee K, Tung Z et al (2020) Retrieval augmented language model pre-training. In: III HD, Singh A (eds) Proceedings of the 37th international conference on machine learning. PMLR, pp 3929–3938
Wu B, Zhang Z, Zhao H (2021) Graph-free multi-hop reading comprehension: a select-to-guide strategy. CoRR. arXiv:2107.11823
Li X-Y, Lei W-J, Yang Y-B (2022) From easy to hard: two-stage selector and reader for multi-hop question answering. CoRR. arXiv:2205.11729
Guan Y, Li Z, Leng J et al (2021) Block-skim: efficient question answering for transformer. CoRR. arXiv:2112.08560
Zhou X, Hu B, Chen Q, Wang X (2018) Recurrent convolutional neural network for answer selection in community question answering. Neurocomputing 274:8–18. https://doi.org/10.1016/j.neucom.2016.07.082
Cohen D, Yang L, Croft WB (2018) WikiPassageQA: a benchmark collection for research on non-factoid answer passage retrieval. In: 41st international ACM SIGIR conference on research and development in information retrieval, SIGIR 2018. Association for Computing Machinery, Inc, pp 1165–1168
Zhang X, Li S, Sha L, Wang H (2017) Attentive interactive neural networks for answer selection in community question answering. In: Proceedings of the AAAI conference on artificial intelligence, vol 31, no 1
Bian W, Li S, Yang Z et al (2017) A compare-aggregate model with dynamic-clip attention for answer selection. In: International conference on information and knowledge management, proceedings. Association for Computing Machinery, pp 1987–1990
Yoon S, Dernoncourt F, Kim DS et al (2019) A compare-aggregate model with latent clustering for answer selection. In: Proceedings of the 28th ACM international conference on information and knowledge management, pp 2093–2096
Peng Y, Liu B (2018) Attention-based neural network for short-text question answering. In: ACM International conference proceeding series. Association for Computing Machinery, pp 21–26
Yu AW, Dohan D, Luong M-T et al (2018) QANet: combining local convolution with global self-attention for reading comprehension. CoRR. arXiv:1804.09541
Miller A, Fisch A, Dodge J et al (2016) Key-value memory networks for directly reading documents. In: Proceedings of the 2016 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 1400–1409
Yang L, Ai Q, Guo J, Croft WB (2016) aNMM: ranking short answer texts with attention-based neural matching model. In: International conference on information and knowledge management, proceedings. Association for Computing Machinery, pp 287–296
Shao T, Guo Y, Chen H, Hao Z (2019) Transformer-based neural network for answer selection in question answering. IEEE Access 7:26146–26156. https://doi.org/10.1109/ACCESS.2019.2900753
Sukhbaatar S, Szlam A, Weston J, Fergus R (2015) End-to-end memory networks. In: Advances in neural information processing systems, pp 2440–2448
Kumar A, Irsoy O, Ondruska P et al (2016) Ask me anything: dynamic memory networks for natural language processing. In: International conference on machine learning, pp 1378–1387
Pan B, Li H, Zhao Z et al (2017) MEMEN: multi-layer embedding with memory networks for machine comprehension. In: AAAI conference on artificial intelligence, (AAAI-18), the 30th innovative applications of artificial intelligence (IAAI-18), and the 8th AAAI symposium on educational advances in artificial intelligence (EAAI-18)
Back S, Yu S, Indurthi SR et al (2018) MemoReader: large-scale reading comprehension through neural memory controller. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 2131–2140
Xiong C, Zhong V, Socher R (2016) Dynamic coattention networks for question answering. In: International conference on learning representations (ICLR)
Wang S, Yu M, Chang S, Jiang J (2018) A co-matching model for multi-choice reading comprehension. In: Association for computational linguistics (ACL), pp 746–751
Xiong C, Zhong V, Socher R (2017) DCN+: mixed objective and deep residual coattention for question answering. CoRR. arXiv:1711.00106
McCann B, Keskar NS, Xiong C, Socher R (2018) The Natural language decathlon: multitask learning as question answering. CoRR. arXiv:1806.08730
Wang W, Yan M, Wu C (2018) Multi-granularity hierarchical attention fusion networks for reading comprehension and question answering. In: Proceedings of the 56th annual meeting of the association for computational linguistics, vol 1: long papers, pp 1705–1714
Tay Y, Tuan LA, Hui SC (2018) Multi-cast attention networks. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery and data mining. ACM, New York, NY, USA, pp 2299–2308
Min S, Seo M, Hajishirzi H (2017) Question answering through transfer learning from large fine-grained supervision Data. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 2: short papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 510–517
Golub D, Huang P-S, He X, Deng L (2017) Two-stage synthesis networks for transfer learning in machine comprehension. In: Proceedings of the 2017 conference on empirical methods in natural language processing. association for computational linguistics, Stroudsburg, PA, USA, pp 835–844
Seo M, Kembhavi A, Farhadi A, Hajishirzi H (2016) Bidirectional attention flow for machine comprehension. In: International conference on learning representations (ICLR)
Liu X, Shen Y, Duh K, Gao J (2018) Stochastic answer networks for machine reading comprehension. In: Proceedings of the 56th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 1694–1704
Xiong W, Yu M, Guo X et al (2019) Simple yet effective bridge reasoning for open-domain multi-hop question answering. In: Proceedings of the 2nd workshop on machine reading for question answering. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 48–52
Hermann KM, Kočiský T, Grefenstette E et al (2015) Teaching machines to read and comprehend. In: Cortes C, Lawrence N, Lee D, Sugiyama M, Garnett R (eds) Advances in neural information processing systems, vol 28. Curran Associates Inc, Red Hook, NY
Kadlec R, Schmid M, Bajgar O, Kleindienst J (2016) Text understanding with the attention sum reader network. In: Proceedings of the 54th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 908–918
Trischler A, Ye Z, Yuan X et al (2016) Natural language comprehension with the EpiReader. In: Proceedings of the 2016 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 128–137
Wang S, Jiang J (2017) Machine comprehension using match-LSTM and answer pointer. In: International conference on learning representations (ICLR), pp 1–15
Cui Y, Chen Z, Wei S et al (2017) Attention-over-attention neural networks for reading comprehension. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 593–602
Dhingra B, Liu H, Yang Z et al (2017) Gated-attention readers for text comprehension. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 1832–1846
Wang W, Yang N, Wei F et al (2017) Gated self-matching networks for reading comprehension and question answering. In: ACL 2017—55th annual meeting of the association for computational linguistics, proceedings of the conference (long papers). Association for Computational Linguistics (ACL), pp 189–198
Liu R, Wei W, Mao W, Chikina M (2017) Phase conductor on multi-layered attentions for machine comprehension. CoRR. arXiv:1710.10504
Huang H-Y, Zhu C, Shen Y, Chen W (2017) FusionNet: fusing via fully-aware attention with application to machine comprehension. CoRR. arXiv:1711.07341
Zhu H, Wei F, Qin B, Liu T (2018) Hierarchical attention flow for multiple-choice reading comprehension. In: Hierarchical Attention flow for multiple-choice reading comprehension, vol 32, no 1
Kundu S, Ng HT (2018) A question-focused multi-factor attention network for question answering. In: Proceedings of the AAAI conference on artificial intelligence, vol 32, no 1
Tan C, Wei F, Yang N et al (2018) S-Net: from answer extraction to answer synthesis for machine reading comprehension. In: Proceedings of the AAAI conference on artificial intelligence, vol 32, no 1
Zhu C, Zeng M, Huang X (2018) SDNet: contextualized attention-based deep network for conversational question answering. CoRR. arXiv:1812.03593
LeeKim HH (2020) GF-Net: improving machine reading comprehension with feature gates. Pattern Recognit Lett 129:8–15. https://doi.org/10.1016/j.patrec.2019.10.030
Huang X, Zhang J, Li D, Li P (2019) Knowledge graph embedding based question answering. In: WSDM 2019—proceedings of the 12th ACM international conference on web search and data mining. Association for Computing Machinery, Inc, pp 105–113
Chen Y, Wu L, Zaki MJ (2020) GraphFlow: exploiting conversation flow with graph neural networks for conversational machine comprehension. In: Proceedings of the twenty-ninth international joint conference on artificial intelligence. International Joint Conferences on Artificial Intelligence Organization, California, pp 1230–1236
Cao X, Liu Y (2022) Coarse-grained decomposition and fine-grained interaction for multi-hop question answering. J Intell Inf Syst 58:21–41. https://doi.org/10.1007/s10844-021-00645-w
Nishida K, Nishida K, Nagata M et al (2019) Answering while summarizing: multi-task learning for multi-hop QA with evidence extraction. In: Proceedings of the 57th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Florence, Italy, pp 2335–2345
Xiao Y, Qu Y, Qiu L et al (2019) dynamically fused graph network for multi-hop reasoning. In: Proceedings of the 57th annual meeting of the association for computational linguistics, Florence, Italy, pp 6140–6150
Cao Y, Fang M, Tao D (2019) BAG: bi-directional attention entity graph convolutional network for multi-hop reasoning question answering. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL): human language technologies, pp 357–362
Fang Y, Sun S, Gan Z et al (2020) Hierarchical graph network for multi-hop question answering. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), pp 8823–8838
Zhang M, Li F, Wang Y et al (2020) Coarse and fine granularity graph reasoning for interpretable multi-hop question answering. IEEE Access 8:56755–56765. https://doi.org/10.1109/ACCESS.2020.2981134
Tu M, Huang K, Wang G et al (2020) Select, answer and explain: interpretable multi-hop reading comprehension over multiple documents. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, no 05, pp 9073–9080
Zhao C, Xiong C, Rosset C et al (2020) Transformer-XH: multi-evidence reasoning with extra hop attention. In: International conference on learning representations (ICLR)
Zhang X, Bosselut A, Yasunaga M et al (2022) GreaseLM: Graph REASoning Enhanced Language Models for question answering. CoRR. arXiv:2201.08860
Shi J, Cao S, Hou L et al (2021) TransferNet: an effective and transparent framework for multi-hop question answering over relation graph. In: Proceedings of the 2021 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 4149–4158
Zhang Y, Nie P, Ramamurthy A, Song L (2021) Answering any-hop open-domain questions with iterative document reranking. In: SIGIR 2021—proceedings of the 44th international ACM SIGIR conference on research and development in information retrieval. Association for Computing Machinery, Inc, pp 481–490
Ren H, Dai H, Dai B et al (2021) LEGO: latent execution-guided reasoning for multi-hop question answering on knowledge graphs. In: International conference on machine learning, pp 8959–8970
Xiong W, Li XL, Iyer S et al (2020) Answering complex open-domain questions with multi-hop dense retrieval. In: Proceedings of the conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing, EMNLP-IJCNLP 2019, pp 2590–2602
Wu J, Mu T, Thiyagalingam J, Goulermas JY (2020) Building interactive sentence-aware representation based on generative language model for community question answering. Neurocomputing 389:93–107. https://doi.org/10.1016/j.neucom.2019.12.107
Bi B, Wu C, Yan M et al (2019) Incorporating external knowledge into machine reading for generative question answering. In: Conference on empirical methods in natural language processing and international joint conference on natural language processing (EMNLP-IJCNLP)
Bauer L, Wang Y, Bansal M (2018) Commonsense for generative multi-hop question answering tasks. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 4220–4230
Izacard G, Grave E (2021) Leveraging passage retrieval with generative models for open domain question answering. In: Proceedings of the 16th conference of the European chapter of the association for computational linguistics: main volume. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 874–880
Yavuz S, Hashimoto K, Zhou Y et al (2022) Modeling multi-hop question answering as single sequence prediction. In: Proceedings of the 60th annual meeting of the association for computational linguistics, vol 1: long papers. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 974–990
Shen Y, Huang P sen, Gao J, Chen W (2017) ReasoNet: learning to stop reading in machine comprehension. In: Proceedings of the ACM SIGKDD international conference on knowledge discovery and data mining. Association for Computing Machinery, pp 1047–1055
Buck C, Bulian J, Ciaramita M et al (2018) Ask the right questions: active question reformulation with reinforcement learning. In: International conference on learning representations (ICLR)
Xu Y, Liu J, Gao J et al (2017) Dynamic fusion networks for machine reading comprehension. CoRR. arXiv:1711.04964
Hu M, Peng Y, Huang Z et al (2018) Reinforced mnemonic reader for machine reading comprehension. In: Proceedings of the twenty-seventh international joint conference on artificial intelligence. International Joint Conferences on Artificial Intelligence Organization, California, pp 4099–4106
Santoro A, Raposo D, Barrett DGT et al (2017) A simple neural network module for relational reasoning. In: Guyon I, Von Luxburg U, Bengio S, Wallach H, Fergus R, Vishwanathan S, Garnett R (eds) Advances in neural information processing systems, vol 30. Curran Associates Inc, Red Hook, NY
Swayamdipta S, Parikh AP, Kwiatkowski T (2017) Multi-mention learning for reading comprehension with neural cascades. In: International conference on learning representations (ICLR)
Tay Y, Tuan LA, Hui SC (2018) Hyperbolic representation learning for fast and efficient neural question answering. In: WSDM 2018—proceedings of the 11th ACM international conference on web search and data mining. Association for Computing Machinery, Inc, pp 583–591
Seonwoo Y, Kim J-H, Ha J-W, Oh A (2020) Context-aware answer extraction in question answering. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), pp 2418–2428
Wu Y, Zhao S (2021) Community answer generation based on knowledge graph. Inf Sci 545:132–152. https://doi.org/10.1016/j.ins.2020.07.077
Zhou G, Xie Z, Yu Z, Huang JX (2021) DFM: a parameter-shared deep fused model for knowledge base question answering. Inf Sci 547:103–118. https://doi.org/10.1016/j.ins.2020.08.037
He H, Gimpel K, Lin J (2015) Multi-perspective sentence similarity modeling with convolutional neural networks. In: Proceedings of the 2015 conference on empirical methods in natural language processing. Association for Computational Linguistics, pp 1576–1586
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9:1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735
Schuster M, Paliwal KK (1997) Bidirectional recurrent neural networks. IEEE Trans Signal Process 45:2673–2681. https://doi.org/10.1109/78.650093
Wang S, Jiang J (2017) A compare-aggregate model for matching text sequences. In: Proceedings of the 5th international conference on learning representations (ICLR 2017)
Vinyals O, Fortunato M, Jaitly N (2015) Pointer networks. In: Cortes C, Lawrence N, Lee D, Sugiyama M, Garnett R (eds) Advances in neural information processing systems, vol 28. Curran Associates Inc, Red Hook, NY
Gong Y, Bowman SR (2018) Ruminating reader: reasoning with gated multi-hop attention. In: Proceedings of the workshop on machine reading for question answering. Association for Computational Linguistics, pp 1–11
Liu H, Singh P (2004) ConceptNet—a practical commonsense reasoning tool-kit. BT Technol J 22:211–226. https://doi.org/10.1023/B:BTTJ.0000047600.45421.6d
Yang Y, Yih W-T, Meek C (2015) WIKIQA: a challenge dataset for open-domain question answering. In: Proceedings of the 2015 conference on empirical methods in natural language processing, pp 2013–2018
Filice S, Croce D, Moschitti A, Basili R (2016) KeLP at SemEval-2016 task 3: learning semantic relations between questions and answers. In: Proceedings of the 10th international workshop on semantic evaluation (SemEval-2016), pp 1116–1123
Rajpurkar P, Zhang J, Lopyrev K, Liang P (2016) SQuAD: 100,000+ questions for machine comprehension of text. In: Proceedings of the 2016 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 2383–2392
Weston J, Bordes A, Chopra S et al (2015) Towards AI-complete question answering: a set of prerequisite toy tasks. CoRR. arXiv:1502.05698
Lai G, Xie Q, Liu H et al (2017) RACE: large-scale ReAding Comprehension Dataset From Examinations. In: Proceedings of the 2017 conference on empirical methods in natural language processing. Association for Computational Linguistics, Stroudsburg, PA, USA, pp 785–794
Zhang S, Liu X, Liu J et al (2018) ReCoRD: bridging the gap between human and machine commonsense reading comprehension. CoRR. arXiv:1810.12885
Joshi M, Choi E, Weld DS, Zettlemoyer L (2017) TriviaQA: a large scale distantly supervised challenge dataset for reading comprehension. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 1: long papers
Kočiský T, Schwarz J, Blunsom P et al (2018) The NarrativeQA reading comprehension challenge. Trans Assoc Comput Linguist 6:317–328. https://doi.org/10.1162/tacl_a_00023
Nguyen T, Rosenberg M, Song X et al (2016) MS MARCO: a human generated machine reading comprehension dataset. In: CoCo@ NIPS
Yang Z, Qi P, Zhang S et al (2018) HotpotQA: a dataset for diverse, explainable multi-hop question answering. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics
Wang S, Yu M, Chang S, Jiang J (2018) A co-matching model for multi-choice reading comprehension. arXiv preprint. arXiv:1806.04068
Choi E, He H, Iyyer M et al (2018) QuAC: question answering in context. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics Brussels, Belgium, pp 2174–2184
Brown T, Mann B, Ryder N et al (2020) Language models are few-shot learners. In: Larochelle H, Ranzato M, Hadsell R et al (eds) Advances in neural information processing systems. Curran Associates Inc, Red Hook, NY, pp 1877–1901
Smith S, Patwary M, Norick B et al (2022) Using DeepSpeed and megatron to train megatron-turing NLG 530B, a large-scale generative language model. CoRR. arXiv:2201.11990
Ahmed W, Anto BP (2017) Question answering system based on neural networks. Int J Eng Res 6:142–144
Romeo S, da San MG, Belinkov Y et al (2019) Language processing and learning models for community question answering in Arabic. Inf Process Manage 56:274–290. https://doi.org/10.1016/j.ipm.2017.07.003
Aouichat A, Hadj Ameur MS, Geussoum A (2018) Arabic question classification using support vector machines and convolutional neural networks. In: International conference on applications of natural language to information systems, pp 113–125
Park C, Lee C, Hong L et al (2019) S2-Net: machine reading comprehension with SRU-based self-matching networks. ETRI J 41:371–382. https://doi.org/10.4218/etrij.2017-0279
Liu J, Yang Y, Lv S et al (2019) Attention-based BiGRU-CNN for Chinese question classification. J Ambient Intell Humaniz Comput. https://doi.org/10.1007/s12652-019-01344-9
Lai Y, Feng Y, Yu X et al (2019) Lattice CNNs for matching based Chinese question answering. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, no 01, pp 6634–6641
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Abdel-Nabi, H., Awajan, A. & Ali, M.Z. Deep learning-based question answering: a survey. Knowl Inf Syst 65, 1399–1485 (2023). https://doi.org/10.1007/s10115-022-01783-5
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10115-022-01783-5