Abstract
Negation is a fundamental grammatical construct that plays a crucial role in understanding QA tasks. It has been revealed that models trained with SQuAD1 still produce original responses when presented with negated sentences. To mitigate this issue, SQuAD2.0 incorporates a plethora of unanswerable questions to enable pre-trained models to distinguish negative inquiries. In this study, we assess the performance of the model on answerable and unanswerable questions that incorporate negative words and find out that the model’s performance on unanswerable negative questions surpasses the baseline. However, the model’s performance on answerable negative questions falls short of the baseline. This outcome prompts us to surmise that SQuAD2.0 includes a substantial number of unanswerable questions, but the pattern of these questions is typically limited to the addition of negative adverbs such as “never” and “not”. As a result, the trained model tends to produce “unanswerable” responses when confronted with questions that contain negative expressions. To address this issue, we propose a novel framework, called NegT5, which adopts the text-to-text multi-task fine-tuning principle introduced in T5 for making the model able to deal with negation in QA.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Choi, E., et al.: QuAC: question answering in context. In: EMNLP (2018)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: NAACL-HLT (2019)
Joshi, M., Choi, E., Weld, D.S., Zettlemoyer, L.: TriviaQA: a large scale distantly supervised challenge dataset for reading comprehension. In: ACL (2017)
Kassner, N., Schütze, H.: Negated and misprimed probes for pretrained language models: birds can talk, but cannot fly. In: ACL (2020)
Khandelwal, A., Sawant, S.: NegBERT: a transfer learning approach for negation detection and scope resolution. In: LREC (2020)
Konstantinova, N., De Sousa, S.C., Díaz, N.P.C., López, M.J.M., Taboada, M., Mitkov, R.: A review corpus annotated for negation, speculation and their scope. In: LREC, pp. 3190–3195 (2012)
Kwiatkowski, T., et al.: Natural questions: a benchmark for question answering research. Trans. Assoc. Comput. Linguist. 7, 453–466 (2019)
Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1), 5485–5551 (2020)
Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: SQuAD: 100,000+ questions for machine comprehension of text. In: EMNLP (2016)
Sen, P., Saffari, A.: What do models learn from question answering datasets? In: EMNLP (2020)
Szarvas, G., Vincze, V., Farkas, R., Csirik, J.: The bioscope corpus: annotation for negation, uncertainty and their scope in biomedical texts. In: Proceedings of the Workshop on Current Trends in Biomedical Natural Language Processing, pp. 38–45 (2008)
Trischler, A., et al.: NewsQA: a machine comprehension dataset. In: RepL4NLP (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Jin, T., Racharak, T., Nguyen, M.L. (2023). NegT5: A Cross-Task Text-to-Text Framework for Negation in Question Answering. In: Nguyen, N.T., et al. Intelligent Information and Database Systems. ACIIDS 2023. Lecture Notes in Computer Science(), vol 13996. Springer, Singapore. https://doi.org/10.1007/978-981-99-5837-5_23
Download citation
DOI: https://doi.org/10.1007/978-981-99-5837-5_23
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-5836-8
Online ISBN: 978-981-99-5837-5
eBook Packages: Computer ScienceComputer Science (R0)