Abstract
Generating natural language explanations is a challenging task in natural language processing (NLP). With recent advancements in deep learning and language modeling techniques increased attention toward explanation generation. However, generating explanations based on commonsense reasoning remains a distinct and challenging problem. While many researchers have explored automatic explanation generation using deep learning approaches, no work has been done on generating Arabic explanations. This paper addresses this issue by presenting the Arabic machine translation of the Explanations for CommonsenseQA (Arabic-ECQA) and Open Mind Common Sense (Arabic-OMCS) datasets and fine-tuning the pre-trained AraGPT-2 model to automatically generate Arabic explanations. The performance of the fine-tunes AraGPT-2 models is evaluated using STS, METEOR, and ROUGE scores. To the best of the authors knowledge, this work is the first of its kind to generate Arabic explanations. To accelerate research on Arabic NLP, we make the datasets code publicly available (https://github.com/MohamedELGhaly/Arabic-ECQA).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Aggarwal, S., Mandowara, D., Agrawal, V., Khandelwal, D., Singla, P., Garg, D.: Explanations for commonsenseQA: new dataset and models. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 3050–3065 (2021)
Agirre, E., Cer, D., Diab, M., Gonzalez-Agirre, A.: SemEval-2012 task 6: a pilot on semantic textual similarity. In: * SEM 2012: The First Joint Conference on Lexical and Computational Semantics-Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012), pp. 385–393 (2012)
Al-Bashabsheh, E., Al-Khazaleh, H., Elayan, O., Duwairi, R.: Commonsense validation for Arabic sentences using deep learning. In: 2021 22nd International Arab Conference on Information Technology (ACIT), pp. 1–7. IEEE (2021)
Antoun, W., Baly, F., Hajj, H.: AraGPT2: pre-trained transformer for Arabic language generation. arXiv preprint arXiv:2012.15520 (2020)
Banerjee, S., Lavie, A.: METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In: Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation AND/OR Summarization, pp. 65–72 (2005)
Davis, E., Marcus, G.: Commonsense reasoning and commonsense knowledge in artificial intelligence. Commun. ACM 58(9), 92–103 (2015)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)
Do, V., Camburu, O.M., Akata, Z., Lukasiewicz, T.: e-SNLI-VE: corrected visual-textual entailment with natural language explanations. arXiv preprint arXiv:2004.03744 (2020)
Hancock, B., Bringmann, M., Varma, P., Liang, P., Wang, S., RĂ©, C.: Training classifiers with natural language explanations. In: Proceedings of the Conference on Association for Computational Linguistics Meeting, vol. 2018, p. 1884. NIH Public Access (2018)
Jacovi, A., Goldberg, Y.: Towards faithfully interpretable NLP systems: how should we define and evaluate faithfulness? In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 4198–4205. Association for Computational Linguistics, Online (2020). https://doi.org/10.18653/v1/2020.acl-main.386, https://aclanthology.org/2020.acl-main.386
Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. CoRR abs/1910.13461 (2019). http://arxiv.org/abs/1910.13461
Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74–81 (2004)
Ostermann, S., Modi, A., Roth, M., Thater, S., Pinkal, M.: MCScript: a novel dataset for assessing machine comprehension using script knowledge. arXiv preprint arXiv:1803.05223 (2018)
Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)
Rajani, N.F., McCann, B., Xiong, C., Socher, R.: Explain yourself! leveraging language models for commonsense reasoning. arXiv preprint arXiv:1906.02361 (2019)
Roemmele, M., Bejan, C.A., Gordon, A.S.: Choice of plausible alternatives: an evaluation of commonsense causal reasoning. In: AAAI Spring Symposium: Logical Formalizations of Commonsense Reasoning, pp. 90–95 (2011)
Sakaguchi, K., Bras, R.L., Bhagavatula, C., Choi, Y.: WinoGrande: an adversarial winograd schema challenge at scale. Commun. ACM 64(9), 99–106 (2021)
Sap, M., Rashkin, H., Chen, D., LeBras, R., Choi, Y.: SocialIQa: commonsense reasoning about social interactions. arXiv preprint arXiv:1904.09728 (2019)
Singh, P., et al.: The public acquisition of commonsense knowledge. In: Proceedings of AAAI Spring Symposium: Acquiring (and Using) Linguistic (and World) Knowledge for Information Access, vol. 3 (2002)
Talmor, A., Herzig, J., Lourie, N., Berant, J.: CommonsenseQA: a question answering challenge targeting commonsense knowledge. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4149–4158. Association for Computational Linguistics, Minneapolis (2019). https://doi.org/10.18653/v1/N19-1421, https://aclanthology.org/N19-1421
Tawalbeh, S., Al-Smadi, M.: Is this sentence valid? An Arabic dataset for commonsense validation. arXiv preprint arXiv:2008.10873 (2020)
Wang, C., Liang, S., Jin, Y., Wang, Y., Zhu, X., Zhang, Y.: SemEval-2020 task 4: commonsense validation and explanation. arXiv preprint arXiv:2007.00236 (2020)
Wiegreffe, S., Marasovic, A.: Teach me to explain: a review of datasets for explainable natural language processing. In: Vanschoren, J., Yeung, S. (eds.) Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks. vol. 1 (2021). https://datasets-benchmarks-proceedings.neurips.cc/paper/2021/file/698d51a19d8a121ce581499d7b701668-Paper-round1.pdf
Yin, K., Fernandes, P., Pruthi, D., Chaudhary, A., Martins, A.F.T., Neubig, G.: Do context-aware translation models pay the right attention? In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics (2021). https://arxiv.org/abs/2105.06977
Zellers, R., Holtzman, A., Bisk, Y., Farhadi, A., Choi, Y.: HellaSwag: can a machine really finish your sentence? arXiv preprint arXiv:1905.07830 (2019)
Zhang, S., Rudinger, R., Duh, K., Van Durme, B.: Ordinal common-sense inference. Trans. Assoc. Comput. Linguist. 5, 379–395 (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Beheitt, M.E.G., Ben HajHmida, M. (2023). Generation of Arabic Commonsense Explanations. In: Nguyen, N.T., et al. Recent Challenges in Intelligent Information and Database Systems. ACIIDS 2023. Communications in Computer and Information Science, vol 1863. Springer, Cham. https://doi.org/10.1007/978-3-031-42430-4_43
Download citation
DOI: https://doi.org/10.1007/978-3-031-42430-4_43
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-42429-8
Online ISBN: 978-3-031-42430-4
eBook Packages: Computer ScienceComputer Science (R0)