skip to main content
research-article

SE4ExSum: An Integrated Semantic-aware Neural Approach with Graph Convolutional Network for Extractive Text Summarization

Published:01 September 2021Publication History
Skip Abstract Section

Abstract

Recently, advanced techniques in deep learning such as recurrent neural network (GRU, LSTM and Bi-LSTM) and auto-encoding (attention-based transformer and BERT) have achieved great successes in multiple application domains including text summarization. Recent state-of-the-art encoding-based text summarization models such as BertSum, PreSum and DiscoBert have demonstrated significant improvements on extractive text summarization tasks. However, recent models still encounter common problems related to the language-specific dependency which requires the supports of the external NLP tools. Besides that, recent advanced text representation methods, such as BERT as the sentence-level textual encoder, also fail to fully capture the representation of a full-length document. To address these challenges, in this paper we proposed a novel semantic-ware embedding approach for extractive text summarization, called as: SE4ExSum. Our proposed SE4ExSum is an integration between the use of feature graph-of-words (FGOW) with BERT-based encoder for effectively learning the word/sentence-level representations of a given document. Then, the graph convolutional network (GCN) based encoder is applied to learn the global document's representation which is then used to facilitate the text summarization task. Extensive experiments on benchmark datasets show the effectiveness of our proposed model in comparing with recent state-of-the-art text summarization models.

References

  1. T. Wang, P. Chen, and D. Simovici. 2016. A new evaluation measure using compression dissimilarity on text summarization. Applied Intelligence 45, 1 (2016), 127–134. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. O. K. Oyedotun and A. Khashman. 2016. Document segmentation using textural features summarization and feedforward neural network. Applied Intelligence 45, 1 (2016), 198–212. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. W. S. El-Kassas, C. R. Salama, A. A. Rafea, and H. K. Mohamed. 2020. Automatic text summarization: A comprehensive survey. Expert Systems with Applications 113679, 2020.Google ScholarGoogle Scholar
  4. A. See, P. J. Liu, and C. D. Manning. 2017. Get to the point: Summarization with pointer-generator networks. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics 2017.Google ScholarGoogle Scholar
  5. M. Gambhir and V. Gupta. 2017. Recent automatic text summarization techniques: A survey. Artificial Intelligence Review 47, 1 (2017), 1–66. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. D. Bahdanau, K. Cho, and Y. Bengio. 2015. Neural machine translation by jointly learning to align and translate. In 3rd International Conference on Learning Representations (ICLR), 2015.Google ScholarGoogle Scholar
  7. C. Kedzie, K. McKeown, and H. Daume III. 2018. Content selection in deep learning models of summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  8. J. Cheng and M. Lapata. 2016. Neural summarization by extracting sentences and words. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 2016.Google ScholarGoogle Scholar
  9. W. Xu, S. Li, and Y. Lu. 2020. Usr-mtl: an unsupervised sentence representation learning framework with multi-task learning. Applied Intelligence 1–16, 2020.Google ScholarGoogle Scholar
  10. T. Wang, L. Liu, N. Liu, H. Zhang, L. Zhang, and S. Feng. 2020. A multi-label text classification method via dynamic semantic representation model and deep neural network. Applied Intelligence 50, 8 (2020), 2339–2351.Google ScholarGoogle ScholarCross RefCross Ref
  11. M. E. Peters, M. Neumann, M. Iyyer, M. Gardner, C. Clark, K. Lee, and L. Zettlemoyer. 2018. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2018.Google ScholarGoogle Scholar
  12. A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever. 2018. Improving language understanding by generative pre-training. OpenAI, 2018.Google ScholarGoogle Scholar
  13. J. Devlin, M. W. Chang, K. Lee, and K. Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019.Google ScholarGoogle Scholar
  14. Y. Liu and M. Lapata. 2019. Text summarization with pretrained encoders. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019.Google ScholarGoogle Scholar
  15. H. Zhang, J. Xu, and J. Wang. 2019. Pretraining-based natural language generation for text summarization. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), 2019.Google ScholarGoogle ScholarCross RefCross Ref
  16. W. Kryściński, R. Paulus, C. Xiong, and R. Socher. 2018. Improving abstraction in text summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  17. W. Xiao and G. Carenini. 2019. Extractive summarization of long documents by combining global and local context. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019.Google ScholarGoogle ScholarCross RefCross Ref
  18. J. Xu, Z. Gan, Y. Cheng, and J. Liu. 2020. Discourse-aware neural extractive text summarization. In Proceedings of the 58th annual Meeting of the Association for Computational Linguistics 2020.Google ScholarGoogle ScholarCross RefCross Ref
  19. T. N. Kipf and M. Welling. 2017. Semi-supervised classification with graph convolutional networks. In 5th International Conference on Learning Representations, ICLR, 2017.Google ScholarGoogle Scholar
  20. R. Paulus, C. Xiong, and R. Socher. 2018. A deep reinforced model for abstractive summarization. In International Conference on Learning Representations, 2018.Google ScholarGoogle Scholar
  21. Q. Guo, J. Huang, N. Xiong, and P. Wang. 2019. MS-pointer network: Abstractive text summary based on multi-head self-attention. IEEE Access 7, (2019), 138603–138613.Google ScholarGoogle ScholarCross RefCross Ref
  22. S. Gehrmann, Y. Deng, and A. M. Rush. 2018. Bottom-up abstractive summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  23. W. Li, X. Xiao, Y. Lyu, and Y. Wang. 2018. Improving neural abstractive document summarization with explicit information selection modeling. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018.Google ScholarGoogle Scholar
  24. R. Pasunuru and M. Bansal. 2018. Multi-reward reinforced summarization with saliency and entailment. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2018.Google ScholarGoogle Scholar
  25. T. Scialom, S. Lamprier, B. Piwowarski, and J. Staiano. 2019. Answers unite! unsupervised metrics for reinforced summarization models. In 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019.Google ScholarGoogle ScholarCross RefCross Ref
  26. L. Liu, Y. Lu, M. Yang, Q. Qu, J. Zhu, and H. Li. 2018. Generative adversarial network for abstractive text summarization. In Proceedings of the AAAI Conference on Artificial Intelligence, 2018.Google ScholarGoogle Scholar
  27. T. Zhang, H. Ji, and A. Sil. 2019. Joint entity and event extraction with generative adversarial imitation learning. Data Intelligence 1, 2 (2019), 99–120.Google ScholarGoogle ScholarCross RefCross Ref
  28. T. Sabbah, A. Selamat, M. H. Selamat, F. S. Al-Anzi, E. H. Viedma, O. Krejcar, and H. Fujita. 2017. Modified frequency-based term weighting schemes for text classification. Applied Soft Computing 58, (2017), 193–206.Google ScholarGoogle ScholarCross RefCross Ref
  29. R. Nallapati, F. Zhai, and B. Zhou. 2017. Summarunner: A recurrent neural network based sequence model for extractive summarization of documents. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, 2017. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Z. Zhang, Y. Wu, H. Zhao, Z. Li, S. Zhang, X. Zhou, and X. Zhou. Semantics-aware BERT for language understanding. In Proceedings of the AAAI Conference on Artificial Intelligence.Google ScholarGoogle Scholar
  31. C. D. Manning, M. Surdeanu, J. Bauer, J. R. Finkel, S. Bethard, and D. McClosky. 2014. The stanford coreNLP natural language processing toolkit. In Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations 2014.Google ScholarGoogle ScholarCross RefCross Ref
  32. R. Catelli, V. Casola, G. De Pietro, H. Fujita, and M. Esposito. 2021. Combining contextualized word representation and sub-document level analysis through Bi-LSTM+ CRF architecture for clinical de-identification. Knowledge-Based Systems 213, 106649, 2021.Google ScholarGoogle ScholarCross RefCross Ref
  33. J. Howard and S. Ruder. 2018. Universal language model fine-tuning for text classification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, 2018.Google ScholarGoogle Scholar
  34. VASWANI, Ashish et al. Attention is all you need. In Advances in Neural Information Processing Systems. Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. T. Mikolov, K. Chen, G. Corrado, and J. Dean. 2013. Efficient estimation of word representations in vector space. In 1st International Conference on Learning Representations (ICRL), 2013.Google ScholarGoogle Scholar
  36. J. Pennington, R. Socher, and C. D. Manning. 2014. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2014.Google ScholarGoogle ScholarCross RefCross Ref
  37. F. Rousseau, E. Kiagias, and M. Vazirgiannis. 2015. Text categorization as a graph classification problem. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, 2015.Google ScholarGoogle Scholar
  38. L. Yao, C. Mao, and Y. Luo. 2019. Graph convolutional networks for text classification. In Proceedings of the AAAI Conference on Artificial Intelligence 2019.Google ScholarGoogle Scholar
  39. K. Lee, L. He, M. Lewis, and L. Zettlemoyer. 2017. End-to-end neural coreference resolution. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, EMNLP, 2017.Google ScholarGoogle Scholar
  40. G. Durrett, T. Berg-Kirkpatrick, and D. Klein. 2016. Learning-based single-document summarization with compression and anaphoricity constraints. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 2016.Google ScholarGoogle Scholar
  41. C. Y. Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Text Summarization Branches Out 2004.Google ScholarGoogle Scholar
  42. W. Wang and B. Chang. 2016. Graph-based dependency parsing with bidirectional LSTM. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 2016.Google ScholarGoogle Scholar

Index Terms

  1. SE4ExSum: An Integrated Semantic-aware Neural Approach with Graph Convolutional Network for Extractive Text Summarization

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Asian and Low-Resource Language Information Processing
      ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 20, Issue 6
      November 2021
      439 pages
      ISSN:2375-4699
      EISSN:2375-4702
      DOI:10.1145/3476127
      Issue’s Table of Contents

      Copyright © 2021 Association for Computing Machinery.

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 1 September 2021
      • Accepted: 1 May 2021
      • Received: 1 December 2020
      Published in tallip Volume 20, Issue 6

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Refereed

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format