Abstract
According to Aggarwal [6] the extractive summarization is solely about scoring sentences to maximize the topical coverage and minimize redundancy, while coherence and fluency are to be considered only in the case of abstractive summaries. It is a rather strong opinion that this paper aims to argue with by introducing different functions of text summarization, the notion of text coherence and cohesion, and last but not least by offering new methods allowing for both sentence extraction and text fluency (to a certain level). This article aims at offering a model that will be as simple as possible (but no simpler, as would Einstein put that) to satisfy the goal of informative extractive summary with a certain level of cohesion determined by sentence connectives. This abstract has been generated with the algorithm proposed in the paper.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Dunning, T.: Accurate methods for the statistics of surprise and coincidence. Comput. Linguist. 19(1), 61–74 (1993)
Allahyari, M., et al.: Text summarization techniques: a brief survey. Int. J. Adv. Comput. Sci. Appl. 8(10) (2017). https://doi.org/10.14569/IJACSA.2017.081052
Moratanch, N., Chitrakala, S.: A survey on extractive text summarization. In: IEEE International Conference on Computer, Communication, and Signal Processing (2017)
Indu, M., Kavitha, K.V.: Review on text summarization evaluation methods. In: International Conference on Research Advances in Integrated Navigation Systems (2016)
Daume III, H., Marcu, D.: Bayesian query-focused summarizing. In: Proceedings of the 21st International Conference on Computational Linguistics and the 44th Annual Meeting of the Association for Computational Linguistics, pp. 305–312. Association for Computational Linguistics (2006)
Aggarwal, C.C.: Machine Learning for Text. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-73531-3
Luhn, H.P.: The automatic creation of literature abstracts. IBM J. Res. Dev. 2(2), 159–165 (1958)
Parveen, D., Ramsl, H.-M., Strube, M.: Topical coherence for graph-based extractive summarization. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics (2015)
Parveen, D., Strube, M.: Integrating importance, non-redundancy and coherence in graph-based extractive summarization. In: Proceedings of the 24h International Joint Conference on Artificial Intelligence (2015)
Saziyabegum, S., Sajja, P.S.: Literature review on extractive text summarization approaches. Int. J. Comput. Appl. 156(12), 28–36 (2016)
McNamara, D.S., Graesser, A.C., McCarthy, P.M., Cai, Z.: Automated Evaluation of Text and Discourse with Coh-Metrix. Cambridge University Press, Cambridge (2014)
Deerwester, S.C., Dumais, S.T., Landauer, T.K., Furnas, G.W., Harshman, R.A.: Indexing by latent semantic analysis. JASIS 41(6), 391–407 (1990)
Mihalcea, R., Tarau, P.: TextRank: bringing order into texts. In: Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics (2014)
Gong, Y., Liu, X.: Generic text summarizing using relevance measure and latent semantic analysis. In: Proceedings of the 24th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 19–25. ACM (2001)
Radev, D.R., Jing, H., Styś, M., Tam, D.: Centroid-based summarization of multiple documents. Inf. Process. Manag. 40(6), 919–938 (2004)
Halliday, M.A.K., Hasan, R.: Cohesion in English. Longman Group Limited, London (1976)
Alonso i Alemany, L., Fort, M.F.: Integrating cohesion and coherence for Automatic Summarization (2003). https://www.aclweb.org/anthology/E03-3002.pdf. Accessed 28 Jan 2020
McNamara, D.S., Louwerse, M.M., McCarthy, P.M., Graesser, A.C.: Coh-metrix: capturing linguistic features of cohesion. Discourse Process. 47(4), 292–330 (2010). https://doi.org/10.1080/01638530902959943
Wang, D., Zhu, S., Li, T., Gong, Y.: Multidocument summarization using sentence-based topic models. In: Proceedings of the ACL-IJCNLP 2009 Conference Short Papers, pp. 297–300. Association for Computational Linguistics (2009)
Kiros, R., et al.: Skip-Thought Vectors (2015). https://arxiv.org/abs/1506.06726. Accessed 28 Jan 2020
Barrios, F., Lopez, F., Argerich, L., Wachenchauzer, R.: Variations of the Similarity Function of TextRank for Automated Summarization (2016). https://arxiv.org/abs/1602.03606. Accessed 28 Jan 2020
Kristina Toutanova, K., Klein, D., Manning, Ch., Singer, Y.: Feature-rich part-of-speech tagging with a cyclic dependency network. In: Proceedings of HLT-NAACL (2003)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Okulska, I. (2020). Team Up! Cohesive Text Summarization Scoring Sentence Coalitions. In: Rutkowski, L., Scherer, R., Korytkowski, M., Pedrycz, W., Tadeusiewicz, R., Zurada, J.M. (eds) Artificial Intelligence and Soft Computing. ICAISC 2020. Lecture Notes in Computer Science(), vol 12416. Springer, Cham. https://doi.org/10.1007/978-3-030-61534-5_35
Download citation
DOI: https://doi.org/10.1007/978-3-030-61534-5_35
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-61533-8
Online ISBN: 978-3-030-61534-5
eBook Packages: Computer ScienceComputer Science (R0)