Abstract
With the emergence of deep learning, the attention of researchers has increased significantly towards abstractive text summarization approaches. Though extractive text summarization (ETS) is an important approach, the generated summaries are not always coherent. This paper mainly focuses on the abstractive text summarization (ATS) approach for Telugu language to generate coherent summary. The majority research on ATS approach is conducted in English, while no significant research in Telugu has been documented. An abstractive Telugu text summarization model based on sequence-to-sequence (seq2seq) encoder-decoder architecture is proposed in this paper. The seq2seq model is implemented with bidirectional long short-term memory (Bi-LSTM) based encoder and long short-term memory (LSTM) based decoder. The existing ATS approaches have some drawbacks such as they cannot handle out vocabulary words, attention deficiency issue arising while handling long text sequence and repetition problem. To overcome these issues, some operating mechanisms like pointer generator network, temporal attention mechanism and coverage mechanism are also integrated in the proposed model. Besides, diverse beam search decoding algorithm is also employed to increase the diversity of generated summary. Thus, the proposed seq2seq model is the combination of Bi-LSTM and LSTM based encoder-decoder, pointer generator network, temporal attention mechanism, coverage mechanism and diverse beam search decoding algorithm. The performance of the proposed work is evaluated using the ROUGE toolkit in terms of F-measure, recall and precision. The experimental results of the proposed scheme are evaluated with other existing methods to show that the proposed ATS model outperforms existing Telugu text summarization models.
Similar content being viewed by others
Data availability
Data will be shared on the reasonable request.
References
Allahyari M, Pouriyeh S, Assefi M, Safaei S, Trippe ED, Gutierrez JB, Kochut K (2017) Text summarization techniques: a brief survey. arXiv. https://doi.org/10.1177/1010428317692226
Alquliti WH, Abdul Ghani NB (2019) Convolutional neural network based for automatic text summarization. Int J Advan Comput Sci Applic (IJACSA) 10(4). https://doi.org/10.14569/IJACSA.2019.0100424
Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate. ArXiv:1409–0473
Bengio Y, Simard P, Frasconi P (1994) Learning long-term dependencies with gradient descent is difficult. IEEE Trans Neural Netw 5(2):157–166. https://doi.org/10.1109/72.279181
Chung J, Gulcehre C, Cho K and Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling In NIPS 2014 Workshop on Deep Learning: https://doi.org/10.1109/TVCG.2013.272
Cibils A, Musat C, Hossman A, Baeriswyl M (2018) Diverse beam search for increased novelty in abstractive summarization. arXiv preprint arXiv:1802.01457
Gimpel K, Batra D, Dyer C, Shakhnarovich G (2013) A systematic exploration of diversity in machine translation. In proceedings of the 2013 conference on empirical methods in natural language processing. 1100–1111: https://doi.org/10.1016/j.cbpa.2013.03.038
Google colab: https://colab.research.google.com/notebooks/intro.ipynb#recent=true.
Gu J, Lu Z, Li H, Li VOK (2016) Incorporating copying mechanism in sequence-to-sequence learning. https://arxiv.org/abs/1603.06393, https://doi.org/10.13703/j.0255-2930.2016.11.022
Gulati AN, Sawarkar SD (2017) A novel technique for multidocument Hindi text summarization. 2017 international conference on nascent Technologies in Engineering (ICNTE). https://doi.org/10.1109/icnte.2017.7947890
Hernandez-Castaneda A, Garcia-Hernandez RA, Ledeneva Y, Millan-Hernandez CE (2020) Extractive automatic text summarization based on lexical-semantic keywords. IEEE Access, 1–1. https://doi.org/10.1109/access.2020.2980226
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735
Joudar NE, Ettaouil M (2022) KRR-CNN: kernels redundancy reduction in convolutional neural networks. Neural Comput Applic 34(3):2443–2454
Kallimani J, Srinivasa K, Eswara B (2011) Information extraction by an abstractive text summarization for an Indian regional language. 319–322. https://doi.org/10.1109/NLPKE.2011.6138217
Kanitha DK, Mubarak DMN, Shanavas SA (2018) Malayalam Text Summarization Using Graph Based Method. Int J Comput Sci Info Technol 9(2):40–44
Krause J, Johnson J, Krishna R, Fei-Fei L (2017) A hierarchical approach for generating descriptive image paragraphs. In proceedings of the IEEE conference on computer vision and pattern recognition. 317–325.
Latha YM, Sudha DN (2020) Multi-document abstractive text summarization through semantic similarity matrix for Telugu language. Int J Advanc Sci Technol 29(1):513–521 http://sersc.org/journals/index.php/IJAST/article/view/3105
Lin CY (2004) 2004. ROUGE, A package for automatic evaluation of summaries. Text Summarization Branches Out: https://doi.org/10.1179/cim.2004.5.Supplement-1.132
Loper E, Bird S (2002). NLTK: The Natural Language Toolkit. CoRR. cs.CL/0205028. https://doi.org/10.3115/1118108.1118117
Mamidala KK, Sanampudi S (2021) Text summarization on Telugu e-news based on long-short term memory with rectified Adam optimizer. Int J Comput Digital Syst 11(1)
Manjari KU (2020) Extractive summarization of Telugu documents using TextRank algorithm. 2020 fourth international conference on I-SMAC (IoT in social, Mobile, analytics and cloud) (I-SMAC). https://doi.org/10.1109/i-smac49090.2020.9243568
Mohammad Masum AK, Abujar S, Islam Talukder MA, Azad Rabby AS, Hossain SA (2019) Abstractive method of text summarization with sequence to sequence RNNs. 2019 10th international conference on computing, Commun Networking Technol (ICCCNT) https://doi.org/10.1109/icccnt45670.2019.8944620.
Mohan Bharath B, Aravindh Gowtham B, Akhil M (2022) Neural abstractive text summarizer for Telugu language. In soft computing and signal processing (pp. 61–70). Springer, Singapore
Naidu R, Bharti D, Babu K, Mohapatra R (2017) Text Summarization with Automatic Keyword Extraction in Telugu e-Newspapers: https://doi.org/10.21037/cdt.2017.08.14
Nallapati R, Zhou B, Santos CD, Gulçehre CG, Xiang B (2016) Abstractive text summarization using sequence-to-sequence RNNs and beyond. CoNLL 2016:280
Norouzi R, Baziyad H, Aknondzadeh Noghabi E, Albadvi A (2022) Developing tourism users’ profiles with data-driven explicit information. Math Probl Eng 2022:1–14
Pan HX, Liu H, Tang Y (2019) A sequence-to-sequence text summarization model with topic based attention mechanism. In: Ni W, Wang X, Song W, Li Y (eds) Web Information Systems and Applications. WISA 2019. Lecture notes in computer science, vol. 11817. Springer, Cham. https://doi.org/10.1007/978-3-030-30952-7_29.
Paulus R, Xiong C, Socher R (2017) A deep reinforced model for abstractive summarization. arXiv preprint arXiv:1705.04304: https://doi.org/10.3389/fpsyg.2017.01779
Priyadharshan T, Sumathipala S (2018) Text summarization for Tamil online sports news using NLP. 2018 3rd international conference on information technology research (ICITR). https://doi.org/10.1109/icitr.2018.8736154
Rodrigues S, Fernandes S, Pai A (2019) "Konkani Text Summarization By Sentence Extraction," 2019 10th international conference on computing, communication and networking technologies (ICCCNT), Kanpur, India, pp. 1–6, https://doi.org/10.1109/ICCCNT45670.2019.8944575
Rush AM, Chopra S, Weston J (2015) A neural attention model for abstractive sentence summarization. In proceedings of the 2015 conference on empirical methods in natural language processing. 379–389.
Rush AM, Chopra S, Weston J (2015) A neural attention model for abstractive sentence summarization. ArXiv:1509–00685
Sarwadnya VV, Sonawane SS (2018) Marathi extractive text summarizer using graph based model. 2018 fourth international conference on computing communication control and automation (ICCUBEA). https://doi.org/10.1109/iccubea.2018.8697741
See A, Liu PJ, Manning CD (2017) Get to the point: summarization with pointer-generator networks. In proceedings of the 55th annual meeting of the Association for Computational Linguistics (volume 1: long papers). Assoc Comput Linguistics. 1073–1083
Shi T, Keneshloo Y, Ramakrishnan N, Reddy CK (2020) Neural abstractive text summarization with sequence-to-sequence models. ACM Trans, Data Sci
Song S, Huang H, Ruan T (2018) Abstractive text summarization using LSTM-CNN based deep learning. Multimed Tools Appl 78:857–875. https://doi.org/10.1007/s11042-018-5749-3
Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. In: Proc. Adv Neural Inf Process Syst, 3104–3112
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. Advance Neural Inform Process Syst:6000–6010
Vinyals O, Fortunato M, Jaitly N (2015) Pointer networks. In: Proc. Adv. Neural Inf. Process. Syst., pp. 2692–2700
Wang S, Zhao X, Li B, Ge B, Tang D (2017) Integrating extractive and abstractive models for long text summarization. 2017 IEEE international congress on big data (BigData congress). https://doi.org/10.1109/bigdatacongress.2017.46
Zhang Y, Xiao W (2018) Keyphrase generation based on deep Seq2seq model. IEEE access, 1–1. https://doi.org/10.1109/access.2018.2865589
Zhang Y, Kampffmeyer M, Liang X, Zhang D, Tan M, Xing EP (2019) Dilated temporal relational adversarial network for generic video summarization. Multimed Tools Appl 78(24):35237–35261. https://doi.org/10.1007/s11042-019-08175-y
Zhang Y, Liang X, Zhang D, Tan M, Xing EP (2020) Unsupervised object-level video summarization with online motion auto-encoder. Pattern Recogn Lett 130:376–385. https://doi.org/10.1016/j.patrec.2018.07.030
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
I confirm that I have read, understand and agreed to the submission guidelines, policies and submission `declaration of the journal.
I confirm that the manuscript is the authors’ original work and the manuscript has not received prior publication and is not under consideration for publication elsewhere.
I confirm that all authors listed on the title page have contributed significantly to the work, have read the manuscript, attest to the validity and legitimacy of the data and its interpretation, and agree to its submission.
I confirm that the paper now submitted is not copied or plagiarized version of some other published work.
I declare that I shall not submit the paper for publication in any other Journal or Magazine till the decision is made by journal editors.
I understand that submission of false or incorrect information/undertaking would invite appropriate penal actions as per norms/rules of the journal.
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Babu, G.L.A., Badugu, S. Deep learning based sequence to sequence model for abstractive telugu text summarization. Multimed Tools Appl 82, 17075–17096 (2023). https://doi.org/10.1007/s11042-022-14099-x
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-022-14099-x