Abstract
In today’s rapidly evolving digital era, educational institutions and professors are increasingly sharing video lectures online to support students. While these lectures effectively teach topics from scratch, they pose challenges for quick revisions. Viewers struggle to control the pace, often interrupting playback to navigate the content. Additionally, finding specific information within the video and skimming the unstructured transcript for relevant content can be cumbersome, hindering efficient learning. The objective of this paper is to provide a framework for summarizing video lectures to facilitate quick revisions. We further aim to help students with topic-wise preparation by summarizing specific lecture sections relevant to them, saving them the extra effort of manually searching through the entire lectures. We employ shot detection, speech transcription, transcript summarization, and ontology tree to generate topic-wise slideshow summaries. The summarization models: BART, T5, GPT3, Extractive, and Seq2Seq are evaluated on the VT-Ssum dataset, and their ROUGE scores are compared to select the best model.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Andra, M.B., Usagawa, T.: Automatic lecture video content summarization with attention-based recurrent neural network. In: 2019 International Conference of Artificial Intelligence and Information Technology (ICAIIT), pp. 54–59. IEEE (2019)
Brown, T., et al.: Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 33, 1877–1901 (2020)
Che, X., Yang, H., Meinel, C.: Lecture video segmentation by automatically analyzing the synchronized slides. In: Proceedings of the 21st ACM International Conference on Multimedia, pp. 345–348 (2013)
Chen, Y.N., Huang, Y., Yeh, C.F., Lee, L.S.: Spoken lecture summarization by random walk over a graph constructed with automatically extracted key terms. In: Twelfth Annual Conference of the International Speech Communication Association (2011)
Chopra, S., Auli, M., Rush, A.M.: Abstractive sentence summarization with attentive recurrent neural networks. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 93–98 (2016)
Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7871–7880 (2020)
Lv, T., Cui, L., Vasilijevic, M., Wei, F.: VT-SSum: a benchmark dataset for video transcript segmentation and summarization. arXiv preprint arXiv:2106.05606 (2021)
Ma, D., Agam, G.: Lecture video segmentation and indexing. In: Document Recognition and Retrieval XIX, vol. 8297, pp. 238–245. SPIE (2012)
Nallapati, R., Zhou, B., Gu?lçehre, C., Xiang, B.: Abstractive text summarization using sequence-to-sequence RNNs and beyond. In: Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning, pp. 280–290 (2016)
Paulus, R., Xiong, C., Socher, R.: A deep reinforced model for abstractive summarization. In: 6th International Conference on Learning Representations, ICLR 2018, Conference Track Proceedings (2018)
Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1), 5485–5551 (2020)
Rush, A.M., Chopra, S., Weston, J.: A neural attention model for abstractive sentence summarization. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 379–389 (2015)
Salatino, A.A., Thanapalasingam, T., Mannocci, A., Birukou, A., Osborne, F., Motta, E.: The computer science ontology: a comprehensive automatically-generated taxonomy of research areas. Data Intell. 2(3), 379–416 (2020)
See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Volume 1: Long Papers), pp. 1073–1083 (2017)
Shi, T., Keneshloo, Y., Ramakrishnan, N., Reddy, C.K.: Neural abstractive text summarization with sequence-to-sequence models. ACM Trans. Data Sci. 2(1), 1–37 (2021)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Khetarpaul, S., Jain, L., Goyal, K., Tej, P.V. (2024). Lecture Video Summarization Using Deep Learning. In: Nguyen, N.T., et al. Recent Challenges in Intelligent Information and Database Systems. ACIIDS 2024. Communications in Computer and Information Science, vol 2145. Springer, Singapore. https://doi.org/10.1007/978-981-97-5934-7_9
Download citation
DOI: https://doi.org/10.1007/978-981-97-5934-7_9
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-97-5933-0
Online ISBN: 978-981-97-5934-7
eBook Packages: Computer ScienceComputer Science (R0)