Skip to main content

Focus-Based Text Summarisation with Hybrid Embeddings

  • Conference paper
  • First Online:
Book cover AI 2021: Advances in Artificial Intelligence (AI 2022)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 13151))

Included in the following conference series:

Abstract

Text summarisation has been recognised as an important Natural Language Processing task, attracting great attention from both researchers and practitioners. It has been widely adopted in various domains. For example, text summarisation of news, articles and book chapters can produce a short text, assisting the readers with grasping the main idea rapidly. In the medical domain, it is also applied to summarise the patients’ questions. However, it is very challenging to control the summariser output by producing domain-specific summaries since the focus of domain-specific information may be ignored. In this paper, we propose a novel summarisation model aiming at producing summaries by focusing on the domain-specific knowledge, where hybrid embeddings, i.e., focus, domain and context embeddings, are utilised. We conduct extensive experiments to evaluate our novel model by using the MeQSum dataset. The experimental results demonstrate that our model outperforms state-of-the-art algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 89.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 119.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://github.com/abachaa/MeQSum.

  2. 2.

    https://colab.research.google.com/.

References

  1. Abacha, A.B., M’rabet, Y., Zhang, Y., Shivade, C., Langlotz, C., Demner-Fushman, D.: Overview of the MEDIQA 2021 shared task on summarization in the medical domain. In: Proceedings of the 20th Workshop on Biomedical Language Processing, pp. 74–85 (2021)

    Google Scholar 

  2. Balumuri, S., Bachina, S., Kamath, S.: Sb_nitk at MEDIQA 2021: leveraging transfer learning for question summarization in medical domain. In: Proceedings of the 20th Workshop on Biomedical Language Processing, pp. 273–279 (2021)

    Google Scholar 

  3. Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, vol. 1 (Long and Short Papers), pp. 4171–4186 (2019)

    Google Scholar 

  4. Fiszman, M., Rindflesch, T.C., Kilicoglu, H.: Abstraction summarization for managing the biomedical research literature. In: Proceedings of the Computational Lexical Semantics Workshop at HLT-NAACL 2004, pp. 76–83 (2004)

    Google Scholar 

  5. Gehrmann, S., Deng, Y., Rush, A.M.: Bottom-up abstractive summarization. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 4098–4109 (2018)

    Google Scholar 

  6. Gu, J., Lu, Z., Li, H., Li, V.O.: Incorporating copying mechanism in sequence-to-sequence learning. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1631–1640 (2016)

    Google Scholar 

  7. Guan, W., Smetannikov, I., Tianxing, M.: Survey on automatic text summarization and transformer models applicability. In: 2020 International Conference on Control, Robotics and Intelligent System, pp. 176–184 (2020)

    Google Scholar 

  8. He, Y., Chen, M., Huang, S.: damo_nlp at MEDIQA 2021: knowledge-based preprocessing and coverage-oriented reranking for medical question summarization. In: Proceedings of the 20th Workshop on Biomedical Language Processing, pp. 112–118 (2021)

    Google Scholar 

  9. Hermann, K.M., et al.: Teaching machines to read and comprehend. Adv. Neural. Inf. Process. Syst. 28, 1693–1701 (2015)

    Google Scholar 

  10. Kim, D., et al.: A neural named entity recognition and multi-type normalization tool for biomedical text mining. IEEE Access 7, 73729–73740 (2019)

    Article  Google Scholar 

  11. Lee, J., et al.: BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36(4), 1234–1240 (2020)

    Google Scholar 

  12. Lee, J., Dang, H., Uzuner, O., Henry, S.: MNLP at MEDIQA 2021: fine-tuning Pegasus for consumer health question summarization. In: Proceedings of the 20th Workshop on Biomedical Language Processing, pp. 320–327 (2021)

    Google Scholar 

  13. Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461 (2019)

  14. Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74–81 (2004)

    Google Scholar 

  15. Luhn, H.P.: The automatic creation of literature abstracts. IBM J. Res. Dev. 2(2), 159–165 (1958)

    Article  MathSciNet  Google Scholar 

  16. Luong, T., Pham, H., Manning, C.D.: Effective approaches to attention-based neural machine translation. In: EMNLP (2015)

    Google Scholar 

  17. Nallapati, R., Zhai, F., Zhou, B.: SummaRuNNER: a recurrent neural network based sequence model for extractive summarization of documents. In: Thirty-First AAAI Conference on Artificial Intelligence (2017)

    Google Scholar 

  18. Narayan, S., Cohen, S.B., Lapata, M.: Don’t give me the details, just the summary! topic-aware convolutional neural networks for extreme summarization. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 1797–1807 (2018)

    Google Scholar 

  19. Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(140), 1–67 (2020). http://jmlr.org/papers/v21/20-074.html

  20. Reeve, L.H., Han, H., Brooks, A.D.: The use of domain-specific concepts in biomedical text summarization. Inf. Process. Manage. 43(6), 1765–1776 (2007)

    Article  Google Scholar 

  21. Roberts, K., Kilicoglu, H., Fiszman, M., Demner-Fushman, D.: Decomposing consumer health questions. In: Proceedings of BioNLP 2014, pp. 29–37. Association for Computational Linguistics, Baltimore, Maryland, June 2014. https://doi.org/10.3115/v1/W14-3405, https://aclanthology.org/W14-3405

  22. Sarkar, K.: Using domain knowledge for text summarization in medical domain. Int. J. Recent Trends Eng. 1(1), 200 (2009)

    Google Scholar 

  23. See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1073–1083 (2017)

    Google Scholar 

  24. Wang, D., Liu, P., Zhong, M., Fu, J., Qiu, X., Huang, X.: Exploring domain shift in extractive text summarization. arXiv preprint arXiv:1908.11664 (2019)

  25. Xu, H., Liu, B., Shu, L., Philip, S.Y.: Double embeddings and CNN-based sequence labeling for aspect extraction. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 592–598 (2018)

    Google Scholar 

  26. Zhang, J., Zhao, Y., Saleh, M., Liu, P.: Pegasus: pre-training with extracted gap-sentences for abstractive summarization. In: International Conference on Machine Learning, pp. 11328–11339. PMLR (2020)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Weihua Li .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Shi, J., Hellesoe, L., Wang, G., Li, W., Bai, Q. (2022). Focus-Based Text Summarisation with Hybrid Embeddings. In: Long, G., Yu, X., Wang, S. (eds) AI 2021: Advances in Artificial Intelligence. AI 2022. Lecture Notes in Computer Science(), vol 13151. Springer, Cham. https://doi.org/10.1007/978-3-030-97546-3_57

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-97546-3_57

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-97545-6

  • Online ISBN: 978-3-030-97546-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics