Skip to main content

Learning to Consider Relevance and Redundancy Dynamically for Abstractive Multi-document Summarization

  • Conference paper
  • First Online:
Natural Language Processing and Chinese Computing (NLPCC 2020)

Abstract

As one of the most essential tasks for information aggregation, multi-document summarization is faced with information redundancy of source document clusters. Recent works have attempted to avoid redundancy while generating summaries. Most state-of-the-art multi-document summarization systems are either extractive or abstractive with an external extractive model. In this paper, we propose an end-to-end abstractive model based on Transformer to generate summaries, considering relevance and redundancy dynamically and jointly. Specifically, we employ sentence masks and design a sentence-level transformer layer for learning sentence representations in a hierarchical manner. Then we use a dynamic Max Marginal Relevance (MMR) model to discern summary-worthy sentences and modify the encoder-decoder attention. We also utilize the pointer mechanism, taking the mean attention of all transformer heads as the probability to copy words from the source text. Experimental results demonstrate that our proposed model outperforms several strong baselines. We also conduct ablation studies to verify the effectiveness of our key mechanisms.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Angelidis, S., Lapata, M.: Summarizing opinions: aspect extraction meets sentiment prediction and they are both weakly supervised. In: EMNLP, pp. 3675–3686. Association for Computational Linguistics (2018)

    Google Scholar 

  2. Antognini, D., Faltings, B.: Learning to create sentence semantic relation graphs for multi-document summarization. In: EMNLP-IJCNLP 2019, vol. 32 (2019)

    Google Scholar 

  3. Cao, Z., Li, W., Li, S., Wei, F.: Improving multi-document summarization via text classification. In: AAAI, pp. 3053–3059. AAAI Press (2017)

    Google Scholar 

  4. Carbonell, J.G., Goldstein, J.: The use of MMR, diversity-based reranking for reordering documents and producing summaries. In: SIGIR, pp. 335–336. ACM (1998)

    Google Scholar 

  5. Cho, S., Lebanoff, L., Foroosh, H., Liu, F.: Improving the similarity measure of determinantal point processes for extractive multi-document summarization. In: ACL, vol. 1, pp. 1027–1038. Association for Computational Linguistics (2019)

    Google Scholar 

  6. Christensen, J., Mausam, Soderland, S., Etzioni, O.: Towards coherent multi-document summarization. In: HLT-NAACL, pp. 1163–1173. The Association for Computational Linguistics (2013)

    Google Scholar 

  7. Chu, E., Liu, P.J.: MeanSum: a neural model for unsupervised multi-document abstractive summarization. In: Proceedings of Machine Learning Research, PMLR, vol. 97, pp. 1223–1232. ICML (2019)

    Google Scholar 

  8. Erkan, G., Radev, D.R.: LexRank: graph-based lexical centrality as salience in text summarization. J. Artif. Intell. Res. 22, 457–479 (2004)

    Article  Google Scholar 

  9. Fabbri, A.R., Li, I., She, T., Li, S., Radev, D.R.: Multi-news: a large-scale multi-document summarization dataset and abstractive hierarchical model. In: ACL, 1, pp. 1074–1084. Association for Computational Linguistics (2019)

    Google Scholar 

  10. Gehrmann, S., Deng, Y., Rush, A.M.: Bottom-up abstractive summarization. In: EMNLP, pp. 4098–4109. Association for Computational Linguistics (2018)

    Google Scholar 

  11. Hao, T.: Overview of DUC 2005. In: Proceedings of the Document Understanding Conference (DUC 2005) (2005)

    Google Scholar 

  12. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: ICLR (Poster) (2015)

    Google Scholar 

  13. Lebanoff, L., Song, K., Liu, F.: Adapting the neural encoder-decoder framework from single to multi-document summarization. In: EMNLP, pp. 4131–4141. Association for Computational Linguistics (2018)

    Google Scholar 

  14. Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74–81 (2004)

    Google Scholar 

  15. Liu, P.J., et al.: Generating wikipedia by summarizing long sequences. In: ICLR (Poster). OpenReview.net (2018)

    Google Scholar 

  16. Liu, Y., Lapata, M.: Hierarchical transformers for multi-document summarization. In: ACL, vol. 1, pp. 5070–5081. Association for Computational Linguistics (2019)

    Google Scholar 

  17. Louviere, J.J., Flynn, T.N., Marley, A.A.J.: Best-Worst Scaling: Theory Methods and Applications. Cambridge University Press, Cambridge (2015)

    Book  Google Scholar 

  18. Luo, W., Liu, F., Liu, Z., Litman, D.J.: Automatic summarization of student course feedback. In: HLT-NAACL, pp. 80–85. The Association for Computational Linguistics (2016)

    Google Scholar 

  19. Mihalcea, R., Tarau, P.: TextRank: bringing order into text. In: EMNLP, pp. 404–411. ACL (2004)

    Google Scholar 

  20. See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: ACL, vol. 1, pp. 1073–1083. Association for Computational Linguistics (2017)

    Google Scholar 

  21. Vaswani, A., et al.: Attention is all you need. In: NIPS, pp. 5998–6008 (2017)

    Google Scholar 

  22. Vinyals, O., Fortunato, M., Jaitly, N.: Pointer networks. In: NIPS, pp. 2692–2700 (2015)

    Google Scholar 

  23. Yang, Z., Dai, Z., Yang, Y., Carbonell, J.G., Salakhutdinov, R., Le, Q.V.: XLNet: generalized autoregressive pretraining for language understanding. In: NeurIPS, pp. 5754–5764 (2019)

    Google Scholar 

  24. Yasunaga, M., Zhang, R., Meelu, K., Pareek, A., Srinivasan, K., Radev, D.R.: Graph-based neural multi-document summarization. In: CoNLL, pp. 452–462. Association for Computational Linguistics (2017)

    Google Scholar 

  25. Zhang, J., Tan, J., Wan, X.: Adapting neural single-document summarization model for abstractive multi-document summarization: a pilot study. In: INLG, pp. 381–390. Association for Computational Linguistics (2018)

    Google Scholar 

Download references

Acknowledgments

This research work has been funded by the National Natural Science Foundation of China (Grant No. 61772337, U1736207).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Gongshen Liu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Liu, Y., Fan, X., Zhou, J., He, C., Liu, G. (2020). Learning to Consider Relevance and Redundancy Dynamically for Abstractive Multi-document Summarization. In: Zhu, X., Zhang, M., Hong, Y., He, R. (eds) Natural Language Processing and Chinese Computing. NLPCC 2020. Lecture Notes in Computer Science(), vol 12430. Springer, Cham. https://doi.org/10.1007/978-3-030-60450-9_38

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-60450-9_38

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-60449-3

  • Online ISBN: 978-3-030-60450-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics