Skip to main content
Log in

Distant context aware text generation from abstract meaning representation

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Text generation from abstract meaning representation is a fundamental task in natural language generation. An interesting challenge is that distant context could influence the surface realization for each node. In the previous encoder-decoder based approaches, graph neural networks have been commonly used to encode abstract meaning representation graphs and exhibited superior performance over the sequence and tree encoders. However, most of them cannot stack numerous layers, thus being too shallow to capture distant context. In this paper, we propose solutions from three aspects. Firstly, we introduce a Transformer based graph encoder to embed abstract meaning representation graphs. This encoder can stack more layers to encode larger context, while without performance degrading. Secondly, we expand the receptive field of each node, i.e. building direct connections between node pairs, to capture the information of its distant neighbors. We also exploit relative position embedding to make the model aware of the original hierarchy of graphs. Thirdly, we encode the linearized version of abstract meaning representation with the pre-trained language model to get the sequence encoding and incorporate it into graph encoding to enrich features. We conduct experiments on LDC2015E86 and LDC2017T10. Experimental results demonstrate that our method outperforms previous strong baselines. Especially, we investigate the performance of our model on large graphs, finding a larger performance gain. Our best model achieves 31.99 of BLEU and 37.02 of METEOR on LDC2015E86, 34.21 of BLEU, and 39.26 of METEOR on LDC2017T10, which are new states of the art.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

Notes

  1. https://propbank.github.io

  2. Both the ancestor and descendant of each node are based on forward edges in the following.

  3. Most of the above hyper parameters are same with [16].

  4. All the results are achieved by the models that trained without additional training data.

References

  1. Chen B, Barzilay R, Jaakkola T (2019) Path-augmented graph transformer network. In: Learning and reasoning with graph-structured representations (ICML workshop)

  2. Bahdanau D, Cho K, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: 3rd International conference on learning representations

  3. Banarescu L, Bonial C, Cai S, Georgescu M, Griffitt K, Hermjakob U, Knight K, Koehn P, Palmer M, Schneider N (2013) Abstract meaning representation for sembanking. In: Proceedings of the 7th Linguistic annotation workshop and interoperability with discourse, pp 178–186

  4. Banerjee S, Lavie A (2005) Meteor: An automatic metric for mt evaluation with improved correlation with human judgments. In: Proceedings of the acl workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization, pp 65–72

  5. Beck D, Haffari G, Cohn T (2018) Graph-to-sequence learning using gated graph neural networks. In: Proceedings of the 56th Annual meeting of the association for computational linguistics, pp 273–283

  6. Cai D, Lam W (2020) Graph transformer for graph-to-sequence learning. In: Proceedings of The Thirty-Fourth AAAI conference on artificial intelligence (AAAI)

  7. Cao K, Clark S (2019) Factorising amr generation through syntax. In: Proceedings of the 2019 Conference of the North American chapter of the association for computational linguistics: human language technologies, pp 2157–2163

  8. Chen Y, Wu L, Zaki MJ (2020) Reinforcement learning based graph-to-sequence model for natural question generation. In: International conference on learning representations

  9. Cho K, van Merrienboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using rnn encoder–decoder for statistical machine translation. In: Proceedings of the 2014 conference on empirical methods in natural language processing, pp 1724–1734

  10. Damonte M, Cohen SB (2019) Structural neural encoders for amr-to-text generation. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, pp 3649–3658

  11. Devlin J, Chang MW, Lee K, Toutanova K (2019) Bert: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American chapter of the association for computational linguistics: human language technologies, vol 1, pp 4171–4186

  12. Choi E, Xu Z, Li Y, Dusenberry MW, Flores G, Xue Y, Dai AM (2020) Learning the graphical structure of electronic health records with graph convolutional transformer. In: Proceedings of the thirty-fourth AAAI conference on artificial intelligence (AAAI)

  13. Flanigan J, Dyer C, Smith NA, Carbonell J (2016) Generation from abstract meaning representation using tree transducers. In: Proceedings of the 2016 conference of the north american chapter of the association for computational linguistics, pp 731–739

  14. Gao H, Ji S (2019) Graph U-nets. In: Proceedings of the 36th international conference on machine learning

  15. Gu J, Lu Z, Li H, Li VO (2016) Incorporating copying mechanism in sequence-to-sequence learning. In: Proceedings of the 54th annual meeting of the association for computational linguistics, pp 1631–1640

  16. Guo Z, Zhang Y, Teng Z, Lu W (2019) Densely connected graph convolutional networks for graph-to-sequence learning. Trans Assoc Comput Linguistics 7:297–312

    Article  Google Scholar 

  17. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: 2016 IEEE Conference on computer vision and pattern recognition, CVPR, pp 770–778

  18. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Computation 9 (8):1735–1780

    Article  Google Scholar 

  19. Hu Z, Dong Y, Wang K, Sun Y (2020) Heterogeneous graph transformer. In: Proceedings of the 2020 World Wide Web Conference

  20. Huang G, Liu Z, van der Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: 2017 IEEE conference on computer vision and pattern recognition, CVPR 2017. IEEE Computer Society, pp 2261–2269

  21. Kasper RT (1989) A flexible interface for linking applications to Penman’s sentence generator. In: Speech and natural language: proceedings of a workshop held at Philadelphia, Pennsylvania, February 21-23, 1989

  22. Kipf TN, Welling M (2017) Semi-supervised classification with graph convolutional networks. In: Proceedings of the 5th international conference on learning representations

  23. Koncel-Kedziorski R, Bekal D, Luan Y, Lapata M, Hajishirzi H (2019) Text generation from knowledge graphs with graph transformers. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, pp 2284–2293

  24. Konstas I, Iyer S, Yatskar M, Choi Y, Zettlemoyer L (2017) Neural amr: Sequence-to-sequence models for parsing and generation. In: Proceedings of the 55th annual meeting of the association for computational linguistics, pp 146–157

  25. Li Q, Han Z, Wu XM (2018) Deeper insights into graph convolutional networks for semi-supervised learning. In: Thirty-Second AAAI conference on artificial intelligence

  26. Liao K, Lebanoff L, Liu F (2018) Abstract meaning representation for multi-document summarization. In: Proceedings of the 27th international conference on computational linguistics, pp 1178–1190

  27. Manuel M, Ramón FA, Tahira N, Md Arafat S, Young-Suk L, Radu F, Salim R (2020) Gpt-too: A language-model-first approach for amr-to-text generation. In: Proceedings of the 58th annual meeting of the association for computational linguistics. Association for computational linguistics, Seattle, USA

  28. Papineni K, Roukos S, Ward T, Zhu WJ (2002) Bleu: a method for automatic evaluation of machine translation. In: Proceedings of the 40th annual meeting on association for computational linguistics. Association for Computational Linguistics, pp 311–318

  29. Petar V, Guillem C, Arantxa C, Adriana R, Pietro L, Yoshua B (2018) Graph attention networks. In: 6th International conference on learning representations, ICLR 2018

  30. Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I (2019) Language models are unsupervised multitask learners. OpenAI Blog 1(8)

  31. Rao S, Marcu D, Knight K, Daumé III H (2017) Biomedical event extraction using abstract meaning representation. In: BioNLP 2017. Association for Computational Linguistics, Vancouver, pp 126–135

  32. Perozzi B, Kapoor A, Harutyunyan H, Alipourfard N, Lerman K, Ver Steeg G, Galstyan A (2019) Mixhop: Higher-order graph convolutional architectures via sparsified neighborhood mixing. In: Proceedings of the 36th international conference on machine learning

  33. See A, Liu PJ, Manning CD (2017) Get to the point: Summarization with pointer-generator networks. In: Proceedings of the 55th annual meeting of the association for computational linguistics, vol 1, pp 1073–1083

  34. Song L, Gildea D, Zhang Y, Wang Z, Su J (2019) Semantic neural machine translation using amr. Trans Assoc Comput Linguistics 7:19–31

    Article  Google Scholar 

  35. Song L, Peng X, Zhang Y, Wang Z, Gildea D (2017) Amr-to-text generation with synchronous node replacement grammar. In: Proceedings of the 55th Annual meeting of the association for computational linguistics, pp 7–13

  36. Song L, Zhang Y, Wang Z, Gildea D (2018) A graph-to-sequence model for amr-to-text generation. In: Proceedings of the 56th annual meeting of the association for computational linguistics, pp 1616–1626

  37. Takase S, Suzuki J, Okazaki N, Hirao T, Nagata M (2016) Neural headline generation on abstract meaning representation. In: Proceedings of the 2016 conference on empirical methods in natural language processing, pp 1054–1059

  38. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems, pp 5998–6008

  39. Wang T, Wan X, Jin H (2020) Amr-to-text generation with graph transformer. Trans Assoc Comput Linguistics 8:19–33

    Article  Google Scholar 

  40. Xu K, Li C, Tian Y, Sonobe T, Kawarabayashi K, Jegelka S (2018) Representation learning on graphs with jumping knowledge networks. In: Proceedings of the 35th international conference on machine learning, vol 80, pp 5449–5458

  41. Xu K, Wu L, Wang Z, Feng Y, Sheinin V (2018) Graph2seq: Graph to sequence learning with attention-based neural networks. arXiv:1804.00823

  42. Yao S, Wang T, Wan X (2020) Heterogeneous graph transformer for graph-to-sequence learning. In: Proceedings of the 58th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Online

  43. Yun S, Jeong M, Kim R, Kang J, Kim HJ (2019) Graph transformer networks. In: Advances in neural information processing systems, vol 32. Curran Associates Inc, pp 11983–11993

  44. Zhao Y, Chen L., Chen Z, Cao R, Zhu S, Yu K (2020) Line graph enhanced AMR-to-text generation with mix-order graph attention networks. In: Proceedings of the 58th Annual meeting of the association for computational linguistics, pp 732–741. Association for Computational Linguistics, Online. https://doi.org/10.18653/v1/2020.acl-main.67. https://www.aclweb.org/anthology/2020.acl-main.67

  45. Zhu J, Li J, Zhu M, Qian L, Zhang M, Zhou G (2019) Modeling graph structure in transformer for better AMR-to-text generation. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pp 5459–5468

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sen Yang.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Yang, S., Feng, D., Liu, Y. et al. Distant context aware text generation from abstract meaning representation. Appl Intell 52, 1672–1685 (2022). https://doi.org/10.1007/s10489-021-02431-1

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-021-02431-1

Keywords

Navigation