Abstract
AMR-to-text generation aims to recover a text containing the same meaning as an input AMR graph. Current research develops increasingly powerful graph encoders to better represent AMR graphs, with decoders based on standard language modeling being used to generate outputs. We propose a decoder that back predicts projected AMR graphs on the target sentence during text generation. As the result, our outputs can better preserve the input meaning than standard decoders. Experiments on two AMR benchmarks show the superiority of our model over the previous state-of-the-art system based on graph Transformer.
Cite
CITATION STYLE
Bai, X., Song, L., & Zhang, Y. (2020). Online back-parsing for AMR-to-text generation. In EMNLP 2020 - 2020 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference (pp. 1206–1219). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.emnlp-main.92
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.