Abstract
This paper investigates the construction of a strong baseline based on general purpose sequence-to-sequence models for constituency parsing. We incorporate several techniques that were mainly developed in natural language generation tasks, e.g., machine translation and summarization, and demonstrate that the sequence-to-sequence model achieves the current top-notch parsers’ performance without requiring explicit task-specific knowledge or architecture of constituent parsing.
Cite
CITATION STYLE
Suzuki, J., Takase, S., Kamigaito, H., Morishita, M., & Nagata, M. (2018). An empirical study of building a strong baseline for constituency parsing. In ACL 2018 - 56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 2, pp. 612–618). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p18-2097
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.