Abstract
The state-of-the-art solutions for extracting multiple entity-relations from an input paragraph always require a multiple-pass encoding on the input. This paper proposes a new solution that can complete the multiple entity-relations extraction task with only one-pass encoding on the input corpus, and achieve a new state-of-the-art accuracy performance, as demonstrated in the ACE 2005 benchmark. Our solution is built on top of the pre-trained self-attentive models (Transformer). Since our method uses a single-pass to compute all relations at once, it scales to larger datasets easily; which makes it more usable in real-world applications.
Cite
CITATION STYLE
Wang, H., Tan, M., Yu, M., Chang, S., Wang, D., Xu, K., … Potdar, S. (2020). Extracting multiple-relations in one-pass with pre-trained transformers. In ACL 2019 - 57th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (pp. 1371–1377). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p19-1132
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.