Constructing Code-mixed Universal Dependency Forest for Unbiased Cross-lingual Relation Extraction

2Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

Latest efforts on cross-lingual relation extraction (XRE) aggressively leverage the language-consistent structural features from the universal dependency (UD) resource, while they may largely suffer from biased transfer (e.g., either target-biased or source-biased) due to the inevitable linguistic disparity between languages. In this work, we investigate an unbiased UD-based XRE transfer by constructing a type of code-mixed UD forest. We first translate the sentence of the source language to the parallel target-side language, for both of which we parse the UD tree respectively. Then, we merge the source-/target-side UD structures as a unified code-mixed UD forest. With such forest features, the gaps of UD-based XRE between the training and predicting phases can be effectively closed. We conduct experiments on the ACE XRE benchmark datasets, where the results demonstrate that the proposed code-mixed UD forests help unbiased UD-based XRE transfer, with which we achieve significant XRE performance gains.

Cite

CITATION STYLE

APA

Fei, H., Zhang, M., Zhang, M., & Chua, T. S. (2023). Constructing Code-mixed Universal Dependency Forest for Unbiased Cross-lingual Relation Extraction. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 9395–9408). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.599

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free