Abstract
Linguistic metrics based on syntactic and semantic information have proven very effective for Automatic MT Evaluation. However, no results have been presented so far on their performance when applied to heavily ill-formed low quality translations. In order to glean some light into this issue, in this work we present an empirical study on the behavior of a heterogeneous set of metrics based on linguistic analysis in the paradigmatic case of speech translation between non-related languages. Corroborating previous findings, we have verified that metrics based on deep linguistic analysis exhibit a very robust and stable behavior at the system level. However, these metrics suffer a significant decrease at the sentence level. This is in many cases attributable to a loss of recall, due to parsing errors or to a lack of parsing at all, which may be partially ameliorated by backing off to lexical similarity.
Cite
CITATION STYLE
Giḿenez, J., & M`arquez, L. (2009). On the Robustness of Syntactic and Semantic Features for Automatic MT Evaluation. In EACL 2009 - 4th Workshop on Statistical Machine Translation, Proceedings of theWorkshop (pp. 250–258). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1626431.1626479
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.