Correlating human and automatic evaluation of a German surface realiser

15Citations
Citations of this article
97Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We examine correlations between native speaker judgements on automatically generated German text against automatic evaluation metrics. We look at a number of metrics from the MT and Summarisation communities and find that for a relative ranking task, most automatic metrics perform equally well and have fairly strong correlations to the human judgements. In contrast, on a naturalness judgement task, the General Text Matcher (GTM) tool correlates best overall, although in general, correlation between the human judgements and the automatic metrics was quite weak. © 2009 ACL and AFNLP.

Cite

CITATION STYLE

APA

Cahill, A. (2009). Correlating human and automatic evaluation of a German surface realiser. In ACL-IJCNLP 2009 - Joint Conf. of the 47th Annual Meeting of the Association for Computational Linguistics and 4th Int. Joint Conf. on Natural Language Processing of the AFNLP, Proceedings of the Conf. (pp. 97–100). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1667583.1667615

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free