Different flavors of gum: Evaluating genre and sentence type effects on multilayer corpus annotation quality

11Citations
Citations of this article
66Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Genre and domain are well known covariates of both manual and automatic annotation quality. Comparatively less is known about the effect of sentence types, such as imperatives, questions or fragments, and how they interact with text type effects. Using mixed effects models, we evaluate the relative influence of genre and sentence types on automatic and manual annotation quality for three related tasks in English data: POS tagging, dependency parsing and coreference resolution. For the latter task, we also develop a new metric for the evaluation of individual regions of coreference annotation. Our results show that while there are substantial differences between manual and automatic annotation in each task, sentence type is generally more important than genre in predicting errors within our data.

Cite

CITATION STYLE

APA

Zeldes, A., & Simonson, D. (2016). Different flavors of gum: Evaluating genre and sentence type effects on multilayer corpus annotation quality. In LAW 2016 - 10th Linguistic Annotation Workshop, held in conjuncion with ACL 2016 - Workshop Proceedings (pp. 68–78). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w16-1709

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free