Abstract
Genre and domain are well known covariates of both manual and automatic annotation quality. Comparatively less is known about the effect of sentence types, such as imperatives, questions or fragments, and how they interact with text type effects. Using mixed effects models, we evaluate the relative influence of genre and sentence types on automatic and manual annotation quality for three related tasks in English data: POS tagging, dependency parsing and coreference resolution. For the latter task, we also develop a new metric for the evaluation of individual regions of coreference annotation. Our results show that while there are substantial differences between manual and automatic annotation in each task, sentence type is generally more important than genre in predicting errors within our data.
Cite
CITATION STYLE
Zeldes, A., & Simonson, D. (2016). Different flavors of gum: Evaluating genre and sentence type effects on multilayer corpus annotation quality. In LAW 2016 - 10th Linguistic Annotation Workshop, held in conjuncion with ACL 2016 - Workshop Proceedings (pp. 68–78). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w16-1709
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.