A cost sensitive part-of-speech tagging: Differentiating serious errors from minor errors

14Citations
Citations of this article
95Readers
Mendeley users who have this article in their library.

Abstract

All types of part-of-speech (POS) tagging errors have been equally treated by existing taggers. However, the errors are not equally important, since some errors affect the performance of subsequent natural language processing (NLP) tasks seriously while others do not. This paper aims to minimize these serious errors while retaining the overall performance of POS tagging. Two gradient loss functions are proposed to reflect the different types of errors. They are designed to assign a larger cost to serious errors and a smaller one to minor errors. Through a set of POS tagging experiments, it is shown that the classifier trained with the proposed loss functions reduces serious errors compared to state-of-the-art POS taggers. In addition, the experimental result on text chunking shows that fewer serious errors help to improve the performance of subsequent NLP tasks. © 2012 Association for Computational Linguistics.

Cite

CITATION STYLE

APA

Song, H. J., Son, J. W., Noh, T. G., Park, S. B., & Lee, S. J. (2012). A cost sensitive part-of-speech tagging: Differentiating serious errors from minor errors. In 50th Annual Meeting of the Association for Computational Linguistics, ACL 2012 - Proceedings of the Conference (Vol. 1, pp. 1025–1034). https://doi.org/10.5391/ijfis.2012.12.1.6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free