A general and multi-lingual phrase chunking model based on masking method

21Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Several phrase chunkers have been proposed over the past few years. Some state-of-the-art chunkers achieved better performance via integrating external resources, e.g., parsers and additional training data, or combining multiple learners. However, in many languages and domains, such external materials are not easily available and the combination of multiple learners will increase the cost of training and testing. In this paper, we propose a mask method to improve the chunking accuracy. The experimental results show that our chunker achieves better performance in comparison with other deep parsers and chunkers. For CoNLL-2000 data set, our system achieves 94.12 in F rate. For the base-chunking task, our system reaches 92.95 in F rate. When porting to Chinese, the performance of the base-chunking task is 92.36 in F rate. Also, our chunker is quite efficient. The complete chunking time of a 50K words document is about 50 seconds. © Springer-Verlag Berlin Heidelberg 2006.

Cite

CITATION STYLE

APA

Wu, Y. C., Chang, C. H., & Lee, Y. S. (2006). A general and multi-lingual phrase chunking model based on masking method. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 3878 LNCS, pp. 144–155). Springer Verlag. https://doi.org/10.1007/11671299_17

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free