In this work, we address the out-of-vocabulary (OOV) problem in sequence labeling using only training data of the task. A typical solution in this field is to represent an OOV word using the mean-pooled representations of its surrounding words at test time. However, such a pipeline approach often suffers from the error propagation problem, since training of the supervised model is independent of the mean-pooling operation. In this work, we propose a novel training strategy to address the error propagation problem suffered by this solution. It designs to mimic the OOV situation in the process of model training and trains the supervised model to fit the OOV word representations generated by the mean-pooling operation. Extensive experiments on different sequence labeling tasks, including part-of-speech tagging (POS), named entity recognition (NER), and chunking verified the effectiveness of our proposed method.
CITATION STYLE
Xing, X., Peng, M., Zhang, Q., Liu, Q., & Huang, X. (2020). Learning to Generate Representations for Novel Words: Mimic the OOV Situation in Training. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12430 LNAI, pp. 321–332). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-60450-9_26
Mendeley helps you to discover research relevant for your work.