Zero-shot sequence labeling: Transferring knowledge from sentences to tokens

43Citations
Citations of this article
192Readers
Mendeley users who have this article in their library.

Abstract

Can attention-or gradient-based visualization techniques be used to infer token-level labels for binary sequence tagging problems, using networks trained only on sentence-level labels? We construct a neural network architecture based on soft attention, train it as a binary sentence classifier and evaluate against tokenlevel annotation on four different datasets. Inferring token labels from a network provides a method for quantitatively evaluating what the model is learning, along with generating useful feedback in assistance systems. Our results indicate that attention-based methods are able to predict token-level labels more accurately, compared to gradient-based methods, sometimes even rivaling the supervised oracle network.

Cite

CITATION STYLE

APA

Rei, M., & Søgaard, A. (2018). Zero-shot sequence labeling: Transferring knowledge from sentences to tokens. In NAACL HLT 2018 - 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Conference (Vol. 1, pp. 293–302). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/n18-1027

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free