Learning what’s easy: Fully differentiable neural easy-first taggers

21Citations
Citations of this article
118Readers
Mendeley users who have this article in their library.

Abstract

We introduce a novel neural easy-first decoder that learns to solve sequence tagging tasks in a flexible order. In contrast to previous easy-first decoders, our models are end-to-end differentiable. The decoder iteratively updates a “sketch” of the predictions over the sequence. At its core is an attention mechanism that controls which parts of the input are strategically the best to process next. We present a new constrained softmax transformation that ensures the same cumulative attention to every word, and show how to efficiently evaluate and backpropagate over it. Our models compare favourably to BILSTM taggers on three sequence tagging tasks.

Cite

CITATION STYLE

APA

Martins, A. F. T., & Kreutzer, J. (2017). Learning what’s easy: Fully differentiable neural easy-first taggers. In EMNLP 2017 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 349–362). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d17-1036

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free