Preserving distributional information in dialogue act classification

12Citations
Citations of this article
103Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper introduces a novel training/decoding strategy for sequence labeling. Instead of greedily choosing a label at each time step, and using it for the next prediction, we retain the probability distribution over the current label, and pass this distribution to the next prediction. This approach allows us to avoid the effect of label bias and error propagation in sequence learning/decoding. Our experiments on dialogue act classification demonstrate the effectiveness of this approach. Even though our underlying neural network model is relatively simple, it outperforms more complex neural models, achieving state-of-the-art results on the MapTask and Switchboard corpora.

Cite

CITATION STYLE

APA

Tran, Q. H., Zukerman, I., & Haffari, G. (2017). Preserving distributional information in dialogue act classification. In EMNLP 2017 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 2151–2156). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d17-1229

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free