Calibrate your listeners! Robust communication-based training for pragmatic speakers

2Citations
Citations of this article
51Readers
Mendeley users who have this article in their library.
Get full text

Abstract

To be good conversational partners, natural language processing (NLP) systems should be trained to produce contextually useful utterances. Prior work has investigated training NLP systems with communication-based objectives, where a neural listener stands in as a communication partner. However, these systems commonly suffer from semantic drift where the learned language diverges radically from natural language. We propose a method that uses a population of neural listeners to regularize speaker training. We first show that language drift originates from the poor uncertainty calibration of a neural listener, which makes high-certainty predictions on novel sentences. We explore ensemble- and dropoutbased populations of listeners and find that the former results in better uncertainty quantification. We evaluate both population-based objectives on reference games, and show that the ensemble method with better calibration enables the speaker to generate pragmatic utterances while scaling to a large vocabulary and generalizing to new games and listeners.

Cite

CITATION STYLE

APA

Wang, R. E., White, J., Mu, J., & Goodman, N. D. (2021). Calibrate your listeners! Robust communication-based training for pragmatic speakers. In Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 (pp. 977–984). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-emnlp.83

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free