Abstract
We present a model for contrastively describing scenes, in which context-specific behavior results from a combination of inference-driven pragmatics and learned semantics. Like previous learned approaches to language generation, our model uses a simple feature-driven architecture (here a pair of neural “listener” and “speaker” models) to ground language in the world. Like inference-driven approaches to pragmatics, our model actively reasons about listener behavior when selecting utterances. For training, our approach requires only ordinary captions, annotated without demonstration of the pragmatic behavior the model ultimately exhibits. In human evaluations on a referring expression game, our approach succeeds 81% of the time, compared to 69% using existing techniques.
Cite
CITATION STYLE
Andreas, J., & Klein, D. (2016). Reasoning about pragmatics with neural listeners and speakers. In EMNLP 2016 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 1173–1182). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d16-1125
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.