Unlike traditional over-the-phone spoken dialog systems (SDSs), modern dialog systems tend to have visual rendering on the device screen as an additional modality to communicate the system's response to the user. Visual display of the system's response not only changes human behavior when interacting with devices, but also creates new research areas in SDSs. Onscreen item identification and resolution in utterances is one critical problem to achieve a natural and accurate humanmachine communication. We pose the problem as a classification task to correctly identify intended on-screen item(s) from user utterances. Using syntactic, semantic as well as context features from the display screen, our model can resolve different types of referring expressions with up to 90% accuracy. In the experiments we also show that the proposed model is robust to domain and screen layout changes.
CITATION STYLE
Celikyilmaz, A., Feizollahi, Z., Hakkani-Tur, D., & Sarikaya, R. (2014). Resolving referring expressions in conversational dialogs for natural user interfaces. In EMNLP 2014 - 2014 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference (pp. 2094–2104). Association for Computational Linguistics (ACL). https://doi.org/10.3115/v1/d14-1223
Mendeley helps you to discover research relevant for your work.