Abstract
To improve the robustness in multimodal input interpretation, this paper presents a new salience driven approach. This approach is based on the observation that, during multimodal conversation, information from deictic gestures (e.g., point or circle) on a graphical display can signal a part of the physical world (i.e., representation of the domain and task) of the application which is salient during the communication. This salient part of the physical world will prime what users tend to communicate in speech and in turn can be used to constrain hypotheses for spoken language understanding, thus improving overall input interpretation. Our experimental results have indicated the potential of this approach in reducing word error rate and improving concept identification in multimodal conversation. © 2005 Association for Computational Linguistics.
Cite
CITATION STYLE
Chai, J. Y., & Qu, S. (2005). A salience driven approach to robust input interpretation in multimodal conversational systems. In HLT/EMNLP 2005 - Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference (pp. 217–224). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1220575.1220603
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.