Abstract
We present the publicly-available Robot Open Street Map Instructions (ROSMI) corpus: a rich multimodal dataset of map and natural language instruction pairs that was collected via crowdsourcing. The goal of this corpus is to aid in the advancement of state-of-the-art visual-dialogue tasks, including reference resolution and robot-instruction understanding. The domain described here concerns robots and autonomous systems being used for inspection and emergency response. The ROSMI corpus is unique in that it captures interaction grounded in map-based visual stimuli that is both human-readable but also contains rich metadata that is needed to plan and deploy robots and autonomous systems, thus facilitating human-robot teaming.
Author supplied keywords
Cite
CITATION STYLE
Katsakioris, M. M., Konstas, I., Mignotte, P. Y., & Hastie, H. (2020). ROSMI: A Multimodal Corpus for Map-based Instruction-Giving. In ICMI 2020 - Proceedings of the 2020 International Conference on Multimodal Interaction (pp. 680–684). Association for Computing Machinery, Inc. https://doi.org/10.1145/3382507.3418861
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.