ROSMI: A Multimodal Corpus for Map-based Instruction-Giving

0Citations
Citations of this article
18Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present the publicly-available Robot Open Street Map Instructions (ROSMI) corpus: a rich multimodal dataset of map and natural language instruction pairs that was collected via crowdsourcing. The goal of this corpus is to aid in the advancement of state-of-the-art visual-dialogue tasks, including reference resolution and robot-instruction understanding. The domain described here concerns robots and autonomous systems being used for inspection and emergency response. The ROSMI corpus is unique in that it captures interaction grounded in map-based visual stimuli that is both human-readable but also contains rich metadata that is needed to plan and deploy robots and autonomous systems, thus facilitating human-robot teaming.

Cite

CITATION STYLE

APA

Katsakioris, M. M., Konstas, I., Mignotte, P. Y., & Hastie, H. (2020). ROSMI: A Multimodal Corpus for Map-based Instruction-Giving. In ICMI 2020 - Proceedings of the 2020 International Conference on Multimodal Interaction (pp. 680–684). Association for Computing Machinery, Inc. https://doi.org/10.1145/3382507.3418861

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free