Photogrammetry and VR for Comparing 2D and Immersive Linguistic Data Collection

0Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.

Abstract

The overarching goal of this work is to enable the collection of language describing a wide variety of objects viewed in virtual reality. We aim to create full 3D models from a small number of 'keyframe' images of objects found in the publicly available Grounded Language Dataset (GoLD) using photogrammetry. We will then collect linguistic descriptions by placing our models in virtual reality and having volunteers describe them. To evaluate the impact of virtual reality immersion on linguistic descriptions of the objects, we intend to apply contrastive learning to perform grounded language learning, then compare the descriptions collected from images (in GoLD) versus our models.

Cite

CITATION STYLE

APA

Rubinstein, J., Matuszek, C., & Engel, D. (2023). Photogrammetry and VR for Comparing 2D and Immersive Linguistic Data Collection. In Proceedings of the 37th AAAI Conference on Artificial Intelligence, AAAI 2023 (Vol. 37, pp. 16312–16313). AAAI Press. https://doi.org/10.1609/aaai.v37i13.27016

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free