The interpretation of spatial references is highly contextual, requiring joint inference over both language and the environment. We consider the task of spatial reasoning in a simulated environment, where an agent can act and receive rewards. The proposed model learns a representation of the world steered by instruction text. This design allows for precise alignment of local neighborhoods with corresponding verbalizations, while also handling global references in the instructions. We train our model with reinforcement learning using a variant of generalized value iteration. The model outperforms state-of-the-art approaches on several metrics, yielding a 45% reduction in goal localization error.
CITATION STYLE
Janner, M., Narasimhan, K., & Barzilay, R. (2018). Representation Learning for Grounded Spatial Reasoning. Transactions of the Association for Computational Linguistics, 6, 49–61. https://doi.org/10.1162/tacl_a_00004
Mendeley helps you to discover research relevant for your work.