Semantic modeling of places using objects

5Citations
Citations of this article
70Readers
Mendeley users who have this article in their library.

Abstract

While robot mapping has seen massive strides recently, higher level abstractions in map representation are still not widespread. Maps containing semantic concepts such as objects and labels are essential for many tasks in manmade environments as well as for human-robot interaction and map communication. In keeping with this aim, we present a model for places using objects as the basic unit of representation. Our model is a 3D extension of the constellation object model, popular in computer vision, in which the objects are modeled by their appearance and shape. The 3D location of each object is maintained in a coordinate frame local to the place. The individual object models are learned in a supervised manner using roughly segmented and labeled training images. Stereo range data is used to compute 3D locations of the objects. We use the Swendsen-Wang algorithm, a cluster MCMC method, to solve the correspondence problem between image features and objects during inference. We provide a technique for building panoramic place models from multiple views of a location. An algorithm for place recognition by comparing models is also provided. Results are presented in the form of place models inferred in an indoor environment.We envision the use of our place model as a building block towards a complete object-based semantic mapping system.

Cite

CITATION STYLE

APA

Ranganathan, A., & Dellaert, F. (2008). Semantic modeling of places using objects. In Robotics: Science and Systems (Vol. 3, pp. 1–8). MIT Press Journals. https://doi.org/10.15607/rss.2007.iii.001

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free