Co-inference for multi-modal scene analysis

28Citations
Citations of this article
79Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

We address the problem of understanding scenes from multiple sources of sensor data (e.g., a camera and a laser scanner) in the case where there is no one-to-one correspondence across modalities (e.g., pixels and 3-D points). This is an important scenario that frequently arises in practice not only when two different types of sensors are used, but also when the sensors are not co-located and have different sampling rates. Previous work has addressed this problem by restricting interpretation to a single representation in one of the domains, with augmented features that attempt to encode the information from the other modalities. Instead, we propose to analyze all modalities simultaneously while propagating information across domains during the inference procedure. In addition to the immediate benefit of generating a complete interpretation in all of the modalities, we demonstrate that this co-inference approach also improves performance over the canonical approach. © 2012 Springer-Verlag.

Cite

CITATION STYLE

APA

Munoz, D., Bagnell, J. A., & Hebert, M. (2012). Co-inference for multi-modal scene analysis. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7577 LNCS, pp. 668–681). https://doi.org/10.1007/978-3-642-33783-3_48

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free