Hand-engineered definitions of spatial categories are increasingly seen as brittle and spatial concepts in human interactions may need to learn these in terms of perceptually grounded "image schemas". Here, we present a developmental approach for the acquisition of grounded spatial schemas in a perceptual agent. We assume a capability for dynamic visual attention, and perceptual notions of wholeness and proximity. We first learn perceptual-object to linguistic-name mappings from simple 2D multi-agent visual streams co-occurring with word-separated utterances. Mutual information based statistical measures are seen to be sufficient to identify nominal participants in a simple discourse, based on a synthetic model of dynamic visual attention. Next, we use this knowledge of nominais to ground the semantics of spatial relations in language. We show that a notion of proximity between perceptual objects is sufficient to obtain a pre-verbal notion of graded spatial poses. Once linguistic data is superimposed on this, simple associative structures lead to distinctions such as "in" or "out". Finally we also show how this can lead to a model of actions, where verbs are learned along with the associated argument structures. © Springer-Verlag Berlin Heidelberg 2007.
CITATION STYLE
Mukerjee, A., & Sarkar, M. (2007). Grounded perceptual schemas: Developmental acquisition of spatial concepts. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 4387 LNAI, pp. 210–228). Springer Verlag. https://doi.org/10.1007/978-3-540-75666-8_13
Mendeley helps you to discover research relevant for your work.