Vision and the representation of the surroundings in spatial memory

  • Tatler B
  • Land M
  • 142

    Readers

    Mendeley users who have this article in their library.
  • 46

    Citations

    Citations of this article.

Abstract

One of the paradoxes of vision is that the world as it appears to us and the image on the retina at any moment are not much like each other. The visual world seems to be extensive and continuous across time. However, the manner in which we sample the visual environment is neither extensive nor continuous. How does the brain reconcile these differences? Here, we consider existing evidence from both static and dynamic viewing paradigms together with the logical requirements of any representational scheme that would be able to support active behaviour. While static scene viewing paradigms favour extensive, but perhaps abstracted, memory representations, dynamic settings suggest sparser and task-selective representation. We suggest that in dynamic settings where movement within extended environments is required to complete a task, the combination of visual input, egocentric and allocentric representations work together to allow efficient behaviour. The egocentric model serves as a coding scheme in which actions can be planned, but also offers a potential means of providing the perceptual stability that we experience.

Author-supplied keywords

  • Allocentric
  • Egocentric
  • Natural tasks
  • Object memory
  • Position memory
  • Saccade

Get free article suggestions today

Mendeley saves you time finding and organizing research

Sign up here
Already have an account ?Sign in

Find this document

Authors

  • Benjamin W. Tatler

  • Michael F. Land

Cite this document

Choose a citation style from the tabs below

Save time finding and organizing research with Mendeley

Sign up for free