Hierarchical temporal prediction captures motion processing along the visual pathway

8Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Visual neurons respond selectively to features that become increasingly complex from the eyes to the cortex. Retinal neurons prefer flashing spots of light, primary visual cortical (V1) neurons prefer moving bars, and those in higher cortical areas favor complex features like moving textures. Previously, we showed that V1 simple cell tuning can be accounted for by a basic model implementing temporal prediction - representing features that predict future sensory input from past input (Singer et al., 2018). Here, we show that hierarchical application of temporal prediction can capture how tuning properties change across at least two levels of the visual system. This suggests that the brain does not efficiently represent all incoming information; instead, it selectively represents sensory inputs that help in predicting the future. When applied hierarchically, temporal prediction extracts time-varying features that depend on increasingly high-level statistics of the sensory input.

Cite

CITATION STYLE

APA

Singer, Y., Taylor, L., Willmore, B. D. B., King, A. J., & Harper, N. S. (2023). Hierarchical temporal prediction captures motion processing along the visual pathway. ELife, 12. https://doi.org/10.7554/eLife.52599

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free