Local dendritic balance enables learning of efficient representations in networks of spiking neurons

16Citations
Citations of this article
56Readers
Mendeley users who have this article in their library.
Get full text

Abstract

How can neural networks learn to efficiently represent complex and high-dimensional inputs via local plasticity mechanisms? Classical models of representation learning assume that feedforward weights are learned via pairwise Hebbian-like plasticity. Here, we show that pairwise Hebbian-like plasticity works only under unrealistic requirements on neural dynamics and input statistics. To overcome these limitations, we derive from first principles a learning scheme based on voltage-dependent synaptic plasticity rules. Here, recurrent connections learn to locally balance feedforward input in individual dendritic compartments and thereby can modulate synaptic plasticity to learn efficient representations. We demonstrate in simulations that this learning scheme works robustly even for complex high-dimensional inputs and with inhibitory transmission delays, where Hebbian-like plasticity fails. Our results draw a direct connection between dendritic excitatory–inhibitory balance and voltage-dependent synaptic plasticity as observed in vivo and suggest that both are crucial for representation learning.

Cite

CITATION STYLE

APA

Mikulasch, F. A., Rudelt, L., & Priesemann, V. (2021). Local dendritic balance enables learning of efficient representations in networks of spiking neurons. Proceedings of the National Academy of Sciences of the United States of America, 118(50). https://doi.org/10.1073/pnas.2021925118

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free