Using gradient descent, we propose a new backpropagation learning algorithm for spiking neural networks with multi-layers, multisynapses between neurons, and multi-spiking neurons. It adjusts synaptic weights, delays, and time constants, and neurons’ thresholds in output and hidden layers. It guarantees convergence to minimum error point, and unlike SpikeProp and its extensions, does not need a one-to-one correspondence between actual and desired spikes in advance. So, it is stably and widely applicable to practical problems.
CITATION STYLE
Matsuda, S. (2016). BPSpike II: A new backpropagation learning algorithm for spiking neural networks. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9948 LNCS, pp. 56–65). Springer Verlag. https://doi.org/10.1007/978-3-319-46672-9_7
Mendeley helps you to discover research relevant for your work.