Sequence Learning with Hidden Units in Spiking Neural Networks
Abstract
We consider a statistical framework in which recurrent networks of spiking neurons learn to generate spatio-temporal spike patterns. Given biologically realistic stochastic neuronal dynamics we derive a tractable learning rule for the synaptic weights towards hidden and visible neurons that leads to optimal recall of the training sequences. We show that learning synaptic weights towards hidden neurons significantly improves the storing capacity of the network. Furthermore, we derive an approximate online learning rule and show that our learning rule is consistent with Spike-Timing Dependent Plasticity in that if a presynaptic spike shortly precedes a postynaptic spike, potentiation is induced and otherwise depression is elicited.
Cite
Text
Brea et al. "Sequence Learning with Hidden Units in Spiking Neural Networks." Neural Information Processing Systems, 2011.Markdown
[Brea et al. "Sequence Learning with Hidden Units in Spiking Neural Networks." Neural Information Processing Systems, 2011.](https://mlanthology.org/neurips/2011/brea2011neurips-sequence/)BibTeX
@inproceedings{brea2011neurips-sequence,
title = {{Sequence Learning with Hidden Units in Spiking Neural Networks}},
author = {Brea, Johanni and Senn, Walter and Pfister, Jean-pascal},
booktitle = {Neural Information Processing Systems},
year = {2011},
pages = {1422-1430},
url = {https://mlanthology.org/neurips/2011/brea2011neurips-sequence/}
}