We investigate under what conditions a neuron can learn by experimentally supported rules for spike timing dependent plasticity (STDP) to predict the arrival times of strong teacher inputs to the same neuron. It turns out that in contrast to the famous Perceptron Convergence Theorem, which predicts
Theoretical Analysis of Learning with Reward-Modulated Spike-Timing-Dependent Plasticity
โ Scribed by Legenstein R., Pecevski D., Maass W.
- Tongue
- English
- Leaves
- 8
- Category
- Library
No coin nor oath required. For personal study only.
โฆ Synopsis
Reward-modulated spike-timing-dependent plasticity (STDP) has recently
emerged as a candidate for a learning rule that could explain how local learning
rules at single synapses support behaviorally relevant adaptive changes in complex
networks of spiking neurons. However the potential and limitations of this
learning rule could so far only be tested through computer simulations. This article
provides tools for an analytic treatment of reward-modulated STDP, which
allow us to predict under which conditions reward-modulated STDP will be able
to achieve a desired learning effect. In particular, we can produce in this way
a theoretical explanation and a computer model for a fundamental experimental
finding on biofeedback in monkeys.
โฆ Subjects
ะะฝัะพัะผะฐัะธะบะฐ ะธ ะฒััะธัะปะธัะตะปัะฝะฐั ัะตั ะฝะธะบะฐ;ะัะบััััะฒะตะฝะฝัะน ะธะฝัะตะปะปะตะบั;ะะตะนัะพะฝะฝัะต ัะตัะธ
๐ SIMILAR VOLUMES
<p></p><p>This monograph provides a concise point of research topics and reference for modeling correlated response data with time-dependent covariates, and longitudinal data for the analysis of population-averaged models, highlighting methods by a variety of pioneering scholars. While the models pr
<p><p></p><p>This monograph provides a concise point of research topics and reference for modeling correlated response data with time-dependent covariates, and longitudinal data for the analysis of population-averaged models, highlighting methods by a variety of pioneering scholars. While the models
We show that under suitable assumptions (primarily linearization) a simple and perspicuous online learning rule for Information Bottleneck optimization with spiking neurons can be derived. This rule performs on common benchmark tasks as well as a rather complex rule that has previously been proposed
This volume highlights the links between model theory and algebra. The work contains a definitive account of algebraically compact modules, a topic of central importance for both module and model theory. Using concrete examples, particular emphasis is given to model theoretic concepts, such as axiom