A Reinforcement Learning Framework for Spiking Networks with Dynamic Synapses

An integration of both the Hebbian-based and reinforcement learning (RL) rules is presented for dynamic synapses. The proposed framework permits the Hebbian rule to update the hidden synaptic model parameters regulating the synaptic response rather than the synaptic weights. This is performed using...

Full description

Bibliographic Details
Main Authors: Karim El-Laithy, Martin Bogdan
Format: Article
Language:English
Published: Hindawi Limited 2011-01-01
Series:Computational Intelligence and Neuroscience
Online Access:http://dx.doi.org/10.1155/2011/869348
id doaj-0d49f016dd9f46199d1755bb955b56c6
record_format Article
spelling doaj-0d49f016dd9f46199d1755bb955b56c62020-11-24T20:46:22ZengHindawi LimitedComputational Intelligence and Neuroscience1687-52651687-52732011-01-01201110.1155/2011/869348869348A Reinforcement Learning Framework for Spiking Networks with Dynamic SynapsesKarim El-Laithy0Martin Bogdan1Department of Computer Engineering, Faculty of Mathematics and Computer Science, Johannisgasse 26, 04103 Leipzig, GermanyDepartment of Computer Engineering, Faculty of Mathematics and Computer Science, Johannisgasse 26, 04103 Leipzig, GermanyAn integration of both the Hebbian-based and reinforcement learning (RL) rules is presented for dynamic synapses. The proposed framework permits the Hebbian rule to update the hidden synaptic model parameters regulating the synaptic response rather than the synaptic weights. This is performed using both the value and the sign of the temporal difference in the reward signal after each trial. Applying this framework, a spiking network with spike-timing-dependent synapses is tested to learn the exclusive-OR computation on a temporally coded basis. Reward values are calculated with the distance between the output spike train of the network and a reference target one. Results show that the network is able to capture the required dynamics and that the proposed framework can reveal indeed an integrated version of Hebbian and RL. The proposed framework is tractable and less computationally expensive. The framework is applicable to a wide class of synaptic models and is not restricted to the used neural representation. This generality, along with the reported results, supports adopting the introduced approach to benefit from the biologically plausible synaptic models in a wide range of intuitive signal processing.http://dx.doi.org/10.1155/2011/869348
collection DOAJ
language English
format Article
sources DOAJ
author Karim El-Laithy
Martin Bogdan
spellingShingle Karim El-Laithy
Martin Bogdan
A Reinforcement Learning Framework for Spiking Networks with Dynamic Synapses
Computational Intelligence and Neuroscience
author_facet Karim El-Laithy
Martin Bogdan
author_sort Karim El-Laithy
title A Reinforcement Learning Framework for Spiking Networks with Dynamic Synapses
title_short A Reinforcement Learning Framework for Spiking Networks with Dynamic Synapses
title_full A Reinforcement Learning Framework for Spiking Networks with Dynamic Synapses
title_fullStr A Reinforcement Learning Framework for Spiking Networks with Dynamic Synapses
title_full_unstemmed A Reinforcement Learning Framework for Spiking Networks with Dynamic Synapses
title_sort reinforcement learning framework for spiking networks with dynamic synapses
publisher Hindawi Limited
series Computational Intelligence and Neuroscience
issn 1687-5265
1687-5273
publishDate 2011-01-01
description An integration of both the Hebbian-based and reinforcement learning (RL) rules is presented for dynamic synapses. The proposed framework permits the Hebbian rule to update the hidden synaptic model parameters regulating the synaptic response rather than the synaptic weights. This is performed using both the value and the sign of the temporal difference in the reward signal after each trial. Applying this framework, a spiking network with spike-timing-dependent synapses is tested to learn the exclusive-OR computation on a temporally coded basis. Reward values are calculated with the distance between the output spike train of the network and a reference target one. Results show that the network is able to capture the required dynamics and that the proposed framework can reveal indeed an integrated version of Hebbian and RL. The proposed framework is tractable and less computationally expensive. The framework is applicable to a wide class of synaptic models and is not restricted to the used neural representation. This generality, along with the reported results, supports adopting the introduced approach to benefit from the biologically plausible synaptic models in a wide range of intuitive signal processing.
url http://dx.doi.org/10.1155/2011/869348
work_keys_str_mv AT karimellaithy areinforcementlearningframeworkforspikingnetworkswithdynamicsynapses
AT martinbogdan areinforcementlearningframeworkforspikingnetworkswithdynamicsynapses
AT karimellaithy reinforcementlearningframeworkforspikingnetworkswithdynamicsynapses
AT martinbogdan reinforcementlearningframeworkforspikingnetworkswithdynamicsynapses
_version_ 1716812789739159552