- Open Access
Gradient estimation in dendritic reinforcement learning
The Journal of Mathematical Neuroscience volume 2, Article number: 2 (2012)
We study synaptic plasticity in a complex neuronal cell model where NMDA-spikes can arise in certain dendritic zones. In the context of reinforcement learning, two kinds of plasticity rules are derived, zone reinforcement (ZR) and cell reinforcement (CR), which both optimize the expected reward by stochastic gradient ascent. For ZR, the synaptic plasticity response to the external reward signal is modulated exclusively by quantities which are local to the NMDA-spike initiation zone in which the synapse is situated. CR, in addition, uses nonlocal feedback from the soma of the cell, provided by mechanisms such as the backpropagating action potential. Simulation results show that, compared to ZR, the use of nonlocal feedback in CR can drastically enhance learning performance. We suggest that the availability of nonlocal feedback for learning is a key advantage of complex neurons over networks of simple point neurons, which have previously been found to be largely equivalent with regard to computational capability.
Except for biologically detailed modeling studies, the overwhelming majority of works in mathematical neuroscience have treated neurons as point neurons, i.e., a linear aggregation of synaptic input followed by a nonlinearity in the generation of somatic action potentials was assumed to characterize a neuron. This disregards the fact that many neurons in the brain have complex dendritic arborization where synaptic inputs may be aggregated in highly nonlinear ways . From an information processing perspective sticking with the minimal point neuron may nevertheless seem justified since networks of such simple neurons already display remarkable computational properties: assuming infinite precision and noiseless arithmetic a suitable network of spiking point neurons can simulate a universal Turing machine and, further, impressive information processing capabilities persist when one makes more realistic assumptions such as taking noise into account (see  and the references therein). Such generic observations are underscored by the detailed compartmental modeling of the computation performed in a hippocampal pyramidal cell . There it was found that (in a rate coding framework) the input-output behavior of the complex cell is easily emulated by a simple two layer network of point neurons.
If the computations of complex cells are readily emulated by relatively simple circuits of point neurons, the question arises why so many of the neurons in the brain are complex. Of course, the reason for this may be only loosely related to information processing proper, it might be that maintaining a complex cell is metabolically less costly than the maintenance of the equivalent network of point neurons. Here, we wish to explore a different hypothesis, namely that complex cells have crucial advantages with regard to learning. This hypothesis is motivated by the fact that many artificial intelligence algorithms for neural networks assume that synaptic plasticity is modulated by information which arises far downstream of the synapse. A prominent example is the backpropagation algorithm where error information needs to be transported upstream via the transpose of the connectivity matrix. But in real axons any fast information flow is strictly downstream, and this is why algorithms such as backpropagation are widely regarded as a biologically unrealistic for networks of point neurons. When one considers complex cells, however, it seems far more plausible that synaptic plasticity could be modulated by events which arise relatively far downstream of the synapse. The backpropagating action potential, for instance, is often capable of conveying information on somatic spiking to synapses which are quite distal in the dendritic tree [4, 5]. If nonlinear processing occurred in the dendritic tree during the forward propagation, this means that somatic spiking can modulate synaptic plasticity even when one or more layers of nonlinearities lie between the synapse and the soma. Thus, compared to networks of point neurons, more sophisticated plasticity rules could be biologically feasible in complex cells.
To study this issue, we formalize a complex cell as a two layer network, with the first layer made up of initiation zones for NMDA-spikes (Figure 1). NMDA-spikes are regenerative events, caused by AMPA mediated synaptic releases when the releases are both near coincident in time and spatially co-located on the dendrite [6–8]. Such NMDA-spikes boost the effect of the synaptic releases, leading to increases in the somatic potential which are stronger as well as longer compared to the effect obtained from a simple linear superposition of the excitatory post synaptic potentials from the individual AMPA releases. Further, we assume that the contribution of NMDA-spikes from different initiation zones combine additively in contributing to the somatic potential and that this potential governs the generation of somatic action potentials via an escape noise process. While we would argue that this provides an adequate minimal model of dendritic computation in basal dendritic structures, one should bear in mind that our model seems insufficient to describe the complex interactions of basal and apical dendritic inputs in cortical pyramidal cells [9, 10].
We will consider synaptic plasticity in the context of reinforcement learning, where the somatic action potentials control the delivery of an external reward signal. The goal of learning is to adjust the strength of the synaptic releases (the synaptic weights) so as to maximize the expected value of the reward signal. In this framework, one can mathematically derive plasticity rules [11, 12] by assuming that weight adaption follows a stochastic gradient ascent procedure in the expected reward . Dopamine is widely believed to be the most important neurotransmitter for such reward modulated plasticity [14–16]. A simple minded application of the approach in  leads to a learning rule where, except for the external reward signal, plasticity is determined by quantities which are local to each NMDA-spike initiation zone (NMDA-zone). Using this rule, NMDA-zones learn as independent agents which are oblivious of their interaction in generating somatic action potentials, with the external reward signal being the only mechanism for coordinating plasticity between the zones. hence we shall refer to this rule as zone reinforcement (ZR). Due to its simplicity, ZR would seem biologically feasible even if the network were not integrated into a single neuron. On the other hand, this approach to multi-agent reinforcement often leads to a learning performance which deteriorates quickly as the number of agents (here, NMDA-zones) increases since it lacks an explicit mechanism for differentially assigning credit to the agents [17, 18]. By algebraic manipulation of the gradient formula leading to the basic ZR-rule, we derive a class of learning rules where synaptic plasticity is also modulated by somatic responses, in addition to reward and quantities local to the NMDA-zone. Such learning rules will be referred to as cell reinforcement (CR), since they would be biologically unrealistic if the nonlinearities where not integrated into a single cell. We present simulation result showing that one rule in the CR-class results in learning which is much faster than for the ZR-rule. This provides evidence for the hypothesis that enabling effective synaptic plasticity rules may be one evolutionary advantage conveyed by dendritic nonlinearities.
2 Stochastic cell model of a neuron
We assume a neuron with initiation zones for NMDA-spikes, indexed by . An NMDA-zone is made up of synapses, with synaptic strength (), where releases are triggered by presynaptic spikes. We denote by the set of times when presynaptic spikes arrive at synapse . In each NMDA-zone, the synaptic releases give rise to a time varying local membrane potential which we assume to be given by a standard spike response equation
Here, X denotes the entire presynaptic input pattern of the neuron, (arbitrary units) is the resting potential, and the postsynaptic response kernel ϵ is given by
We use for the membrane time constant, for the synaptic rise time, and Θ is the Heaviside step function.
The local potential controls the rate at which what we call NMDA-events are generated in the zone - in our model NMDA-events are closely related to the onset of NMDA-spikes as described in detail below. Formally, we assume that NMDA-events are generated by an inhomogeneous Poisson process with rate function , choosing
with and . We adopt the symbol to denote the set of NMDA-event times in zone ν. For future use, we recall the standard result  that the probability density of an event-train generated during an observation period running from to T satisfies
where is the δ-function representation of .
Conceptually, it would be simplest to assume that each NMDA-event initiates a NMDA-spike. But we need some mechanism for refractoriness, since NMDA-spikes have an extended duration (20-200 ms) and there is no evidence that multiple simultaneous NMDA-spikes can arise in a single NMDA-zone. Hence, we shall assume that, while a NMDA-event occurring in temporal isolation causes a NMDA-spike, a rapid succession of NMDA-events within one zone only leads to a somewhat longer but not to a stronger NMDA-spike. In particular, we will assume that a NMDA-spike contributes to the somatic potential during a period of after the time of the last preceding NMDA-event. Hence, if a NMDA-event is followed by a second one with a delay, the first event initiates a NMDA-spike which lasts for due to the second NMDA-event. Formally, we denote by the time of the last NMDA-event up to time t and model the somatic effect of an NMDA-spike by the response kernel
The main motivation for modeling the generation of NMDA-spikes in this way is that it proves mathematically convenient in the calculations below. Having said this, it is worthwhile mentioning that treating NMDA-spikes as rectangular pulses seems reasonable, since their rise and fall times are typically short compared to the duration of the spike. Also, there is some evidence that increased excitatory presynaptic activity extends the duration of a NMDA-spike but does not increase its amplitude [7, 8]. Qualitatively, the above model is in line with such findings.
For specifying the somatic potential U of the neuron, we denote by Y the vector of all NMDA-event trains and by Z the set of times when the soma generates action potentials. We then use
for the time course of the somatic potential, where the reset kernel κ is given by
This is a highly stylized model of the somatic potential since we assume that NMDA-zones contribute equally to the somatic potential (with a strength controlled by the positive parameter a) and that, further, the AMPA-releases themselves do not contribute directly to U. Even if these restrictive assumptions may not be entirely unreasonable (for instance, AMPA-releases can be much more strongly attenuated on their way to the soma than NMDA-spikes) we wish to point out that, while becoming simpler, the mathematical approach below does not rely on these restrictions.
Somatic firing is modeled as an escape noise process with an instantaneous rate function where
with and . As shown in , for the probability density of responding to the NMDA-events with a somatic spike train Z during the observation period this implies
3 Reinforcement learning
In reinforcement learning, one assumes a scalar reward function providing feedback about the appropriateness of the somatic response Z to the input X. The goal of learning is to adapt the synaptic strengths so as to obtain appropriate somatic responses. For our neuronal model, the expected value of the reward signal is
where is the probability density of the input spike patterns and . The goal of learning can now be formalized as finding a w maximizing and synaptic plasticity rules can be obtained using stochastic gradient ascent procedures for this task.
In stochastic gradient ascent, , and Z are sampled at each trial and every weight is updated by
where is the learning rate and is an (unbiased) estimator of . Under mild regularity conditions, convergence to a local optimum is guaranteed if one uses an appropriate schedule for decreasing η towards 0 during learning . In biological modeling, one usually simply assumes a small but fixed learning rate.
The derivative of with respect to the weight of synapse can be written as
Hence, a simple choice for the gradient estimator is
with given by Equation 3. Note that the conditional probability does not explicitly appear in the estimator, so the update is oblivious of the architecture of the model neuron, i.e., of how NMDA-events contribute to somatic spiking. Since the only learning mechanism for coordinating the responses of the different NMDA-zones is the global reward signal , we refer to the update given by Equation 10 as ZR.
Better plasticity rules can be obtained by algebraic manipulations of Equations 8 and 9 which yield gradient estimators which have a reduced variance compared to Equation 10 - this should lead to faster learning. A simple and well-known example for this is adjusting the reinforcement baseline by choosing a constant c and replacing with in Equation 10; this amounts to adding c to and hence does not change the gradient. But a judicious choice of c can reduce the variance of the gradient estimator. More ambitiously, one could consider analytically integrating out Y in Equation 8, yielding an estimator which directly considers the relationship between synaptic weights and somatic spiking because it is based on . While actually doing the integration analytically seems impractical, we shall obtain estimators below from a partial realization of this program.
4 From zone reinforcement to cell reinforcement
Due to the algebraic symmetries of our model cell, it suffices to give explicit plasticity rules only for one synaptic weight. To reduce clutter we will thus focus on the first synapse in the first NMDA-zone.
4.1 Notational simplifications
Let denote the vector of all NMDA-event trains but the first and the collection of synaptic weights in all but the first NMDA-zone. We rewrite the expected reward as
Since in Equation 11 only r depends on we just need to consider . Hence, we can regard X and as fixed and suppress them in the notation. This allows us to write the somatic potential (Equation 5) simply as
using Y as shorthand for the NMDA-event train of the first zone and, further, incorporating into a time varying base potential the following contributions in Equation 5: (i) the resting potential, (ii) the influence of , i.e., NMDA-events in the other zones, (iii) any reset caused by somatic spiking. Similarly, the notation for the local membrane potential of the first NMDA-zone becomes
where w stands for the strength of the first synapse, , and the effect of the other synapses impinging on the zone is absorbed into . Finally, the w-dependent contribution r to the expected reward (Equation 11) can be written as
4.2 Cell reinforcement
To simplify the manipulation of Equation 14, we replace the Poisson process generating Y by a discrete time process with step-size . We assume that NMDA-events in Y can only occur at times where k runs from 1 to and introduce K independent binary random variables to record whether or not a NMDA-event occurred. For the probability of not having a NMDA-event at time we use
With this definition, we can recover the original Poisson process by taking the limit . We use to denote the entire response of the NMDA-zone and, to make contact with the set-based description of the NMDA-trains, we denote by the set of NMDA-event times in y, i.e., . Next, the discrete time version of Equation 14 is
where . In the end, we will recover r from by taking δ to zero.
The derivative of Equation 17 is
and to focus on the contributions to from each time bin we set
We now exploit the trivial fact that we can think of as a function linear in , simply because is binary. As a consequence, we can decompose into two terms: one which depends on and one which does not. For this, we pick a scalar μ and rewrite as
Rearranging terms in , we get
Now, , hence
The two equations above encapsulate our main idea for improving on ZR. In showing that we summed over the two outcomes , thus identifying a noise contribution in the ZR estimator for which vanishes through the averaging by the sampling procedure. Note that the remaining contribution has as factor , a term which explicitly reflects how a NMDA-event at time contributes to the generation of somatic action potentials. In going from Equation 20 to Equation 21, we assumed that the parameter μ was constant. However, a quick perusal of the above derivation shows that this is not really necessary. For justifying Equation 21, one just needs that μ does not depend on , so that is indeed independent of . In the sequel, it shall turn to be useful to introduce a value of μ which depends on somatic quantities.
A drawback of Equations 20 and 21 is that they do not immediately lend themselves to Monte-Carlo estimation by sampling the process generating neuronal events. The reason being the missing term in the formula for . To reintroduce the term, we set
Hence, is an unbiased estimator of and, since gives the contribution to from the k th time step,
is an unbiased estimator of . Note that, while unavoidable, the above recasting of the gradient calculation as an estimation procedure does seem risky. Due to the division by in introducing , Equation 22, rare somatic spike trains Z can potentially lead to large values of the estimator .
To obtain a CR estimator for the expected reward in our original problem, we now just need to take δ to 0 in Equation 23 and tidy up a little. The detailed calculations are presented in Appendix 1, here we just display the final result:
In contrast to the ZR-estimator, depends on somatic quantities via which assesses the effect of having a NMDA-event at time t on the probability of the observed somatic spike train. This requires the integration over the duration Δ of a NMDA-spike.
The CR-rule can be written as the sum of two terms, a time-discrete one depending on the NMDA-events , and a time-continuous one depending on the instantaneous NMDA-rate, both weighted by the effect of an NMDA-event on the probability of producing the somatic spike train:
5 Performance of zone and cell reinforcements
To compare the two plasticity rules, we first consider a rudimentary learning scenario where producing a somatic spike during a trial of duration is deemed an incorrect response, resulting in reward . The correct response is not to spike () and this results in a reward of 0. With these reward signals, synaptic updates become less frequent as performance improves. This compensates somewhat for having a constant learning rate instead of the decreasing schedule which would ensure proper convergence of the stochastic gradient procedure. We use for the NMDA-spike strength in Equation 5, so that just 2-3 concurrent NMDA-spikes are likely to generate a somatic action potential. The input pattern X is held fixed and initial weight values are chosen so that correct and incorrect responses are equally likely before learning. Simulation details are given in Appendix 2. Given our choice of a and the initial weights, dendritic activity is already fairly low before learning and decreasing it to a very low level is all that is required for good performance in this simple task (Figure 2).
Simulations for ZR and CR (with a constant value of ) are shown in panel 6A. Given the sophistication of the rule, the performance of CR is disappointing, yielding on average only a modest improvement over ZR. The histogram in panel 6B shows that in most cases CR does in fact learn substantially faster than ZR but, in contrast to ZR, CR spectacularly fails on some runs. Performance in a bad run of the CR-rule is shown in panel 6C, revealing that performance can deteriorate in a single trial. In this trial, a very unlikely somatic response was observed (panel 6D), resulting in a large value of , thus leading to an excessively large change in synaptic strength.
The finding that large fluctuations in the CR-estimator can arise from rare somatic events, confirms the suspicion in Section 4.2 that recasting Equation 20 as a sampling procedure can lead to problems. Luckily, this can be addressed using the additional degree of freedom provided by the parameter μ in the CR-rule. To dampen the effect of the fluctuations in , we set μ to the time-dependent value
Note that μ is independent of whether or not . Hence, in view of our remark following Equation 21, this is in fact a valid choice for μ. The specific form of Equation 25 is to some extent motivated by the aesthetic considerations. It simplifies the first line of Equation 24 to
We refer to this estimator as balanced cell reinforcement (bCR) (Figure 3).
From the third line of Equation 24, one sees that the somato-dendritic interaction term in Equation 26 can be written as . This highlights the terms role as assessing the relevance to the produced somatic spike train of having an NMDA-event at time t. In this, it is analogous to the terms in the CR-rule. But in contrast to these terms, is bounded. In ZR, plasticity is driven by the exploration inherent in the stochasticity of NMDA-event generation. Formally, this is reflected by the difference entering as a factor in Equation 15, which represents the deviation of the sampled NMDA-events from the expected rate. In bCR, this difference has become a sum. Hence, exploration at the NMDA-event level is only of minor importance for the bCR-rule, where the essential driving force for plasticity is the somatic exploration entering through the factor .
Due to the modification, bCR consistently and markedly improves on ZR, as demonstrated by panel 5A which compares the learning curves for the same task as in panel 6A. The performance improvement seems to become even larger for more demanding tasks. This is highlighted by panel 5B showing the performance when not just one but four different stimulus-response associations have to be learned. For two of the patterns, the correct somatic response was to emit at least one spike, for the other two patterns the correct response was to stay quiescent. One of the four stimulus-response associations was randomly chosen on each trial and, as before, correct somatic responses lead to a reward signal of whereas incorrect responses resulted in . The inset to panel 5B shows the distribution of NMDA-spike durations after learning the four stimulus-response associations with bCR. Over 70% of the NMDA-spikes last for just a little longer than the minimal length of . Further nearly all of the spikes are shorter than , thus staying well within a physiologically reasonable range.
Panels 5C and 5D show results in a task where reward delivery is contingent on an appropriate temporal modulation of the firing rate. Also, in this second output coding paradigm, the bCR-update is found to be much more efficient in estimating the gradient of the expected reward.
We have derived a class of synaptic plasticity rules for reinforcement learning in a complex neuronal cell model with NMDA-mediated dendritic nonlinearities. The novel feature of the rules is that the plasticity response to the external reward signal is shaped by the interaction of global somatic quantities with variables local to the dendritic zone where the nonlinear response to the synaptic release arises. Simulation results show that such so-called CR rules can strongly enhance learning performance compared to the case where the plasticity response is determined just from quantities local to the dendritic zone.
In the simulations, we have considered only a very simple task with a single complex cell learning stimulus-response associations. The results, however, show that compared to ZR the bCR rule provides a less noisy procedure for estimating the gradient of the log-likelihood of the somatic response given the neuronal input (). Estimating this gradient for each neuron is also the key step for reinforcement learning in networks of complex cells . Further, simply memorizing the gradient estimator with an eligibility trace until reward information becomes available, yields a learning procedure for partially observable Markov decision processes, i.e., tasks where the somatic response may have an influence on which stimuli are subsequently encountered and where reward delivery may be contingent on producing a sequence of appropriate somatic responses [22–24]. The quality of the gradient estimator is a crucial factor also in these cases. Hence, it is safe to assume that the observed performance advantage of the bCR rules carries over to learning scenarios which are much more complex than the ones considered here.
In this investigation, we have adopted a normative perspective, asking how the different variables arising in a complex neuronal model should interact in shaping the plasticity response - striving for maximal mathematical transparency and not for maximal biological realism. Ultimately, of course, we have to face the question of how instructive the obtained results are for modeling biological reality. The question has two aspects which we will address in turn: (A) Can the quantities shaping the plasticity response be read-out at the synapse? (B) Is the computational structure of the rules feasible?
(A) The global quantities in CR are the timing of somatic spikes as well as the value of the somatic potential. The fact that somatic spiking can modulate plasticity is well established by STDP experiments (spike timing-dependent plasticity). In fact such experiments can also provide phenomenological evidence for the modulation of synaptic plasticity by the somatic potential, or at least by a low-pass filtered version thereof. The evidence arises from the fact that the synaptic change for multiple spike interactions is not a linear superposition of the plasticity found when pairing a single pre-synaptic and a somatic spike. Explaining the discrepancy seems to require the introduction of the somatic potential as an additional modulating factor .
In CR-learning, however, we assume that the somatic potential U (Equation 5) can differ substantially from a local membrane potential (Equation 1) and both potentials have to be read-out by a synapse located in the ν th dendritic zone. In a purely electrophysiological framework, this is nonsensical. The way out is to note that what a synapse in CR-learning really needs is to differentiate between the total current flow into the neuron and the flow resulting from AMPA-releases in its local dendritic NMDA-zone. While the differential contribution of the two flows is going to be indistinguishable in any local potential reading, the difference could conceivably be established from the detailed ionic composition giving rise to the local potential at the synapse. A second, perhaps more likely, option arises when one considers that NMDA-spiking is widely believed to rely on the pre-binding of Glutamate to NMDA-receptors . Hence, could simply be the level of such NMDA-receptor bound Glutamate, whereas U is relatively reliably inferred from the local potential. Such a reinterpretation does not change the basic structure of our model, although it might require adjusting some of the time constants governing the build up of .
(B) The plasticity rules considered here integrate over the duration T corresponding to the period during which somatic activity determines eventual reward delivery. But synapses are unlikely to know when such a period starts and ends. As in previous works [12, 18], this can be addressed by replacing the integral by a low-pass filter with a time constant matched to the value of T. The CR-rules, however, when evaluating to assess the effect of an NMDA-spike, require a second integration extending from time t into the future up to . The acausality of integrating into the future can be taken care of by time shifting the integration variable in the first line of Equation 24, and similarly for Equation 26. But the time shifted rules would require each synapse to buffer an impressive number of quantities. Hence, further approximations seem unavoidable and, in this regard, the bCR-rule (Equation 26) seem particularly promising due to its relatively simple structure. Approximating the hyperbolic tangent in the rule by a linear function yields an update which can be written as a proper double integral. This is an important step in obtaining a rule which can be implemented by a biologically reasonable cascade of low-pass filters.
The derivation of the CR-rules presented above builds on previous work on reinforcement learning in a population of spiking point neurons [18, 24, 26]. But in contrast to neuronal firings, NMDA-spikes have a non-negligible extended duration and this makes the plasticity problem in our complex cell model more involved. The previous works introduced a feedback signal about the population decision which has a role similar to the somatic feedback in the present CR-rules. A key difference, however, is that the population feedback had to be temporally coarse grained since possible delivery mechanisms such as changing neurotransmitters levels are slow. In a complex cell model, however, a close to instantaneous somatic feedback can be assumed. As a consequence, the CR-rules can now support reinforcement learning also when the precise timing of somatic action potentials is crucial for reward delivery. Yet, if the soma only integrates NMDA-spikes which extend across or more, it appears to be difficult to reach a higher temporal precision in the somatic firing. In real neurons, the temporal precision is likely to result from the interaction of NMDA-spikes with AMPA-releases, with the NMDA-spikes determining periods of heightened excitability during which AMPA-releases can easily trigger a precise somatic action potential. While important in terms of neuronal functionality, incorporating the direct somatic effect of AMPA-releases into the model poses no mathematical challenge, just yielding additional plasticity terms similar to the ones for point neurons . To focus on the main mathematical issues, we have not considered such direct somatic effects here.
We first obtain a more explicit form for . In view of Equation 22, if , whereas if there is NMDA-triggering at time . Hence, setting
Further, from Equation 16,
Hence, taking the limit , we obtain
equivalent to the first equation in Equation 24.
We next note that times s outside of the interval do not contribute to the above integrals since for such s. Further, for . Hence,
For the term in square brackets we note that, since is zero or one, . Hence, finally,
which gives the last line of (Equation 24).
Here, we provide the remaining simulation details.
An input pattern has a duration of and is made up from 150 fixed spike trains chosen independently from a Poisson process with a mean firing rate of (independent realizations are used for each pattern). We think of the input as being generated by an input layer with 150 sites, with each NMDA-zone having a 50% probability of being connected to one of the sites. Hence, on average a NMDA-zone receives 75 input spike trains and 37.5 spike trains are shared between any two NMDA-zones.
A roughly optimized learning rate was used for all tasks and learning rules. Roughly, optimized means that the used learning rate yields a performance which is better that when using or .
In obtaining the learning curves, for each run a moving average of the actual trial by trial performance was computed using an exponential filter with time constant 0.1. Mean learning curves where subsequently obtained by averaging over 40 runs. The exception to this is the single run learning curve in panel 6C. There, subsequently to each learning trial, 100 non-learning trials were used for estimating mean performance.
Initial weights for each run were picked independently from a Gaussian with mean and variance equal to 0.5. Euler’s method with a time step of 0.2 ms was used for numerically integrating the differential equations.
Polsky A, Mel BW, Schiller J: Computational subunits in thin dendrites of pyramidal cells. Nat Neurosci 2004,7(Jun):621–627.
Maass W: Computation with spiking neurons. In The Handbook of Brain Theory and Neural Networks. Edited by: Arbib MA. MIT Press, Cambridge; 2003:1080–1083.
Poirazi P, Brannon T, Mel BW: Pyramidal neuron as two-layer neural network. Neuron 2003,37(Mar):989–999.
Nevian T, Larkum ME, Polsky A, Schiller J: Properties of basal dendrites of layer 5 pyramidal neurons: a direct patch-clamp recording study. Nat Neurosci 2007,10(Feb):206–214.
Zhou WL, Yan P, Wuskell JP, Loew LM, Antic SD: Dynamics of action potential backpropagation in basal dendrites of prefrontal cortical pyramidal neurons. Eur J Neurosci 2008,27(Feb):923–936.
Schiller J, Major G, Koester HJ, Schiller Y: NMDA spikes in basal dendrites of cortical pyramidal neurons. Nature 2000,404(Mar):285–289.
Schiller J, Schiller Y: NMDA receptor-mediated dendritic spikes and coincident signal amplification. Curr Opin Neurobiol 2001,11(Jun):343–348.
Major G, Polsky A, Denk W, Schiller J, Tank DW: Spatiotemporally graded NMDA spike/plateau potentials in basal dendrites of neocortical pyramidal neurons. J Neurophysiol 2008,99(May):2584–2601.
Larkum ME, Zhu JJ, Sakmann B: A new cellular mechanism for coupling inputs arriving at different cortical layers. Nature 1999,398(Mar):338–341.
Larkum ME, Nevian T, Sandler M, Polsky A, Schiller J: Synaptic integration in tuft dendrites of layer 5 pyramidal neurons: a new unifying principle. Science 2009,325(Aug):756–760.
Seung H: Learning in spiking neural networks by reinforcement of stochastic synaptic transmission. Neuron 2003, 40: 1063–1073. 10.1016/S0896-6273(03)00761-X
Fremaux N, Sprekeler H, Gerstner W: Functional requirements for reward-modulated spike-timing-dependent plasticity. J Neurosci 2010,30(Oct):13326–13337.
Williams R: Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach Learn 1992, 8: 229–256.
Matsuda Y, Marzo A, Otani S: The presence of background dopamine signal converts long-term synaptic depression to potentiation in rat prefrontal cortex. J Neurosci 2006, 26: 4803–4810. 10.1523/JNEUROSCI.5312-05.2006
Seol G, Ziburkus J, Huang S, Song L, Kim I, Takamiya K, Huganir R, Lee H, Kirkwood A: Neuromodulators control the polarity of spike-timing-dependent synaptic plasticity. Neuron 2007, 55: 919–929. Erratum in: Neuron56:754. Erratum in: Neuron56:754. 10.1016/j.neuron.2007.08.013
Pawlak V, Kerr JN: Dopamine receptor activation is required for corticostriatal spike-timing-dependent plasticity. J Neurosci 2008,28(Mar):2435–2446.
Werfel J, Xie X, Seung HS: Learning curves for stochastic gradient descent in linear feedforward networks. Neural Comput 2005, 17: 2699–2718. 10.1162/089976605774320539
Urbanczik R, Senn W: Reinforcement learning in populations of spiking neurons. Nat Neurosci 2009, 12: 250–252. 10.1038/nn.2264
Dayan P, Abbott L: Theoretical Neuroscience. MIT Press, Cambridge; 2001.
Pfister J, Toyoizumi T, Barber D, Gerstner W: Optimal spike-timing-dependent plasticity for precise action potential firing in supervised learning. Neural Comput 2006, 18: 1318–1348. 10.1162/neco.2006.18.6.1318
Bertsekas DP, Tsitsiklis JN: Parallel and Distributed Computation: Numerical Methods. Prentice-Hall, Englewood Cliffs; 1989.
Baxter J, Bartlett P: Infinite-horizon policy-gradient estimation. J Artif Intell Res 2001, 15: 319–350.
Baxter J, Bartlett P, Weaver L: Experiments with infinite-horizon, policy-gradient estimation. J Artif Intell Res 2001, 15: 351–381.
Friedrich J, Urbanczik R, Senn W: Spatio-temporal credit assignment in neuronal population learning. PLoS Comput Biol 2011.,7(Jun):
Clopath C, Büsing L, Vasilaki E, Gerstner W: Connectivity reflects coding: a model of voltage-based STDP with homeostasis. Nat Neurosci 2010,13(Mar):344–352.
Friedrich J, Urbanczik R, Senn W: Learning spike-based population codes by reward and population feedback. Neural Comput 2010, 22: 1698–1717. 10.1162/neco.2010.05-09-1010
This study was supported by the Swiss National Science Foundation (SNSF, sinergia grant CRSIKO 122697/1) and a grant of the Swiss SystemsX.ch initiative (Neurochoice, evaluated by the SNSF).
The authors declare that they have no competing interests.
About this article
Cite this article
Schiess, M., Urbanczik, R. & Senn, W. Gradient estimation in dendritic reinforcement learning. J. Math. Neurosc. 2, 2 (2012). https://doi.org/10.1186/2190-8567-2-2
- Dendritic computation
- reinforcement learning
- spiking neuron