Abstract
New sensory stimuli can be learned with a single or a few presentations. Similarly, the responses of cortical neurons to a stimulus have been shown to increase reliably after just a few repetitions. Long-term memory is thought to be mediated by synaptic plasticity, but in vitro experiments in cortical cells typically show very small changes in synaptic strength after a pair of presynaptic and postsynaptic spikes. Thus, it is traditionally thought that fast learning requires stronger synaptic changes, possibly because of neuromodulation. Here we show theoretically that weak synaptic plasticity can, in fact, support fast learning, because of the large number of synapses N onto a cortical neuron. In the fluctuation-driven regime characteristic of cortical neurons in vivo, the size of membrane potential fluctuations grows only as
SIGNIFICANCE STATEMENT Long-term memory is thought to rely on the strengthening of coactive synapses. This physiological mechanism is generally considered to be very gradual, and yet new sensory stimuli can be learned with just a few presentations. Here we show theoretically that this apparent paradox can be solved when there is a tight balance between excitatory and inhibitory input. In this case, small synaptic modifications applied to the many synapses onto a given neuron disrupt that balance and produce a large effect even for modifications induced by a single stimulus. This effect makes fast learning possible with small synaptic changes and reconciles physiological and behavioral observations.
Introduction
New sensory stimuli can be learned with just a few presentations (Seitz, 2010). For example, children can acquire a new word with a single presentation (Carey and Bartlett, 1978). In vision, fast perceptual learning has also been shown to occur over 10–100 presentations for different tasks such as shape discrimination (Rubin et al., 1997), hyperacuity (Poggio et al., 1992; Fahle et al., 1995), and face identification (Hussain et al., 2012). In audition, 500 ms tokens of white noise can be discriminated after just a few presentations in an implicit learning task, and the memory remains after several weeks (Agus et al., 2010). In auditory (Fritz et al., 2003) and visual (Yao et al., 2007) cortex, responses of neurons can also change rapidly after exposure to sensory stimuli. In particular, when a natural movie is presented a few times, cortical neurons start to fire reliably at specific times, and the specific response persists after other visual stimuli have been presented (Yao et al., 2007).
Long-term memory is thought to be mediated by synaptic plasticity (Whitlock et al., 2006; Nabavi et al., 2014). But in contrast with behavioral and in vivo physiological observations, in vitro experiments on synaptic plasticity in cortical cells reveal very small changes after a pair of presynaptic and postsynaptic spikes: the size of a postsynaptic potential typically increases by around 0.5–5% when a postsynaptic spike shortly follows a presynaptic spike (Markram et al., 1997; Bi and Poo, 1998; Sjöström et al., 2001; Wang et al., 2005; Caporale and Dan, 2008; for the average over a large number of pairings, see Discussion). In addition, only synapses where a presynaptic spike occurs shortly before the postsynaptic spike are potentiated (≃20 ms; Gerstner et al., 1996; Markram et al., 1997; Bi and Poo, 1998), which may concern only a small fraction of them if most inputs are asynchronous during alert states (Renart et al., 2010; Tan et al., 2014). Finally, most cortical neurons fire at a low firing rate (Griffith and Horn, 1966; Koch and Fuster, 1989; Laughlin, 2001; Koulakov et al., 2009), and spontaneous or evoked activities do not normally exceed a few tens of hertz. Therefore, a single sensory stimulus lasting 500 ms would not be expected to trigger substantial changes in synaptic strength. Accordingly, in computational models of learning with weak synaptic plasticity, neurons start to respond selectively to a repeated stimulus only after a large number of presentations (Song and Abbott, 2001; Masquelier et al., 2008; Masquelier et al., 2009; Klampfl and Maass, 2013).
Thus, it is traditionally thought that fast learning requires stronger synaptic changes than usually observed in vitro, possibly because of neuromodulation (Fusi et al., 2005). Here we show theoretically that weak synaptic plasticity can, in fact, support fast learning. The mechanism relies on the notion that in vivo, total excitatory input to a cortical neuron is large and matched by inhibition (Softky and Koch, 1993; Destexhe and Paré, 1999; Okun and Lampl, 2008) in the so-called balanced regime. A small relative change in excitation at specific times of the sensory input can then produce a significant difference between the excitatory and the inhibitory drives, i.e., a misbalance resulting in high spiking probability. We show that the leverage effect caused by the excitatory–inhibitory balance is of order
Materials and Methods
Neuron model.
Simulations were performed using current-based, leaky integrate-and-fire neurons with membrane time constant τm = 5 ms (Destexhe and Paré, 1999) and resting membrane potential Vrest = −70 mV. When the membrane potential Vm reaches the spiking threshold Vthresh = −55 mV, a spike is generated, and the membrane potential is clamped to the reset potential Vreset = −70 mV during a refractory period of duration τref = 5 ms. Each presynaptic spike produces an exponential decaying current with time constant τexc = 3 ms for excitation and τinh = 10 ms for inhibition. The model equations are thus as follows: where syn ε {exc, inh} and each presynaptic spike produces an instantaneous increase: gsyn → gsyn + λsyn wi, where wi is the synaptic weight and λsyn is a scaling factor calculated so that a synaptic weight of 1 mV produces postsynaptic potentials (PSPs) of peak size 1 mV. Note that gsyn is in units of volt, i.e., the membrane resistance is implicitly included in the variable.
Spike-timing-dependent plasticity.
The weights of excitatory synapses evolve according to an additive spike-timing-dependent plasticity (STDP) rule (Markram et al., 1997; Bi and Poo, 1998; Song and Abbott, 2001). Each pair of presynaptic and postsynaptic spikes produces an increase δw in weight: In addition, weights are clipped between 0 mV and wmax = 2 mV. We set Adep = 1.2 Apot = εwmax, with ε = 1/100, and τpot = τdep = 20 ms.
In recurrent network simulations, inhibitory to excitatory synapses are also plastic, following the results of Vogels et al. (2011). For every neuron k, a synaptic trace xk is assigned such that at each spike xk ← xk + 1, and otherwise
In Figure 9, we show that the results are valid for several learning rules. We consider the case of binary synapses, whose probability of switch at each post or presynaptic spike [respectively, ζpost(t) and ζpre(t)] is a function of the traces Apost(t) and Apre(t), defined by Apre/post ← Apre/post + Apot/dep at each spike, and otherwise
Feedforward simulations.
Each neuron receives Ne = 8000 excitatory inputs and Ni = 2000 inhibitory inputs, firing according to Poisson processes with firing rate Fe = Fi = 1 Hz (Fig. 1A). The initial excitatory weights are uniformly drawn between 0 and wmax, and inhibitory weights are set such that wi = −1 mV (the results were unchanged with heterogeneous weights). Simulations run for 30 h (biological time) before the presentation of patterns, so as to reach equilibrum. Repeated patterns are presented by pregenerating a single 500-ms-long spatiotemporal pattern of spike trains (identical statistical properties) and presenting it every 2 s. In Figure 7, A and B, we added a normally distributed jitter to all the spike times of the pattern, independent between all spikes and between presentations. In Figure 7, C and D, we added Gaussian white noise to the right side of the membrane equation: σv =
Reliability measure.
We measured response reliability with the same method as in the work of Yao et al. (2007). For each neuron, spike trains were binned with a 50 ms time bin, and we reported the correlation coefficient between each repeat ri and the average of its nearest four repeats. Using a smaller time bin (5 ms) did not change the results.
Recurrent simulations.
We used a random balanced network (van Vreeswijk and Sompolinsky, 1996) composed of 8000 excitatory neurons and 2000 inhibitory neurons (see Fig. 8A). Every neuron in the network is connected to 20% of the others, with delays drawn from a uniform distribution between 0 and 5 ms. Within the network, all the excitatory connections are plastic, governed by a classical STDP rule (see Eq. 1). All inhibitory to excitatory connections are also plastic (see corresponding sections; Vogels et al., 2011) to ensure the stability of the balanced regime. All excitatory neurons within the network are stimulated with Next = 8000 external excitatory inputs, as in the feedforward scenario. External inputs project to those neurons with connection probability 20% and uniformly randomized delays, and those connections are also plastic. Initial synaptic weights for recurrent connections are drawn from Gaussian distributions such that 〈we〉 = 0.2 mV, 〈wi〉 = −1 mV. SDs are chosen to be one-fourth of the means. The initial strength of the external synapses is also drawn from a Gaussian distribution such that 〈wext〉 = 〈we〉. To sustain the asynchronous activity, a constant input of Vthresh − Vrest = 15 mV is injected into all neurons as in the study by Vogels et al. (2011).
Theory.
To calculate the Vm distribution, we neglect the contribution of spikes and use Campbell's theorems, which give the mean and variance of a shot noise (corresponding to sums of PSPs with Poisson statistics):
where averages are over synapses. PSPs of the same type vary only by their synaptic strength (defined as the peak value): EPSP = weEPSP* and IPSP = wiIPSP*, where:
where τs is the synaptic time constant and λs is the normalization factor (defined above). We define αe = ∫ EPSP(s)ds = λeτe and βe= ∫ EPSP(s)2ds = λe2
To estimate σ2(Vm) at equilibrium in a neuron with synapses governed by STDP, we assume that (1) all excitatory synapses are either weak (we = 0 mV) or maximally potentiated (we = wmax) and (2) excitation and inhibition are balanced, so that 〈Vm〉 = Vrest. This latter approximation, which we call the balance equation (using Eq. 4), is valid when both excitation and inhibition are strong. Denoting p the proportion of active (nonweak) synapses, the balance equation reads as follows: The substitution in Equation 5 gives the following: This theoretical prediction is displayed in Figure 1D and used in Figure 3C.
In Figure 3A, we consider a neuron with uniform excitatory and inhibitory synapses (a single pair of values for we and wi), and we calculate we and wi so that 〈Vm〉 = Vrest and σ2 (Vm) is a given value. Here the balance equation reads as follows:
The substitution in Equation 5 gives the following:
This equation gives wi as a function of σ2(Vm), so that
Once wi is obtained, we is deduced using the balanced Equation 9, and these values are used in Figure 3A. Note that both wi and we scale as 1/
Simulator.
All simulations were performed using the Brian simulator version 2 (Goodman and Brette, 2009) with a fixed time step of 0.1 ms.
Results
The balanced regime
Cortical neurons receive inputs from about 10,000 synapses (Braitenberg and Schütz, 1991), resulting in an intense bombardment of synaptic activity in vivo (Destexhe et al., 2003). Approximately 80% of those synapses are excitatory, and it is believed that the large excitatory drive is balanced by a similarly large inhibitory drive (Okun and Lampl, 2008). This “balanced regime” (van Vreeswijk and Sompolinsky, 1996; Brunel, 2000; Renart et al., 2010) has several characteristic features observed in vivo. First, the membrane potential (Vm) distribution peaks below threshold [Rossant et al. (2011), their Fig. 7]. Second, neurons fire irregularly when excitation transiently exceeds inhibition (Softky and Koch, 1993; Shadlen and Newsome, 1998). Third, their firing rates are relatively low (Koulakov et al., 2009; Bruno, 2011).
As shown previously (Song and Abbott, 2001), these features can be reproduced by a simple neuron model undergoing STDP (Gerstner et al., 1996; Markram et al., 1997; Bi and Poo, 1998) where long-term depression is larger than long-term potentiation. In Figure 1, we simulated an integrate-and-fire model with Ne = 8000 excitatory inputs and Ni = 2000 inhibitory inputs (Fig. 1A), randomly firing at 1 Hz with no correlations (see Discussion for the impact of correlations). The excitatory synapses undergo STDP (Fig. 1B) with a maximum PSP of 2 mV (time constant τe = 3 ms), whereas inhibitory synapses were not plastic (PSP size, −0.5 mV; time constant τi = 10 ms); we examine inhibitory plasticity in a later part. Results are shown at equilibrium, after 30 h of biological time (Fig. 1C). The neuron then fires irregularly (〈CV〉 ≃1) at a low firing rate (F = 2.17 Hz; Fig. 1D,E). The Vm distribution is approximately Gaussian and peaks near −71 mV, ≃1 mV below the resting potential (Fig. 1D). The excitatory drive is large, corresponding to a mean depolarization of 22 mV (±6.7 mV) in the absence of inhibition, but is matched by a large amount of inhibition so that the mean Vm is well below threshold (Fig. 1F). Note that we modeled synaptic inputs as currents rather than conductances so that we can easily compare with analytical calculations (Fig. 1D, dashed line shows the theoretical prediction for the Vm distribution), but the same basic properties would hold with synaptic conductances (El Boustani and Destexhe, 2009), where the balance equation is the equality of mean excitatory and inhibitory currents at mean Vm (Richardson and Gerstner, 2005).
Synaptic plasticity induced by a single spike
We now examine the effect of a single output spike on synaptic modifications, when the neuron receives random inputs (Fe = Fi = 1 Hz; Fig. 2A). With a LTP window of size τpot = 20 ms, approximately Nsyn = τpot FeNe = 160 synapses are potentiated, on average, and a similar number are depressed. This is a small fraction of the 8000 excitatory synapses, and each synaptic modification represents only ε = 1% of the maximum PSP size (i.e., δw = 20 μV), in accordance with experimental observations (Bi and Poo, 1998; Wang et al., 2005), so that the distribution of synaptic weights is essentially unchanged by this single spike (Fig. 2B). However, the total synaptic modification is a sizable fraction of maximum PSP size (160 × 1% = 160%). If the same pattern of excitatory inputs were presented again, this would result in a corresponding change in the excitatory–inhibitory balance at the time of the previous spike that triggered LTP (Fig. 2C–E), where ΔV is the total change in depolarization attributable to the plastic changes. Such a misbalance has a strong effect on firing probability in the balanced regime (Rossant et al., 2011). Figure 2F shows the probability distribution of Vm at the time of the output spike for the particular pattern of excitatory inputs shown in Figure 2A (here inhibition is the source of variability), before and after synaptic modification. The probability that the neuron fires in response to the excitatory pattern (red area) goes from 22% to 50% because of the synaptic changes. Thus, plasticity induced by a single output spike more than doubles the probability that the neuron fires to the same input.
This phenomenon can be explained theoretically by a leverage effect caused by the large number of synapses that contribute to relatively small fluctuations in Vm in the balanced regime (van Vreeswijk and Sompolinsky, 1996). Assuming inputs are independent, the standard deviation of the excitatory drive (measured as current, conductance, or total depolarization) scales as
To demonstrate this leverage effect, we simulated an integrate-and-fire model receiving random excitatory and inhibitory inputs (Fe = Fi = 1 Hz) through synapses with fixed weights, in a 4:1 ratio (Fig. 3). The synaptic weights for excitation we and inhibition wi were adjusted so that the mean Vm equals the resting potential (imposing a linear relationship between the weights we and wi; see Materials and Methods), and the SD σ(Vm) is 5.2 mV (imposing an affine relationship between we2 and wi2). The value of 5.2 mV is arbitrary, similar to the value measured in the equilibrium obtained in Figure 1. Therefore, to keep the SD of the voltage constant, weights had to be scaled as 1/
There is an additional leverage effect in the situation shown in Figure 2, coming from the large proportion of weak synapses (90% of synaptic weights are smaller than half the maximum; Fig. 2B). This is the result of the STDP mechanism we chose, which produces bimodal distributions of weights (van Rossum et al., 2000; Gütig et al., 2003), and this is consistent with the fact that neurons, in particular hippocampal cells, are known to have a large fraction of weak synapses (Isaac et al., 1995; Rumpel et al., 1998). Weak synapses do not contribute to the Vm fluctuations, but they can, in principle, be potentiated (Fig. 2A,B, green dots). That is, if p is the proportion of nonweak synapses, then σ(Vm) is proportional to
Fast learning, slower forgetting
Because of these two leverage effects, neurons can learn to respond to a repeated pattern, as shown in Figure 4. We simulated the same model as in Figure 2, and we embedded a repeated 500 ms pattern of excitatory inputs into the random stream of spikes, with the same statistics (independent Poisson inputs at Fe = 1 Hz; Fig. 4A). The pattern was repeated every 2 s. The only difference between the pattern and background activity is that it is repeated. The output of 5000 independent neurons responding to the same pattern is shown in Figure 4B. After just 10 presentations, the neurons respond much more to the repeated pattern than to background activity (∼8 Hz vs 2 Hz; Fig. 4B), even though synaptic weights have changed very little (Fig. 4C,D). Figure 4E shows the response of two sample neurons to the repeated pattern over several successive presentations. The firing rate appears to increase by inserting spikes at precise times, and response reliability (see Materials and Methods) also tends to increase (Fig. 4F). This observation agrees with in vivo experiments in which response reliability was seen to increase in response to repeated presentations of natural stimulus (Yao et al., 2007).
In contrast with learning, forgetting is not impacted by the leverage effects discussed above and therefore is much slower. In Figure 5A, we presented the model with 10 repetitions of the same excitatory input pattern, presented random inputs for 2 min, and presented the repeated pattern again. After the 2 min pause, neurons respond to the pattern with about the same firing rate as they did just before the pause. This observation mimics experimental observations in vivo, where enhanced cortical responses to natural stimulus persisted after the presentation of unrelated stimuli (Yao et al., 2007; Bermudez Contreras et al., 2013). Why is forgetting much slower than learning? This is simply because erasing a memory requires modifying the same synapses that have been potentiated, but a given synapse has a low probability of being modified by a random input. For a given output spike, the probability that an excitatory input spike falls in the temporal window of depression (τdep = 20 ms here) is τdepFe = 0.02. If the neuron spikes at a rate Fpost = 2 Hz, then a given synapse is depressed randomly at rate τdepFeFpost = 0.04 Hz, i.e., once every 25 s. In addition, that synapse may be depressed, but it may also be potentiated. The net average modification is the integral of the STDP function (Fig. 1B), so the rate of modification is Fe Fpost(Apot τpot − Adep τdep, where Apot is the peak LTP and Adep is the peak LTD), which must be compared with a previous potentiation of approximately Apot. Thus, the relative rate of synaptic decay is FeFpost(τpot − τdepAdep/Apot) = 0.8%/s here. This decay corresponds to a retention time of ∼2 min. In Figure 5B, we show the firing rate of the neurons just after the pause as a function of pause duration for different initial numbers of presentations of the pattern. It appears that the retention time, defined as the time when the response to the stimulus has decayed by 50%, is of the order of 15 min (Fig. 5C). The theoretical prediction underestimates the retention time because the calculation implicitly assumes that inputs and outputs are independent, which is only true for weak synapses. As can be seen in Figure 5D, the average modification for the weight change (solid red curve) matches the theoretical prediction (dashed line) for weak synapses, but not for strong synapses. For the latter ones, weights tend to increase because they can cause output spikes and thus induce LTP, which increases the retention time calculated above. We note that although forgetting is slower than learning, the mechanism we describe explains the retention of memories over minutes or tens of minutes, but not over days or weeks. We assume that longer time scales involve additional consolidation mechanisms (see Discussion).
Interference between patterns
What happens when two patterns are presented? In Figure 6, we consider two different scenarios. In the first scenario, two independent patterns are repeatedly presented in an interleaved fashion, 10 times each (Fig. 6A). Neurons fire increasingly to both patterns (Fig. 6B). When the two patterns are presented again after a 110 s pause, neurons fire again at the same rate as in the last presentation before the pause. In the second scenario, a pattern is repeatedly presented 10 times, then a second pattern is presented 10 times, and the first pattern is presented again 10 times (Fig. 6C). It appears that there is no interference between the two patterns (Fig. 6D). That is, when the second pattern is presented, neurons fire increasingly as they did for the first pattern; when the first pattern is presented again, neurons start firing at about the same rate as in the last presentation of that pattern. This lack of interference is expected because an output spike triggers potentiation for a small proportion of synapses (on the order of Fe τpot = 0.5%), and therefore there should be little overlap in the potentiated synapses between two independent patterns.
Requirements for fast learning
The effect we have shown relies mainly on the neuron being in the balanced regime, where spikes are caused by transient increases in excitation relative to inhibition (or decrease of inhibition), and on the leverage effect because of the large number of synapses. Another implicit requirement is that the precise spike times of the stimulus are reproducible. Indeed, the synaptic modifications caused by a single spike can only contribute to an increased response to the stimulus if the same synapses are simultaneously coactivated. More precisely, the relative spike times or at least the synchrony events must be reproducible. This is shown in Figure 7, A and B, where we introduced random jitter in the input spike times of our model. When the jitter is small, the firing rate in response to the stimulus increases with the number of presentations, but the effect vanishes when the jitter exceeds the membrane time constant (τm = 5 ms here). Synaptic failures have no impact on spike timing and, therefore, may quantitatively decrease the rate of learning but have no consequence on the basic leverage effect.
In addition, to provide a more complete picture of how noise can perturb the learning process, we added a Gaussian noise with a fixed voltage variance σV2 to the membrane potential dynamics of all neurons, representing intrinsic noise or external stimulus-unrelated noise. As can be seen in Figure 7, C and D, adding noise results in a global increase in the firing rate but has little effect on the rate increase with repeated pattern presentations.
Our analysis was done in a feedforward structure. Would the phenomenon still occur in a recurrently connected neural network? To answer this question, we considered a randomly connected network of 8000 excitatory neurons and 2000 inhibitory neurons, with connection probability 20% between any pair of neurons and conduction delays drawn uniformly between 0 and 5 ms (Fig. 8A). Excitatory neurons also receive external excitatory inputs from 8000 neurons, with 20% connection probability and random delays (see Materials and Methods). Feedforward external synapses are plastic as before, recurrent excitatory synapses within the network are also subject to classical STDP, and inhibitory to excitatory synapses within the network are subject to STDP according to a previously described mechanism (Vogels et al., 2011). We first run the model with random external inputs (same as in the feedforward model) until the network has reached equilibrium (500 s; Fig. 8B, left). Then, during a period of 100 s, the same pattern is presented every 2 s, as in the feedforward simulation (Fig. 8B, middle and right). After this period of 100 s, we present random input to the network during 100 s, then we present the pattern again (Fig. 8B, inset, average firing rate of the network over 2 s time bins during the whole simulation). The response of the network to the pattern increases rapidly during the first 100 s of presentation (Fig. 8B, middle), and this specific response is preserved after the 100 s pause, as in the feedforward situation (Fig. 8B, inset). Figure 8C shows the response to the pattern (first half) and to the random input (second half) after learning. At equilibrium, neurons fire irregularly and receive substantial excitatory input from both external (〈ΣEPSPext〉 = 10 mV) and recurrent (〈ΣEPSPrec〉 = 8.1 mV) connections (Fig. 8D). The balanced state is maintained by inhibitory plasticity (Vogels et al., 2011), such that inhibition and excitation approximately cancel each other. Figure 8E shows the distribution of synaptic weights, with a bimodal distribution for excitatory weights, both from recurrent wee and external wext connections (inset; most synapses are weak). Note that, also as in the feedforward scenario, the reliability of the response to the pattern increases with the number of presentations (Fig. 8F). Thus, the phenomenon of fast learning is essentially conserved in a recurrently connected network, and we expect a stronger effect in networks with more synapses (here we used only Nsyn = 1600 external plastic synapses for computational reasons).
Extension to other plasticity rules
To illustrate the generality of our results, we also show in Figure 9 the results of simulations performed with two other plasticity rules, a pair-based plasticity rule with binary synapses and a triplet rule (Pfister and Gerstner, 2006). In experiments, changes in synaptic strength are reported for a large number of pairings, and we have implicitly assumed that each pre–post pairing yields a small change in synaptic strength. Alternatively, each pre–post pairing could yield a large change in synaptic strength, with low probability. This alternative interpretation makes little difference to our theoretical proposition, which relies on the effect on a large number of synapses (i.e., the average effect is what matters). This is demonstrated in Figure 9B, where we simulated a plasticity model with binary synapses (with either zero or maximal weight; see Materials and Methods), and pairings trigger low probability switches in the synapse state (average synaptic modifications were the same as in the original model). Specifically, plastic modifications are determined by the same presynaptic and postsynaptic traces Apre/post(t) and as in the previous model, but the traces determine the probability of state switching rather than the change in synaptic strength. As can be seen in Figure 9B, the same effect can be observed (although with more variability).
Figure 9C shows the results of a simulation performed with a triplet rule adapted from Pfister and Gerstner (2006). This learning rule is able to capture more complex phenomena such as the frequency dependence of STDP (Sjöström et al., 2001). As shown in Figure 9, the results are qualitatively similar compared with those in Figure 2A. A few presentations of the pattern can trigger a significant change in the responses of the neurons. This is because the leverage effect observed here, whether with pair-based STDP or with the triplet learning rule, only depends on the fact that many synapses are potentiated by a single output spike (of order N), whereas the magnitude of Vm fluctuations represents only a small number of PSPs (of order
Discussion
This theoretical study has demonstrated that a fast increase in the response of a neuron to a new sensory stimulus, as observed in vivo (Yao et al., 2007), is consistent with the very small changes in synaptic strength observed in vitro (Markram et al., 1997; Bi and Poo, 1998; Wang et al., 2005). This apparent paradox is mainly caused by a leverage effect of order
The main requirements for the effect are (1) a large number of synapses, (2) a fluctuation-driven regime, and (3) precise spike timing (within the neuron's temporal window of integration; Fig. 7). Pyramidal cells of the cortex and hippocampus receive, on average, ∼10,000 synapses (Braitenberg and Schütz, 1991), giving a leverage effect of order 100. Evidence that cortical neurons operate in vivo in the fluctuation-driven regime comes from the irregularity of spike trains (Shadlen and Newsome, 1998; Nawrot et al., 2008) and from the characteristic shape of Vm distributions with a peak below threshold, reflecting a subthreshold mean current [Rossant et al. (2011), their Fig. 7] and measurements in barrel cortex (Crochet and Petersen, 2006), primary auditory cortex (DeWeese and Zador, 2006), primary visual cortex (Azouz and Gray, 1999), frontal cortex (Léger et al., 2005), and primary motor cortex (Brecht et al., 2004). There is additional evidence that inhibition is finely matched to excitation (Froemke et al., 2007; Okun and Lampl, 2008). Evidence for precise spike timing is more debated, but it should be noted that the requirement is not the reproducibility of absolute spike timing (relative to the stimulus) but of relative spike timing (or, more precisely, stimulus-specific synchrony). Many studies have reported such experimental evidence in response to natural stimuli presented in vivo (Petersen et al., 2001; Gollisch and Meister, 2008; Haider and McCormick, 2009; Bruno, 2011). Finally, there is evidence that neural activity is not stationary, with correlated firing in groups of neurons (Miller et al., 2014), contrary to what we considered in this study. However, those correlations were observed on a relatively long time scale (∼250 ms), whereas only widespread correlations at the time scale of the integration time window (a few milliseconds) would have an impact on our assumptions (fluctuations of order
An implicit assumption in our study is that single pre–post pairings should yield a change in synaptic strength. In experiments, changes in synaptic strength are reported for a large number of pairings, and here we have considered that the change is linear in the number of pairings. Linearity has been observed for 10–60 pairings (Froemke et al., 2006; Wittenberg and Wang, 2006), with saturation for larger numbers of pairings. Usually, observations of synaptic plasticity need several repetitions of a single pairing to observe an effect (Bi and Poo, 1998; Wang et al., 2005). This could be because the change for a single pairing is too small to be detected. Alternatively, it could be that there is a low probability of change in synaptic weight for a single pairing, rather than a small change in synaptic weight for each pairing. This would mean that on most trials with a single pairing, the synaptic weight would not change at all. As we have shown in Figure 9B, this alternative interpretation does not substantially alter our results. In vitro studies on synaptic plasticity also do not always report changes for low-frequency pairings (Sjöström et al., 2001). However, these negative results may be attributable to the neuron's voltage being at resting potential. In a depolarized state characteristic of cortical neurons in vivo, synaptic modifications are indeed seen with low-frequency pairings (Cassenaer and Laurent, 2007; Harvey and Svoboda, 2007). Nonetheless, we cannot entirely rule out the possibility that a single pairing does not produce synaptic changes, in which case changes in neural responses after a single presentation would be attributable to some other mechanism than synaptic plasticity.
We have also shown that an additional leverage effect is provided by the presence of weak synapses that have small strength but can undergo the same plastic changes as active synapses. The proportion of weak synapses in cortical pyramidal cells or hippocampal neurons is thought to be substantial (Montgomery and Madison, 2002; Kerchner and Nicoll, 2008), and it has indeed been shown that the magnitude of LTP is similar for weak and strong synapses (Bi and Poo, 1998). In this study, we used a simple additive model of STDP (Bi and Poo, 1998; Song and Abbott, 2001). The actual physiological mechanism is likely to be more complex in neurons; for example, synaptic modifications may be influenced by voltage (Ngezahayo et al., 2000), pairing frequency (Sjöström et al., 2001), triplet effects (Wang et al., 2005), and initial synaptic strength (Bi and Poo, 1998). Several models may be able to capture those effects (Clopath et al., 2010; Yger and Harris, 2013), but even if this complexity would probably have a quantitative impact on the results, the main phenomenon does not depend on it (Fig. 9). The leverage effect observed here only depends on the fact that many synapses are potentiated by a single output spike (of order N), whereas the magnitude of Vm fluctuations represents only a small number of PSPs (of order
In a former neuronal network study with recurrent plasticity (Morrison et al., 2007), it was observed that synchronous stimulation resulted in de-balancing the network, which did not occur in this study for three reasons. First, we found no significant difference with and without plastic recurrent excitatory synapses in our study. Second and most importantly, delays were modeled as dendritic in the aforementioned study so that simultaneous spikes in two connected neurons result in a pre–post sequence at the synapse, triggering potentiation. In our simulations, delays were modeled as axonal, and the same configuration leads to depression. Third, inhibitory plasticity, which was included in our simulations, tends to restore the balance between excitation and inhibition (Vogels et al., 2011).
We also modeled synaptic inputs as currents rather than conductances, a simplifying assumption that makes analytical calculations simpler. The theory can be readily extended to the conductance case by replacing the balance equation with the equality of mean excitatory and inhibitory current at mean Vm (Richardson and Gerstner, 2005). As the excitatory reversal potential is high, including conductances should make little difference on the effect of excitatory inputs, but the effective membrane time constant of neurons would be lower. In our study, we implicitly included this effect by using a short membrane time constant (5 ms). We also assumed that the inputs were uncorrelated. The presence of stimulus-specific correlations between excitatory inputs would be expected to increase the effect we have described. Indeed, in the fluctuation-driven regime, very small pairwise correlations (of order 1/N) result in a large increase in output firing rate, simply because there are many pairs of inputs (Rossant et al., 2011). Thus, we expect that stimulus-specific correlations would increase the learning rate. Nonstimulus-specific correlations, e.g., attributable to shared anatomical inputs, might be canceled by excitatory–inhibitory correlations (Renart et al., 2010).
Finally, our study addresses the retention of learned stimuli on the time scale of minutes or tens of minutes, but not on the time scale of days or weeks. Understanding how memory can be kept stable in the presence of spontaneous activity is still an open question. It has been shown that in networks with standard STDP, such memories cannot last (Morrison et al., 2007; Vogels et al., 2011; Litwin-Kumar and Doiron, 2014), except if some additional mechanisms are added such as nonlinearities (El Boustani et al., 2012) or bistability (Higgins et al., 2014). Therefore, longer time scales are thought to involve additional mechanisms such as synaptic tagging (Redondo and Morris, 2011), reflected in the late phase of LTP (Frey and Morris, 1997), which consolidate the initial changes, or neuromodulation (Matsuda et al., 2006). Moreover, the phenomenon discussed in this study may also underlie the reverberation of evoked activity patterns in spontaneous activity as observed in vivo (Kenet et al., 2003; Ikegaya et al., 2004; Yao et al., 2007; Han et al., 2008; Bermudez Contreras et al., 2013), which is thought to play a role in the consolidation of memories during sleep (Peyrache et al., 2009).
Footnotes
This work was supported by the Fondation pour la Recherche Médicale and Agence Nationale de la Recherche (Grant ANR-14-CE13-0003).
The authors declare no competing financial interests.
- Correspondence should be addressed to Pierre Yger, Institut de la Vision, INSERM U968, CNRS UMR7210, 75012 Paris, France. pierre.yger{at}inserm.fr