 Research
 Open access
 Published:
Identification of Criticality in Neuronal Avalanches: I. A Theoretical Investigation of the Nondriven Case
The Journal of Mathematical Neuroscience volume 3, Article number: 5 (2013)
Abstract
In this paper, we study a simple model of a purely excitatory neural network that, by construction, operates at a critical point. This model allows us to consider various markers of criticality and illustrate how they should perform in a finitesize system. By calculating the exact distribution of avalanche sizes, we are able to show that, over a limited range of avalanche sizes which we precisely identify, the distribution has scale free properties but is not a power law. This suggests that it would be inappropriate to dismiss a system as not being critical purely based on an inability to rigorously fit a power law distribution as has been recently advocated. In assessing whether a system, especially a finitesize one, is critical it is thus important to consider other possible markers. We illustrate one of these by showing the divergence of susceptibility as the critical point of the system is approached. Finally, we provide evidence that power laws may underlie other observables of the system that may be more amenable to robust experimental assessment.
1 Introduction
A number of in vitro and in vivo studies [1–4] have shown neuronal avalanches—cascades of neuronal firing—that may exhibit power law statistics in the relationship between avalanche size and occurrence. This has been used as prima facie evidence that the brain may be operating near, or at, criticality [5, 6]. In turn, these results have generated considerable interest because a brain at or near criticality would have maximum dynamic range [7–10] enabling it to optimally react and adapt to the dynamics of the surrounding environment [5, 11] whilst maintaining balanced neuronal activity [12–14]. Neuropathological states (e.g., epileptic seizures) could then be conceptualised as a breakdown of, or deviation from, the critical state; see [15], for example. Furthermore, these findings have led to the notion that the brain may selforganise to a critical state [16], i.e., its dynamics would be driven toward the critical regime by some intrinsic mechanism and not be dependent on external tuning. In support of this view, Levina and colleagues [17] showed analytically and numerically that activitydependent depressive synapses could lead to parameterindependent criticality.
The interpretation that neuronal activity is poised at a critical state appears to be mostly phenomenological whereby an analogy has been developed between the propagation of spikes in a neuronal network and models of percolation dynamics [18] or branching processes [19, 20]. Remarkable qualitative similarities between the statistical properties of neuronal activity and the above models have given credence to this analogy, however, the question remains as to whether it is justified. Indeed, various key assumptions underlying percolation dynamics and branching processes are typically violated in the neuroscience domain. For example, full sampling, which is required in order to assess selforganised criticality, is unattainable even in the most simple in vitro scenario and yet it has been shown that subsampling can have profound effects on the distribution of the resulting observations [21]. On a related note, and more generally, the formal definition of a critical system as one operating at, or near, a secondorder (continuous) phase transition is problematic since the concept of phase transition applies to systems with infinite degrees of freedom [22]. Many neuroscience authors address this by appealing to the concept of finite size scaling and many published reports implicitly assume that distributions are power law with truncation to account for the socalled finite size effect. Typically, such reports adopt an approach whereby (a) scale invariance is assessed through finite size scaling analysis, confirming that upon rescaling the event size, the curves collapse to a power law with truncation at system size (but see below regarding the definition of system size); (b) the parameters of statistical models are estimated, typically over a range of event size values that are rarely justified; and (c) the best model is determined by model selection, in which power law and exponentially truncated power law are compared to alternatives such as exponential, lognormal and gamma distributions; see [23] for a typical example. Whilst greater rigour in the statistical treatment of the assessment of the presence of power laws has been attained following Clauset and colleagues’ influential paper [24], what seems to be lacking is a rigorous treatment as to why a power law should be assumed to begin with. Although this question is particularly pertinent to the neurosciences, it should be noted that similar questions remain open in the field of percolation theory (e.g., [25, 26]), namely: (i) how does the critical transition behaviour emerge from the behaviour of large finite systems and what are the features of the transition? (ii) what is the location of the scaling window in which to determine the critical parameters?
This paper specifically seeks to address the following questions:

1.
Assuming that the whole brain, or indeed a region of interest defined by where data can be obtained, is operating at criticality, can we reasonably expect power law statistics in neural data coming from a very small (possibly subsampled) subset of the system? If not, what would be the expected distribution? Sornette [27] states that the Gamma distribution is “found in critical phenomena in the presence of a finite size effect or at a finite distance from the critical point.” Jensen [28] claims that finitesize systems often show an exponential cutoff below the system size. However, we are not aware of any study in which the distribution of event sizes in a finitesize system set to operate at a critical regime has been investigated.

2.
In a finitesize system, is it reasonable/possible to perform a robust statistical assessment of power law statistics? Even the application of a rigorous model selection approach will lead to different results depending on the choice of the range of event sizes and the number of samples being considered [29]. The issue of range selection is of particular interest. Whilst the notion of system size is clear in models of criticality such as the Abelian sandpile where (i) there is full sampling, (ii) the number of sites is finite, and (iii) there is dissipation at the edges, system size is much less obvious where reentrant connections exist, making it possible, in principle, for avalanches to be of infinite size. Here, the counting measure which leads to the definition of an avalanche is important. Counting the number of neurones involved in an avalanche will lead to a clearly defined system size, whereas counting the total number of activations—the de facto standard, e.g., [12, 17, 30]—will not. Furthermore, it should also be noted that the presence of reentrant connections invalidates the standard theory of branching processes [20], and makes a rigorous determination of the branching parameter σ problematic if not impossible, e.g., in the presence of avalanches merging.

3.
Are there other markers of criticality that may be more amenable to characterisation and that should be considered instead of, or in addition to, the statistics of event sizes? The need for such markers in neuroscience has been recognised (see [29] for example) and a number of studies have investigated longrange temporal correlations (powerlaw decay of the autocorrelation function) in amplitude fluctuations [31] and in interburst intervals [32, 33]. However, a theoretical account of how those may relate to one another is lacking (although see the recent work in [34]). Other markers of criticality (or markers of transitions) have been associated with critical physical systems, e.g., divergence of susceptibility and slowing of the recovery from perturbations near the critical point [27], however, we are not aware of any theoretical or empirical study investigating them in a neuroscience context.
One way to address these questions more rigorously is to use simplified, but therefore more tractable conceptual models (e.g., [35]). In this paper, we use a model of a purely excitatory neuronal system with simple stochastic neuronal dynamics that can be tuned to operate at, or near, a secondorder phase transition (specifically, a transcritical bifurcation). The simplicity of the model allows us to analytically calculate the exact distribution of avalanche sizes, which we confirm through simulations of the system’s dynamics. We study our model at the critical point and compare our exact distribution to the explicit but approximate solution proposed by Kessler [36] in an analogous problem of modelling disease dynamics. We confirm that Kessler’s approximate solution converges to our exact result. Importantly, we show that, in the proposed finitesize system, this distribution is not a power law, thus highlighting the necessity of considering other markers of criticality. We thus analyse two potential markers of criticality: (i) the divergence of susceptibility that arises in the model as we approach the critical point, (ii) the slowing down of the recovery time from small disturbances as the system approaches the critical point. Finally, we speculate on a sufficient but not necessary condition under which our exact distribution could converge to a true power law in the limit of the system size.
2 The Stochastic Model
We start from the stochastic model of Benayoun et al. [12], which we simplify to the most trivial of models. A fully connected network of N neurones is considered with purely excitatory connections (as opposed to the excitatory and inhibitory networks considered in [12]). Within the network, neurones are considered to be either quiescent (Q) or active (A). Taking a small time step dt and allowing dt\to 0 the transition probabilities between the two states are then given by:
where {s}_{i}(t)={\sum}_{j}\frac{{w}_{ij}}{N}{a}_{j}(t)+{h}_{i} is the input to the neurone. Here, f is an activation function, {h}_{i} is an optional external input, {w}_{ij} is the connection strength from neurone i to neurone j, and {a}_{j}(t)=1 if neurone j is active at time t and zero otherwise. α is the deactivation rate and, therefore, controls the refractory period of the neurone.
To allow tractability, we further make the following simplifications:

1.
We assume that all synaptic weightings are equal ({w}_{ij}=w).

2.
We assume there is no external input. The driven case will be explored theoretically and empirically in a companion manuscript.

3.
We assume the linear identity activation function f(x)=x. Although it is more common to use sigmoid activation functions, we note that the identity function can just be thought of as a suitably scaled tanh function over the desired range.
As the network is fully connected, we can write the following mean field equation for active neurones:
where we have appealed to the fact that the system is closed, and thus A+Q=N. This ODE is analogous to the much studied [37] susceptible → infectious → susceptible (SIS) model used in mathematical epidemiology and we can appeal to some of the known results in studying its behaviour. Primarily, we can use simple stability analysis. The nonzero steady state is given by {A}^{\ast}=N(1\alpha /w). Setting g(A)=\frac{dA}{dt}, this equilibrium point is stable if {g}^{\prime}({A}^{\ast})<0. Thus,
Borrowing from epidemiology, we define the threshold {R}_{0}=\frac{w}{\alpha}. If {R}_{0}>1, we see that {g}^{\prime}({A}^{\ast})=\alpha w<0 and the nonzero steady state is stable. Figure 1 illustrates the differing behaviour of the solution to the above ODE for {R}_{0}<1 (subcritical), {R}_{0}=1 (critical), and {R}_{0}>1 (supercritical).
2.1 Firing Neurones and Avalanches
Instead of focussing on the average activity level across the network, we will instead look at the size distribution of firing neurones following one firing event. To do this, we begin with a fully quiescent network and initially activate just one neurone. We then record the total number of neurones that fire (the number of quiescent to active transitions) until the network returns to the fully quiescent state. We use this process of sequential activation as our definition of an avalanche and our main interest is the distribution of the avalanche sizes. Unfortunately, the simple ODE approach will not provide us with this distribution. To calculate this distribution, we use the semianalytic approach described in the following section.
2.2 Tree Approach to the Avalanche Distribution
We begin by defining {q}_{i} as the probability the next transition is a recovery (from A to Q) given i active neurones (i>0). The probability the next transition is an activation is then 1{q}_{i} and we have:
In order to calculate the avalanche size distribution, we adopt a recursive approach. We begin by considering the process unfolding in a treelike manner with 1 initially active neurone. The tree can be divided into levels based on the number of transitions that have occurred and how the process is unfolding. Let level j contain the possible number of active neurones after j transitions. The recursive tree approach relates the probability of transition between levels to the final avalanche size. Figure 2 illustrates the initial levels of this process.
To continue we define {p}_{j}^{i} as the probability of having i active neurones on level j with i=0,1,2,\dots ,N and j\in {\mathbb{N}}_{0}. Assuming initially only one active neurone, we immediately see that {p}_{0}^{1}=1, {p}_{1}^{2}=1{q}_{1} and {p}_{1}^{0}={q}_{1}. To proceed, we will consider the probability of having a particular number of active neurones on an arbitrary level. First, we note the following relation between levels:
We now define:
We can now write \mathbf{p}(l+1)=\mathbf{A}\cdot \mathbf{p}(l) where matrix A is given by the following tridiagonal matrix:
with {b}_{i}=(1{q}_{i1}) and {c}_{i}={q}_{i+1}.
On the j th level of the tree, the probability of only 1 neurone being active is given by {p}_{j}^{1}. As on level 0, we began with only a single active neurone then for j odd, {p}_{j}^{1} is always equal to zero. For j even, say j=2k, then as we began with only one active neurone on level 0, to have only one active neurone on level j means that k firings must have occurred. We can then calculate the probability of zero active neurones after k firings as {q}_{1}{p}_{2k}^{1}; this is thus the probability, P(k+1), of having an avalanche of size k+1 (or size k if we were not to include the initial active neurone). Setting e={(1,0,0,\dots ,0)}^{T} and noting that {p}_{2k}^{1}={e}^{T}{A}^{2k}e we have P(k+1)={q}_{1}{e}^{T}{A}^{2k}e. To calculate the distribution, we implemented this recursive method of calculation in the MATLAB^{®} environment. Whilst this result is exact, and will be referred to as such henceforth, it can only be calculated numerically via recursion and cannot be given in the form of a closed expression.
2.3 Simulations of Neuronal Avalanches
In order to check the validity of our method, we performed simulations of the firing neurones using the Gillespie algorithm [38]. Due to the network being fully connected the algorithm is relatively straightforward:

As earlier, let A be the number of active neurones in the network (Q the number of quiescent). Given that an individual neurone becomes quiescent at rate α then the total rate of (Active → Quiescent) transitions is given by {r}_{aq}=A\alpha. Similarly, the total rate of (Quiescent → Active) transitions is given by {r}_{qa}=f({s}_{i})Q=f({s}_{i})(NA).

Let r={r}_{aq}+{r}_{qa} and generate a timestep dt from an exponential distribution of rate r.

Generate a random number n between 0 and 1. If n<\frac{{r}_{aq}}{r} an active neurone turns quiescent, otherwise a quiescent neurone is activated (fires). This event is said to occur at time t+dt and the network is updated accordingly.
2.4 Exact Solution Compared to Simulation
Values of the threshold, {R}_{0}, were chosen less than, equal to and finally above 1. We will refer to these regimes as subcritical, critical, and supercritical, respectively. Figure 3 illustrates the, as expected, good agreement between the simulations and the exact result for the three different regimes of {R}_{0}.
2.5 Comparing the Exact Solution to a Closed Form Approximation
In [36], Kessler proposed a closed solution to the analogous susceptibleinfectedsusceptible (SIS) problem where he was interested in the number of infections (including reinfections) occurring over the course of an epidemic. For small avalanche sizes where the number of infectives is negligible compared to the network size, the transition probabilities can be approximated as
In the critical regime {R}_{0}=1, the problem reduces to calculating the distribution of first passage times of a random walk with equal transition probabilities. Thus, for avalanche sizes in the range, 1\ll n\ll \sqrt{N}, Kessler [36] gave the following distribution based on Stirling’s approximation:
We note however that the range over which the distribution can be shown to be a power law is rather limited and for small networks will not hold. Using the theory of random walks and a Fokker–Planck approximation, Kessler also derived the following closed solution to the probability distribution of infections in the critical regime ({R}_{0}=1) for larger sizes:
Figure 4 plots this approximation against our exact solution for a network of size N=800. To more formally assess the convergence of the approximate solution to that of our exact solution, we considered the probabilities of avalanches from size N/10 to 20N and measured the difference between the distributions using two different metrics. Letting {P}_{e}(n) be the exact probability of an avalanche of size n and {P}_{k}(n) be the Kessler approximation to this, we first considered the standard mean square error given by
Secondly, we considered a more stringent measure of the error by looking at the supremum of difference between the same range of avalanches
Figure 5 illustrates the two errors for increasing network size and both show how the proposed closed solution is indeed converging to that of the exact.
3 ScaleFree Behaviour in the {R}_{0}=1 Regime
Whilst Eq. (1) gives a power law, this equation only holds over a limited range. Equation (2), in turn, is neither a power law nor a truncated power law. Here, we assess the range over which the distribution of sizes can be said to exhibit scalefree behaviour. For a rigorous assessment of this range, we employ a subset of the model selection approach described by Clauset and colleagues [24]. Specifically, we consider 100\text{,}000 of the simulated avalanches described earlier, and fit a truncated power law distribution of the form P(x)=C{x}^{\alpha} to avalanches up to size {x}_{max}=\frac{9}{10}N (the choice of this upper bound will be justified in the following section) by using the maximum likelihood method (here C is a normalising constant to keep the sum of the distribution between [{x}_{min},{x}_{max}] equal to 1). We do this by finding values of α and {x}_{min} that maximise the probability of obtaining our simulated avalanches given the fitted distribution. Next, we randomly generate 1\text{,}000 data sets from the fitted distribution and compute the difference between these synthetic data sets and the fitted form (using the Kolmogorov–Smirnov statistic). Similarly, we compute the difference between our simulations and the fitted power law. The pvalue is then calculated as the proportion of synthetic data sets that are further away from the theoretical distributions than our simulations. As per [24], the hypothesis (that the data comes from a power law) is rejected if the pvalue is less than 0.1. Note that in the model selection approach, should the hypothesis not be rejected, then one should test alternative models and use an information criterion to identify the best model. However, our focus here is purely on assessing whether our distribution can be said to behave like a power law distribution (we know it is not actually a power law) and therefore alternative models were not tested. With 100\text{,}000 avalanches, we obtained a pvalue of 0.382 leading us not to reject the hypothesis that the distribution was power law (see Fig. 6). Since the distribution is not a power law, we would expect that upon considering a larger number of avalanches, this hypothesis should be rejected [23]. Indeed, using data from 1\text{,}000\text{,}000 avalanches yielded a pvalue of 0, i.e., the truncated power law is not an appropriate model for the distribution.
The fact that the truncated power law was a plausible fit for the fewer number of avalanches (note that 100\text{,}000 is of the same order of magnitude as the number of avalanches typically reported in in vitro or in vivo studies of neuronal avalanches) is indicative of the partial scalefree behaviour the model exhibits. The appeal of the concept of critical brain is that the critical regime is the one in which longrange correlations keep the system poised between too highly correlated states of no behavioural value and too weakly correlated states that prevent information flow [39]. Thus, the actual nature of the distribution of the avalanche size matters less than any indication of the presence of long range correlations. In other words, neuronal avalanches need not precisely follow a power law, they just need to exhibit similar behaviour. It is important to appreciate this distinction. As the exact solution to the distribution of avalanche sizes is known, we can then compare it visually with a fit of a truncated power law over avalanche sizes from \frac{1}{10}N to \frac{9}{10}N. This is done in Fig. 7, which confirms that over a limited range of sizes the distribution is well approximated by a power law.
4 Origin of the Distribution’s Truncation
The fact that we have an exact form for the distribution allows us to make further important observations about some of its characteristics. Here, we explore the origin of the distribution’s truncation. Let {\lambda}_{1},{\lambda}_{2},\dots ,{\lambda}_{N} be the eigenvalues of A with the corresponding eigenvectors {u}_{1},{u}_{2},\dots ,{u}_{N}. The initial condition can then be given as \mathbf{p}(0)={c}_{1}{u}_{1}+{c}_{2}{u}_{2}+\cdots +{c}_{N}{u}_{N}. As the matrix A is similar to a symmetric tridiagonal matrix with real entries (consider the diagonal similarity transformation matrix D, with {D}_{1}=1 and {D}_{j}=\sqrt{({b}_{j}{b}_{j1}\cdots {b}_{2})/({c}_{j1}{c}_{j2}\cdots {c}_{1})}), we know that its eigenvalues are real.
Using the property \mathbf{A}{u}_{j}={\lambda}_{j}{u}_{j} we then obtain \mathbf{p}(k)={c}_{1}{\lambda}_{1}^{k}{u}_{1}+{c}_{2}{\lambda}_{2}^{k}{u}_{2}+\cdots +{c}_{N}{\lambda}_{N}^{k}{u}_{N}. This calculation leads to the probability of an avalanche being of size n being:
where {q}_{1} is the probability that the next transition is a recovery (from A to Q) given 1 active neurone (as defined earlier), {\lambda}_{i} are the eigenvalues of the transition matrix A and {d}_{i} are specified by the eigenvectors of the transition matrix and the initial conditions. We note that the earlier equation, \mathbf{p}(0)={c}_{1}{u}_{1}+{c}_{1}{u}_{1}+\cdots +{c}_{N}{u}_{N}, can be solved to obtain {c}_{i}. Using this, we can then calculate {d}_{i} as the first entry of the vector {c}_{i}{u}_{i}. Equation (3), which is exact, thus demonstrates that the distribution of avalanche sizes is a linear combination of exponentials.
The structure of A (namely the all zero diagonal) means that if u=({u}_{1},{u}_{2},\dots ,{u}_{N1},{u}_{N}) is an eigenvector with corresponding eigenvalue {\lambda}_{u}, then v=({u}_{1},{u}_{2},\dots ,{(1)}^{N}{u}_{N1},{(1)}^{N+1}{u}_{N}) is an independent eigenvector with corresponding eigenvalue {\lambda}_{u} (here, and in all that follows, we are assuming N is even; for N odd there is an additional zero eigenvalue). Setting \tilde{N}=\frac{N}{2} and {e}_{i}={d}_{i}+{d}_{Ni+1} allows us to rewrite Eq. (3) as
Assuming the lead eigenvalue is denoted by {\lambda}_{1}, then for all i, {\lambda}_{i}<{\lambda}_{1} and we have
Taking the limit as n increases, we find
Hence, P(n)\sim {q}_{1}{e}_{1}{\lambda}_{1}^{2n} and for larger avalanche sizes we have the leading eigenvalue dominating thus giving the exponential cutoff observed. We illustrate this convergence in Fig. 8 where we plot the exact avalanche distribution, P(n), against {q}_{1}{e}_{1}{\lambda}_{1}^{2n}. This figure also illustrates that the leading eigenvalue begins to dominate for avalanches just over the system size. It is for this reason that we chose an upper bound of \frac{9N}{10} when fitting a power law to the distribution of avalanche sizes in the previous section.
Such a distribution as (4) could converge to a true power law under two important conditions:

1.
the eigenvalues are well approximated by a geometric distribution, i.e., they are in the form {\lambda}_{i}=K{e}^{(\mu /2)i},

2.
the constants, {e}_{i}, are well approximated by {e}_{i}=L{i}^{q},
where K, μ, L, and q can be inferred via a numerical fitting procedure. In such a scenario, Eq. (4) can be rewritten to give
where C is a given constant. In the limit of an infinite network size, we then have
While P(n) can be found based on standard mathematical arguments, we have chosen to use results derived in the context of the Ztransform. The standard results for integer values of q give
where D is an operator such that D(f(z))=z\frac{d(f(z))}{dz}. For a fixed integer value of q, an approximation for P(n) can be obtained by simply applying the operator as many times as necessary and then substituting z={e}^{\mu n}. For q=1, for example, P(n)\propto \frac{{e}^{\mu n}}{{({e}^{\mu n}1)}^{2}} which for small values of μ is well approximated by \frac{1}{{\mu}^{2}}\frac{1}{{n}^{2}}.
These results only hold for integer values of q so an alternative approach is to approximate the sum for P(n) in terms of an integral. Taking into account the special form for the eigenvalues and constants, P(n) can be approximated as follows:
The latter integral can be interpreted as a Laplace transform of {x}^{q} , and thus yields
It is worth noting that this result is consistent with that obtained for integer values of q.
For a simple empirical verification of this conjecture, we determined the values of K, μ, L, and q in the above conditions through numerical fitting of the first 23 eigenvalues and e constants of the exact distribution for a network of size N=800 (see Fig. 9(a), (b)) and compared the resulting probability distribution with the exact distribution. Whilst the lesser valued eigenvalues and larger e values were not fitted well, Fig. 9(c) shows there is still remarkable agreement between both curves over a broad range of values, including the range [\frac{1}{10}N,\frac{9}{10}N] over which a power law like behaviour was established earlier (see Fig. 7). This result clearly illustrates the dominance of the larger eigenvalues and, given that the fitted distribution converges to a power law, gives support to the conjecture that the exact distribution would do so in the limit of an infinite network.
5 Other Markers of Criticality
Since the distribution of avalanche sizes in the finitesize critical system does not necessarily follow a true power law, the application of robust statistical testing in experimental conditions could well lead to rejecting the hypothesis that the data may come from a system operating in the critical regime. Therefore, in this section, we consider two experimentally testable markers of criticality: critical slowing down and divergence of susceptibility. We will define those concepts below but first we briefly summarise Van Kampen’s system size expansion [40], which we use to illustrate those markers on our system.
5.1 System Size Expansion
For generality, we now assume that each neurone receives a constant external input and that the activation function can take forms other than the simple identity function. We define the probability that the number of neurones active at time t is n as {P}_{n}(t). Then the master equation can be given as
The idea of the system size expansion is to now model the number of active neurones as the sum of a deterministic component scaled by N and a stochastic perturbation scaled by \sqrt{N}, i.e.,
A more detailed explanation of this can be found in [12] and [40], but importantly what is obtained is the following set of equations for μ (which is the solution to the mean field equation of the proportion of active neurones), \u3008\xi \u3009 (the expected value of the fluctuations) and {\sigma}^{2}=\u3008{\xi}^{2}\u3009{\u3008\xi \u3009}^{2} (the variance of the fluctuations)
Here, \stackrel{\u02c6}{f}=f(w\mu +h) and {\stackrel{\u02c6}{f}}^{\prime}={f}^{\prime}(w\mu +h). These equations, in turn, give the following equations for the mean, A, and variance, {A}_{\sigma}, of the number of active neurones:
We make use of these equations in the following two sections.
5.1.1 Critical Slowing Down
In dynamical systems theory, a number of bifurcations, including the transcritical bifurcation in our system, involve the dominant eigenvalue characterising the rates of changes around the equilibrium crossing zero. As a consequence, the characteristic return time to the equilibrium following a perturbation increases when the threshold is approached [41]. This increases has led to the notion of critical slowing down as a marker of critical transitions [42]. Here, we illustrate the critical slowing down of our model with the analytic derivation of the rate of convergence to the steady state (this derivation has been previously shown by [43]). We first begin by calculating the analytic solution to Eq. (10) for the percentage of active neurones. We again consider the case where f is the identity function and can thus write
Assuming zero external input (h=0), we have
We are interested in the solution of this equation and consider the result for different values of α. Firstly, we consider \alpha \ne w. In this case, we have
Integrating this using separation of variables and the initial condition \mu (0)={\mu}_{0}, we find
The solution to this depends on whether \alpha <w or \alpha >w ({R}_{0}>1 and {R}_{0}<1, respectively). If \alpha <w, then as t\to \mathrm{\infty}, \mu \to \frac{w\alpha}{w}. If \alpha >w then as t\to \mathrm{\infty}, \mu \to 0. Note that in both cases, convergence of the number of active neurones to the steady state solution is exponential.
Now we consider the solution when \alpha =w, i.e., the critical regime
Thus, as t\to \mathrm{\infty} we find \mu (t)\to 0. However, unlike for {R}_{0}\ne 1, convergence to the steady state exhibits a power law dependence on time [43].
5.1.2 Divergence of Susceptibility
A correlate of the phenomenon of critical slowing down is that of the divergence of susceptibility of the system as the system approaches the bifurcation [42]. In this section, we investigate the behaviour of the equation for the variance. For simplicity, we consider again the case of the identity activation function and a nondriven system h=0. First, we use Eq. (12) to calculate the variance in the percentage of active neurones:
Setting this equal to zero and rearranging, we obtain
Here, we note that unlike the equation for μ where there was only the single bifurcation parameter {R}_{0}, we now have the additional dependence on w. To maintain consistency with earlier results, we now set w=1 to obtain
Using Eq. (14), we obtain
Figure 10 illustrates the jump to a nonzero steady state when the critical value {R}_{0}=1 is approached from below, and the divergence in variance when it is approached from above.
Here, it should be noted that any finitesize network has a zero absorbing state so that eventually all activity will die out irrespective of the value of {R}_{0}. However, it has been shown that the ODE limit is a valid approximation to the solution of the master equation for reasonably sized systems with values of {R}_{0} greater than 1 and only over a finite time horizon (see [44] for further discussion). Defining the true (i.e., calculated directly from the master equation for P(n)) expected value of active neurones at time t as \tilde{A}(t), the convergence of the ODE approximation for A(t) given by Eq. (13) is such that for any t\ge 0{lim}_{N\to \mathrm{\infty}}A(t)\tilde{A}(t)=0[45].
6 Discussion
Over the last decade or so, the search for evidence that the brain may be a critical system has been the focus of much research. This is because it is thought that a critical brain would benefit from maximised dynamic range of processing, fidelity of information transmission and information capacity [46]. Whilst support for the critical brain hypothesis has emerged from comparing brain dynamics at various scales with the dynamics of physical systems at criticality (e.g., [31, 34, 47–50]), in this paper, we focus on the important body of work that has relied on characterising power laws in the distributions of size of neuronal avalanches [8, 30]. Our focus on this scale is motivated by empirical considerations regarding how one can go about demonstrating the above functional properties. Shew and Plenz [46] remark that any research strategy to test whether these properties are optimal near criticality will have to achieve two criteria: a means of altering the overall balance of interactions between neurones and a means of assessing how close to criticality the cortex is operating. As argued by these authors, the study of neuronal avalanches offers the greatest likelihood of achieving those two criteria.
The importance of a robust assessment of the statistical properties of the avalanche size is therefore twofold: on the one hand, it is about ascertaining the extent to which the system being studied has the statistical properties expected of a system operating at, or near, criticality; on the other hand, it is about being able to confirm that a manipulation/perturbation of the system aimed to push the system away from this critical regime has been effective. This consideration therefore puts a lot of importance on the description of the statistics one should expect in such a system. In the current literature, the assumption of the distribution of avalanche sizes taking a power law functional form relies on an analogy between the propagation of spikes in a neuronal network and models of percolation dynamics or branching processes for which exact power laws have been demonstrated in the limit of system size. As a result of the importance of having a robust assessment of the expected presence of a power law, greater emphasis has recently been put on using a sound statistical testing framework, e.g., [24]. Whilst we are unaware of any study in which the criticality hypothesis was rejected due to failure of rigorous statistical testing (which we suspect is due to the necessarily small number of observations, as we will argue below), there is clear evidence that many authors are now using the methods of Clauset et al. [24] to confirm the criticality of their experimental findings, e.g., [12, 23, 29]. As a result, we feel that it is all the more important to confirm that the assumed power law functional form is indeed a sensible representation of what one should expect in in vivo and in vitro recordings, which, unlike the physical systems considered when deriving the power law statistics, are finitesize systems. The aim of the paper was therefore to consider a model of neuronal dynamics that would be simple enough to allow the derivation of analytical or semianalytical results whilst (i) giving us a handle on the parameter controlling the fundamental principle thought to underlie criticality in the brain, namely, the balancing between processes that enhance and suppress activity (note that we are intentionally not referring to excitation and/or inhibition—we will return to this below) and (ii) allowing us to determine its distribution of avalanche sizes when operating in the critical regime. Note that because we are using a finitesize system, we are appealing to a normal form of standard bifurcation, here, a transcritical bifurcation, because it embodies all that needs to be known about the ‘critical’ transition (Sornette, private communication).
Our semianalytic derivation of the true distribution of avalanche sizes in a finitesize system suggests that, even though it is approximately scale free over a limited range, the distribution is not a true power law. First, this has important implications for the interpretation of results from a robust statistical assessment of the distribution. Indeed, as has been discussed by Klaus and Plenz [23], with a large number of samples, any distribution that deviates from the expected distribution by more than noise due to sampling, will eventually yield a pvalue such that the power law hypothesis will be rejected, thus leading to the potentially incorrect conclusion that the system is not critical. This is the case in our scenario where using 10^{6} avalanches lead to a rejection of the criticality hypothesis even though the system is tuned to the critical regime. In contrast, with 10^{5} avalanches (which is consistent with empirical observations), a pvalue above threshold leads to not rejecting the hypothesis that the distribution is a power law even though we established it is not one.^{Footnote 1} This finding therefore provides an important counterpart to the analytical results of Touboul and colleagues [29] who showed that thresholded stochastic processes could generically yield apparent power laws that only stringent statistical testing will reject. Whilst the stringent testing will reject the hypothesis of criticality for a system that is not necessarily critical, it may also reject the hypothesis of criticality for a system that is critical only because the actual distribution is not actually a power law. This ambiguity of the avalanche distribution in the finitesize system therefore requires that one should carefully consider to what fundamental property the idea of a critical brain actually appeals to. We suggest that the key appeal is that the brain can exhibit longrange correlations between neurones without it ever experiencing an over saturation of activity or long periods of inactivity. It then follows that the importance is not in the exact distribution obtained but in the approximately scalefree behaviour it exhibits. In turn, this highlights the importance of looking at other markers of criticality (which we will discuss below).
Another important result of this work is to provide the beginning of a mechanistic explanation for an often alluded to (e.g., [51]) but never properly treated (as far as we are aware) observation that whereas avalanches in a critical system with reentrant connections could in principle be arbitrarily long, and certainly, exceeding the number of recording sites, neuronal avalanches in in vitro or in vivo systems (and many computational models of selforganised criticality) often show a cutoff at the number of sites. Our work suggests that the lead eigenvalue of the transition matrix between states fully determine the location of this cutoff, which turns out indeed to be at about the system size, even if avalanches of up to 20 times the system size can be observed. This finding therefore provides some justification for setting, or accepting, a bound within which to apply a Clausettype methodology (we note that various reports use different ranges, e.g., 80 % of system size in [17], roughly system size in [51]). It is worth remembering that the number of recording sites can have profound implications on the nature of the distribution observed [21].
In addition to providing results on the distribution of avalanche sizes, we also sought to explore other potential markers of criticality. We provided results on two other markers of criticality—critical slowing down and divergence of susceptibility—both of which again follow from a dynamical systems appreciation of a critical bifurcation, i.e., the behaviour of a system whose lead eigenvalue crosses zero. The appeal of those markers, which have been documented in many other natural processes, e.g., [42, 52], but seldom at the mesoscopic brain level^{Footnote 2} (see [53] for a rare example) is that (a) they strengthen the assessment of the system being critical and (b) may contribute to achieving the second criterion of Shew and Plenz [46]. Although the authors are not in a position to provide explicit recommendations for an experimental design, we believe that these markers are amenable to robust experimentation, e.g., through pharmacological manipulation.
Whilst we hope we have convinced the reader of the potential importance of these findings, we also need to recognise that the very simplicity that makes analytical work possible does also raise questions regarding how physiologically plausible such a model is and, therefore, whether its conclusions should be expected to hold. Below, we address a few of the points worthy of further consideration.
6.1 Validity of Inferring Criticality in a Finite Network
In using the meanfield equations, it is important to understand how well they capture the behaviour and bifurcation structure of the stochastic process they are approximating. Whilst it is known that on the complete graph (see [54] for instance) and in the limit N\to \mathrm{\infty} the steady state solution of the ODE will converge to the expected value of the comparable stochastic process, it is unclear whether the critical point of the infinite system corresponds to that in the finite system. Furthermore, it is unclear whether a finite system can truly have a critical point and we must be cautious in claiming one exists. Importantly, however, it has been shown in [55] that for a complete graph, {R}_{0}\approx 1 (the paper proves the result for α fixed as 1 but the result is generalisable for any α) is the threshold below which the disease will die out quickly (expected time to extinction O(log(n))), and above which it dies out slowly (expected time to extinction O({n}^{a}) for some a). Simulating the steady state of the network for increasing {R}_{0} also shows (see Fig. 11) the characteristic feature of a secondorder phase transition found at a critical point. For these reasons, whilst acknowledging the problem of inferring criticality in a finite regime, we feel justified in claiming {R}_{0}=1 as the critical point for the process unfolding on our finite network.
6.2 Validity of a Purely Excitatory Network
In this paper, we have used a purely excitatory neuronal model. This not only simplifies the system but is also an important characteristic of the brain during early development. Experimental results have shown that during early development, before birth, GABAergic neurones (i.e., neurones which will later be inhibitory) have a depolarising effect on their postsynaptic neighbours [56–58]. Thus, our model might be considered as representative of early development. Power law statistics have been observed in early development at a time when networks are thought to be purely excitatory [32, 59]. It should be noted that this approach has the benefit of casting a new light on the question of what is the minimum requirement for a neuronal system to show criticality. To a large extent, the current literature has been focused on a form of homeostasis resulting from either a fine balance between excitation and inhibition, e.g., [12, 13] or some relatively complex dynamical processes at synaptic level, e.g., [17]. Our results show that a purely excitatory system can show the exact same behaviour such that on average each active neurone only activates one postsynaptic neurone. Here, this balanced state is achieved through a tradeoff between the rates at which neurones become active and quiescent. It should be noted that this formulation of the problem leads to interesting parallels with classical models of mathematical epidemiology, which the authors intend to continue exploring.
6.3 Spatial Structure
To make use of the analytic tractability of the mean field equation it was necessary to consider a fully connected network. While this is not true of the whole brain, it may be closer to the reality of the kind of in vitro systems typically considered in studies of neuronal avalanches. For example, Hellwig et al. [60] report up to 80 % connection probability in local connectivity between pyramidal neurones in layers 2/3 of the rat visual cortex. Extending the work presented here to consider the effect of network topology on the system’s dynamics and the resulting distribution of event sizes would be of particular interest from a developmental viewpoint (see, for instance, Larremore et al. [61], who have considered the avalanche distribution of general treelike networks with discrete dynamics). As networks mature, there is not only a switch to inhibition by a proportion of the neurones (the socalled GABA switch), but also a subsequent pruning of synaptic connections [62]. The level of pruning is high, with a 40 % reduction in the number of synaptic connections between early childhood and adulthood [62]. Thus, a developing network may be more readily approximated by a fully connected network than an adult neural network would be.
The lack of a spatial embedding of our model is in contrast with many classical models of criticality, and also with physiological systems. Accordingly, our model cannot display another important marker of criticality, namely, the divergence of correlation lengths in space. A spatial embedding is not needed for our system to be critical and to exhibit a distribution of avalanche size similar to that observed in physiological neuronal avalanches. It therefore begs the question of the exact role of spatial embedding in the dynamics of neuronal avalanches. It may well be that, just like balanced activity in our model comes about from a tradeoff between excitation and refractoriness rather than between excitation and inhibition, specific spatial embeddings may enable balanced activity without the need for plastic mechanisms. Kaiser and Hilgetag [63] showed that hierarchical modular networks can lead to limited sustained activity whereby the activity of neural populations in the network persists between the extremes of either quickly dying out or activating the whole network. Roxin and colleagues [64] observed selfsustained activity in excitable integrateandfire neurones in a smallworld network, whose dynamics depends sensitively on the propagation velocity of the excitation.
6.4 Nondriven Case
Finally, in this paper, we have focused on the nondriven case h=0. Whilst this constraint allowed the derivation of analytical results, it obviously contrasts with the reality of a physiological system unless one considers that any ‘external’ input operates at such a slower timescale that one could assume separation of time scales (an important assumption in the selforganised criticality framework). However, the fact that binning is required for identifying avalanches in physiological recordings suggests that this separation of time scales is unlikely. Whilst the introduction of a nonzero h in our model does not affect the results obtained using finite size expansion, it does effectively make it impossible for the system to operate at {R}_{0}=1. A thorough investigation of the driven case (h>0) will be the subject of the companion paper.
Notes
As the power law is not a sufficient condition of criticality, one should not infer from this that the system is indeed critical, however, this step is commonly taken in published reports and that is worth mentioning here.
Strictly speaking the notion of critical slowing in neurones firing near firing threshold appeals to the same notion.
References
Beggs JM, Plenz D: Neuronal avalanches in neocortical circuits. J Neurosci 2003, 23(35):11167–11177.
Beggs JM, Plenz D: Neuronal avalanches are diverse and precise activity patterns that are stable for many hours in cortical slice cultures. J Neurosci 2004, 24(22):5216–5229. 10.1523/JNEUROSCI.054004.2004
Petermann T, Thiagarajan TC, Lebedev MA, Nicolelis MAL, Chialvo DR, Plenz D: Spontaneous cortical activity in awake monkeys composed of neuronal avalanches. Proc Natl Acad Sci USA 2009, 106(37):15921–15926. 10.1073/pnas.0904089106
Hahn G, Petermann T, Havenith MN, Yu S, Singer W, Plenz D, Nikolic D: Neuronal avalanches in spontaneous activity in vivo. J Neurophysiol 2010, 104(6):3312–3322. 10.1152/jn.00953.2009
Chialvo DR: Emergent complex neural dynamics. Nat Phys 2010, 6(10):744–750. 10.1038/nphys1803
Sethna JP, Dahmen KA, Myers CR: Crackling noise. Nature 2001, 410(6825):242–250. 10.1038/35065675
Kinouchi O, Copelli M: Optimal dynamical range of excitable networks at criticality. Nat Phys 2006, 2(5):348–352. 10.1038/nphys289
Shew WL, Yang H, Petermann T, Roy R, Plenz D: Neuronal avalanches imply maximum dynamic range in cortical networks at criticality. J Neurosci 2009, 29(49):15595–15600. 10.1523/JNEUROSCI.386409.2009
Buckley CL, Nowotny T: Multiscale model of an inhibitory network shows optimal properties near bifurcation. Phys Rev Lett 2011., 106(23): Article ID 238109 Article ID 238109
Larremore DB, Shew WL, Restrepo JG: Predicting criticality and dynamic range in complex networks: effects of topology. Phys Rev Lett 2011., 106: Article ID 058101 Article ID 058101
LinkenkaerHansen K, Nikouline VV, Palva JM, Ilmoniemi RJ: Longrange temporal correlations and scaling behavior in human brain oscillations. J Neurosci 2001, 21(4):1370–1377.
Benayoun M, Cowan JD, van Drongelen W, Wallace E: Avalanches in a stochastic model of spiking neurons. PLoS Comput Biol 2010., 6(7): Article ID e1000846 Article ID e1000846
Magnasco MO, Piro O, Cecchi GA: Selftuned critical antiHebbian networks. Phys Rev Lett 2009., 102(25): Article ID 258102 Article ID 258102
Meisel C, Storch A, HallmeyerElgner S, Bullmore E, Gross T: Failure of adaptive selforganized criticality during epileptic seizure attacks. PLoS Comput Biol 2012., 8: Article ID e1002312 Article ID e1002312
Milton JG: Neuronal avalanches, epileptic quakes and other transient forms of neurodynamics. Eur J Neurosci 2012, 36(2):2156–2163. 10.1111/j.14609568.2012.08102.x
Bak P, Tang C, Wiesenfeld K: Selforganized criticality: an explanation of the 1/ f noise. Phys Rev Lett 1987, 59(4):381–384. 10.1103/PhysRevLett.59.381
Levina A, Herrmann JM, Geisel T: Dynamical synapses causing selforganized criticality in neural networks. Nat Phys 2007, 3: 857–860. 10.1038/nphys758
Essam JW: Percolation theory. Rep Prog Phys 1980, 43: 833–912. 10.1088/00344885/43/7/001
Beggs JM: Neuronal avalanche. Scholarpedia 2006., 2: Article ID 1344 Article ID 1344
Harris TE: The Theory of Branching Processes. Springer, Berlin; 1963.
Priesemann V, Munk MHJ, Wibral M: Subsampling effects in neuronal avalanche distributions recorded in vivo. BMC Neurosci 2009., 10: Article ID 40 Article ID 40
Deco G, Senden M, Jirsa V: How anatomy shapes dynamics: a semianalytical study of the brain at rest by a simple spin model. Front Comput Neurosci 2012., 6: Article ID 68 Article ID 68
Klaus A, Yu S, Plenz D: Statistical analyses support power law distributions found in neuronal avalanches. PLoS ONE 2011., 6(5): Article ID e19779 Article ID e19779
Clauset A, Shalizi CR, Newman MEJ: Powerlaw distributions in empirical data. SIAM Rev 2009, 51(4):661–703. 10.1137/070710111
Ziff RM: Correctiontoscaling exponent for twodimensional percolation. Phys Rev E 2011., 83: Article ID 020107 Article ID 020107
Borgs C, Chayes JT, Kesten H, Spencer J: The birth of the infinite cluster: finitesize scaling in percolation. Commun Math Phys 2001, 224: 153–204. 10.1007/s002200100521
Sornette D: Critical Phenomena in Natural Sciences. 2nd edition. Springer, Berlin; 2006.
Jensen H: Selforganized Criticality: Emergent Complex Behavior in Physical and Biological Systems. Cambridge University Press, Cambridge; 1998.
Touboul J, Destexhe A: Can powerlaw scaling and neuronal avalanches arise from stochastic dynamics? PLoS ONE 2010., 5(2): Article ID e8982 Article ID e8982
Beggs JM, Plenz D: Neuronal avalanches in neocortical circuits. J Neurosci 2003, 23(35):11167–11177.
LinkenkaerHansen K, Nikouline VV, Palva JM, Ilmoniemi RJ: Longrange temporal correlations and scaling behavior in human brain oscillations. J Neurosci 2001, 21(4):1370–1377.
Hartley C, Berthouze L, Mathieson SR, Boylan GB, Rennie JM, Marlow N, Farmer SF: Longrange temporal correlations in the EEG bursts of human preterm babies. PLoS ONE 2012., 7(2): Article ID e31543 Article ID e31543
Segev R, Benveniste M, Hulata E, Cohen N, Palevski A, Kapon E, Shapira Y, BenJacob E: Long term behavior of lithographically prepared in vitro neuronal networks. Phys Rev Lett 2002., 88(11): Article ID 118102 Article ID 118102
Poil SS, Hardstone R, Mansvelder HD, LinkenkaerHansen K: Criticalstate dynamics of avalanches and oscillations jointly emerge from balanced excitation/inhibition in neuronal networks. J Neurosci 2012, 32(29):9817–9823. 10.1523/JNEUROSCI.599011.2012
Droste F, Do AL, Gross T: Analytical investigation of selforganized criticality in neural networks. J R Soc Interface 2012., 10(78): Article ID 20120558 Article ID 20120558
Kessler DA: Epidemic size in the SIS model of endemic infections. J Appl Probab 2008, 45(3):757–778. 10.1239/jap/1222441828
Allen LJ: Introduction to stochastic epidemic models. Lecture Notes in Mathematics 1945. In Mathematical Epidemiology. Springer, Berlin; 2008:81–130.
Gillespie DT: Exact stochastic simulation of coupled chemical reactions. J Phys Chem 1977, 81(25):2340–2361. 10.1021/j100540a008
Chialvo R: Critical brain networks. Physica A 2004, 340: 756–765. 10.1016/j.physa.2004.05.064
Van Kampen NG: Stochastic Processes in Physics and Chemistry. 3rd edition. NorthHolland, Amsterdam; 2007.
Wissel C: A universal law of the characteristic return time near thresholds. Oecologia 1984, 65: 101–107. 10.1007/BF00384470
Scheffer M, Bascompte J, Brock WA, Brovkin V, Carpenter SR, Dakos V, Held H, van Nes EH, Rietkerk M, Sugihara G: Earlywarning signals for critical transitions. Nature 2009, 461(7260):53–59. 10.1038/nature08227
Stollenwerk N, Jansen VA: Criticality in epidemiology. World Scientific Lecture Notes in Complex Systems 7. In Complex Population Dynamics: Nonlinear Modelling in Ecology, Epidemiology and Genetics. Edited by: Blasius B, Stone L, Kurths J. World Scientific, Singapore; 2007:159–188.
Nåsell I: The quasistationary distribution of the closed endemic SIS model. Adv Appl Probab 1996, 28(3):895–932. 10.2307/1428186
Simon P, Taylor M, Kiss I: Exact epidemic models on graphs using graphautomorphism driven lumping. J Math Biol 2011, 62(4):479–508. 10.1007/s002850100344x
Shew WL, Plenz D: The functional benefits of criticality in the cortex. Neuroscientist 2012, 19(1):88–100.
Plenz D, Chialvo DR: Scaling properties of neuronal avalanches are consistent with critical dynamics. [arXiv:0912.5369] Plenz D, Chialvo DR: Scaling properties of neuronal avalanches are consistent with critical dynamics. [arXiv:0912.5369]
Expert P, Lambiotte R, Chialvo DR, Christensen K, Jensen HJJ, Sharp DJ, Turkheimer F: Selfsimilar correlation function in brain restingstate functional magnetic resonance imaging. J R Soc Interface 2011, 8(57):472–479. 10.1098/rsif.2010.0416
Friedman N, Ito S, Brinkman BAW, Shimono M, DeVille REL, Dahmen KA, Beggs JM, Butler TC: Universal critical dynamics in high resolution neuronal avalanche data. Phys Rev Lett 2012., 108: Article ID 208102 Article ID 208102
Ribeiro TL, Copelli M, Caixeta F, Belchior H, Chialvo DR, Nicolelis MAL, Ribeiro S: Spike avalanches exhibit universal dynamics across the sleepwake cycle. PLoS ONE 2010., 5(11): Article ID e14129 Article ID e14129
Rubinov M, Sporns O, Thivierge JP, Breakspear M: Neurobiologically realistic determinants of selforganized criticality in networks of spiking neurons. PLoS Comput Biol 2011., 7(6): Article ID e1002038 Article ID e1002038
Kelso JAS: Haken–Kelso–Bunz model. Scholarpedia 2008., 3(10): Article ID 1612 Article ID 1612
SteynRoss ML, SteynRoss DA, Sleigh JW, Whiting DR: Theoretical predictions for spatial covariance of the electroencephalographic signal during the anestheticinduced phase transition: increased correlation length and emergence of spatial selforganization. Phys Rev E 2003., 68(2 Pt 1): Article ID 021902 Article ID 021902
Simon PL, Kiss IZ: From exact stochastic to meanfield ODE models: a new approach to prove convergence results. IMA J Appl Math 2012. doi:10.1093/imamat/hxs001 doi:10.1093/imamat/hxs001
Ganesh A, Massoulie L, Towsley D: The effect of network topology on the spread of epidemics. INFOCOM 2005. Proceedings of the 24th Annual Joint Conference of the IEEE Computer and Communications Societies. Volume 2 2005, 1455–1466.
Cherubini E, Gaiarsa JL, BenAri Y: GABA: an excitatory transmitter in early postnatal life. Trends Neurosci 1991, 14(12):515–519. 10.1016/01662236(91)90003D
Rivera C, Voipio J, Payne JA, Ruusuvuori E, Lahtinen H, Lamsa K, Pirvola U, Saarma M, Kaila K: The K+/Cl− cotransporter KCC2 renders GABA hyperpolarizing during neuronal maturation. Nature 1999, 397(6716):251–255. 10.1038/16697
BenAri Y: Excitatory actions of GABA during development: the nature of the nurture. Nat Rev, Neurosci 2002, 3(9):728–739. 10.1038/nrn920
Gireesh ED, Plenz D: Neuronal avalanches organize as nested theta and beta/gammaoscillations during development of cortical layer 2/3. Proc Natl Acad Sci USA 2008, 105(21):7576–7581. 10.1073/pnas.0800537105
Hellwig B: A quantitative analysis of the local connectivity between pyramidal neurons in layers 2/3 of the rat visual cortex. Biol Cybern 2000, 82(2):111–121. 10.1007/PL00007964
Larremore DB, Carpenter MY, Ott E, Restrepo JG: Statistical properties of avalanches in networks. Phys Rev E 2012., 85: Article ID 066131 Article ID 066131
Huttenlocher PR, Dabholkar AS: Regional differences in synaptogenesis in human cerebral cortex. J Comp Neurol 1997, 387(2):167–178. 10.1002/(SICI)10969861(19971020)387:2<167::AIDCNE1>3.0.CO;2Z
Kaiser M, Hilgetag CC: Optimal hierarchical modular topologies for producing limited sustained activation of neural networks. Front Neuroinform 2010., 4: Article ID 8 Article ID 8
Roxin A, Riecke H, Solla S: Selfsustained activity in a smallworld network of excitable neurons. Phys Rev Lett 2004., 92(19): Article ID 198101 Article ID 198101
Acknowledgements
Timothy Taylor is funded by a PGR studentship from MRC, and the Departments of Informatics and Mathematics at University of Sussex. Caroline Hartley is funded through CoMPLEX (Centre for Mathematics and Physics in the Life Sciences and Experimental Biology), University College London. Istvan Z. Kiss acknowledges support from EPSRC (EP/H001085/1). Péter L. Simon acknowledges support from OTKA (grant no. 81403) and from the European Union and the European Social Fund (financial support to the project under the grant agreement no. TÁMOP4.2.1/B09/1/KMR).
Author information
Authors and Affiliations
Corresponding author
Additional information
Competing Interests
The authors declare that they have no competing interests.
Authors’ Contributions
TT carried out analysis and numerical simulations for tree approach, finite size expansion, critical slowing, comparison to Kessler’s approximate solution. TT and LB wrote the manuscript. CH carried out additional calculations and numerical simulations. PS and IK contributed the tree approach, and the derivation of the power law in the limit of the system size. LB conceived of the analysis and of the overall goals of the study, participated in the implementation and analysis of the different simulations. All authors read and approved the final manuscript.
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License ( https://creativecommons.org/licenses/by/2.0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Taylor, T.J., Hartley, C., Simon, P.L. et al. Identification of Criticality in Neuronal Avalanches: I. A Theoretical Investigation of the Nondriven Case. J. Math. Neurosc. 3, 5 (2013). https://doi.org/10.1186/2190856735
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/2190856735