Figures
Abstract
Irregularly timed action potentials, or spikes, are pervasively observed in the brain activity of awake mammals. However, the role of this temporal irregularity in neural computation is still not well understood. In canonical network models irregular spiking emerges via balanced, fluctuating input currents, leading to collective responses that track inputs linearly. How networks characterized by irregular spiking could support flexible nonlinear dynamics needed for general-purpose computation remains under ongoing debate. Here we characterize the dynamics of networks whose elementary unit is not a single neuron but a small group of neurons, with distinct tunings, that compete at each timestep via a winner-take-all (WTA) interaction. While WTA has long been proposed as an elementary functional motif in the brain and represents a powerful computational primitive, how large networks of such units behave has received less investigation. We show that these networks, like classic excitatory-inhibitory balanced networks, exhibit a chaotic fluctuation-driven regime characterized by sustained irregular activity resembling realistic cortical spiking, which we interpret as a multidimensional balance spread over several competing neural populations with different tunings. We develop a mean-field theory for the network, which shows how irregular spiking sustained by time-varying input fluctuations can support flexible nonlinear collective dynamics. Using the theory we predict and verify network regimes in which input fluctuations alone yield multistability, stable sequence generation, or complex heterogeneous firing rate dynamics—three core dynamical primitives thought to underlie memory-dependent neural computation—via consistent Poisson-like spiking produced through chaos. Thus, networks of WTA units support a chaotic fluctuation-driven regime characterized by irregular spiking that can power complex nonlinear collective dynamics. This represents a new model of brain activity capable of simultaneously reproducing realistic spike trains and diverse nonlinear firing rate patterns well posed for flexible computation, and which can be trained or fit to data.
Author summary
The irregular timing of action potentials, or “spikes,” is one of the most pervasive features of neural activity in awake mammals, observed in a wide range of brain areas and conditions. However, the role of this striking temporal irregularity in neural information processing is still not understood. In network models, it has been particularly challenging to reconcile irregular spiking with flexible nonlinear “firing rate” dynamics, which are thought to be crucial for general-purpose neural computation. Here we present a network model, based on winner-take-all units akin to small groups of neurons, in which irregular spiking and flexible nonlinear dynamics go hand-in-hand and can be related via a concise mathematical theory. The result is a new model of brain activity capable of simultaneously capturing realistic spiking and firing rate dynamics, which is well posed for flexible computation and can be trained or fit to data.
Citation: Pang R (2025) Balanced state of networks of winner-take-all units. PLoS Comput Biol 21(6): e1013081. https://doi.org/10.1371/journal.pcbi.1013081
Editor: Jonathan Rubin, University of Pittsburgh, UNITED STATES OF AMERICA
Received: December 11, 2024; Accepted: April 23, 2025; Published: June 11, 2025
Copyright: © 2025 Rich Pang. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All code was written in Python 3 and is available at https://github.com/rkp8000/balanced-wta. Demos of spiking network simulations producing multistability and sequences can be run on a CPU and require about 8 Gb of RAM.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Computation in the brain is thought to be orchestrated through the activity dynamics of populations of neurons, which respond to and transform inputs to process information [1–10]. Yet while this framework has provided transformative insights into neural computation, we still lack a comprehensive account of the role of the microscopic activity of the neurons involved, which in many brain areas and conditions manifests as sparse spike trains with high temporal irregularity and weak correlations [11–16]. This Poisson process-like spiking regime, termed the asynchronous irregular regime, has been most commonly observed in cerebral cortex of awake mammals, for example in primates performing working memory tasks [12, 16], and has motivated substantial theoretical work investigating its origin and function [11, 17–24]. Reconciling the asynchronous irregular regime with more complex, flexible collective dynamics in network models, however, has proved challenging [11]. Resolving this challenge will be crucial to understanding whether irregular spiking largely emerges from noisy biological processes, possibly co-opted for certain benefits (e.g. sampling input representations [25] or state transitions [26–28]), or if it plays a more central role in computation in general.
One core obstacle to resolving this problem is mathematical. At a cellular level irregular spiking is thought to arise via strongly fluctuating input currents [11, 13, 29, 30], which foundational theoretical work showed could arise in strongly coupled random networks in which a balance of excitatory and inhibitory inputs emerges dynamically through network interactions—the classic balanced state [17–19]. In this model excitatory-inhibitory balance is central to irregular activity as excess inhibition or excitation would lead to quiescence or regular (non-Poissonian) spiking, respectively. (Note that excitatory-inhibitory balance can exist in networks that do not produce irregular spikes [31–33], and irregular spiking can emerge without balance, e.g. when driving neurons with noise; however, most spiking models of cortex assume that balance and irregular activity are tightly linked [11].) Strikingly, despite the spiking nonlinearity in the classic balanced state, collective responses—population-averaged firing rates—are linear in their inputs, converging to the form of in the limit of large numbers of synaptic inputs, where r are the population firing rates, u is external input, and A is a matrix [11, 17–19]. This emergent linearity substantially limits the model’s dynamical and computational abilities, as well as its ability to reproduce slow, complex, nonlinear dynamics observed in cortex [34–38]. Much work has thus sought to extend the classic balanced state to nonlinear or slow computation, for example via finite-size theories [11, 24, 39], clustered connectivity [26, 27], “semi-balance” from excess inhibition [23], or additional synaptic nonlinearities [40]. Nonetheless, how to construct robust network models in which flexible nonlinear dynamics co-exist with or potentially exploit irregular spiking remains a topic of ongoing debate [11, 23, 24, 41–43].
Here we present an alternative to the classic balanced state that reveals how irregular activity can not only coexist with but directly support nonlinear firing rate dynamics. At the heart of the model is not a neuronal thresholding nonlinearity—by far the most common single-neuron nonlinearity used in spiking network models [17, 19, 42, 44])—but instead a multi-neuron winner-take-all (WTA) nonlinearity. Specifically, we consider large networks of elementary units each corresponding to a small group of neurons with distinct tunings (external stimulus preferences), where within each unit the neurons compete with each other to activate at each time step. WTA has long been proposed as an elementary functional motif in the brain [45, 46] and represents a powerful computational primitive [47], and WTA-like operations can be implemented quickly and efficiently in groups of model spiking neurons [48]. WTA operations also resemble the well-known computation of divisive normalization, in which neural activities are scaled by the summed activity of neurons with competing stimulus preferences [49]. Yet while the emergence of WTA competition or divisive normalization and their influence on sensory encoding have been studied extensively [48–55], how a multi-neuron WTA nonlinearity shapes global dynamics when repeated as a functional motif throughout a large network has received less attention. While fixed points of networks of WTA units have been examined [56], along with computational abilities of WTA units in specialized networks [57–59], the existence and properties of a more general and biologically realistic fluctuation-driven regime are not known.
Here we show that such a regime exists in networks of tuned WTA units. We find that in the high-gain or hard-WTA limit activity resembles realistic Poisson-like irregular spiking, and moreover supports flexible nonlinear firing rate dynamics. We refer to this dynamical regime as the balanced state of the network, in analogy with the classic excitatory-inhibitory balanced state [11, 17–19]. In the network of WTA units, however, balance exists not between excitatory and inhibitory populations but among multiple populations of neurons with different tunings—hence the balance is “multi-dimensional.” Individual WTA units receive effectively random, yet statistically balanced inputs from all active populations; but within a unit only one neuron can spike at each time step, resulting in Poisson-like activation patterns. Crucially, the multi-dimensional nature of this balance allows flexible, nonlinear collective dynamics to unfold simultaneous with and sustained by consistent Poisson-like spiking activity. Thus, networks of WTA units can support fluctuation-driven irregular activity capable of powering flexible nonlinear collective dynamics.
Model
To approach this problem theoretically we consider an idealized network of N elementary “units,” each corresponding to a group of D neurons. Within each unit we let each neuron be “tuned” to a discrete external input , with each unit containing exactly one neuron per tuning; hence we can label any neuron in any unit by its tuning
. For instance, a tuning d might represent a neuron’s preferred stimulus orientation [60, 61]. For simplicity we assume there is no overlap in the tunings, i.e. each neuron is tuned to exactly one of the D external inputs.
Terminology
Each group of D neurons we refer to as a unit. All N neurons with the same tuning d we refer to as a co-tuned population, or simply population. Thus, in total the network contains neurons, and there are D different tunings or populations.
Note also that we will refer to the high-gain limit of the network, in which each neuron is either “on” or “off,” as spiking. This contrasts with more common definitions of spiking networks, which usually connote leaky integrate-and-fire neurons that operate in continuous time and integrate their inputs over a finite timescale (the membrane time constant), emitting spikes when the integrated input crosses a spiking threshold [19]. For simplicity the network we study operates in discrete time, but intuitively one can think of each “on” event as corresponding to a spike or action potential, and the timestep of the update rule as being on the order of a typical neuronal membrane time constant (10–20 ms [62]).
Our study also involves two classes of dynamical systems, which we will frequently compare. The first class (Eq 1–4) describes the “microscopic” activity dynamics of each neuron in each WTA unit, which we will typically refer to as the full or full spiking (in the high-gain limit) dynamics. The second class (Eq 5–10) directly describes the dynamics of the “macroscopic” firing rates of the D populations, where the firing rate of a population is the average spiking activity of the N neurons in that population at time t—these dynamics are derived as a mean-field description of the full spiking dynamics, and which we will refer to as the mean-field dynamics for
.
Update rule
Let the activation state of unit i at time t be given by a vector , where
is the activity level of neuron d in the unit. Within each unit we let the D neurons “compete” with each other in a WTA-like fashion, such that
for all t. This also means that every unit is always active—the only thing that changes is which neuron(s) are active at each timestep. This is supported by neural population recordings in monkey prefrontal cortex and monkey and mouse visual cortex, where, e.g., the mean or total activity across the population remains close to constant across task epochs even though individual neurons may differ [38, 63].
We implement the WTA competition within each unit via a Softmax function. Given input , the activation of unit i at time t is given by
where
with the parameter g acting as a gain term. In the limit the Softmax function becomes a hard WTA nonlinearity, such that
becomes a 1-hot vector indicating which of the D neurons in unit i received the largest input at t.
Inputs to neurons arise through network interactions (Fig 1a). Specifically, the inputs to unit i depend on the previous network state through
(a) Network schematic. Each unit contains D neurons that compete at each timestep through a Softmax nonlinearity. Inputs to each neuron in each unit can come from any other neuron in any other unit in the network. (b) Example activity dynamics of two example units (color = tuning) for three different gain levels (top to bottom). Dashed gray line indicates activity level of 1. (c) Numerically estimated autocovariance functions of 4 units in a network with infinite gain (N=1000, D=5). (d) Critical gain (estimated by identifying when a near-infinitesimal perturbation switched from being contracted to being expanded by the network dynamics) vs D and N.
where is the weight from neuron
in unit j onto neuron d in unit i, and
is external input. In general, a single neuron can receive inputs from any neuron in any population. Equations 1–3 specify the complete update rule for the model. Unless stated otherwise, there is no noise—all variability in activity arises through deterministic network interactions.
Note that the total activity for all t, which in the
limit corresponds to exactly N neurons spiking at each timestep. Constant WTA-like activity, as a functional consequence of inhibitory feedback, was also proposed in an earlier “synfire chain” model of cortical dynamics [64–66], and is supported by neural recordings in which mean activity across the total recorded population is constant across task epochs, even under stimulation [38, 63]. This stability of total activity also means the network is spontaneously active even without external input.
Network weights
The network weights are given by a 4-tensor , whose elements we take to be i.i.d. (independent and identically distributed) Gaussian, except with tuning-dependent means and variances. Specifically, we let
where and
specify the means and variances of synaptic weights from neurons with tuning
onto neurons with tuning d. As we will see, the 1/N scaling of the variances promotes fluctuation-driven dynamics. Note also that we often refer to
, which is just the square root of
.
Results
Unstructured network
To study the baseline behavior of the network we first investigate the case of unstructured connectivity, in which and
are independent of pre- and post-synaptic tuning, and absent external input (
). For sufficiently low gain (e.g. g = 0.5), the network quickly evolves to a fixed point, reflected in the constant activity levels of individual units (Fig 1b, top). As g is increased, however, autonomous temporal fluctuations emerge (Fig 1b, middle), and in the
limit individual neurons produce sustained highly irregular switches between on and off states (Fig 1b, bottom). The temporal irregularity of single-neuron activity in the
limit is quantified in the autocovariance functions estimated from numerical simulations (Fig 1c). Estimating (numerically) whether microscopic perturbations (swapping the activity levels of two neurons within one unit) expand or contract suggests that the network exhibits a transition to chaos at a critical gain level that decreases with D, and as D grows becomes independent of N (Fig 1d; presumably for sufficiently large N). Note that strictly speaking for finite N the
limit is not chaotic, since the dynamics unfold deterministically on a discrete space so must eventually fall into a periodic orbit; however, they are effectively chaotic, as measured by sensitivity to small perturbations (S1 Fig), and we expect them to approximate true chaos more closely as
[22, 67–69]. Thus, similar to transitions to chaos in other random networks [1, 22, 70], the network of WTA units becomes chaotic above a critical gain, producing activity patterns suggestive of the irregular activity patterns observed in single neurons in the brain.
Spiking network representation
In the limit, the model admits a representation as a spiking neural network. At each time t each neuron is either on (
) or off (
), allowing us to treat the collective dynamics as population spike trains. A raster plot of a subset of neurons in such a network reveals an irregular spiking pattern (Fig 2a) strongly resembling the well-known asynchronous irregular state of cortical firing [11, 13–15], and visually similar to activity produced by classic balanced networks [17–19]. Numerical estimates reveal approximately exponential single-cell inter-spike interval (ISI) distributions (Fig 2b), a key signature of spikes emitted from a homogeneous Poisson process. A concise measure of how similar a spike train is to a Poisson process is the ISI coefficient of variation (ISI CV), or its non-stationary equivalent ISI CV2 [14, 16] (Methods), both of which are near unity for a homogeneous Poisson spiking process; ISI CV2 (but not ISI CV) will also be near unity for an inhomogeneous Poisson process. Indeed, the ISI CV2 distributions in the unstructured network are near unity (Fig 2c), recapitulating a key signature of irregular spiking in the brain [15, 16]. The network also produces a right-skewed distribution of time-averaged firing rates across each population (Fig 2d), resembling firing rate distributions observed in biological spike trains [71]. Finally, the mean firing rate decreases as 1/D, suggesting no lower limit on baseline firing rate (Fig 2e).
(a) Raster plot for one network simulation with D = 5, showing a few example neurons with each tuning (color = tuning). (b) Interspike interval (ISI) distribution for one example unit from each tuning. (c) Distribution of (non-stationary) inter-spike interval coefficient of variation ISI CV2 for each co-tuned population. (d) Distribution of time-averaged single-neuron firing rates (number of spikes divided by simulation length [3000]) for each co-tuned population (colored curves, D = 5). Gray curves show equivalent for a network with D = 12 different tunings (not colored). (e) Population firing rates vs D for a network with 1000 units. Firing rates are computed as averages across neurons within each co-tuned population. Solid line shows theory (1/D); scatter points and error bars show simulation results. Error bars are standard deviations over tunings and time.
Thus, in the high-gain limit the network produces activity equivalent to irregular spiking, recapitulating a fundamental dynamical feature of realistic neural activity, and with no apparent minimum firing rate. In analogy to the well-studied excitatory-inhibitory balanced state [11, 17–19], the multi-dimensional balance of inputs from different co-tuned populations unlocks a fluctuation-driven dynamical regime in which input fluctuations perpetuate themselves through irregular spiking.
Theory
To understand the capacity of this network to generate computationally beneficial collective dynamics we develop a mean-field theory, in the limit, for the time evolution of the population firing rates. To this end, although the true evolution of the neural activity is deterministic Eq. 1–3, we assume it is sufficiently chaotic to be described statistically. Specifically, we treat the activity probability distribution of a single, “typical” representative unit
as statistically representing the collective state. We are then interested in the time evolution of
, the categorical distribution over which of the D neurons in
spikes at t. We can define the population-averaged firing rates of the D co-tuned populations as
, with
the expected number of spikes per
of a neuron in population d at time t (assuming a neuron can spike at most once per
). For simplicity we choose units where
(reflecting the 1-timestep update rule of Eq. 1–3) so that
.
As the firing rate vector is equivalent to a discrete probability distribution, it evolves on the regular (D-1)-simplex
(which has D corners) (Fig 3a and 3b). Accordingly, it can be shown (S1 Appendix) that the firing rates
evolve according to:
(a) Schematic of the mean-field variable (the population firing rates ), which is equivalent to a categorical probability distribution
over which neuron (defined by its tuning) is active in the representative unit
at time t. (b)
evolves on a (D-1)-dimensional simplex
(which has D corners). (c) Simulations of mean-field dynamics for unstructured networks (D = 5) of three different sizes, with a 1-timestep external input provided to tunings 1 and 2 at t = 25. (d) Full network simulations corresponding to (C) (except that the large network uses N = 5000). (e) Mean-field predictions and full simulation (N = 5000) estimations of input distributions to neurons with tunings 1 and 4 at the time when the external input is applied. (f) As in e except at a time when no external input is applied. (g) Time-averaged response of the first co-tuned population to constant external input applied to population 1, populations 1 and 2, populations 1–4, or all co-tuned populations. Circles give averages across three initial conditions of the full network simulation (N = 1000).
where is the cumulative normal density and
and
are the input means and variances to neurons with different tunings, which depend on the previous firing rate vector
. Intuitively, Eq 5 gives the probability that a sample from the Gaussian with the d-th mean and variance is larger than samples from all the others, reflecting the WTA competition. Note that the theory does not incorporate the heterogeneity of firing rates within each population (Fig 2d), which may account for some of the small errors in predicting the full network simulations.
The theoretical input means and variances are given by
and
where and
are the mean and variance of the external input to neurons with tuning d, and
is the number of neurons with tuning
that spiked at t–1. For finite networks the latter is effectively stochastic:
Thus, one expects fluctuations in the population firing rates even when the firing rates at the previous timestep are known exactly. For large networks, however, the input means and variances approach
and
such that the time-evolution of becomes deterministic, even when the underlying spiking dynamics remain irregular. Thus, in total, the mean-field equations for the firing rate dynamics are parameterized by D, N (for finite-size simulations), and two
matrices
and
. In contrast to the linear steady-state equations of the classic balanced state [17, 18], the mean-field equations governing our network are generically nonlinear with respect to both external inputs and the previous network state, even when
. This suggests the network may be able to exhibit more expressive collective dynamics than a linear system alone.
Input response
Simulating the mean-field dynamics (Eq 5–10) lets us predict the behavior and input responses of the full spiking network (Eq. 1–4). For instance, the theory predicts that in infinitely large unstructured networks spontaneous activity will be evenly balanced across the D populations and perturbations immediately corrected (Fig 3c), which is recapitulated in full simulations of large networks (Fig 3d, N = 5000). Notably, the theory also lets us predict the dynamics of finite networks (Eq. 6–7), which similarly balance activity across all populations and immediately correct the perturbation (Fig 3c), but atop a blanket of noise that decreases with N, which is reflected in the full simulations with matched N (Fig 3d). Eq 9 and Eq 10 also give an exact description of the (Gaussian) input distribution when the previous state and external input mean and variance are known, which is corroborated by the full simulation (Fig 3e and 3f).
We can also predict the response of the unstructured network to constant external input, a common experimental stimulus [72]. For simplicity we set the external input variances . In general, the network response depends on the total number of co-tuned populations in the network along with how many are driven by the external input (Fig 3g). In all cases, the WTA nonlinearity (Eq. 1) bounds the total firing rate of any population at 1 (since the maximum rate is one spike per timestep). When K out of D populations are driven, the firing rate of each individual population saturates at 1/K, and as expected the network shows no response when all populations are driven (since this just shifts the baseline atop which the WTA competition takes place). When only a few out of a large number of populations are driven, however, the driven populations exhibit an additional supralinear response at weak inputs (Fig 3g, right), before eventually reaching the sublinear saturating regime at strong inputs.
One feature of this system is that neurons’ responses to their preferred stimulus are suppressed in the presence of non-preferred stimuli. In Fig 3g (D=4), for instance, population 1 exhibits a strong response when driven alone but a weaker response when population 2 is also driven. This is similar to the phenomenon of cross-orientation suppression observed in visual cortex, in which presentation of a non-preferred orientation suppresses a neuron’s response to its preferred orientation [34]. Intuitively, this reflects the within-unit WTA competition, which influences the population response. This behavior cannot emerge in the classic balanced state, which responds linearly to external inputs [17, 18], although nonlinear responses can emerge if populations are allowed to be silenced via excess inhibition (“semi-balance”) [23]. The input responses of our network are instead closer to those of networks operating in the “loose balance” regime, which also exhibits cross-orientation suppression [11, 39], except here our network’s response is computed in the limit of an infinite number of pre-synaptic inputs.
Stability of the uniform distribution
The return to evenly distributed activity (across all populations) after external input is removed (Fig 3c and 3d) can be understood using our theory. In unstructured networks, where and
, we see that absent external inputs the uniform distribution is stable. This is because
(Eq 9), and from Eq 10:
Thus, becomes the distribution over the argmax of samples from D identical Gaussians (Eq 5), which by symmetry must be uniform, so
. Therefore all network states are attracted back to the uniform fixed point in a single time step. Stability does not depend on D, hence uniform distributions over a large number of tunings are also stable, even though individual neurons spike at a mean rate of 1/D. In agreement with simulation, this means firing rates can be made arbitrarily low, yet nonzero, simply by increasing D (Fig 2e).
Structured networks
More interesting dynamics arise if we introduce structure into the network connectivity. To this end we now explore network implementations of three core dynamical primitives thought to support memory-dependent neural computation in the brain: (1) multistability, which allows information to persist over time in macroscopic attractor states of the network [73, 74]; (2) sequence generation, in which activity evolves along a 1-dimensional manifold, which is thought to support decision-making and time-keeping [4, 75, 76]; and (3) slow (relative to the single-neuron response timescale), complex, nonlinear heterogeneous firing rate dynamics, which are thought to facilitate learning of flexible nonlinear temporal computations and motor outputs [2, 77].
Notably, our theory lets us predict how such macroscopic dynamics could emerge via multiple different network mechanisms. First, we can adjust only the mean weights among the different populations. Alternatively, we can adjust only the weight variances
—in this scenario all mean inputs are always zero (Eq 6, Eq 9), hence macroscopic dynamics can only emerge from changes in the input variances over time, i.e. the dynamics are forced to be fluctuation-driven, which promotes irregular spiking. In general, one can also adjust combinations of the weight means and variances. Below we demonstrate advantages of fluctuation-driven over mean-driven implementations of both multistability and sequence generation, then show how the fluctuation-driven regime alone can support complex, nonlinear firing rate dynamics mediated by irregular spiking. In each setting, we first study the mean-field dynamics (Eq 5–10) of the D-dimensional vector
of firing rates of the D populations, then simulate the corresponding full spiking network (Eq 1–4) governing the dynamics of the single-neuron activities
.
Multistability
A long-standing hypothesis for the basis of slow neural dynamics supporting working memory is multistability of the network state through persistent spiking [12, 16, 26, 27, 78, 79]. As multistability does not emerge in randomly coupled balanced spiking networks of excitatory and inhibitory neurons without fine-tuning [17, 80, 81], how to achieve multistability in networks that exhibit irregular spiking in all stable states remains an open question. In particular, introducing positive excitatory feedback among neurons (e.g. in clustered balanced networks) produces multistability, but spiking becomes less irregular [26, 27], in contrast with experiments. Multistability can also emerge in semi-balanced networks, in which certain populations are allowed to be silent [23]. Alternatively, one study showed how certain synaptic nonlinearities could produce two stable states characterized by different input variances [40], recapitulating the near-unity ISI CVs observed across task epochs in cortex [16]; while this model required additional synaptic nonlinearities via short-term plasticity rules, it demonstrates the consistency of irregular spiking with nonlinear global dynamics flowing through input variances.
Mean-driven multistability.
To test our network’s ability to exhibit multistability, we first explored “mean-driven” multistability. To do so we increased the mean weights within each co-tuned population (S2 Fig). This reflects traditional working memory models, in which network states are stabilized by positive feedback, but typically at the cost of less irregular hence less realistic spiking [26, 27, 73, 82]. This was also reflected in our model—as we increased the diagonal of the mean weights , our theory predicted a sharp transition from monostability (of the uniform distribution) to exactly one of the populations being maximally active and the rest silent (S2 Fig). The theoretical predictions were recapitulated in the full simulations (S2 Fig), which additionally revealed the consequent regularity of spiking, which is inconsistent with experimental observations [16].
Using our theory we were able to explain the mono-to-multistable transition of the mean-driven network mathematically. In our model multistability arises when the uniform distribution () destabilizes, which forces the macroscopic state elsewhere, in the mean-driven case to a corner of the simplex on which the dynamics unfold (Fig 3b). Our theory let us predict when this would occur, which corresponds to the maximum eigenvalue
of the Jacobian of the mean-field dynamics (Eq 5), evaluated at
, exceeding 1. This occurred when
was sufficiently large and when D was sufficiently small (S3 Fig)—increasing D in a multistable network can make the network monostable, as there are more co-tuned populations to compete with. This in turn let us predict the expected number of active populations, along with the information content of the network state (
of the number of active populations), which was approximately recapitulated in the full network simulations (S3 Fig). Because in the mean-driven multistable network only 1 population was active at a time, the information content of a stable state is bounded at
bits.
Fluctuation-driven multistability.
Next, we explored the possibility of fluctuation-driven multistability in our network. To this end, we attempted to induce multistability by increasing the diagonal of the weight variances, (Fig 4a), with all off-diagonal weight variances set to
, which led to substantially different macroscopic behavior from the mean-driven case. Our theory predicted that, in contrast to the mean-driven regime, the fluctuation-driven regime should exhibit a much smoother transition to multistability in which not one but several populations are active and the rest quiescent (but not necessarily silent) (Fig 4b), which was recapitulated in full simulations that also suggested the presence of highly irregular spiking (Fig 4c). Thus, the network can produce multistability in a fluctuation-driven regime as well.
(a) Schematic of network parameters (means and variances). (b) Mean-field simulations () of firing rate dynamics for 6 different
. (c) Example firing rate dynamics and spike trains from a full simulation of the fluctuation-driven multistable network (
, N = 3000, D = 16).
Computing for the fluctuation-driven case also revealed a transition to multistability (via destabilization of the uniform distribution) given sufficient feedback (here through increased weight variances) (Fig 5a). Unlike the mean-driven case (S3 Fig), however, multistability in the fluctuation-driven network emerges only when D is sufficiently large, after which a subset of the D populations remain active in each stable state. Because multiple populations are active in each network state, similar to the clustered balanced networks of [26, 27], the network can hold much more information than if only one population were active (Fig 5b). Curiously, in the
limit when
, the uniform distribution destabilizes when D = 11, leading to exactly 10 active populations for
. The number of populations at which this transition occurs, however, can be controlled by changing the diagonal
or off-diagonal
mean weights (Fig 5c and 5d), such that the information capacity of the network is not limited by only 10 populations being active. In general, the number of active populations increases as
decreases (Fig 5c). Further, when more populations are active the firing rate of neurons in those populations decreases, suggesting stable states where neurons are active at low firing rates.
(a) Maximum eigenvalue of the Jacobian of the
mean-field dynamics evaluated at the uniform distribution
, as a function of D.
for the bottom two traces. (b) Number of active populations (top) or equivalently information content (bottom) vs D (
). Solid lines are theory; circles are full simulation results (N = 3000). (c) Theoretical predictions of the number of active populations vs the diagonal term
, when
, with circles depicting median over 15 initializations (N = 2000). (d) Theoretical phase portrait depicting when the network is multistable. (e) Theoretical (lines) and simulation predictions (squares) of input distributions to active and quiescent populations in the fluctuation-driven multistable network. (f) ISI CV2 distributions for two co-tuned populations using extended runs (1500 timesteps) of the fluctuation-driven multistable simulations in Fig 4c. Dashed traces show ISI CV.
The fluctuation-driven multistable network behaves very differently from the mean-driven network. For instance, examining the fluctuation-driven simulations more carefully revealed that different stable states are distinguished by different input variances, rather than means (Fig 5e). Moreover, spiking was highly irregular (near-unity ISI CV and ISI CV2) in both the active and quiescent states of a population (Fig 5f), which was also observed in monkey prefrontal cortex during a working memory task [16], with irregular spiking in the model emerging through chaos (S4 Fig). Thus, our network admits a fluctuation-driven multistable regime in which spiking is consistently irregular regardless of the network state, suggesting a new solution to multistability in a brain-like dynamical regime. In our model these phenomena emerge ultimately through the multi-dimensional WTA nonlinearity and collective interactions of the neurons, without introducing any other single-neuron or synaptic timescales beyond the 1-timestep update rule (Eq 1–3). This also suggests a fundamentally different form of multistability from traditional models, in which activity is stabilized via positive excitatory feedback [73, 78, 80, 83–85]. Here activity is stabilized via increased input variances (similar to [40], except without requiring additional synaptic nonlinearities), corresponding to co-tuned neurons with higher variance, rather than higher mean, recurrent connection weights.
Sequence generation
A second common dynamical motif thought to serve memory- and time-dependent computation in the brain is the generation of stereotyped neural activity sequences [4, 75, 76, 86, 87]. Unlike multistability, sequence generation allows information to be not only held but transformed over time. To assess our network’s capacity to generate sequences we destabilized the multistable network, in a similar spirit as [87]. First we placed each population on a ring, such that the D populations were evenly spaced between and
(we focus on periodic sequences). Starting with either the mean- or fluctuation-driven multistable network we then introduced an asymmetric change to the network connectivity, corresponding to unidirectional increases in the weights (either in the mean or variance) from each population to its clockwise neighbor.
Mean-driven sequence generation.
As with the multistable network, we first explored the consequences of unidirectionally increasing the mean weights linking neighboring populations (S5 Fig). Similar to the multistable case, our theory predicts a transition from the network producing no sequential activity (one population saturated and the rest silent) to producing very fast sequences, with one population approximately saturated at each time point (S5 Fig). This was corroborated in the full simulations, where fast sequential firing rate dynamics were accompanied by regular spike trains (S5 Fig). Nonetheless, the transition was not perfectly sharp, and the propagation speed of the sequence could be accurately predicted from the mean-field equations (S6 Fig). Examination of the activity distribution of the mean-driven sequence-generating network also suggests that the network “hops” between metastable states (S6 Fig), limiting the effective resolution of the network (i.e. the network can effectively be in one of D states; S6 Fig). Thus, asymmetrically increasing mean weights between adjacent populations in an originally multistable network can produce rapid, stereotyped sequences accompanied by regular spiking. This possibly reflects a dynamical regime observed in neural circuits underlying song production in songbirds, where spike trains are thought to be highly regular and stereotyped [88]; however, it does not reproduce the irregular spiking of mammalian cortex.
Fluctuation-driven sequence generation.
Sequences produced in the fluctuation-driven regime behaved quite differently. Indeed, unidirectionally increasing the weight variances between adjacent populations (Fig 6a) of a fluctuation-driven multistable network also led to the production of sequences. As with the multistable case, our theory predicted a much smoother transition to sequence generation than in the mean-driven network (Fig 6b). This was corroborated by the full simulations, in which the propagation of the sequence could be easily controlled by setting
, and which additionally suggested consistently irregular spiking patterns throughout the sequence evolution (Fig 6c). The theory accurately predicted the speed of propagation in full simulations of the fluctuation-driven network (Fig 7a). Curiously, unlike the mean-driven case, the theory suggested that the activity pattern of the network should not hop between metastable states, but instead should “flow” more continuously around the ring (Fig 7b). This was corroborated in the full simulation, suggesting that the resolution of information stored in the sequence location may be higher than in the mean-driven network (Fig 7c). Finally, estimating the ISI CV2 distribution from spike trains in the full simulation revealed high, Poisson-like irregularity (Fig 7d), suggesting that this network does not collapse to an unrealistic regular spiking regime when it produces sequences. As in the fluctuation-driven multistable case, irregular spiking in the fluctuation-driven sequence network emerges through chaos (S7 Fig). Thus, fluctuation-driven dynamics mediated by weight variances both produce and may advantage sequence generation, while also yielding consistently brain-like irregular spiking. Similar to the multistable network, this suggests that stable neural activity sequences can be produced without increased positive excitatory connections between neurons—increased variances in connectivity can suffice instead, and are consistent with sequences mediated by irregular spiking.
(a) Schematic of network weights for fluctuation-driven sequence generation. (b) mean-field simulations of firing rate dynamics in fluctuation-driven sequence generation, for 5 different values of
, with
, and D = 16. (c) Population firing rates and spike trains from an example full network simulation producing sequences at two different speeds (different
, with
).
(a) Theoretical (mean-field; ) predictions and full simulations (
) of sequence propagation speed in the fluctuation-driven network. (b) Heatmaps of the evolution of macroscopic network activity in the fluctuation-driven sequence-generating network, with
, via either the
theory or the equivalent full simulation (N = 2500). (c) Histogram of circular means of macroscopic activity distribution in the fluctuation-driven sequence-generating network in (b). (d) ISI CV2 distributions for neurons in two different populations in the fluctuation-driven sequence-generating network for two different propagation speeds.
Heterogeneous fluctuation-driven firing rate dynamics
Finally, we asked whether the network could produce more complex firing rate dynamics than multistability and sequences alone, yet still simultaneous with irregular spiking. Complex, time-varying firing rates are thought to provide a reservoir of temporal patterns that can be composed into flexible computations [2, 28, 42, 89, 90]. Typically, however, such dynamics are studied in “rate network” models of neural populations that do not explicitly account for spikes; most commonly, a D-dimensional vector of population firing rates,
, evolves according to
, where h are private variables, u are inputs,
is a time constant, W is a
matrix of interaction weights between populations, and
is an element-wise nonlinearity such as a tanh or sigmoid [1, 2, 5, 33, 90, 91]. In contrast, our equations for
(Eq 5–10), which also describe the time-varying firing rates of D populations, operate fundamentally differently—besides being in discrete time, Eq 5–10 are parameterized by two
matrices,
and
, and are mediated by a non-element-wise nonlinearity (Eq 5). Thus, it is unclear how similarly they behave to the more common rate network equations. Note also that while the equations for
are derived (as the mean-field description of the full network dynamics) under the assumption of irregular spiking in
, irregular single-neuron activity does not imply irregular or complex dynamics in
; for instance, in the unstructured network the spikes are irregular but the population firing rates are constant (Fig 2 and 3). Here, we investigate whether irregular activity in
can co-exist with complex, time-varying dynamics of the population firing rates
.
To address this problem, we simulated the mean-field equations for (Eq 5, 9, 10;
limit), fixing
and only allowing ourselves to vary the matrix parameter
. This procedure keeps the network in the fluctuation-driven regime (since all mean inputs are 0; Eq. 6, 9), which is in turn expected to produce irregular spiking in the corresponding full simulations. A challenge of this approach, however, is that whereas the matrix W in classic rate networks can usually contain both positive and negative values [1, 2, 5, 90, 91], all elements of
are non-negative by definition. (Note that in our full model, the weights
between individual neurons can be both positive and negative [Eq 4]—the non-negative matrix
specifies the variances of
, in both the mean-field equations for
and the full model, and is the only parameter we vary in the present simulations of
.) Given this constraint, it is not obvious that Eq 5–10 could produce activity with the same level of complexity as can arise in classic rate networks. We hypothesized, however, that if
were a sparse random matrix, the normalization of the network activity through Eq 5 would play a similar role as the inhibitory feedback in classic rate networks.
As with the sequence-generating simulations, to test our network’s ability to generate complex firing rate dynamics we started with a multistable network, then destabilized it. Specifically, we first set all diagonal entries of to 1 and the rest to 0 (which yields multistability), then outside the diagonal introduced
nonzero elements in each row, each with a value of
(Fig 8a). We predicted that when the density of nonzero entries in the weight variance matrix
was sufficiently high, but still sparse relative to the total number of populations D (i.e.
), the network would be forced into a regime in which strong fluctuation-driven feedback competes with this normalization to produce chaotic-like dynamics, in loose analogy to how tanh or sigmoidal “squashing” nonlinearities transform unstable linear dynamics into chaos in classic rate networks [1].
(a) Schematic of network configuration. (b). Simulations of mean-field dynamics for for different parameter sets corresponding to the architecture shown in (a). (c) Extended mean-field simulation (
) of fluctuation-driven firing rate dynamics produced by an example network starting from 2 different initial conditions that lead to complex rate dynamics. (d) Example initial condition from the same network in (c) leading to a fixed point in the mean-field dynamics. (e) Full spiking network simulation showing firing rate dynamics and spike train dynamics with macroscopic parameters equivalent to the simulation in (c–d). Spike trains are shown for two example neurons per population.
Simulating the mean-field equations for for a variety of D,
, and
revealed a rich repertoire of firing rate dynamics. For instance, for D = 100 and small
and
, the network remained multistable, with multiple firing rate levels corresponding to active states (Fig 8b, left). For a larger D = 200, and stronger recurrent weight variances (
), certain initial conditions additionally produced extended firing rate dynamics in which different populations became active in a haphazard fashion, although fixed points could still arise (Fig 8b, middle). When
was even larger, the firing rate dynamics exhibited consistently strong spatiotemporal heterogeneity (Fig 8b, right). Thus, similar to classic rate networks [1, 2], strong recurrent feedback can also produce highly variable, heterogeneous firing rate dynamics—the mechanism and corresponding dynamical system are different in our model, however, since all firing rate dynamics follow from dynamics of the input variances
, with mean inputs
always zero (Eq 6–10).
Finally, we ran the full spiking network simulations (Eq 1–4) corresponding to an example parameter regime in which the equations for produced temporally complex firing rate dynamics. We first identified an example set of parameters (using the mean-field dynamics [Eq 5, 9, 10]) in which random initial conditions frequently produced extended complex firing rate dynamics in the
case (Fig 8c). This particular network instantiation also sometimes exhibited fixed points in the dynamics (Fig 8d), suggesting the possible coexistence of chaos (or at least complex transients) with fixed points in a single network, although further analyses will be required to characterize this in detail. We then ran the corresponding full spiking network simulation, which produced similar dynamics as the mean-field predictions (Fig 8e), exhibiting slow, heterogeneous changes in firing rates that continued over many timesteps. Both the mean-field and full spiking network simulations yielded firing rates with wide autocovariance functions (Fig 9a and 9b), extending far beyond the single-neuron memory timescale (
) in the microscopic update rule (Eq 1, Eq 3). Note that in the full spiking simulations some of the dynamics may arise from finite-size effects, similar to how finite numbers of neurons in balanced clustered networks allow spontaneous fluctuations to drive transitions between metastable states [26–28]. However, because the slow timescales persist in our mean-field simulations (computed in the
limit), the slowness and complexity of the firing rate dynamics are likely not exclusively a finite-size effect in N. Intuitively, the slowness is inherited from the multistable network, with the additional couplings between the mean-field populations via the nonzero entries of
destabilizing the multistability. Further, the existence of slow firing rate dynamics given different network parameters (Fig 8b) suggests that fine-tuning is not required.
(a) Sample firing rate autocovariance function from the infinite mean-field simulations in Fig 8c, averaged across all co-tuned populations (for 8 example initializations of the same network). (b) As in (a) but computed from the full spiking network simulation (N = 2000) (c). ISI CV2 distributions for several example co-tuned populations from Fig 8e. (d). Firing rate vs ISI CV2 for all neurons in the network. (e) Spike-train autocovariance function, averaged across all neurons (each of the 8 traces shows a different initialization, although the traces are very similar).
To assess whether the model also produced irregular spikes we examined the temporal patterning of the spikes generated in the full network simulations. As we constructed our mean-field simulations for to operate in the fluctuation-driven regime, the spikes generated in the corresponding full network simulation of
were highly irregular (Fig 9c). Although some neurons substantially deviated from the Poissonian signature of ISI
, these neurons spiked rarely in the simulation, hence their spike trains were probably undersampled (Fig 9d). The autocovariances of the spike trains were a combination of a delta-function-like peak at a zero time-lag, indicative of Poisson-like firing, together with the wider base reflecting the slow evolution of the firing rates (Fig 9e). Thus, the fluctuation-driven regime of our network can produce brain-like irregular spike trains together with complex, heterogeneous, and slowly varying firing rate dynamics; which our theory suggests should persist as
. A detailed analysis of the dependence of this regime on D and further features of
and
, as well as the possibility of chaotic firing rate dynamics, will be an important direction for future work.
Discussion
We have shown how recurrent networks of winner-take-all (WTA) units can produce brain-like irregular spiking capable of powering flexible nonlinear collective dynamics. In contrast to the more typical spike-threshold nonlinearity, the WTA nonlinearity causes the network to avoid the emergent linear dynamics of classic balanced networks [11, 17, 18]. In turn, programming the network by adjusting the weight variances alone causes firing rate dynamics to follow the dynamics in the input variances, rather than the means, which in turn yields consistently irregular spiking. Our model thus reveals the dynamical consequences of WTA as a repeated functional motif in a large network and represents an alternative model class (compared to excitatory-inhibitory balanced networks [1, 11, 17, 19, 26, 27, 40, 41, 44, 92]) for studying biological neural activity, which unifies Poisson-like temporal irregularity with flexible nonlinear collective dynamics.
Our mean-field equations for the firing rates describe deterministic, nonlinear dynamics even in the
limit (Eq 5, 9–10). Thus, the network’s nonlinear behavior holds when neurons receive an infinite number of synaptic inputs. This contrasts with finite-size theories, e.g. the theory of “loose balance,” which reproduces empirical nonlinear phenomena as a consequence of neurons receiving finite synaptic inputs [11, 39, 93]. A related model [24] showed how single-neuron response-onset and saturation nonlinearities could partially propagate to the network level if recurrent interactions were weak, but this network also relied on finite synaptic inputs or weak couplings. Nonlinearity can also emerge in “semi-balanced” networks, in which populations can be temporarily silenced by excess inhibition, leading them to behave as rectified linear units [23]; our network, however, does not strictly require any population to be silent to produce nonlinear behavior (Fig 4b). Clustered balanced networks can produce multi- or metastable dynamics [26, 27], but active clusters typically generate more regular spikes, whereas the active populations in our multistable network produce highly irregular spikes (ISI CV
1). Certain short-term plasticity processes can yield bistable collective activity, with irregular spiking persisting in active states—as in our model, irregular spiking follows from state-dependent input variances [40]; however, the role such plasticity processes play in cortex is not well established, and it is moreover unclear how to generalize [40] to produce more complex dynamics such as sequences (Fig 6). An alternative model showed how correlated noise could drive sequences of state transitions matching empirical temporal variability [87], but this network did not include spikes—it will be interesting to explore whether correlated noise can facilitate dynamics in our network as well. In general, it will be interesting to compare these and other balanced network models [11, 20, 21, 23, 24, 40, 41, 43, 94, 95] with our network in terms of the collective dynamics they can produce and their relationship to empirical neural dynamics.
Biological basis
The key feature of our network is the Softmax/WTA nonlinearity operating within each unit, in which neurons with different tunings compete to activate [46]. This competition was inspired by biology [46] and could emerge through a variety of mechanisms, such as lateral inhibition, recurrent amplification, or synaptic depression [46, 48, 49, 55, 96]. Another possibility is via axo-axonic inhibition, in which one excitatory cell projects directly onto an inhibitory terminal onto a second excitatory cell, providing a means for inhibition without requiring the inhibitory cell to spike [97]. In principle, competition within a unit could also occur between different activity patterns rather than different neurons, with these patterns competing in a WTA-like fashion; or similarly between attractor-like states in small cortical patches [98]. An important consideration for more detailed network models is the timescales over which the key interactions unfold. In particular, we have modeled the integration of presynaptic inputs and the WTA operation as occurring within one timestep, roughly corresponding to one membrane time constant (10-20 ms [62]). Indeed, neurons integrate their presynaptic inputs over this timescale [62], and it has been shown that WTA-like operations can occur at or faster than this timescale in networks of integrate-and-fire neurons [48], suggesting the microscopic timescales in our model are consistent with biology. Formally extending the network to operate with continuous-time integrate-and-fire neurons that compete within small groups while also receiving fluctuating inputs will be important future work.
A second important direction is to determine how the network would be implemented with neurons that are strictly excitatory or inhibitory. In our model any neuron can send both positive and negative projections to other neurons in the network. While this violates Dale’s law (the restriction that each neuron should be strictly excitatory or inhibitory [99]), such an approach is often used as a first-pass approximation when studying collective neural dynamics [1, 5, 70, 100, 101]. The key assumption in our theoretical treatment is that inputs to any neuron at any timestep are i.i.d. Gaussian—therefore we expect our results to also hold when each neuron is strictly excitatory or inhibitory, so long as connectivity is sparse enough that inputs across different cells are nearly uncorrelated [19] and the network is large enough that the central limit theorem holds. Biologically, our network model is also consistent with multiple roles for inhibition, mediating (1) interactions between units as well as (2) the competition within a unit, and which might be supported biologically by different inhibitory cell types [102–104].
Training and fitting to data
The network we have presented can produce flexible macroscopic dynamics, depending on the matrices
and
. Thus, beyond the specific regimes we have studied here, the network’s many parameters (
in total) can also be generically adjusted to reproduce specific target dynamics that perform computations, possibly via e.g. gradient methods like back-propagation through time [105] or reservoir-computing methods [2, 89, 106]. The network’s flexibility suggests it may in turn share key features with traditional non-spiking recurrent neural networks [1, 2], notably the ability to transform input time-series into rich, high-dimensional nonlinear representations to facilitate flexible learning. The tendency of the fluctuation-driven regime to avoid saturation, however, may endow the network with additional abilities that challenge traditional networks, such as the handling of long-term dependencies [107]. A detailed comparison of the computational abilities of our network, traditional RNNs, as well as other modern recurrent neural networks [108, 109] specializing in handling long timescales, will be an exciting future research direction.
In principle one can also fit our network to data using maximum likelihood methods. Recent experimental technologies now allow for large-scale recordings of population spike trains over extended periods [110–112]. Modeling the spiking produced by our network as i.i.d. Bernoulli conditioned on firing rates (a slight simplification since firing rates are heterogeneous within a population and neurons in individual units negatively correlate), we can write the conditional probability of a binarized population spike train recording , of length T from n neurons, as
where are the D population firing rates from time 1 till T, t and k index time and neurons, and
indicates the population to which neuron k belongs. The probability of a population recording s, i.e. the marginal likelihood of the network parameters, is then
Here plays the role of a latent variable, much like the hidden state of a Hidden Markov Model (HMM) or Kalman filter [113–115]—in fact, the dynamics of the firing rates (Eq 5–10) given the network parameters are Markovian, since they depend only on the previous timestep. The model could therefore potentially be fit with a variation of Expectation Maximization [116], which alternates between inferring the latent sequence (here
) and learning the parameters (here
), and which is commonly applied to HMMs. Note that here one also requires a procedure for learning or marginalizing over the hyperparameters D, N, and
.
Thus, our proposed network represents a new statistical model for analyzing population spike train data. A key advantage over most latent variable models, which usually treat spikes as noisy readouts of abstract latent dynamics [9, 10, 117–119] (although see [120]), is that the fit model can be directly translated into a self-contained spiking network. Therefore any computational dynamics inferred by fitting the network to data are automatically compatible with a mechanistic model operating at the level of individual spiking neurons (grouped into WTA units), and which as we have shown is well-suited for implementing complex dynamics through irregular spike trains.
Limitations
Our model as instantiated has important limitations that should be addressed in future work. First, although our fluctuation-driven results (Fig 4–9) suggest that these dynamics, most notably the slow timescales, do not require fine-tuning but rather change gracefully as the parameters are varied, the model may still be sensitive to other features of its construction. It will be particularly useful to examine robustness against perturbations of the competitive WTA nonlinearity, e.g. if competition is not symmetric among neurons in a unit but spreads across units, possibly asymmetrically; if not every unit contains every tuning; or if neurons have continuous and/or overlapping tunings rather than discrete tunings. Second, the firing rates that drive nonlinear computation are averages over neurons within a population—this stands in contrast to approaches for constructing spiking networks where single rate units are converted to single spiking neurons [42, 121–123], which can in principle make more efficient use of individual neurons. Thus, an interesting future direction will be to examine how heterogeneity of activity within each population may support additional computational dynamics [124], such that the effective dimensionality of the system computation extends beyond the number of populations and toward the number of neurons.
Predictions
Despite its limitations and idealized nature our model makes testable predictions. First, we predict that neurons with correlated slow changes in firing rates (i.e. with similar “tuning”) should have higher variance functional couplings at millisecond timescales, but not necessarily higher mean couplings, as in our model the former are the source of slow timescales. This could be tested by fitting a network-generalized-linear-model [125] to population spike trains and comparing the inferred coupling filters between neuron pairs with the strength of correlation of their firing rates. A second prediction is that in large neural recordings computational dynamics will unfold primarily without affecting the mean firing rate across the full population of recorded neurons (which was fixed at 1/D in our model), which has been observed in monkey prefrontal cortex [38] and in mouse and monkey visual cortex [63]. Third, we predict that nonlinear firing rate dynamics underlying neural computations are significantly fluctuation-driven, rather than only mean-driven. While this prediction could be tested via in vivo intracellular recordings, one could more easily test it by fitting the network model to large-scale population spike trains [110–112], where we predict that the goodness of fit will rely heavily on structure in the weight variances. Our final core prediction is that empirical neural dynamics, in particular in awake cortical recordings, will be better described by rate equations closer to our mean-field equations (Eqs 5–10) than to more common model classes, e.g. of the form , where
are the firing rates and
is an element-wise monotonic nonlinearity [1, 2, 5, 33, 90, 91]. This could be tested by fitting both models to data, treating the firing rates as latent variables that produce observed spikes through a Poisson process, and comparing the fits.
Methods
Mean-field theory summary
Here we give a summary/intuition behind the derivation of the mean-field dynamics for the firing rates (Eq 5–10). Briefly, given , this means that approximately
neurons in population
just spiked at t–1. We suppose that these
neurons are uniformly sampled from that population (note that this assumption ignores heterogeneity of firing rates within the population). The total input to a postsynaptic neuron with tuning d from population
is then approximately a sum of
i.i.d. Gaussian inputs with mean
and variance
, since we have now randomly sampled and summed
elements from the quenched weight matrix
, which were constructed as i.i.d. (Eq 4), so the resulting sum is itself Gaussian. The total input to a postsynaptic neuron will then be the sum of D such inputs from the D populations, with possibly different means and variances, which is again Gaussian. Thus, at time t each neuron d within a unit receives a Gaussian input with a mean and variance determined by
, and
, and the probability of that neuron spiking is the probability that its input is larger than the input to every other neuron in the unit. This probability is equal to the firing rate
of population d, which once computed for all D populations gives us
. A full derivation can be found in S1 Appendix.
Simulations
For the most part, networks were simulated directly using Eqs 1–4. For very large networks, however, we approximate the network connectivity where
by storing a single connectivity block
in memory and generating the remaining blocks at run time via quenched random permutations of J0 and systematic scaling/shifting for each
. Details can be found in S2 Appendix.
Estimation of eigenvalues
In Fig 4 we numerically estimated the eigenvalue (of the Jacobian of the mean-field dynamics evaluated at the uniform firing rate distribution
), using a step size of
.
Nonstationary coefficient of variation of interspike intervals
The nonstationary coefficient of variation of interspike intervals, or ISI CV2, quantifies how similar interspike intervals are to those of an inhomogeneous Poisson process [14]. It is given by
where ISIl is the l-th interspike interval in a spike train. A value of is a signature of Poissonian spiking.
Numerical estimation of transition to chaos
We estimated the transition to chaos in Fig 1d by (1) running a simulation of the full network for 10 timesteps, (2) creating a copy of the network and its activity state at t = 10, (3) picking a single WTA unit in the copied network and swapping the activity levels of the most and least active neuron in that unit (a microscopic perturbation), (4) running both networks forward for 20 more timesteps, (5) computing the Euclidean distance between the complete states of the two networks as a function of time, (6) labeling the network chaotic if the distance increased following the perturbation and non-chaotic if it decreased.
Supporting information
S1 Fig. Sensitivity of the uniform balanced state to microscopic perturbations.
a. Root-mean-square distance between the microscopic network state () vs time since a single spike-swap perturbation is introduced. Parameters used were: D = 16, N = 2000;
;
. b-c. Firing rates corresponding to the reference run (b) vs the perturbed run (c, in which the spike swap is introduced at t=0).
https://doi.org/10.1371/journal.pcbi.1013081.s001
(PDF)
S2 Fig. Mean-driven multistability.
a. Schematic of network weights for mean-driven multistability. b. Mean-field simulations () of firing rate dynamics for 6 different
. c. Example firing rate dynamics and spike trains from a full simulation of the mean-driven multistable network (
, N = 3000, D = 16).
https://doi.org/10.1371/journal.pcbi.1013081.s002
(PDF)
S3 Fig. Properties of mean-driven multistable network.
a. Maximum eigenvalue of the Jacobian of the
mean-field dynamics evaluated at the uniform distribution
as a function of D and
. b. Number of active populations (top) or equivalently information content (bottom) for the mean-driven network (
). Solid lines are theory; circles are full simulation results (N = 3000).
https://doi.org/10.1371/journal.pcbi.1013081.s003
(PDF)
S4 Fig. Sensitivity to spike perturbations in the fluctuation-driven multistable network.
a. Root-mean-square distance between the microscopic network state () vs time since a single spike-swap perturbation is introduced, in the fluctuation-driven multistable network. Trace and error bars show mean and standard deviation, respectively, over 10 trials. Parameters used were N = 2000, D = 16,
,
. b. Evolution of firing rates for a reference simulation and a perturbed simulation (created by a single spike swap, relative to the reference network, at time 0). c. As in b, but for another trial.
https://doi.org/10.1371/journal.pcbi.1013081.s004
(PDF)
S5 Fig. Mean-driven sequence generation.
a. Schematic of network weights for mean-driven sequence generation. b. mean-field simulations of firing rate dynamics in mean-driven sequence generation, for 5 different values of
, with
, and D = 16. c. Population firing rates and spike trains from an example full network simulation with
and
.
https://doi.org/10.1371/journal.pcbi.1013081.s005
(PDF)
S6 Fig. Properties of mean-driven sequence-generating network.
a. Theoretical predictions and full simulations () of sequence propagation speed in the mean-driven network. b. Heatmaps of the evolution of macroscopic network activity (firing rates) in the mean-driven sequence-generating network, via either the
theory or a full simulation (
). c. Histogram of circular means of macroscopic activity distribution in the mean-driven sequence-generating network.
https://doi.org/10.1371/journal.pcbi.1013081.s006
(PDF)
S7 Fig. Sensitivity to spike perturbations in the fluctuation-driven sequence network
a. Root-mean-square distance between the microscopic network state () vs time since a single spike-swap perturbation is introduced, in the fluctuation-driven sequence-generating network. Trace and error bars show mean and standard deviation, respectively, over 10 trials. Parameters used were N = 2000, D = 16,
,
,
. b. Evolution of firing rates for a reference simulation and a perturbed simulation (created by a single spike swap, relative to the reference network, at time 0). c. As in b, but for another trial.
https://doi.org/10.1371/journal.pcbi.1013081.s007
(PDF)
S1 Appendix. Full derivation of the mean-field dynamics for the firing rates.
https://doi.org/10.1371/journal.pcbi.1013081.s008
(PDF)
S2 Appendix. Approximate simulations for large networks.
https://doi.org/10.1371/journal.pcbi.1013081.s009
(PDF)
Acknowledgments
I would like to thank Agostina Palmigiano, Tatiana Engel, Rainer Engelken, Lee Susman, Luca di Carlo, Kamesh Krishnamurthy, Yiqin Gao, Tankut Can, Daniel Weilandt, Nicolas Lenner, and Manuel Schottdorf for insightful conversations about this project, as well as other members of the Engel lab. Computational resources, office space, and conference funding to present earlier versions of this work were provided by the National Science Foundation through the Center for the Physics of Biological Function (PHY-1734030). This work was conceived during the 2019 IBRO-Simons Computational Neuroscience Imbizo.
References
- 1. Sompolinsky H, Crisanti A, Sommers H. Chaos in random neural networks. Phys Rev Lett. 1988;61(3):259–62. pmid:10039285
- 2. Sussillo D, Abbott LF. Generating coherent patterns of activity from chaotic neural networks. Neuron. 2009;63(4):544–57. pmid:19709635
- 3. Churchland MM, Cunningham JP, Kaufman MT, Foster JD, Nuyujukian P, Ryu SI, et al. Neural population dynamics during reaching. Nature. 2012;487(7405):51–6. pmid:22722855
- 4. Rajan K, Harvey CD, Tank DW. Recurrent network models of sequence generation and memory. Neuron. 2016;90(1):128–42. pmid:26971945
- 5. Mastrogiuseppe F, Ostojic S. Linking connectivity, dynamics, and computations in low-rank recurrent neural networks. Neuron. 2018;99(3):609-623.e29. pmid:30057201
- 6. Remington ED, Narain D, Hosseini EA, Jazayeri M. Flexible sensorimotor computations through rapid reconfiguration of cortical dynamics. Neuron. 2018;98(5):1005–1019.e5. pmid:29879384
- 7. Vyas S, Golub MD, Sussillo D, Shenoy KV. Computation through neural population dynamics. Annu Rev Neurosci. 2020;43:249–75. pmid:32640928
- 8. Duncker L, Sahani M. Dynamics on the manifold: identifying computational dynamical activity from neural population recordings. Curr Opin Neurobiol. 2021;70:163–70. pmid:34837752
- 9. Galgali AR, Sahani M, Mante V. Residual dynamics resolves recurrent contributions to neural computation. Nat Neurosci. 2023;26(2):1–13.
- 10. Chadwick A, Khan AG, Poort J, Blot A, Hofer SB, Mrsic-Flogel TD, et al. Learning shapes cortical dynamics to enhance integration of relevant sensory input. Neuron. 2023;111(1):106-120.e10. pmid:36283408
- 11. Ahmadian Y, Miller KD. What is the dynamical regime of cerebral cortex? Neuron. 2021;109(21):3373–91. pmid:34464597
- 12. Funahashi S, Bruce CJ, Goldman-Rakic PS. Mnemonic coding of visual space in the monkey’s dorsolateral prefrontal cortex. J Neurophysiol. 1989;61(2):331–49. pmid:2918358
- 13. Softky WR, Koch C. The highly irregular firing of cortical cells is inconsistent with temporal integration of random EPSPs. J Neurosci. 1993;13(1):334–50. pmid:8423479
- 14. Holt GR, Softky WR, Koch C, Douglas RJ. Comparison of discharge variability in vitro and in vivo in cat visual cortex neurons. J Neurophysiol. 1996;75(5):1806–14. pmid:8734581
- 15. Shadlen MN, Newsome WT. The variable discharge of cortical neurons: implications for connectivity, computation, and information coding. J Neurosci. 1998;18(10):3870–96. pmid:9570816
- 16. Compte A, Constantinidis C, Tegner J, Raghavachari S, Chafee MV, Goldman-Rakic PS, et al. Temporally irregular mnemonic persistent activity in prefrontal neurons of monkeys during a delayed response task. J Neurophysiol. 2003;90(5):3441–54. pmid:12773500
- 17. van Vreeswijk C, Sompolinsky H. Chaos in neuronal networks with balanced excitatory and inhibitory activity. Science. 1996;274(5293):1724–6. pmid:8939866
- 18. van Vreeswijk C, Sompolinsky H. Chaotic balanced state in a model of cortical circuits. Neural Comput. 1998;10(6):1321–71. pmid:9698348
- 19. Brunel N. Dynamics of sparsely connected networks of excitatory and inhibitory spiking neurons. J Comput Neurosci. 2000;8(3):183–208. pmid:10809012
- 20. Vogels TP, Abbott LF. Gating multiple signals through detailed balance of excitation and inhibition in spiking networks. Nat Neurosci. 2009;12(4):483–91. pmid:19305402
- 21. Vogels TP, Sprekeler H, Zenke F, Clopath C, Gerstner W. Inhibitory plasticity balances excitation and inhibition in sensory pathways and memory networks. Science. 2011;334(6062):1569–73. pmid:22075724
- 22. Monteforte M, Wolf F. Dynamic flux tubes form reservoirs of stability in neuronal circuits. Phys Rev X. 2012;2(4).
- 23. Baker C, Zhu V, Rosenbaum R. Nonlinear stimulus representations in neural circuits with approximate excitatory-inhibitory balance. PLoS Comput Biol. 2020;16(9):e1008192. pmid:32946433
- 24. Sanzeni A, Histed MH, Brunel N. Response nonlinearities in networks of spiking neurons. PLoS Comput Biol. 2020;16(9):e1008165. pmid:32941457
- 25. Ma WJ, Beck JM, Latham PE, Pouget A. Bayesian inference with probabilistic population codes. Nat Neurosci. 2006;9(11):1432–8. pmid:17057707
- 26. Litwin-Kumar A, Doiron B. Slow dynamics and high variability in balanced cortical networks with clustered connections. Nat Neurosci. 2012;15(11):1498–505. pmid:23001062
- 27. Mazzucato L, Fontanini A, La Camera G. Dynamics of multistable states during ongoing and evoked cortical activity. J Neurosci. 2015;35(21):8214–31. pmid:26019337
- 28. Huang C, Doiron B. Once upon a (slow) time in the land of recurrent neuronal networks…. Curr Opin Neurobiol. 2017;46:31–8. pmid:28756341
- 29. Destexhe A, Rudolph M, Paré D. The high-conductance state of neocortical neurons in vivo. Nat Rev Neurosci. 2003;4(9):739–51. pmid:12951566
- 30. Okun M, Lampl I. Instantaneous correlation of excitation and inhibition during ongoing and sensory-evoked activities. Nat Neurosci. 2008;11(5):535–7. pmid:18376400
- 31. Ingrosso A, Abbott LF. Training dynamically balanced excitatory-inhibitory networks. PLoS One. 2019;14(8):e0220547. pmid:31393909
- 32.
Kadmon J, Timcheck J, Ganguli S. Predictive coding in balanced neural networks with noise, chaos and delays. In: Advances in Neural Information Processing Systems 33. 2020, pp. 16677–88.
- 33. Engelken R, Ingrosso A, Khajeh R, Goedeke S, Abbott LF. Input correlations impede suppression of chaos and learning in balanced firing-rate networks. PLoS Comput Biol. 2022;18(12):e1010590. pmid:36469504
- 34. Priebe NJ, Ferster D. Mechanisms underlying cross-orientation suppression in cat visual cortex. Nat Neurosci. 2006;9(4):552–61. pmid:16520737
- 35. Warden MR, Miller EK. Task-dependent changes in short-term memory in the prefrontal cortex. J Neurosci. 2010;30(47):15801–10. pmid:21106819
- 36. Rigotti M, Barak O, Warden MR, Wang X-J, Daw ND, Miller EK, et al. The importance of mixed selectivity in complex cognitive tasks. Nature. 2013;497(7451):585–90. pmid:23685452
- 37. Fyall AM, El-Shamayleh Y, Choi H, Shea-Brown E, Pasupathy A. Dynamic representation of partially occluded objects in primate prefrontal and visual cortex. Elife. 2017;6:e25784. pmid:28925354
- 38. Murray JD, Bernacchia A, Roy NA, Constantinidis C, Romo R, Wang X-J. Stable population coding for working memory coexists with heterogeneous neural dynamics in prefrontal cortex. Proc Natl Acad Sci U S A. 2017;114(2):394–9. pmid:28028221
- 39. Ahmadian Y, Rubin DB, Miller KD. Analysis of the stabilized supralinear network. Neural Comput. 2013;25(8):1994–2037. pmid:23663149
- 40. Mongillo G, Hansel D, van Vreeswijk C. Bistability and spatiotemporal irregularity in neuronal networks with nonlinear synaptic transmission. Phys Rev Lett. 2012;108(15):158101. pmid:22587287
- 41. Boerlin M, Machens CK, Denève S. Predictive coding of dynamical variables in balanced spiking networks. PLoS Comput Biol. 2013;9(11):e1003258. pmid:24244113
- 42. Abbott LF, DePasquale B, Memmesheimer R-M. Building functional networks of spiking model neurons. Nat Neurosci. 2016;19(3):350–5. pmid:26906501
- 43. Khajeh R, Fumarola F, Abbott LF. Sparse balance: excitatory-inhibitory networks with small bias currents and broadly distributed synaptic weights. PLoS Comput Biol. 2022;18(2):e1008836. pmid:35139071
- 44. Renart A, de la Rocha J, Bartho P, Hollender L, Parga N, Reyes A, et al. The asynchronous state in cortical circuits. Science. 2010;327(5965):587–90. pmid:20110507
- 45.
Yuille A, Geiger D. Winner-take-all networks. In: Arbib MA, editor. The handbook of brain theory and neural networks. MIT Press; 2003, pp. 1228–31.
- 46. Douglas RJ, Martin KAC. Neuronal circuits of the neocortex. Annu Rev Neurosci. 2004;27:419–51. pmid:15217339
- 47. Maass W. On the computational power of winner-take-all. Neural Comput. 2000;12(11):2519–35. pmid:11110125
- 48. Jin DZ, Seung HS. Fast computation with spikes in a recurrent neural network. Phys Rev E Stat Nonlin Soft Matter Phys. 2002;65(5 Pt 1):051922. pmid:12059608
- 49. Carandini M, Heeger DJ. Normalization as a canonical neural computation. Nat Rev Neurosci. 2011;13(1):51–62. pmid:22108672
- 50. Heeger DJ. Normalization of cell responses in cat striate cortex. Vis Neurosci. 1992;9(2):181–97. pmid:1504027
- 51. Bonin V, Mante V, Carandini M. The suppressive field of neurons in lateral geniculate nucleus. J Neurosci. 2005;25(47):10844–56. pmid:16306397
- 52. Berger TK, Perin R, Silberberg G, Markram H. Frequency-dependent disynaptic inhibition in the pyramidal network: a ubiquitous pathway in the developing rat neocortex. J Physiol. 2009;587(Pt 22):5411–25. pmid:19770187
- 53. Busse L, Wade AR, Carandini M. Representation of concurrent stimuli by population activity in visual cortex. Neuron. 2009;64(6):931–42. pmid:20064398
- 54. Olsen SR, Bhandawat V, Wilson RI. Divisive normalization in olfactory population codes. Neuron. 2010;66(2):287–99. pmid:20435004
- 55. Heeger DJ, Zemlianova KO. A recurrent circuit implements normalization, simulating the dynamics of V1 activity. Proc Natl Acad Sci U S A. 2020;117(36):22494–505. pmid:32843341
- 56. Rutishauser U, Douglas RJ, Slotine J-J. Collective stability of networks of winner-take-all circuits. Neural Comput. 2011;23(3):735–73. pmid:21162667
- 57. Nessler B, Pfeiffer M, Buesing L, Maass W. Bayesian computation emerges in generic cortical microcircuits through spike-timing-dependent plasticity. PLoS Comput Biol. 2013;9(4):e1003037. pmid:23633941
- 58. Klampfl S, Maass W. Emergence of dynamic memory traces in cortical microcircuit models through STDP. J Neurosci. 2013;33(28):11515–29. pmid:23843522
- 59. Kappel D, Nessler B, Maass W. STDP installs in Winner-Take-All circuits an online approximation to hidden Markov model learning. PLoS Comput Biol. 2014;10(3):e1003511. pmid:24675787
- 60. Hubel DH, Wiesel TN. Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J Physiol. 1962;160(1):106–54. pmid:14449617
- 61. Anderson JS, Carandini M, Ferster D. Orientation tuning of input conductance, excitation, and inhibition in cat primary visual cortex. J Neurophysiol. 2000;84(2):909–26. pmid:10938316
- 62.
Kandel ER, Schwartz JH, Jessell TM, Siegelbaum S, Hudspeth AJ, Mack S, et al. Principles of neural science, vol. 4. New York: McGraw-Hill; 2000.
- 63. Sanzeni A, Palmigiano A, Nguyen TH, Luo J, Nassi JJ, Reynolds JH, et al. Mechanisms underlying reshuffling of visual responses by optogenetic stimulation in mice and monkeys. Neuron. 2023;111(24):4102-4115.e9. pmid:37865082
- 64. Bienenstock E. A model of neocortex. Network. 1995;6(2):179–224.
- 65. Diesmann M, Gewaltig MO, Aertsen A. Stable propagation of synchronous spiking in cortical neural networks. Nature. 1999;402(6761):529–33. pmid:10591212
- 66. Abeles M, Hayon G, Lehmann D. Modeling compositionality by dynamic binding of synfire chains. J Comput Neurosci. 2004;17(2):179–201. pmid:15306739
- 67. Jahnke S, Memmesheimer R-M, Timme M. Stable irregular dynamics in complex neural networks. Phys Rev Lett. 2008;100(4):048102. pmid:18352336
- 68. Monteforte M, Wolf F. Dynamical entropy production in spiking neuron networks in the balanced state. Phys Rev Lett. 2010;105(26):268104. pmid:21231716
- 69.
Engelken R, Monteforte M, Wolf F. Sparse chaos in cortical circuits. arXiv, preprint, 2024. https://arxiv.org/abs/241221188
- 70. Kadmon J, Sompolinsky H. Transition to chaos in random neuronal networks. Phys Rev X. 2015;5(4).
- 71. Koulakov AA, Hromádka T, Zador AM. Correlated connectivity and the distribution of firing rates in the neocortex. J Neurosci. 2009;29(12):3685–94. pmid:19321765
- 72. Adesnik H. Synaptic mechanisms of feature coding in the visual cortex of awake mice. Neuron. 2017;95(5):1147–1159.e4. pmid:28858618
- 73. Compte A, Brunel N, Goldman-Rakic PS, Wang XJ. Synaptic mechanisms and network dynamics underlying spatial working memory in a cortical network model. Cereb Cortex. 2000;10(9):910–23. pmid:10982751
- 74. Khona M, Fiete IR. Attractor and integrator networks in the brain. Nat Rev Neurosci. 2022;23(12):744–66. pmid:36329249
- 75. Harvey CD, Coen P, Tank DW. Choice-specific sequences in parietal cortex during a virtual-navigation decision task. Nature. 2012;484(7392):62–8. pmid:22419153
- 76. Zhou S, Masmanidis SC, Buonomano DV. Neural sequences as an optimal dynamical regime for the readout of time. Neuron. 2020;108(4):651–8.
- 77. Sussillo D, Barak O. Opening the black box: low-dimensional dynamics in high-dimensional recurrent neural networks. Neural Comput. 2013;25(3):626–49. pmid:23272922
- 78. Chaudhuri R, Fiete I. Computational principles of memory. Nat Neurosci. 2016;19(3):394–403. pmid:26906506
- 79. Constantinidis C, Funahashi S, Lee D, Murray JD, Qi X-L, Wang M, et al. Persistent spiking activity underlies working memory. J Neurosci. 2018;38(32):7020–8. pmid:30089641
- 80. Renart A, Moreno-Bote R, Wang X-J, Parga N. Mean-driven and fluctuation-driven persistent activity in recurrent networks. Neural Comput. 2007;19(1):1–46. pmid:17134316
- 81. Wallace E, Maei HR, Latham PE. Randomly connected networks have short temporal memory. Neural Comput. 2013;25(6):1408–39. pmid:23517097
- 82.
Renart A, Brunel N, Wang XJ. Mean-field theory of irregularly spiking neuronal populations and working memory in recurrent cortical networks. In: Gross LJ, Lenhart S, Maini PK, Etheridge AM, Voit EO, Safer H, et al., editors. Computational neuroscience: a comprehensive approach. Chapman & Hall; 2004, pp. 431–90.
- 83. Amit DJ, Brunel N. Model of global spontaneous activity and local structured activity during delay periods in the cerebral cortex. Cereb Cortex. 1997;7(3):237–52. pmid:9143444
- 84. Barak O, Tsodyks M. Working models of working memory. Curr Opin Neurobiol. 2014;25:20–4. pmid:24709596
- 85. Wimmer K, Nykamp DQ, Constantinidis C, Compte A. Bump attractor dynamics in prefrontal cortex explains behavioral precision in spatial working memory. Nat Neurosci. 2014;17(3):431–9. pmid:24487232
- 86. Goldman MS. Memory without feedback in a neural network. Neuron. 2009;61(4):621–34.
- 87. Recanatesi S, Pereira-Obilinovic U, Murakami M, Mainen Z, Mazzucato L. Metastable attractors explain the variable timing of stable behavioral action sequences. Neuron. 2022;110(1):139-153.e9. pmid:34717794
- 88. Hahnloser RHR, Kozhevnikov AA, Fee MS. An ultra-sparse code underlies the generation of neural sequences in a songbird. Nature. 2002;419(6902):65–70. pmid:12214232
- 89.
Jaeger H. The “echo state” approach to analysing and training recurrent neural networks-with an erratum note. 148. Bonn, Germany: German National Research Center for Information Technology; 2001.
- 90. Driscoll LN, Shenoy K, Sussillo D. Flexible multitask computation in recurrent networks utilizes shared dynamical motifs. Nat Neurosci. 2024;27(7):1349–63. pmid:38982201
- 91. Schuessler F, Mastrogiuseppe F, Ostojic S, Barak O. Aligned and oblique dynamics in recurrent neural networks. eLife. 2024;13.
- 92. Roudi Y, Latham PE. A balanced memory network. PLoS Comput Biol. 2007;3(9):1679–700. pmid:17845070
- 93. Rubin DB, Van Hooser SD, Miller KD. The stabilized supralinear network: a unifying circuit motif underlying multi-input integration in sensory cortex. Neuron. 2015;85(2):402–17. pmid:25611511
- 94. Kraynyukova N, Tchumatchenko T. Stabilized supralinear network can give rise to bistable, oscillatory, and persistent activity. Proc Natl Acad Sci U S A. 2018;115(13):3464–9. pmid:29531035
- 95. Podlaski WF, Machens CK. Approximating nonlinear functions with latent boundaries in low-rank excitatory-inhibitory spiking networks. Neural Comput. 2024;36(5):803–57. pmid:38658028
- 96. Abbott LF, Varela JA, Sen K, Nelson SB. Synaptic depression and cortical gain control. Science. 1997;275(5297):220–4. pmid:8985017
- 97. Ren M, Yoshimura Y, Takada N, Horibe S, Komatsu Y. Specialized inhibitory synaptic actions between nearby neocortical pyramidal neurons. Science. 2007;316(5825):758–61. pmid:17478724
- 98. Ryom KI, Treves A. Speed inversion in a Potts glass model of cortical dynamics. PRX Life. 2023;1(1).
- 99. Strata P, Harvey R. Dale’s principle. Brain Res Bull. 1999;50(5–6):349–50. pmid:10643431
- 100. Dubreuil A, Valente A, Beiran M, Mastrogiuseppe F, Ostojic S. The role of population structure in computations through neural dynamics. Nat Neurosci. 2022;25(6):783–94. pmid:35668174
- 101. Beiran M, Meirhaeghe N, Sohn H, Jazayeri M, Ostojic S. Parametric control of flexible timing through low-dimensional neural manifolds. Neuron. 2023;111(5):739–53.
- 102. Pfeffer CK, Xue M, He M, Huang ZJ, Scanziani M. Inhibition of inhibition in visual cortex: the logic of connections between molecularly distinct interneurons. Nat Neurosci. 2013;16(8):1068–76. pmid:23817549
- 103.
Miller KD, Palmigiano A. Generalized paradoxical effects in excitatory/inhibitory networks. bioRxiv, preprint, 2020.
- 104.
Palmigiano A, Fumarola F, Mossing DP, Kraynyukova N, Adesnik H, Miller KD. Common rules underlying optogenetic and behavioral modulation of responses in multi-cell-type V1 circuits. bioRxiv, preprint, 2020, pp. 2020–11.
- 105. Werbos PJ. Backpropagation through time: what it does and how to do it. Proc IEEE. 1990;78(10):1550–60.
- 106. Maass W, Natschläger T, Markram H. Real-time computing without stable states: a new framework for neural computation based on perturbations. Neural Comput. 2002;14(11):2531–60. pmid:12433288
- 107.
Pascanu R, Mikolov T, Bengio Y. On the difficulty of training recurrent neural networks. In: International Conference on Machine Learning. Pmlr; 2013, pp. 1310–8.
- 108. Yu Y, Si X, Hu C, Zhang J. A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput. 2019;31(7):1235–70. pmid:31113301
- 109. Krishnamurthy K, Can T, Schwab DJ. Theory of gating in recurrent neural networks. Phys Rev X. 2022;12(1):011011. pmid:36545030
- 110. Siegle JH, Jia X, Durand S, Gale S, Bennett C, Graddis N, et al. Survey of spiking in the mouse visual system reveals functional hierarchy. Nature. 2021;592(7852):86–92. pmid:33473216
- 111. Steinmetz NA, Aydin C, Lebedeva A, Okun M, Pachitariu M, Bauza M, et al. Neuropixels 2.0: a miniaturized high-density probe for stable, long-term brain recordings. Science. 2021;372(6539):eabf4588. pmid:33859006
- 112.
Laboratory IB, Banga K, Benson J, Bonacchi N, Bruijns SA, Campbell R. Reproducibility of in-vivo electrophysiological measurements in mice. bioRxiv, preprint, 2022, pp. 2022–5.
- 113. Baum LE, Petrie T. Statistical inference for probabilistic functions of finite state Markov chains. Ann Math Statist. 1966;37(6):1554–63.
- 114. Rabiner L, Juang B. An introduction to hidden Markov models. IEEE ASSP Mag. 1986;3(1):4–16.
- 115.
Welch G, Bishop G. An introduction to the Kalman filter. Chapel Hill, NC: University of North Carolina at Chapel Hill; 1995.
- 116. Bilmes JA. A gentle tutorial of the EM algorithm and its application to parameter estimation for Gaussian mixture and hidden Markov models. Int Comput Sci Institute. 1998;4(510):126.
- 117.
Yu BM, Cunningham JP, Santhanam G, Ryu S, Shenoy KV, Sahani M. Gaussian-process factor analysis for low-dimensional single-trial analysis of neural population activity. In: Advances in neural information processing systems, 2008.
- 118. Mante V, Sussillo D, Shenoy KV, Newsome WT. Context-dependent computation by recurrent dynamics in prefrontal cortex. Nature. 2013;503(7474):78–84. pmid:24201281
- 119. Pandarinath C, O’Shea DJ, Collins J, Jozefowicz R, Stavisky SD, Kao JC, et al. Inferring single-trial neural population dynamics using sequential auto-encoders. Nat Methods. 2018;15(10):805–15. pmid:30224673
- 120. Ladenbauer J, McKenzie S, English DF, Hagens O, Ostojic S. Inferring and validating mechanistic models of neural microcircuits based on spike-train data. Nat Commun. 2019;10(1):4933. pmid:31666513
- 121. Ostojic S. Two types of asynchronous activity in networks of excitatory and inhibitory spiking neurons. Nat Neurosci. 2014;17(4):594–600. pmid:24561997
- 122. Harish O, Hansel D. Asynchronous rate chaos in spiking neuronal circuits. PLoS Comput Biol. 2015;11(7):e1004266. pmid:26230679
- 123. Cimeša L, Ciric L, Ostojic S. Geometry of population activity in spiking networks with low-rank structure. PLoS Comput Biol. 2023;19(8):e1011315. pmid:37549194
- 124. Lajoie G, Lin KK, Shea-Brown E. Chaos and reliability in balanced spiking networks with temporal drive. Phys Rev E Stat Nonlin Soft Matter Phys. 2013;87(5):052901. pmid:23767592
- 125. Pillow JW, Shlens J, Paninski L, Sher A, Litke AM, Chichilnisky EJ, et al. Spatio-temporal correlations and visual signalling in a complete neuronal population. Nature. 2008;454(7207):995–9. pmid:18650810