The functional structure of cortical neuronal networks grown in vitro
Luı́s M. A. Bettencourt∗
T-7, Theoretical Division, MS B284, Los Alamos National Laboratory, Los Alamos NM 87545
Greg J. Stephens†
Lewis-Sigler Institute for Integrative Genomics and Center for Study of Brain,
Mind and Behavior, Princeton University, Princeton NJ 08544
arXiv:q-bio/0703018v1 [q-bio.NC] 7 Mar 2007
Michael I. Ham, and Guenter W. Gross
Center for Network Neuroscience, University of North Texas,
Dept. of Biological Sciences P.O. Box 305220 Denton, TX 76203-5220
(Dated: February 9, 2008)
We apply an information theoretic treatment of action potential time series measured with microelectrode arrays to estimate the connectivity of mammalian neuronal cell assemblies grown in vitro.
We infer connectivity between two neurons via the measurement of the mutual information between
their spike trains. In addition we measure higher point multi-informations between any two spike
trains conditional on the activity of a third cell, as a means to identify and distinguish classes of
functional connectivity among three neurons. The use of a conditional three-cell measure removes
some interpretational shortcomings of the pairwise mutual information and sheds light into the functional connectivity arrangements of any three cells. We analyze the resultant connectivity graphs
in light of other complex networks and demonstrate that, despite their ex vivo development, the
connectivity maps derived from cultured neural assemblies are similar to other biological networks
and display nontrivial structure in clustering coefficient, network diameter and assortative mixing.
Specifically we show that these networks are weakly disassortative small world graphs, which differ significantly in their structure from randomized graphs with the same degree. We expect our
analysis to be useful in identifying the computational motifs of a wide variety of complex networks,
derived from time series data.
I.
INTRODUCTION
Understanding and quantifying the dynamical mechanisms used by the nervous system to store and process information remains one of the greatest challenges to contemporary science. At present, broad outlines of the physical
mechanisms that underpin the basic functioning of single neurons and synapses in the brain are understood [1].
However, this detailed knowledge of individual units sheds little light into the origin of the unmatched computational
power of mammalian nervous systems, achieved despite characteristic operating times that are six orders of magnitude
slower than those of modern digital computers.
The computational nature of the brain lies therefore principally in the ensemble properties of neurons, synapses
and their emergent complex, dynamical networks. Over the last few years the interaction structure of many complex
systems has been mapped in terms of graphs, which can in turn be characterized using tools of statistical physics [2].
This approach has revealed broad classes of networks such as small world graphs [3] and scale free networks [4], which
occur across fields of study, from technological networks, such as the internet, to various biological and social systems.
The structural properties of these graphs, such as their degree distribution or their local transitivity, moreover, have
been suggested to result from optimization constraints [5, 6] or network growth dynamics [6], thus connecting graph
structure to operational definitions of function, independent of a system’s details. These lines of research provide
new quantitative insights, connecting the interaction structure of a complex systems to novel definitions of function
in fields where quantitative syntheses have been lacking.
The mammalian nervous system is, arguably, simultaneously the most complex and fascinating of all networks.
Studies that pursue its quantitative understanding in terms of structural, dynamical networks of simple units (neurons
and synapses) [7] are now beginning to be possible. Notwithstanding recent reconstructions for invertebrate cells [8],
which form relatively small numbers of connections to other cells, measuring the structure of living networks of
neurons directly is a difficult problem due to typical huge synaptic densities (about 108 per mm3 in cortex) and large
∗ Electronic
† Electronic
address: lmbett@lanl.gov
address: gstephen@princeton.edu
2
potential connectivity, with degree per cell of order 104 [9]. Much more accessible are electrophysiological signals action potentials or ’spikes’ - generated by neurons. The electro-physiological technology to measure these signals
over hundreds or even thousands of individual cells is now mature. Other new techniques based on fluorescence of
electrically active cells can also be used to map correlated activity, and putative connectivity [10]. Because action
potentials are the signals encoding information in the brain, the study of collective network activity in terms of
their time series is both natural and important. From these types of data ’networks of information’ can be inferred,
revealing how signals are shared between neurons, stored and collectively processed.
Here we construct networks of neurons, based on the information theoretic treatment of time series of action
potentials from cortical neuronal networks grown in vitro. Though these systems may differ substantially from an
intact brain they are interesting both as non-trivial starting points for in vivo analysis, and as examples of living
neuronal networks capable of non-trivial computation [11]. For these reasons neuronal networks in vitro have received
much recent attention in terms of their collective dynamical and statistical properties [12, 13, 14, 15, 16, 17, 18, 19].
Below we develop a methodology, based on the information theoretic treatment of multiple cell spike time series,
to map the structure of networks of neurons. We show that the resulting graph displays non-trivial structure shared
by other complex networks. Our approach also identifies collective functional modules consisting of groups of neurons
that enable different modes of information processing in nervous systems.
The remaining of this paper is organized as follows. In section II we give details of our experimental setting and
of the electrophysiological activity recordings. We also introduce the several information theoretic quantities to be
used subsequently to reconstruct a proxy connectivity network. We show how the use of several correlation and
information theoretic quantities taken together can lead to the identification of specific connectivity structures, when
multiple neurons are considered simultaneously. In section III we give the results of these methods applied to data.
Section IV analyzes the resulting graph in light of the properties of other complex networks. Finally we present the
outlook for the methods proposed here, particularly in determining the dynamical evolution of networks and their
collective response under driven conditions. In addition in Appendix A we prove a relation between definitions of
information theoretic quantities, introduced in section III.
II.
EXPERIMENTS AND DATA
We analyze the time series of mammalian (murine) spiking cortical networks living in culture over microelectrode
arrays [20, 21, 22]. These platforms, along with other types of neuronal culture [15, 16, 19], currently allow the most
extensive coverage of network elecrophysiological activity, in terms of both numbers of cells and length of recordings.
Naturally, cultured cell networks have presumably a simpler structure than their counterparts in vivo. Nevertheless
they are fully active neuronal networks, displaying many of the qualitative properties of developing mammalian tissue
[11]. The topology of networks of neurons formed in vitro remains largely uncharted, although recent work in this
direction has started to appear [8, 10].
Neuronal networks on microelectrode arrays (MEAs) are assemblies of mammalian nerve cells growing on fields
of non-invasive, substrate-integrated microelectrodes, see Fig. 1. The cells remain viable for over six months and
form spontaneously active networks. Tissue is dissected from mouse embryos, dissociated, and then seeded on MEAs.
Neural activity takes the form of action potentials (or spikes): short pulses (at which we record a time stamp) occurring
when a neuronal membrane potential crosses threshold. Each channel (up to 4 per electrode, scanned at 40 KHz) is
associated with a specific neuron via its action potential shape (spike sorting). Cell-electrode coupling is stable so
that wave-shapes can be followed with accuracy, for many channels and over many days at a time.
The results presented below were obtained from five hour recordings taken from three different cortical cultures,
spanning a range of sizes, with 20, 33 and 62 recorded cells. The three networks are mature (see e.g. [13]), with
ages in vitro of 34, 51 and 42 days, respectively. The number of cells recorded are a fraction (estimated roughly at
5-10%) of all active neurons in the culture. The total activity rates for the network were stable over the course of the
recordings. To construct binary states (zero/one) of multiple time series we sampled them at one half the spike rate
of the most active cell. This gives the highest entropy per measured bit for a temporally uncorrelated data stream.
III.
INFORMATION THEORETICAL ANALYSIS OF NEURONAL SPIKE TRAINS
Analyzing spike trains to understand how groups of neurons share and process information is challenging because
neuronal time series typically display strong temporal stochasticity. This is the result of both the spatial sparseness
of the collected signals of individual neurons, which receive and send processes to many other hidden (not recorded)
cells, and of the intrinsically non-linear and stochastic nature of neuronal and synaptic dynamics [1].
3
60
60
50
50
40
40
neuron
neuron
FIG. 1: A neuronal culture (right) over a microelectrode array. The uniform grid of points shows electrodes where neuronal
activity is measured. Left panels show individual neurons. White bars are 50 µm (center-to-center distances between electrodes
is 40 µm).
30
30
20
20
10
10
0
110
0
112
114
116
118
120
113.38
113.4
time [s]
113.42 113.44 113.46 113.48
time [s]
113.5
FIG. 2: Spike time series for the spontaneous (native) activity of a cortical culture with 62 recorded neurons. Much of the
activity takes place as part of network-wide collective events (left, vertical features), known as network bursts. The structure
of each neuron’s spike train observed over short time scales (right, network burst detail) shows strong stochasticity.
The noisy character of neuronal spike trains, see Fig. 2, suggests a probabilistic description of their time series, in
terms of sequences of active spiking periods and silences [23], as strings of zeros and ones. This standard decomposition
allows the analysis of neuronal time series as noisy channels via information theory. In particular we can measure how
the activity of one cell is informed by the state of others and determine their relative configuration, which will allow
us to reconstruct proxy connectivity networks mapping how information is shared, stored and processed collectively
in these ensembles.
A.
Mutual information between neurons
Mutual information provides a natural quantitative measure of the degree of statistical dependence between two
stochastic variables, free of assumptions about the nature of the underlying joint distribution. As such, mutual
information is a non-linear generalization of familiar correlation functions [24]. In the neuronal networks under
4
FIG. 3: Different connectivity arrangements of neurons have distinct roles in short term memory or in information processing.
These can be identified via the information theoretical analysis of multi-cell spike time series, with the former associated with
R > 0 and the latter R < 0, see text.
consideration here, large mutual information between suitably defined time series of two neurons will be taken as an
indication of connectivity between them. Subtleties that arise with this association will also be discussed and, at least
partially, resolved below.
We start by recalling the definition of the mutual information between two stochastic variables X,Y
X
p(x, y)
I(X; Y ) =
p(x, y) log2
,
(1)
p(x)p(y)
where p(x, y) is the joint distribution and p(x), p(y) are the single variable marginals. The mutual information
is semi-positive definite, and equal to zero if and only if the two variables X, Y are statistically independent, i.e.
p(x, y) = p(x)p(y).
The magnitude of the mutual information between two cells is dependent on the information content of the time
series of each cell, in terms of their Shannon entropy H(X), H(Y ). This becomes clear through consideration of the
relations
I(X; Y ) = H(X) − H(X|Y ) = H(Y ) − H(Y |X)
→ I(X; Y ) ≤ min [H(X), H(Y )] .
(2)
This bound, taken together with I(X; Y ) ≥ 0, implies that the normalized mutual information i(X; Y )
i(X; Y ) =
I(X; Y )
min [H(X), H(Y )]
(3)
is always in the range 0 ≤ i(X; Y ) ≤ 1, and can be used to measure the strength of the link between two cells. In
this spirit, more restricted measures of statistical dependence such as cross-correlation have also been used to identify
putative links between simultaneously recorded neurons [26].
B.
Multiple cell functional arrangements
We would like to not only reconstruct a network reflecting putative connectivity between neurons, but also to
identify how information is relayed or processed through neuronal activity. Fig. 3 distinguishes relaying and processing
computations for arrangements of three neurons. Sequential chains of cells can relay information, and are known to
play a part in short term memory in the brain [9, 25]. On the other hand neurons receive inputs from many other
cells, so that their spiking activity may behave as the (non-linear) function of multiple inputs. In this way individual
neurons naturally integrate activity from many sources and are able to process information. Here we show how these
two different types of structures can be distinguished given data streams for multiple cells. Consequently we can
determine which type of functional structure is more prevalent in cortical cultures. While three-cell networks are still
extremely small relative to even a single cortical column (approximately 105 neurons), they are complex enough to
illustrate the elementary properties of network computation.
The computations performed by networks of neurons can be analyzed via the consideration of appropriate multiinformations between groups of cells. To see this consider that the notion of mutual information, Eq. (1), naturally
generalizes to higher dimensions. For example I(X1 , X2 , . . . , Xn ) is the mutual information between the joint distribution of n variables p({xi }) and the product of each single variable’s distribution. This quantity is given explicitly
for n=3, for several examples in table I(b) . Similarly we will write I({X1 , X2 , . . . , Xn }; Y ) to denote the mutual
information between the joint state of n variables {Xi } and variable Y.
5
The use of multi-information between groups of variables is useful because it allows us to examine connectivity
arrangements involving more than just pairs of elements. In particular, for an arrangement of three elements, we
consider the conditional mutual information of two of the variables X, Y , subject to the state of a third Z [27],
I(X; Y |Z) =
X
x,y,z
p(x, y, z) log2
p(x, y|z)
.
p(x|z)p(y|z)
(4)
I(X; Y |Z) is a mutual information and as such is semi-positive definite and zero only when X, Y are independent
stochastic variables (each conditioned on Z). Thus I(X; Y |Z) tells us whether consideration of a third variable
supplies more or less information about the nature of the correlation between X and Y .
This property suggests the consideration of the difference
R(X, Y, Z) = I(X; Y ) − I(X; Y |Z)
(5)
as a compact means to characterize how knowledge of Z informs us of the state of {X, Y }. Although perhaps not
immediately obvious this quantity is identical to that recently proposed by Schneidman et al. [28, 29] as a measure
of redundancy of the knowledge of Z given X, Y . In their definition, for three variables
R(X, Y, Z) = I(X; Z) + I(Y ; Z) − I({X, Y }; Z).
(6)
We prove in Appendix A that the two definitions are equivalent, where we also show that R is fully symmetric under
permutations of its arguments, and consequently can be computed more parsimoniously than its definition may at
first suggest.
It is immediately clear from the definition of R that it can be positive, negative or zero. A positive R is equivalent
to I(X; Y ) > I(X; Y |Z) indicating that knowledge of Z does not improve (and may degrade) knowledge of {X, Y }
correlations. Conversely R < 0 indicates that knowledge of Z leads to more complete information about {X, Y }.
Finally R = 0 indicates that at least one of the three variables, is independent of the others. In this sense R = 0
denotes the absence of a genuine three variable structure. After Schneidman et al. [28] we refer to these three
situations as redundancy (R > 0), synergy (R < 0) and independence (R = 0). To better convey their meaning we
illustrate each of these cases with examples. A summary of what can be learned from these quantities about the
detailed connectivity mapping between variables is given in Table II.
1.
Independence: R = 0
It follows immediately from definition (4) that if Z is independent of {X, Y } then R = 0. Indeed if both X and
Y are statistically independent of Z, so that p(x|z) = p(x), p(y|z) = p(y) and naturally p(x, y|z) = p(x, y). Then
I(x, y|z) = I(x, y), which implies R = 0.
Moreover this is true in general, as expected from symmetry of R, if any of the three variables is independent of
the other two. We see this for the Boolean functions X and Y , of tables (a) and (b). In these cases we would infer
the existence of only one link from the only non-zero binary mutual information, and the absence of a three variable
structure, from R = 0.
2.
Synergy: R < 0
If Z is a function of both variables X, Y then knowledge of Z leads to greater information of the joint state of
{X, Y }. In this case R < 0, and we speak of synergy.
In order to illustrate this point we follow Schneidman et al. [28] in considering a set of Boolean functions of two
binary variables X, Y . Although these functions are much too stylized to be realized between neurons, they supply us
with the simplest useful illustrations. Table (a) shows several of the sixteen possible Boolean functions of two binary
variables.
In all the cases of Table (a), except for CHAIN to be discussed below, X and Y are independent random variables
(p0 = p1 = 1/2) and are thus statistically independent so that I(X; Y ) = 0. Information theoretic quantities for these
functions are given in Table (b).
Considering exclusively the mutual information in Table (b) as a starting point we see that we would infer correctly
that X, Z and Y, Z are connected but X, Y are not for the functions AND, OR (and X ⊣ Y and Y ⊣ X, not shown).
We may find it surprising however that for XOR we would infer no connectivity (all pairwise mutual informations
are zero). We find nevertheless a negative R and a nonzero I(X, Y, Z), which tell us that the three variable joint
6
(a)Examples of the relationships among three Boolean variables, showing
synergy, independence or redundancy, see text. Nontrivial logical circuits
(e.g. OR, AND) are examples of functions for which output is a joint
function of the two inputs and are thus associated with information
processing. Outputs that reflect only one of the inputs are trivial as three
cell structures (independence). CHAIN is an example of a third type of
relation where variables are not independent, but knowledge of any third
cell does not convey new information.
Inputs
x y
0 0
0 1
1 0
1 1
Synergy
Independence Redundancy
AND OR XOR Y⊣X X⊣Y
X Y
CHAINS
0
0
0
0
0
0
0
0
1
0
1
1
0
1
0
1
0
1
1
1
0
1
0
1
1
0
0
0
1
1
1
0
(b)Information theoretic quantities for the Boolean functions of Table (a). Inputs
for X, Y are taken at random so that H(X) = H(Y ) = 1, except for CHAIN, where
x=y, resulting in the only case where there is mutual information between the two
inputs. The functions Y⊣X and X⊣Y give the same results as AND and were thus
omitted for simplicity. Note that there are possible connectivity structures, like
XOR, which have zero mutual information between any two links, but still exist as
synergetic triplets.
Function
H(Z)
I(X;Y)
AND 2 − 34 ln2 (3)
0
OR
2 − 34 ln2 (3)
0
XOR
1
0
X
1
0
Y
1
0
CHAIN
1
1
3
2
3
2
I(X;Z)
I(Y;Z) I(X,Y,Z)
R
− 43 ln2 (3) I(X;Z) 2 − 43 ln2 (3) 1 − 34 ln2 (3)
− 43 ln2 (3) I(X;Z) 2 − 43 ln2 (3) 1 − 34 ln2 (3)
0
0
1
-1
1
0
1
0
0
1
1
0
1
1
2
1
TABLE I: Examples of 3 cell functional arrangements and their information theoretical signatures.
distribution is non-trivial. Thus we infer a genuine three cell object present, despite the absence of pairwise mutual
informations. Such structures would be missed under exclusively pairwise analysis.
Finally we notice that a fully connected diagram, in which variables are not Markovian (see below) is also synergetic
and therefore implies R < 0.
3.
Redundancy: R > 0
Finally we consider cases where knowledge of a third variable Z provides redundant information about to the state
of {X, Y }. The simplest examples are the functions CHAIN of Table (a). CHAIN can be seen as the limit of the
(anti)ferromagnetic interaction considered in Schneidman et al. [28], as the coupling β → ∞.
Clearly the state of the third variable, while not independent, does not add information about the state of the
other two. This situation is generic of connections organized as chains. An important and tractable situation is when
variables are Markovian, in which case a well known result, the information processing inequality, emerges. This
implies in turn that R > 0, generally, so it is worth explicit consideration.
If X, Y , Z form a Markov chain, denoted X → Y → Z, then their joint distribution can be written as [27]
p(x, y, z) = p(x)p(y|x)p(z|y).
(7)
This means in particular that X, Z are independent given Y , i.e. I(X; Z|Y ) = 0. In this situation the information
processing inequality [27] also tells tells us that for a Markov chain X → Y → Z
I(X; Y ) ≥ I(X; Z).
(8)
Now, writing I({X, Y }; Z) as (see Appendix A)
I({X, Y }; Z) = I(X; Z) + I(X; Y |Z) = I(X; Y ) + I(X; Z|Y )
(9)
7
and using I(X; Z|Y ) = 0 we conclude that
R = I(X; Y ) − I(X; Y |Z) = I(X; Z) ≥ 0.
(10)
The same result is obtained for the chain Z → Y → X, which from the point of view of information quantities is
equivalent to its reversed order considered above.
For the chain X → Z → Y , for which I(X; Y |Z) = 0, it follows immediately that R = I(X; Y ) ≥ 0. In particular
if the variables are indeed connected then I(X; Y ) > 0 strictly and R > 0 is a general property of Markov chains.
Finally we note that if Z drives both X and Y , as in X ← Z → Y then R > 0. We may suspect this from the fact
that Z is then a function of a single variable. Nevertheless formally this follows from the form of the joint distribution
p(x, y, z) = p(z)p(x|z)p(y|z) ⇒ p(x, y|z) =
p(x, y, z)
= p(x|z)p(y|z),
p(z)
(11)
which implies I(X; Y |Z) = 0 and therefore that R = I(X; Y ) > 0. In the case of R > 0 we therefore preserve only
the two strongest links among the three cells, a strategy also adopted in [30].
4.
Normalization of R
Just as for the mutual information the absolute value of R is dependent on the magnitude of the various multiinformations involved in its definition. We can place upper and lower bounds on R, so that a normalized quantity
r(X, Y, Z) can be constructed, that factors out these effects.
First note that it follows directly from the definitions (see also appendix A) that
R ≤ min [I(X; Y ), I(X; Z); I(Y ; Z)] .
(12)
So that if R > 0, r = R/min [I((X; Y ), I(X; Z); I(Y ; Z)]. Similarly when R < 0 then
R ≥ −min [I(X; Y |Z), I(X; Z|Y ), I(Y ; Z|X)] .
(13)
R ≥ −min [I({X, Y }; Z), I({X, Z}; Y ), I({Y, Z}; X)] ,
(14)
Note also that
but that the first bound is more stringent (see appendix A). Thus for R < 0 a normalized r is given by
r=
IV.
R
.
min [I(X; Y |Z), I(X; Z|Y ), I(Y ; Z|X)]
(15)
THE STRUCTURE OF NETWORKS OF SPIKING NEURONS IN VITRO
In this section we show how to use the information theoretic diagnostics of the previous section to construct
connectivity graphs for neuronal ensembles.
A.
Graph construction in practice
The joint consideration of R and of pairwise mutual informations leads to a classification of the connectivity
diagrams of any three nodes. The summary of these results, which can be implemented as an algorithm to map the
network connectivity, is summarized in Table II. An efficient algorithm computes first the value of R for a chosen
triplet of variables. This determines how many links to attribute. The strength of these links, whenever appropriate,
are determined by mutual information between pairs of variables. Whenever the consideration of several trios implies
different weights for the same connection between two neurons, the results are averaged.
To determine whether measured values of R and I are significant, we compute their values according to a null model
where every cell i in the network produces spikes at the same average rate λi as observed for each neuron’s actual spike
train, but whose temporal structure is otherwise random. This corresponds to a number of spikes per unit time given
by a Poisson distribution with rate λi or equivalently, as we do in practice, a time interval between consecutive spikes
given by an exponential distribution. Under perfect estimation this Poisson model generates statistically independent
spike trains between any two cells. In practice estimation is imperfect and this procedure leads instead to very small,
but non-zero, values of R and I. Thus for each trio of cells we take the observed R to be significant if it is larger
than three times that measured via the Poisson model. Otherwise the trio is considered statistically independent with
R = 0. We proceed similarly for the mutual information measured for any two cells.
8
No 3-cell connected structure
R=0
No connections:
∀i,j∈{1,2,3} I(Xi , Xj ) = 0
1 connection:
∃!i,j∈{1,2,3} I(Xi , Xj ) 6= 0
There is a 3-cell connected structure
2 connections:
R > 0 (redundancy)
X ↔Y ↔Z
X↔Z↔Y
X←Z→Y
Markov:
Markov:
I(X; Y ) ≥ I(X; Z) I(X; Z) ≥ I(X; Y )
I(X; Z|Y ) = 0
I(X; Y |Z) = 0
I(X; Y |Z) = 0
R = I(X; Z)
R = I(X; Y )
R = I(X; Y )
R 6= 0
3 connections:
R < 0 (synergy)
X → Z ← Y : Z = f (X, Y )
Each variable is a function of all others.
TABLE II: The joint consideration of the values of R and of the binary mutual informations I lead to a classification of the
connectivity arrangements between three stochastic channels. For R > 0 only the two strongest links are adopted. For R < 0
we adopt three links as a general representation of the specific functional interdependence between the three cells.
B.
Graph Theoretic Analysis
Next we characterize the structural properties of the resulting networks in terms of graph theoretic quantities.
Although all three graphs are small in terms of numbers of nodes we show that results about their network structure
are consistent for the different networks. Figure 4 shows the weighted graphs for two of these networks. Figure 5
shows the connectivity weight distribution and the histogram of r for all cell trios.
The distribution of connectivity weights is well-described by an exponential. This implies that most links are weak,
with a few stronger ones tending to inter-connect subsets of cells. The distribution of r shows that most trios are
associated with redundant structures that may be involved in short term memory functions. Nevertheless a number
of trios display r < 0, indicating the existence of non-trivial functional relations between measured cells, that may
play a role in information processing.
To compare the structure of these weighted graphs quantitatively to other complex networks we transform them
into binary graphs, with each link set to zero or one if its strength lies below or above a given threshold, respectively.
Table III shows the average clustering coefficient, diameter and Pearson coefficient for degree-degree correlations for
the three networks as functions of the threshold, and compares them to results from one hundred realizations of the
standard null model, viz. a network with the same node degrees but with randomized connections.
We see that in every case the neuronal graphs display significantly higher clustering than the null networks. They
also show small diameters of the largest component d, of order d ∼ ln(N ), where N is the number of nodes. These
two properties together indicate that cortical networks in vitro, like the neuronal network of the nematode [3], are
small world graphs.
Finally we observe that cortical networks in vitro are weakly disassortative, i.e. that the Pearson coefficient associated with degree-degree correlations is, in most cases, negative but small [31]. This indicates that nodes of dissimilar
degree preferentially connect to each other. Together these measures show that networks of neurons, constructed via
the information shared by groups of cells, are significantly different from randomized graphs, but share instead many
structural properties with other complex networks [3, 5, 6].
V.
DISCUSSION AND CONCLUSIONS
The understanding of the functional network structures used by living nervous systems to store and process information is key to unravelling the mechanisms of their computational power. It is now possible to measure the activity
of networks of individual neurons and to start to identify such informational structures, their development over time
and their role in response to external stimuli (see e.g [32]). Here we have shown how time series for networks of
9
FIG. 4: Weighted graphs for network of a) 62 and c) 33 recorded neurons. Panels b) and d) show the corresponding networks
selecting only the links with strength above 0.4. The graphs are characterized by a large number of weak links that connect
almost every pair of cells, and of fewer stronger links that tend to interconnect subsets of neurons. Node size is drawn
proportionally to its total connectivity weigh.
Size
Threshold Av. Clustering
Diameter
Assortativity
(recorded neurons N)
[0, 1]
Actual Random Actual Random Pearson coeff.
0.1
0.89 0.52± 0.04
2
2.97± 0.03 -0.150 ± 0.025
N=20
0.3
0.57 0.24 ± 0.06
4
4.15 ± 0.55 -0.040 ± 0.048
0.5
0.32 0.05 ± 0.10
1
2.10 ± 0.70 -0.200 ± 0.100
0.1
0.83 0.48± 0.02
3
2.97± 0.03 -0.114 ± 0.001
N=33
0.3
0.61 0.26 ± 0.04
4
3.99 ± 0.30 -0.033 ± 0.026
0.5
0.07 0.00 ± 0.02
2
2.98 ± 0.86 0.064 ± 0.053
0.1
0.82 0.51± 0.01
3
2.97± 0.03 -0.025 ± 0.004
N=62
0.3
0.60 0.37 ± 0.02
4
4.00 ± 0.20 -0.093 ± 0.004
0.5
0.43 0.24 ± 0.02
3
3.93 ± 0.18 -0.102 ± 0.006
TABLE III: Graph theoretic measures for each of three cortical neuronal networks and as functions of the weight threshold,
and their comparisons with the average and standard deviation obtained from a graph with the same degree but randomized
connections. Cultured neuronal networks are clearly distinguishable from randomized networks with the same degree. They
are small world graphs with weakly disassortative degree correlations.
spiking neurons can be treated statistically and analyzed in terms of informational theoretic quantities to reveal how
information is shared, relayed and processed by groups of cells. This analysis generates networks of shared information, which are proxies for the physical connectivity between neurons. In fact, knowledge of this type of connectivity
and its dynamics may well suffice to reveal the underpinnings of collective computation in networks of neurons.
We have also shown that consideration of information theoretic quantities beyond two cells can identify structures
that are naturally higher order, revealing for example a cell whose output is a non-linear function of several others.
The specificity of such structures is lost when we represent them as sets of binary links, as we have done here to
10
0.2
0.25
0.18
0.16
0.2
Probability
Probability
0.14
0.12
0.1
0.08
0.15
0.1
0.06
0.04
0.05
0.02
0
0
0
0.2
0.4
0.6
0.8
1
Weights
-1
-0.5
0
0.5
1
r
FIG. 5: The probability distribution for weights of the inferred network of 62 recorded cells (left). This distribution is well
described by an exponential (dashed line), with average connectivity weight 0.294 ± 0.019. The distribution of normalized r for
the same network (right). Most trios show information redundancy (r > 0), while a smaller number display synergy (r < 0).
The central peak at r = 0, with probability p = 0.091, consists of a fraction of 70% of truly independent trios (R = 0), 22%
with negative r ≥ −0.05 and 8% with positive r ≤ 0.05. The other two networks exhibit quantitatively consistent results, but
with poorer statistics. These results are suggestive that multi neuron connectivity structures associated with both information
processing (R < 0) and short term memory (R > 0) are present in cultured neuronal networks in vitro.
produce standard network analyses. The systematic mapping and characterization of these functional modules goes
beyond the scope of the present work, and will be pursued elsewhere.
Neglecting such functional specificity we have built networks of shared information, which we showed to be weakly
disassortative small world graphs, in agreement with e.g. the detailed neuronal network of C. elegans [3], but in
contrast to networks of correlation between large scale brain regions, identified via fMRI measurements [7]. The
distribution of weights is approximately exponential and multi neuron structures associated with information relaying
dominate over a smaller number of cell arrangements associated with information processing. Regardless of details of
specific proportions for these cellular arrangements we showed, by identifying multi-neuron structures with R > 0 and
R < 0, that neuronal modules associated with short term memory and information processing co-exist in disassociate
neuronal cultures in vitro. How these structures and their generalizations to larger cell numbers may be harnessed to
perform reliable computation remains perhaps the clearest single challenge for future studies of information processing
in neuronal networks living in vitro.
These results raise several other interesting questions for future research. It is natural to expand the analysis
presented here to include multi-information quantities involving more cells, in order to identify the functional role
of more complex neuronal arrangements. It is also interesting to investigate their change over time, and under
electrophysiological stimulation. Finally, this type of analysis can be applied to any distributed information processing
system, natural or artificial, and should prove useful in revealing the computational structures that constitute other
types of nervous tissue, in vivo and in vitro, and in creating mathematical models and engineered systems that mimic
them.
Acknowledgments
This work was carried out in part under the auspices of the National Nuclear Security Administration of the U.S.
Department of Energy at Los Alamos National Laboratory under Contract No. DE-AC52-06NA25396. Additional
support from LDRD-ER project 20050411. LMAB thanks Jim Crutchfield, Aric Hagberg and Pieter Swart for useful
discussions.
Appendix A
In this appendix we prove the equivalence of the definitions of R, Eqs. (5) and (6). This equality establishes that
the measure R of redundancy, proposed by Schneidman et al. in Ref. [28], can be written, for three variables in terms
of the difference between the mutual information of a pair of variables and its mutual information conditional on a
third cell (5).
11
We begin with the definition of R given two variables X, Y, and upon consideration of a third Z, which is a candidate
for having connections with X, Y or may be an external stimulus
R = I(X; Z) + I(Y ; Z) − I({X, Y }; Z),
(16)
where, as usual, the semicolon separates the states being compared, so that the last mutual information is between
the joint state of {X, Y } and Z. We will now prove that R, defined in (16) can also be written as
R = I(X; Y ) − I(X; Y |Z).
(17)
Starting from (16) we use the chain rule for mutual information [27]
I({X1 , X2 , . . . , Xn }; Z) =
n
X
I(Xi ; Z|Xi−1 Xi−2 , . . . , X1 ).
(18)
i=1
Specifically for three variables (n=2)
I({X, Y }; Z) = I(X; Z) + I(Y ; Z|X) = I(Y ; Z) + I(X; Z|Y ),
I({Y, Z}; X) = I(X; Y ) + I(X; Z|Y ) = I(X; Z) + I(X; Y |Z),
I({X, Z}; Y ) = I(Y ; Z) + I(X; Y |Z) = I(X; Y ) + I(Y ; Z|X).
(19)
(20)
(21)
These expression can now be used in (16) to give
I(X; Z) + I(Y ; Z) − I({X, Y }; Z) = I(Y ; Z) − I(Y ; Z|X) = I(X; Z) − I(X; Z|Y ).
(22)
Similarly we can write
I(X; Y ) + I(Z; Y ) − I({X, Z}; Y ) = I(X; Y ) − I(X; Y |Z) = I(Y ; Z) − I(Y ; Z|X),
I(Y ; X) + I(Z; X) − I({Y, Z}; X) = I(Y ; X) − I(Y ; X|Z) = I(Z; X) − I(Z; X|Y ).
(23)
(24)
These identities taken together show that the definition of R is symmetric under variable permutations and in
particular proves our proposition
R = I(X; Z) + I(Y ; Z) − I({X, Y }; Z) = I(X; Y ) − I(X; Y |Z).
(25)
As a corollary we see that R is fully symmetric, i.e.
R
= I(X; Z) + I(Y ; Z) − I({X, Y }; Z) = I(X; Y ) + I(Z; Y ) − I({X, Z}; Y )
= I(Y ; X) + I(X; Z) − I({Y, Z}; X)
= I(X; Y ) − I(X; Y |Z) = I(X; Z) − I(X; Z|Y ) = I(Y ; Z) − I(Y ; Z|X).
(26)
Thus R can be computed efficiently through the most convenient set of estimated distributions.
In the case of more than three variables analogous expression for R in terms of conditional mutual informations can
be written easily, through (18), but become less elegant as the distributions are now conditional on several variables.
[1] See e.g. C. Koch, Biophysics of Computation : Information Processing in Single Neurons, (Oxford University Press, Oxford,
U.K. 1999).
[2] R. Albert and A.-L. Barabsi, Rev. Mod. Phys. 74, 47 (2002)
[3] D. J. Watts and S. H. Strogatz, Nature 393, 440-442 (1998).
[4] A.-L. Barabsi, and R. Albert. Science, 286, 509 (1999).
[5] See collected papers in M. Newman, Albert-Laszlo Barabasi, Duncan J. Watts, The Structure and Dynamics of Networks,
(Princeton University Press Princeton, NJ, USA, 2006).
[6] S. N. Dorogovtsev, J. F. F. Mendes, Evolution of Networks: from biological networks to the Internet and WWW, (Oxford
University Press, Oxford, U.K. 2003)
[7] Activity patterns of the brain over much larger scales are now accessible via fMRI measurements. See a corresponding
network reconstruction in V. M. Eguiluz, D. R. Chialvo, G. A. Cecchi, M. Baliki, and A. V. Apkarian, Phys. Rev. Lett.
94, 018102 (2005).
[8] O. Shefi, I. Golding, R. Segev, E. Ben-Jacob, and A. Ayali, Phys. Rev. E 66, 021905 (2002).
12
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
[24]
[25]
[26]
[27]
[28]
[29]
[30]
[31]
[32]
M. Abeles, Corticonics: Neural Circuits of the Cerebral Cortex (Cambridge Univ. Press, Cambridge, 1991).
L. C. Jia, M. Sano, P.-Y. Lai, and C. K. Chan, Phys. Rev. Lett. 93, 088101 (2004).
S. Marom, and G. Shahaf, Q. Rev. Biophys. 35 63 (2002).
R. Segev, Y. Shapira, M. Benveniste, E. Ben-Jacob, Phys. Rev. E 64 011920 (2001).
T. Tateno, A. Kawana, and Y. Jimbo, Phys. Rev. E 65 051924 (2002).
R. Segev, M. Benveniste, E. Hulata, N. Cohen, A. Palevski, E. Kapon, Y. Shapira, and E. Ben-Jacob, Phys. Rev. Lett. 88
118102 (2002).
J. M. Beggs and D. Plenz, J. Neurosci. 23 11167 (2003).
J. M. Beggs and D. Plenz, J. Neurosci. 24, 5216 (2004).
W. Hulata, I. Baruchi, R. Segev, Y. Shapira, and E. Ben-Jacob, Phys. Rev. Lett. 92 198105 (2004).
R. Segev, I. Baruchi, E. Hulata, and E. Ben-Jacob, Phys. Rev. Lett. 92 118102 (2004).
C. Haldeman, and J. M. Beggs, Phys. Rev. Lett. 94 058101 (2005).
G. W. Gross, and F. U. Schwalm, J. Neurosci. Methods, 52 73 (1994).
E.W. Keefer, A. Gramowski, and G. W. Gross, J. Neurophysiol. 86 3030 (2001).
G. W. Gross, and K. V. Gopal, in Advances in Network Electrophysiology using Microelectrode Arrays, Taketani (Eds.),
(Kluwer Pulishing Company, Dordrecht, Holland, 2006).
F. Rieke, D. Warland, R. R. de Ruyter Van Steveninck, W. Bialek, Spikes: Exploring the Neural Code, (MIT Press,
Cambridge, MA, 1997).
W. Li, J. Stat. Phys. 60, 823 (1990).
M. Diesmann, M. O. Gewaltig, and A. Aertsen, Nature 402, 529 (1999).
D. H. Perkel, G. L. Gerstein and G. P. Moore, ”Neuronal spike trains and stochastic point processes II: Simultaneous
spike trains”, Biophys. J. 7 419 (1967); A. Aertsen and G. L. Gerstein, ”Evaluation of Neuronal Connectivity: Sensitivity
of cross-correlation”, Brain. Res. 340 341 (1985); A. Aertsen, G. L. Gerstein, M. K. Habib and G. Palm, ”Dynamics of
Neuronal Firing Correlation: Modulation of Effective Connectivity”, J. Neurophys. 61 900 (1989); G. Palm, A. Aersten
and G. L. Gerstein, ”On the significance of correlations among neuronal spike trains”, Bio. Cyber. 59 1 (1988); C. D.
Brody, ”Correlations without synchrony”, Neural Comp. 11 1537 (1999).
T. M. Cover, and J. A.. Thomas, Elements of Information Theory (Wiley, New York, 1991).
E. Schneidman, S. Still, M. J. Berry II, and W. Bialek, Phys. Rev. Lett. 91, 238701 (2003).
E. Schneidman, W. Bialek, M. J. Berry II, J. Neurosci. 23, 11539 (2003).
A. A Margolin, I. Nemenman, K. Basso, C. Wiggins, G. Stolovitzky, R. D. Favera, and A. Califano, BMC Bioinformatics,
7 (Suppl 1) S7 (2006).
M. E. J. Newman, Phys. Rev. Lett. 89, 208701 (2002).
E. Schneidman, M. J. Berry II, R. Segev, W. Bialek, Nature 440, 1007-1012 (2006)
View publication stats