Académique Documents
Professionnel Documents
Culture Documents
Habilitationsschrift
zur Erlangung der Lehrbefugnis
fur Theoretische Physik
am Fachbereich 1: Physik/Elektrotechnik
der Universitat Bremen
vorgelegt von
Christian Eurich
1
There is, of course, another reason for this name: Theoretical neurophysics is made by
physicists. By definition, everything a physicist does is called Physics.
2
Citations in the text referring to one of these reprints are marked by an additional L (for
List of reprints) after the year of publication, for example Eurich et al. (1995:L).
iii
iv
Contents
v
vi CONTENTS
Bibliography I-134
I-1
Chapter 1
Introduction
I-3
I-4 CHAPTER 1. INTRODUCTION
can be better understood when asking for what purpose they have evolved. Puta-
tive functions include accurately perceiving the environment, performing mean-
ingful motor actions, and developing abstract mental concepts of future events
and actions.
A paradigm which has proven to be fruitful in the last decades is to regard the
brain as a signal-processing system. As far as perception is concerned, one can look
for neurons, neural populations and brain areas which become active whenever
certain sensory stimuli are presented during an electrophysiological experiment
(Figure 1.1). Within the functional framework, such correlations between stimuli
and neural behavior are interpreted in terms of representation and coding : neural
firing is interpreted as a response to the stimulus in the sense that the neurons
encode the stimulus. Typical question to be addressed are:
Which part of the neural response carries information?
What is the nature of the neural code? Is it a single-neuron code or a
population code?
Which properties of neurons and neural populations (firing rates, tuning
widths, correlations in the activity, etc.) yield a high sensory resolution or
motor accuracy?
What is a good strategy to discriminate simultaneously presented stimuli?
Why do neural responses appear noisy?
Do mental states like arousal or attention modify neural responses and
thereby the encoding properties?
Theoretical work in this field includes the development of data analysis tech-
niques capable of integrating the responses of large neural populations, the mod-
eling of neural populations in order to study their representational properties,
and the evaluation of mathematical (esp. statistical) methods to account for the
corresponding signal processing. Mathematically, the stochastic nature of the
neural responses suggests the use of information theory (e. g., Rieke et al., 1997;
Brunel and Nadal, 1998; Borst and Theunissen, 1999) and statistical estima-
tion theory including maximum-likelihood methods (Pouget et al., 1998), Fisher
information (e. g., Paradiso, 1988; Seung and Sompolinsky, 1993; Brunel and
Nadal, 1998; Zhang and Sejnowski, 1999; Eurich and Wilke, 2000:L; Wilke and
Eurich, 2001:Lb; Bethge et al., 2002), and Bayesian estimation (e. g., Abbott,
1994; Sanger, 1996; Oram et al., 1998; Zemel et al., 1998; Zhang et al., 1998a).
Other mathematical approaches are kernel methods (Stanley et al., 1999; Wess-
berg et al., 2000) and vector-based methods (Georgopoulos et al., 1986; Scott
et al., 2001) to reconstruct stimuli from neural responses, as well as neural net-
work simulations (Eurich et al., 1995:L; Thorpe and Gautrais, 1997; Van Rullen
et al., 1998).
I-6 CHAPTER 1. INTRODUCTION
Sensory
input
}
Neuron #
Correlations?
Motor
behaviour
0 1.5
Time (s)
world.
If the brain is regarded as a dynamical system, a different aspect of brain
processes is emphasized. This becomes clear when studying the well-known com-
pilation of visual (mostly cortical) areas of the macaque monkey by Felleman and
Van Essen (1991); see Figure 1.2. In their scheme, Felleman and Van Essen also
Figure 1.2: Felleman and Van Essen (1991) compiled this scheme of brain areas in
the macaque monkey that were known at that time to process visual information.
Included are also all identified anatomical connections between these areas.
Perception
}
Neuron #
Motor
planning Correlations?
and
output
0 1.5
Time (s)
Mental
states
Figure 1.3: A modified scheme correlates perception, motor planning and execu-
tion, and mental states rather than objects and features of the real world
with neural activity. The image on the lower left symbolizes attention: humans
and animals are able to attend a certain object (gray dot) while neglecting other
objects (black dot).
I-10 CHAPTER 1. INTRODUCTION
Chapter 2
Neural Coding
2.1 Introduction
This chapter deals with the functional approach to phenomena of the nervous
system. The starting point is the framework of Figure 1.3 or in a simpler
interpretation Figure 1.1. Once a correlation between a subset of the internal
world (perception of a certain class of stimuli, a class of motor actions, or some
mental phenomenon) and the neural world (usually some aspect of the spiking
behavior of a single neuron or a population of neurons) has been identified, one
may speak of a neural representation of the considered phenomena of the internal
world.1 For example, the direction of three-dimensional arm movements has
been suggested to be represented in the firing rates of a population of neurons in
monkey motor cortex (Georgopoulos et al., 1986; Wessberg et al., 2000). The set
of neural activities correlating with the subset of the internal world defines an
alphabet, and the neural code is a mapping assigning each element of the subset
of the internal world a neural activity, i. e., a letter of the alphabet. In the case
of the monkey arm movements, the alphabet consists of the set of possible neural
firing rates of the population of neurons in motor cortex, and the code maps
each arm movement to a population activity; the mapping is computed via the
so-called population vector (Georgopoulos et al., 1986; Wessberg et al., 2000).
It should be noted that the interpretation of the issue of neural coding is con-
troversially discussed in the neuroscience community. As outlined in Chapter 1,
it assumes that the brain at least partially represents an outside world. Further-
more, a code makes sense only if encoded messages are decoded somewhere; it is
not clear where and in which form this happens in the nervous system. Last not
least, the notion of a neural code is not used consistently among scientists. See
Section 2.3.4 for further discussion. Overviews over the field of neural coding are
1
This working definition of representation is fairly naive from a philosophical point of view.
However, this is the way it is often used in neuroscience. For a more detailed discussion of the
notion of representation see Peschl (1994); Ziemke and Breidbach (1996).
I-11
I-12 CHAPTER 2. NEURAL CODING
given in Rieke et al. (1997); Abbott and Sejnowski (1999); deCharms and Zador
(2000).
In Sections 2.2 and 2.3, a brief overview of the field of neural coding is given.
Section 2.2 summarizes the most common experimental and theoretical methods,
scientific objectives in the context of coding, as well as optimization principles
and biological constraints. Section 2.3 provides a list of neural codes which have
been suggested in the last decades and discusses the question which code might
be realized in the nervous system. The remainder of this chapter, Section 2.4,
integrates my own work in the given framework and outlines its most important
results.
the maximal neural firing rate which limits the bandwidth of signal trans-
duction;
delays in signal integration and signal conduction which shapes the spatio-
temporal flow of information;
the topology of brain areas which may, for example, determine the topology
of neural maps (Wolf et al., 1996);
the amount of neural tissue, i. e., the number of neurons and glial cells,
the total length of cables, the thickness of myelin sheath, the number
of synapses on a restricted neural surface area etc. These constraints are
related to both energy consumption and brain size;
the size of the brain which is likely to be limited by the size of the head of
a baby during birth.
Apart from the large number of experimental situations that fit the single-
object case, this issue is also important because it may serve as a testbed for
neural encoding strategies under various constraints; for a more detailed discus-
sion of the relevance of single-object coding, see Section 2.4.1.
Questions that are usually addressed are:
Which part of a neural response carries information about the given object?
That is, which code can be identified in the neural system?
How well can the object be localized, given a neural response? That is, what
is the encoding accuracy obtained with a single cell or a cell population?
Theoretical approaches to computing the resolution obtained by a population of
neurons are described in Section 2.4 for two different neural codes.
On the theoretical side, signal detection theory was used to account for dis-
crimination performance (e. g., Balakrishnan (1999); Verghese (2001); Tougaard
(2002); see Young and Calvert (1974); Green and Swets (1988); Ripley (1996);
Duda et al. (2001) for introductory texts). Discrimination thresholds for neu-
ral populations with overlapping receptive fields were considered in Snippe and
Koenderink (1992a,b) for uncorrelated and correlated neural activity, respec-
tively. Woesler (2001) considered the problem of object segmentation in the
salamander visual system. Zemel et al. (1998) employ Bayesian estimation theory
in a scheme for encoding and decoding distributions of stimuli. Eurich (2003:L)
extends a Fisher information formalism for computing the localization accuracy
of a single stimulus to include the simultaneous presentation of multiple stimuli;
see Section 2.4.4 for details.
Neural responses usually appear structured in time. For example, neurons may
show spike frequency adaptation (Smith et al., 2001; Benda and Herz, 2003) or
bursting (Lisman, 1997). Non-uniform behavior in time may be observed for two
different reasons:
I-18 CHAPTER 2. NEURAL CODING
The stimuli themselves may change in time. This is true for most nat-
ural stimuli and has been considered in electrophysiological experiments
(e. g., Bialek et al., 1991; Zhang et al., 1998a; Borst and Theunissen, 1999;
Wessberg et al., 2000).
The two causes for non-stationary behavior may occur concurrently. For example,
Worgotter et al. (1998) present flashed stimuli to anaesthetized cats and record
responses from LGN neurons. They observe phasic bursts followed by tonic re-
sponses. Subsequently, LGN target cells in visual cortex show a fast narrowing
of receptive fields. Such physiological behavior may reflect both time-changing
stimuli and time-changing neural processing.
The temporal behavior of orientation tuning in macaque primary visual cor-
tex has been studied by Ringach et al. (1997a) using reverse correlation in the
orientation domain and later also by Gillespie et al. (2001); the latter find the
same properties of V1 cells (preferred orientation, tuning width) in early and late
reponses. In the auditory system, receptive field changes have been reported by
Edeline et al. (2001) who studied evoked responses in cortical cells of guinea-
pigs during natural sleep. Spatiotemporal receptive field properties in the cat
and monkey visual system are studied by Dinse et al. (1990) and Bredfeldt and
Ringach (2002), respectively. See also Worgotter and Eysel (2000) for a review
article about the dynamics of receptive fields.
While the experiments cited in the previous paragraph are not accompanied
by investigations on neural coding, Victor (1999) reviews temporal aspects of
coding in the retina and LGN. Fairhall et al. (2001) examine the dynamics of a
neural code in the context of stimuli whose statistical properties are themselves
evolving dynamically; they find adaptation processes on time scales ranging from
tens of milliseconds to minutes.
Data analysis techniques for dynamic stimuli include reverse correlation (Hida
and Naka, 1982; Hida et al., 1983; Eckhorn et al., 1993; Ringach et al., 1997a,b;
Theunissen et al., 2001) and the reconstruction of stimuli (e. g., Zhang et al.,
1998a; Brown et al., 1998; Wessberg et al., 2000; Jakel, 2001; Bringmann, 2002;
Wiener and Richmond, 2003). For example, Zhang et al. (1998a) use Bayesian
techniques to reconstruct the path of a rat from the firing of its hippocampal
place cells; a considerable improvement of the reconstruction is obtained in a
two-step algorithm where information on the previous reconstruction step enters
the actual positional estimate.
2.2. TASKS, CONSTRAINTS, OPTIMIZATION PRINCIPLES, I-19
OBJECTIVES AND METHODS
In the last two decades, there has been an increasing interest in neural correlates
beyond simple sensory stimuli or motor actions. It is now well established that
cortical neural activity depends on mental states such as attention or the intention
to solve a certain task. To give an example, Toth and Assad (2002) have found
neurons in monkey lateral intraparietal area that encode the color of a stimulus
only if color is a feature that is relevant for a task at hand; otherwise, color
sensitivity is virtually absent in that brain area. The correlation between mental
states and neural activity is reflected in the extended scheme of Figure 1.3.
Attention. Attention comprises all mechanisms that are used by the brain
to select and modulate behaviorally relevant information for further processing
(Chun and Wolfe, 2001). The selection process involves an enhancement of the
neural representation of the attended stimulus and a suppression of distractor
stimuli (e. g., Kastner et al., 1998). The neural ensemble representing the at-
tented stimulus will therefore be better visible for subsequent processing stages
(Desimone and Duncan, 1995).
A mainstream interpretation of the function of attention is the grouping of
single features to coherent objects (e. g., Treisman and Gelade, 1980; Wolfe
and Bennett, 1997). However, attending an object also increases the spatial
resolution (Yeshurun and Carrasco, 1998, 1999) and accelerates the rate of visual
information processing (Carrasco and McElree, 2001). In the context of such
observations, the traditional view on attention has recently been complemented
by statistical and informational aspects (Dayan et al., 2000; Verghese, 2001).
A number of electrophysiological studies reveal effects of the attentional state
on the response properties of neurons in visual cortex. Earlier work reports the
influence of attention only on later stages of the visual pathway (e. g., Moran
and Desimone, 1985; Maunsell, 1995; Treue and Maunsell, 1996; for a review, see
Groh et al., 1996). Motter (1993), however, and more recent publications find
attentional modulations also in the early visual system (e. g., Luck et al., 1997;
Roelfsema et al., 1998; Reynolds et al., 1999; reviews are given by Posner and
Gilbert, 1999; Olson, 2001; Treue, 2001). The observed attentional effects include
a change in neural synchrony (Steinmetz et al., 2000; Fries et al., 2001) and an
increase of the gain of tuning curves (McAdams and Maunsell, 1999; Treue and
I-20 CHAPTER 2. NEURAL CODING
Trujillo, 1999). A sharpening of tuning curves has also been suggested (Lee et al.,
1999).
Modelling studies referring to the statistical and informational aspects of at-
tention include Bayesian (Dayan and Zemel, 1999) and Fisher information com-
putations (Nakahara et al., 2001; Eurich, 2003:L); the latter is further described
in Section 2.4.4. In a current project, Eurich and Freiwald (2001) employ multi-
cell recordings in monkey visual cortex and reconstruction techniques to identify
neural correlates of object-based attention and to relate them to psychophysical
behavior.
Consciousness. The first topic is the search for neural correlates of conscious-
ness which has been addressed in numerous publications in philosophy and neu-
roscience (e. g., Chalmers, 1995; Shear, 1997; Metzinger, 2000). More recently,
Crick and Koch (2003) have suggested a framework for the neural basis of con-
sciousness whose elements can be tested empirically. Predictions include the
existence of explicit representations (i. e., neural representations of objects by a
small number of neurons) and competing coalitions (assemblies) of neurons.
Accompanying neural modeling may elucidate mechanisms for such phenomena
and contribute to the question under which conditions humans become conscious
of their actions and perceptions.
2.2.3 Methods
2.2.3.1 Introduction
A large number of empirical and theoretical methods have been developed and
used to characterize the relation between neural behavior and sensory stimuli,
motor stimuli, and mental states.
Methods
Psychophysics
Miles, 1991) or even cell cultures (Murray, 1965; Mains and Patterson, 1973;
Dichter, 1978; Potter and DeMarse, 2001).
In order to quantify correlations between neural dynamics and animal or hu-
man perception and behavior, electrophysiological recordings have to be comple-
mented by measurements of sensory performance or accuracy of motor control.
This is done in behavioral experiments including psychophysics. An ideal ex-
perimental setup is of course the combined behavioral and electrophysiological
measurement; this has been achieved only in a few cases such as Georgopoulos
et al. (1986); Britten et al. (1992); Zhang et al. (1998a); Wessberg et al. (2000).
In most cases, electrophysiology and behavior are obtained from separate exper-
imental setups from the same species (Roth, 1982; Wiggers et al., 1995b; Eurich
et al., 1995:L) or even from different species; for example, electropyhsiology in an
2.2. TASKS, CONSTRAINTS, OPTIMIZATION PRINCIPLES, I-23
OBJECTIVES AND METHODS
animal species may be combined with human psychophysics (Dinse et al., 1994,
1995; Eurich et al., 1997b; Dicke et al., 1999).
is most sensitive to changes in the stimulus (see Figure 2 in Eurich and Pawelzik
(2000); Fisher information is further described in Sections 2.2.3.4 and 2.4.3.1).
The concept of a tuning curve has been extended to include the spatio-temporal
neural behavior (Hida and Naka, 1982; Hida et al., 1983; Eckhorn et al., 1993;
Ringach et al., 1997b; Theunissen et al., 2001).
In the case of multi-neuron reponses, the concept of tuning curves becomes
unwieldy: it is neither convenient nor sufficient to provide a list of tuning curves
without further discussing the consequences of the combined neural activities
for perception and action or for the internal signal processing. A framework
frequently employed in the last 20 years is that of neural reconstruction (Figure
2.2): Neural responses (usually those of a whole population; but see Bialek et al.
(1991); Borst and Theunissen (1999) for single-neuron reconstructions) are sup-
posed to encode a sensory stimulus or a motor action, and mathematical methods
are employed to decode these responses, that is, to infer the presented stimulus
or the motor action from the seqences of spike trains of the neurons under con-
sideration. The reconstructed stimulus or motor action can then be compared
Encoding
Figure 2.2: The idea of neu-
Neuron #
to the original stimulus or motor action, respectively. This allows for hypothe-
ses about neural codes and the performance of the neurons under consideration.
In this sense, reconstruction can be regarded as a data analysis method that is
capable of integrating the activity of multiple neurons. For a brief discussion of
the interpretation of this group of methods see Section 2.3.4.
Further approaches for data analysis include detailed investigations of the
correlations in population responses motivated by the question if synchronous
spiking mediates the binding of features into objects (Eckhorn et al., 1988; Gray
et al., 1989). See for example Perkel et al. (1967); Gray et al. (1989); Gruen et al.
(2001a,b); Gutig et al. (2001); for pitfalls in the detection of synchronous spike
events, cf. Brody (1999).
Statistical analysis of neural encoding strategies, the second subcategory in
Figure 2.1 subsumes various methods to study encoding properties of neurons or
neural populations. These methods may include the evaluation of real data such
that there is an overlap with data evaluation techniques; this does especially hold
for reconstructions as outlined above. In many cases, however, responses of hy-
pothetical neural populations are considered to find advantageous physiological
2.2. TASKS, CONSTRAINTS, OPTIMIZATION PRINCIPLES, I-25
OBJECTIVES AND METHODS
properties in the light of the questions raised in Section 2.2.2. For example, the
measure of Fisher information from classical statistics allows for analytical inves-
tigations of the sensory resolution obtained by the concerted firing of neurons;
cf. Section 2.2.3.4 and references therein.
The last subcategory of theoretial methods shown in Figure 2.1 is neural mod-
eling which is defined here as the modeling of neurons or neural populations as
dynamical systems with the goal to evaluate model performance with respect to
issues of neural coding. For example, Herzog et al. (2003:L) model cortical neural
populations and identify neural correlates of the visibility of a vernier element in
psychophysical experiments. Such dynamical models will be the topic of Chap-
ter 3; the issue of combining neural dynamics and neural coding will be briefly
discussed in Chapter 4.
The following sections consider a few theoretical methods used to address
questions like those listed in Section 2.2.2. Section 2.2.3.2 gives an example of a
new approach to estimated tuning curves from noisy data; Sections 2.2.3.32.2.3.5
describe frequently used approaches to analyze the activity of neural populations:
vector-based reconstruction methods, statistical estimation theory and informa-
tion theory, respectively.
For a single parameter value, the average firing rate may vary considerably
upon repetition of the same experiment. There are three reasons for this:
First, due to experimental difficulties such as positioning and maintaining
recording electrodes in place, the number of measurements in neural record-
ings is often very limited. Therefore, the statistical response properties have
to be inferred from a small set of data points, even if the neural system is
assumed to be static throughout the experiment. Second, the trial-to-trial
variability in the neural spike count is sometimes very large, i. e., there is
a considerable amount of noise in the system. Two examples for sets of
noisy data from rat visual cortex are shown in Figure 2.3. Third, a generic
I-26 CHAPTER 2. NEURAL CODING
20
200
15
10
100
0
0 360 0 360
Figure 2.3: Responses of two neurons from area 17 of the rat to moving bar
stimuli. Dots indicate firing rates in single trials. Solid lines are tuning curves
computed according to the method described in Section 4. Left: A cell which
has a statistically significant tuning. Right: A cell where no significant tuning is
found. The wiggled curve is a polynomial of maximal degree as obtained from
the method developed in Etzold et al. (2002, 2003:L, in press); this function can
be simplified which results in a constant tuning function (straight line). Data by
W. Freiwald with kind permission; for the experimental setup, see Freiwald et al.
(2002:L).
First, the least square measure is replaced by a rank method that gives
outlying values smaller weights without completely discarding them.
While details of the method are described in Etzold et al. (2003:L, in press), only
a few results shall be given here. Figure 2.3 shows tuning curves fitted with the
new method from extracellular recordings in rat visual cortex. The cell on the
left is clearly tuned; the cell on the right hand side is first approximated by a
Fourier polynomial of maximal degree (corresponding to half of the number of
stimulus conditions; wiggled curve). The degree of the polynomial, however, can
be reduced using a criterion based on the distribution of residuals; in this case, it
can be reduced to the order zero (constant tuning), indicating that the cell does
not show statistically significant tuning.
The robustness of the method is demonstrated using artifically generated
data in combination with the bootstrap method. Figure 2.4A shows an ideal
tuning curve based on which three bootstrap samples are drawn according to a
mixture distribution. The left column of Figure 2.4B, C, D shows tuning curves
as obtained from a least-square fit, while tuning curves in the right column result
from the new algorithm. The variation in the fits on the left is clearly visible
while rank-based method yields stable results.
The one-step fitting procedure will be especially helpful in situations where
data are limited, stimulus repetitions are not equally distributed among stimulus
conditions, and where neurons tend to be very noisy as it is the case in rat primary
visual cortex (Freiwald, personal communication).
further details and examples on vector reconstruction see Salinas and Abbott,
1994; Vogels, 1990; Blum and Abbott, 1996). However, Foldiak (1993) points
out that this method requires a high number of cells to derive reliable estimates
(see also Zohary, 1992). Scott et al. (2001) demonstrate a bias of the population
vector in the reconstruction of reaching movements from the activity of neurons
in monkey motor cortex. Although the population vector method can be ex-
tended to include the length of the vector rather than its direction only (Abbott
and Blum, 1996; Blum and Abbott, 1996; Zhang et al., 1998a), Bayesian methods
are preferred today due to their superior performance (see Foldiak (1993); Zhang
et al. (1998a) for a discussion of this issue).
The mean square error (or its square root) is the usual measure of the encoding
accuracy of a neural population.
BMSE = E (x x)2 ,
(2.7)
whereby E(. . .) denotes the expected value over the joint distribution p(n1 , . . . , nN , x) =
P (n1 , . . . , nN |x)p(x).6 For the BMSE, the optimal choice of x is the mean of the
conditional distribution,
Z
x = x p(x|n1 , . . . , nN ) dx , (2.8)
X
i. e., it minimizes the mean square error (x)2 (see for example Rieke et al.,
1997). A convenient estimate which can be easily evaluated is the maximum of
the posterior distribution:
Using the optimal estimator (2.8) implies the philosophy that frequently
occurring stimuli should be represented with a high accuracy while rarely
occurring stimuli may have a larger reconstruction error such that the
error is small if averaged over stimuli and noise distribution. However, this
may not be feasible in all cases: A rarely occurring stimulus, for example
the appearance of a predator, may nevertheless be relevant for an animal
in which case some attentional mechanism will be elicited. From a signal
processing point of view, such events could be considered by including an
additional weighting function in the reconstruction algorithm giving more
weight to stimuli that are considered to be relevant.
The results obtained from Bayesian estimation are superior to other meth-
ods such as vector decoding. Zhang et al. (1998a) applied several techniques to
reconstruct the path of a rat running in a maze from the firing of hippocampal
place cells. A straightforward Bayesian reconstruction with a flat prior as out-
lined above proved to be second-best; the best method was an extended Bayesian
scheme where additional information about the position of the rat entered the
computation.
Freiwald et al. (2002:L) document a collaboration of eight institutes to estab-
lish a new recording technique with micro-machined silicon probes for obtaining
spike data from large neural populations, to apply this technique to the rat visual
cortex, and to evaluate the data using estimation theory. The rat primary visual
cortex is not as well studied as corresponding structures in cats or monkeys (see,
for example, van der Togt et al., 1998; Girman et al., 1999), and cells are consid-
ered to be very noisy. We therefore used standard stimuli (whole screen black and
white gratings at various orientations ) to study the encoding accuracy of neural
populations in area V1. Most neurons turned out not to be tuned significantly,
but a few neurons showed direction preferences; Figure 2.6 shows an example.
A Bayesian analysis was subsequently performed; cf. Figure 2.7a for a con-
ditional probability distribution of the tuned cell response shown in Figure 2.6.
To yield an estimate of the orientation resolution obtained with a population of
tuned neurons of rat area V1, a well-tuned cell was mathematically duplicated,
and the resulting tuning curves were distributed over all orientations. A sub-
sequent Bayesian analysis employing the MAP estimator revealed that only few
neurons are sufficient to yield a good resolution of oriented gratings (Figure 2.7b).
Fisher information. Apart from the fact that the natural prior distribution of
stimuli, p(x), is usually not available, Bayesian techniques have the disadvantage
I-34 CHAPTER 2. NEURAL CODING
to allow only for a very limited analytical access: results are usually obtained
through model simulation and subsequent evaluation. In contrast, Fisher infor-
mation, a measure from classical estimation theory, provides a general framework
which yields also an analytical access in some cases.
In brief, for a given characteristics of neural spike counts P (n1 , . . . , nN ; x),
i. e., for a multivariate distribution depending on a D-dimensional parameter x,
2.2. TASKS, CONSTRAINTS, OPTIMIZATION PRINCIPLES, I-35
OBJECTIVES AND METHODS
(a) (b)
The significance of the Fisher information measure lies in the fact that it yields
a lower bound on the expected square estimation error for an arbitrary, unbiased
estimator , whereby the error is computed according to (2.4). This relation is
quantified by the Cramer-Rao inequality :
D
X
2
h(x x) i (J 1 (x))jj (2.13)
j=1
The minimal square error will henceforth be denoted as 2min ; Equation 2.13 can
then be written as
XD
2
min (x) = (J 1 (x))jj . (2.14)
j=1
A proof of the Cramer-Rao inequality can be found in Cover and Thomas (1991).
When applied to neural coding, the minimal square estimation error provides
a bound on the accuracy with which a stimulus can be reconstructed, given the
spike activity of a neural population. Fisher information has been applied in a
number of contributions (Paradiso, 1988; Seung and Sompolinsky, 1993; Brunel
and Nadal, 1998; Abbott and Dayan, 1999; Zhang and Sejnowski, 1999; Wilke
and Eurich, 1999; Pouget et al., 1999; Yoon and Sompolinsky, 1999; Eurich and
Wilke, 2000:L; Eurich et al., 2000:L; Karbowski, 2000; Wilke and Eurich, 2001:Lb;
Nakahara et al., 2001; Wilke and Eurich, 2001:La; Bethge et al., 2002; Wilke and
Eurich, 2002; Wu et al., 2002; Eurich, 2003:L), most of which deal with neural
encoding strategies.
There are two things that have to be taken into account when employing the
measure of Fisher information:
The existence of a lower bound min (x) does not imply the existence of
an estimator that reaches this bound, i. e., of a so-called efficient estima-
tor . Zhang et al. (1998a) compares Bayesian reconstruction results with a
Fisher information calculation in the encoding of the position of a rat by
hippocampal place cells. Despite several simplifying assumptions on the
neural firing, the Bayesian estimator achieves a value of twice the Cramer-
Rao bound min with only 25 neurons. A neural decoding mechanism based
on Bayesian principles might therefore extract the signal content in the neu-
ral spike events rather efficiently. A theoretical account of this issue is given
in Bethge et al. (2002) who faithom the validity of the Fisher information
measure, in particular with respect to the observation time window . A
basic result is that has to be sufficiently long (or equivalently, firing rates
have to be sufficiently high) for Fisher information to provide the inverse
of an existing efficient estimator.
Fisher information is concerned with unbiased estimators only. There is an
extension to biased estimators (cf. for example Cover and Thomas, 1991, p.
2.2. TASKS, CONSTRAINTS, OPTIMIZATION PRINCIPLES, I-37
OBJECTIVES AND METHODS
335). However, in this case, the bound includes the bias of a given estimator
and its derivate; therefore, is is no longer as universal as in (2.13) which
refers to arbitrary unbiased estimators. If biased estimators are taken into
account, estimation errors may fall below this Cramer-Rao bound.
Sections 2.4.3 and 2.4.4 discuss neural encoding strategies for a single object and
for multiple objects, respectively, under the assumption that Fisher information
yields a good measure of the available signal content of spike activity. It turns out
that simple dichotomies like small vs. broad tuning are by no means sufficient
to characterize the performance of a neural population.
A neuron which has background activity (as do, for example, most cortical
neurons) responds with a significant rapid increase in its firing rate. A
typical case is a phasic burst shortly after stimulus onset (e. g., Ahmed et al.,
1998 and Lisman, 1997 for an account of bursting behavior in neurons).
Second, the notion of a rate may also refer to a population rate, i. e., the
number of spikes fired by a given set of neurons within a certain time inter-
val. This quantity contains less information than the set of single-neuron
rates in the first case, because the spikes are not attributed to specific neu-
rons. However, the dynamics of population rates can be treated analytically
I-42 CHAPTER 2. NEURAL CODING
(e. g., Gerstner, 2000; Fourcaud and Brunel, 2002) and therefore yields an
access also for the area of neural coding.
Rate coding may be realized in motor cortex (e. g., Evarts, 1967), but also in
other cortical areas, for example, in the motion-sensitive visual area MT (Britten
et al., 1992; Shadlen and Newsome, 1998). The latter argue that cortical neurons
are unlikely to transmit information in the temporal pattern of spike discharge,
based on observations that neural responses exhibit a large variability (see also
Softky and Koch, 1992, 1993). Oram et al. (1999) argue that exactly timed
spike triplets and quadruplets observed in monkey LGN and V1 do not carry any
information beyond that already contained in the firing rate.
Major objections against a rate code are based on the fact that it takes a long
time to reliably estimate a single number, the rate, from a neural response, and
that the estimate is obtained with many energy-consuming spikes (Gautrais and
Thorpe, 1998). The temporal constraint can be evaded by coding with a popula-
tion of neurons; this requires, however, a large number of neurons. For example,
to obtain an estimate as precise as 100 10 Hz in 10 ms from neurons firing
according to a Poisson statistics, no less than 281 redundant and independent
neurons are needed (Gautrais and Thorpe, 1998).
Shadlen and Newsome (1998) consider networks of balanced excitation and
inhibition and suggest that neurons transmit information through rates in en-
sembles of 50 100 neurons, providing reliable rate estimates in 10 50 ms, i. e.,
in one interspike interval. Rapid state switching has also been found in other
studies of balanced networks (Tsodyks and Sejnowski, 1995; van Vreeswijk and
Sompolinsky, 1996). Likewise, in his theoretical account of the dynamics of a
homogeneous infinite population of spiking neurons, Gerstner (2000) finds an im-
mediate change in the population activity upon changing the input. These results
suggest that rate coding may be feasible also under time constraints.
A prerequisite for the existence of a temporal code is the neurons ability to pro-
duce spike discharges with a precise timing. Cortical spike activity may show
a large variability including spontaneous discharges; however, the cause of this
behavior is not completely clarified. Arieli et al. (1996); Tsodyks et al. (1999)
link evoked cortical responses and spontaneous activity to the underlying func-
tional architecture and the ongoing activity in the neighborhood of the neurons
2.3. NEURAL CODES I-43
under consideration. This leaves the possibility that the neural behavior which
appears noisy to an observer may reflect more deterministic dynamical processes
in the cortical tissue. Mainen and Sejnowski (1995) studied the reliability of spike
timing in single neurons upon current injection: whereas spike times appeared
variable for constant current input, the reliability increased dramatically upon
the injection of noise-like currents that resemble the large number of synaptic
inputs of a neuron in vivo. In this case, the standard deviation in spike timing
was less than 1 ms. A similar accuracy of spike timing was found in retinal gan-
glion cells of rabbits and salamanders (Berry et al., 1997). Thus, neurons seem
to be able to reliably produce action potentials. This is further corroborated
by the discovery of physiological processes whose dynamics depends on an exact
spike timing: spike-dependent Hebbian plasticity (e. g., Bi and Poo, 1998, 1999;
Egger et al., 1999; Bi and Poo, 2001) and dynamic synapses (e. g., Markram and
Tsodyks, 1996; Tsodyks and Markram, 1997; Tsodyks et al., 1998).
A prominent example of a system where the timing of signals plays a ma-
jor role is the auditory system of barn owls. Based on an idea put forward by
Jeffress (1948), the horizontal sound localization is accomplished by the evalua-
tion of phases differences of sounds arriving at the two ears. Phase-locked action
potentials run down axons that serve as delay lines to arrive at an array of postsy-
naptic coincidence detectors. Those cells that become active due to synchronous
input from both ears code for the position of the sound. This principle was later
validated in the barn owl (Konishi, 1986; Carr and Konishi, 1990) and seems
to be realized also in other sensory systems such as electrosensation in weakly
electric fish and echolocation in bats (Konishi, 1991; Carr, 1993; Carr and Fried-
man, 1999).10 Theoretical papers on delays in neural networks and their putative
adaptability were published by Tank and Hopfield (1987); Gerstner et al. (1996);
Eurich et al. (1999:L).
Spike timing has been shown to play a role also in other systems. Studies
revealed that the timing of individual spikes can represent the time structure of
rapidly varying stimuli (Bialek et al., 1991; Borst and Theunissen, 1999) but also
encodes nontemporal properties, such as the location of a stimulus applied to a
single whisker in rats (Panzeri et al., 2001; Petersen et al., 2001). The authors
find that the time of the single neurons first poststimulus spikes is crucial for
encoding stimulus location.
Panzeri and Schultz (2001) recently described a unified approach to temporal
coding and rate coding by expanding the mutual information between spike count
and stimulus as a Taylor series in the experimental time window. The connection
between rate coding and temporal coding was also addressed by Ahissar et al.
(2000) and Mehta et al. (2002); the latter propose a mechanism based on oscilla-
tory inhibition to transform a rate code into a temporal code. Similarly, Hopfield
10
Jeffress model, however, does not seem to apply to the mammalian auditory system
(McAlpine and Grothe, 2003).
I-44 CHAPTER 2. NEURAL CODING
The second type of temporal coding employs correlations between neurons; here,
we consider noise correlations11 with an accuracy of about 1 ms.
Several studies have found noise correlations in a variety of neural systems, for
example in monkey areas IT (Gawne and Richmond, 1993), MT (Zohary et al.,
1994; Cardoso de Oliveira et al., 1997), somatosensory cortex (Steinmetz et al.,
2000), motor cortex (Lee et al., 1998; Hatsopoulos et al., 1998; Maynard et al.,
1999), and parietal cortex (Lee et al., 1998), as well as in cat LGN (Dan et al.,
1998) and visual cortex (DeAngelis et al., 1999). In cultured cortical neurons,
Tateno and Jimbo (1999) find an enhancement in the reliability of correlated
spike timings upon external stimulation.
The functional role of these correlations is not completely understood so far.
Some investigations indicate that correlations depend on the attentional state of
the animal (Cardoso de Oliveira et al., 1997; Steinmetz et al., 2000). Dan et al.
(1998) use information theory to analyse the activity of LGN cells upon white
noise input; they found that considerably more information can be extracted
from two cells if temporal correlations between them are considered. Maynard
et al. (1999) classified the direction of arm movements from the concerted firing
of motor cortical neurons and also found a significant improvement as compared
to the case of independent encoders.
Another suggestion relating correlated neural activity to the encoding of ob-
jects was put forward by Singer (e. g., Gray et al., 1989; Engel et al., 1991; Singer,
1993; Singer and Gray, 1995; Roelfsema et al., 1997) and Eckhorn (e. g., Eckhorn
et al., 1988, 1989, 1990; Frien et al., 1994). According to this so-called temporal
correlation hypothesis, neurons encoding for the same object synchronize at zero
phase lag, whereas neurons encoding for different objects have a nonzero phase
lag. The temporal correlation hypothesis is a proposed solution of the binding
problem. An extended discussion of this issue is published as a series of review
articles in the journal Neuron, introduced by Roskies (1999).
For theoretical considerations on the correlated activity of neural populations,
see the discussion in Section 2.4.3.3.
11
See Section 2.2.3.4 for the distinction between signal correlations and noise correlations.
2.3. NEURAL CODES I-45
Another encoding scheme proposed by Thorpe and Gautrais (1997, 1998) is mo-
tivated by studies demonstrating a high speed of processing in the human visual
system (Thorpe et al., 1996). In these experiments, unknown photographs were
briefly (i. e., for 20 ms) presented to subjects who had to categorize the pictures
as to whether they contained an animal or not. During the trials, event-related
potential were recorded demonstrating that differences in the potentials between
the animal vs. no-animal cases were visible a mere 150 ms after stimulus onset.
The authors suggest that a great deal of visual processing must have been com-
pleted before this time. Delorme et al. (1999, 2000) trained monkeys to perform
similar categorization tasks (animal vs. no-animal and food vs. no-food) based
on both color and black-and-white pictures; the same task was also done by hu-
man observers. The measured accuracy in monkeys which have reaction times
between 250 ms and 300 ms and are faster than humans do not depend on
color information, just as for the fastest humans. The accuracy impairment with
achromatic stimuli increased progressively in human subjects with longer reaction
times. The interpretation of these findings is that fast categorization of natural
stimuli, i. e., the first wave of processing in the visual system, is color blind.
The fast neural processing poses a problem for the rate coding scheme (Thorpe
and Gautrais, 1997, 1998; Gautrais and Thorpe, 1998; Thorpe et al., 2001); the
evaluation of a firing rate would either take too much time or require large neural
populations. For example, to obtain an estimate as precise as 100 10 Hz in
10 ms needs 281 redundant and independent neurons, which seems a high cost
for the transmission of a single analog number. Instead, the autors suggest a code
based on the order in which spike from different neurons arrive at a postsynaptic
layer. Such a code would be very efficient in terms of transmitted information
(Thorpe and Gautrais, 1998). Moreover, it is robust with respect to the exact
spike timing, as long as the order of signals is conserved. Van Rullen et al. (1998);
Van Rullen and Thorpe (2001) model face processing with a feedforward neural
network employing the rank order code; detection and localization of human faces
in natural images succeeds with only one spike per neuron. The order of spike
arrivals can be evaluated by neural populations. So far, however, it is not clear
how a neural system can self-organize to establish the code: the rank order is
global information whereas neural adaptation rules are local.
The literature on neural coding has largely focused on codes based on the spike
activity of neurons. However, there are further processes in neural tissue that are
likely to contribute to signal processing.
I-46 CHAPTER 2. NEURAL CODING
The second answer to the above question goes in the direction of scientific method-
ology. A meaningful theoretical model employs as few parameters as possible in
order to identify those ingredients which are necessary for the phenomena to be
explained. In other words, in good scientific practice models are as simple as
possible; they apply to a given set of observations while necglecting other aspects
of the system under study. As a consequence of this methodogical approach,
the model cannot be said to represent (or even to be identical with) the system
in every respect. The model rather yields a minimalistic description, while the
true system may well be more complex with respect to the phenomena under
consideration but also to further phenomena not captured by the model. As an
example, consider the binary code described above. We used this code to account
for the observation that the receptive fields of neurons in the salamander optic
tectum are large (average diameter > 40 ) and yet allow for a high accuracy in
localizing prey (Eurich et al., 1995:L; Wiggers et al., 1995a,b; Eurich and Schwe-
gler, 1997:L; Eurich et al., 1997:L; Eurich and Wiggers, 1997). What follows from
these studies is the fact that a binary neural response is sufficient to account for
the observed electrophyiology and behavior. It does not state that the neurons
are actually binary: first, the biological neurons are of course more complex than
the model neurons; second, the all-or-none response might not be evaluated by
subsequent neural structures in the salamander brain (see the first answer above);
third, high sensory resolution may well be achieved also with other, more complex
neural codes.
The third aspect to be considered in view of the large variety of possible neural
codes is that different neural systems may use different codes. The experimental
studies which have inspired the codes were conducted in different species, for
example blowflies (Bialek et al., 1991), salamanders (Wiggers et al., 1995b; Meis-
ter et al., 1995), turtles (Wilke et al., 2001:L), electric fish (Heiligenberg, 1987),
rats (Zhang et al., 1998a; Freiwald et al., 2002:L), cats (Sengpiel et al., 1995;
Oram et al., 1998), macaque monkeys (Georgopoulos et al., 1986; Scott et al.,
2001; Wessberg et al., 2000), and humans (Thorpe et al., 1996; Delorme et al.,
1999). Likewise, data were obtained from different systems including the retina
(Meister et al., 1995; Wilke et al., 2001:L), midbrain visual and auditory areas
(Wiggers et al., 1995b; Knudsen and Konishi, 1978), hippocampus (Zhang et al.,
1998a) visual cortex (Sengpiel et al., 1995), motor cortex (Georgopoulos et al.,
1986; Scott et al., 2001; Wessberg et al., 2000), and the fly visual system (Bialek
et al., 1991; Rieke et al., 1997) which is not homologous to the vertebrate visual
system. One cannot expect to find the same principles of signal processing in
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-49
such diverse systems. For example, in the blowfly visual system, a single neuron,
the so-called H1 neuron, encodes for rigid horizontal movements over the entire
visual field (see Rieke et al., 1997, and references therein). On the other hand,
neural codes for sensory stimuli or motor actions in vertebrates are likely to be
based on large populations of neurons in most cases (e. g., Georgopoulos et al.,
1986; Eurich et al., 1995:L; Dan et al., 1998; Maynard et al., 1999; see, however,
Britten et al., 1992 for a contrary argumentation). When deciding about a neural
code, one must therefore look carefully at the specific anatomical and physiolog-
ical properties of the system under study.
Finally, neural behavior is so complex that it is not unlikely that a neural sys-
tem uses various codes simultaneously . For example, the binary response (firing
vs. non-firing or large activity vs. background activity) may be used to
infer the position of an object in the visual field whereas the firing rate gives
information about its properties (color, spatial frequencies, etc.). The precise
timing of the spikes may in addition distinguish the object from other, simulta-
neously presented, objects. Such an encoding scheme seems to be very efficient
because it will be fast (several aspects of the scene can be evaluated simultane-
ously) and energy-efficient (various codes are hidden in a single response). I want
to call this type of encoding palimpsest coding .
X S m Rn , (2.15)
While an overview of the literature on single-object coding has been given under
various viewpoints in Sections 2.2 and 2.3, we will now focus on the issue of the
encoding accuracy of a neural population. Section 2.4.2 addresses this problem by
using a model of binary coding, whereas Section 2.4.3 investigates properties of
firing rate coding using the statistical measure of Fisher information. In Section
2.4.4, the latter formalism is extended to include multiple objects.
(x) 0 (2.16)
Z
(x) dx = N, (2.17)
X
14
The phenomenon that broadly tuned receptors allow for a high resolution is a well-known
phenomenon in vision called hyperacuity : human observers can determine the position of an
object with a precision that corresponds to only a fraction of the diameter of a photoreceptor
(e. g., Wachtler et al., 1996). Another example is color vision: we can distinguish roughly 2
million colors with only three types of broadly tuned color-sensitive photoreceptors; publications
on this issue date back to von Helmholtz (1867).
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-53
w( )
0.04
0.03
0.02
0.01
0.0
0 20 40 60 80 100 120 140 160 180
2 (deg)
(b)
Tyto alba:
MLD neurons
(Knudsen and Konishi, 1978))
Figure 2.8: Examples for large receptive fields in the visual and the auditory
system. (a) Distribution w(%) of receptive field sizes of neurons of class T1 in the
optic tectum of the tongue-projecting salamander Hydromantes italicus; % is the
half angular field diameter (adapted from Wiggers et al., 1995b). (b) Sketch of
receptive fields in the midbrain MLD nucleus of the barn owl, Tyto alba (adapted
from Knudsen and Konishi, 1978).
where the approximation results from the fact that the number of real receptive
fields is integer while the integral on the left side takes continuous values. The
resolution Ae (x0 ) at some point x0 X is defined to be
1
Ae (x0 ) = (2.19)
Le (x0 )
shows the angular resolution (i. e., the resolution in direction localization) as
a function of the receptive field half angular diameter %. The resolution has an
optimal value for % = 90 , i. e., for receptive fields covering half of the sphere! The
resolution in this case is 1.8 . A simple explanation of this observation is that
these large receptive fields have the largest possible receptive field boundaries (the
boundaries are great circles) yielding a maximal density of boundaries, which in
turn corresponds to an optimal resolution.
20
Angular resolution
for N=100 neurons
15 Figure 2.10: Angular resolution (%)
as a function of the receptive field half
a(r) (deg)
80 a
60
(deg)
40
20
-20
-40
-150 -100 -50 0 50 100 150
(deg)
1000 1
900
800
Figure 2.12: Resolution of direction lo-
700
calization along the horizontal visual di-
rection as obtained from the electrophys-
600 2
iological data using the binary coding
500
scheme. For details, see text (adapted
400
from Eurich et al., 1997:L).
-150 -100 -50 0 50 100 150
Results are shown in Figure 2.12. N e () is the angular resolution for the
horizontal direction e as a function of the horizontal angle , whereby = 0
corresponds to the center of the visual field. Curve 1 considers the distribution of
receptive field positions as shown in Figure 2.11 and assumes identical receptive
fields with a diameter corresponding to the mean of the distribution in Figure 2.8a
for all x S 2 . As expected, the resolution reaches its optimum in the middle of
the visual field. The corresponding value is 385 . That is, with N neurons an
angular resolution of about 385 /N is achieved, because the density of receptive
field boundaries is proportional to N . In addition, curve 2 employs the estimate of
w(%|x) as described above and therefore the particularly large receptive fields in
the lateral visual field. The calculations reveal the following result: the existence
of large receptive fields in the periphery of the visual field leads to a lateral
increase in resolution compared to the situation of curve 1, where all receptive
fields are assumed to have the same size. These findings suggest that the large
receptive field neurons which are found in urodeles as well as in anurans (Grusser-
Cornehls and Himstedt, 1976; Roth and Himstedt, 1978; Grusser-Cornehls, 1984)
contribute to the localization of small objects such as prey and are no mere
predator detectors signaling the emergence of large objects somewhere in the
visual field (e. g., Ewert, 1984).
Conversely, the number of neurons necessary to yield the angular resolution
observed in behavioral experiments can now be estimated. We assume that the
salamander can localize a 0.05 cm sized prey at the distance of the maximal
protraction length of the tongue, which is 5 cm in Hydromantes italicus (Roth,
1976; Wiggers, 1991). This corresponds to an angular resolution of 0.57 . Curve 2
in Figure 2.12 yields an angular resolution per neuron of about 450 at = 0.
Thus, the necessary number of neurons is N 450/0.57 790. This is in good
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-57
1 3
Visual field
2 4
rostral
Optic Tectum
2 x 50 Tectum neurons
caudal
1:1
Medulla oblongata /
Medulla spinalis
1:1 1:1
Muscles
Pop.code
M. intertransversi
M1 M3 capitis superiores
Pop.code
M2 M4 M. recti cervicis
exctatory parts
inhibitory parts
Figure 2.13: The Simulander I network for the orienting movement of salamanders
(from Eurich et al., 1995:L).
decoding takes place only at the level of the muscles while the internal signal
processing is exclusively based on the activities of unspecific neurons.
Artificial lesion experiments in the network demonstrate the importance of
large receptive field neurons as compared to neurons with smaller receptive fields.
In addition, the model correctly reproduces the behavior of monocular salaman-
ders (see Eurich et al., 1995:L for details).
The Simulander II model (Eurich et al., 1997:L; Eurich and Wiggers, 1997) in-
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-59
vestigates depth perception and the activation of the projectile tongue to catch
prey at the correct distance.
Depth perception in urodeles and anurans is mainly due to the evaluation of
binocular disparity, whereas mechanisms such as accomodation, vergence, and
motion parallax are either absent or play only a minor role (e. g., Ingle, 1976;
Collett, 1977; Roth, 1987). In tongue-projecting salamanders, binocular neurons
in the rostral optic tectum receive direct input from both retinae; part of them
project to the brainstem and spinal cord and activate motor networks responsible
for the control of the projectile tongue.
For simplicity, a binocular receptive field is defined as the overlap of the two
monocular receptive fields of a binocular neuron. A suitable object in this re-
gion of the visual field elicits a strong reaction in the neuron. From the size
distribution of monocular fields (Figure 2.8) it becomes clear that binocular re-
ceptive fields also tend to be large. The question arises whether an exact distance
determination can be achieved at all with these physiological properties.
Figure 2.14 shows the two-dimensional projections of three typical binocular
receptive fields; they occupy a high percentage of the binocular visual field. A
discrimination of small regions of space is nevertheless possible with an ensemble
coding. This becomes clear from the following qualitative example. Consider
the small rhombus V in Figure 2.14. An object in the visual field is positioned
within V if and only if the neuron which corresponds to receptive field 1 fires,
but the neurons which correspond to receptive fields 2 and 3 do not fire. This
suggests that a non-firing neuron conveys as much information as a firing one:
the symbols 0 and 1 are of equal relevance in information theory.
retraction of the tongue, respectively. After training the network with an evolu-
tion strategy, the network successfully activates the tongue. Apart from a small
region close to the maximal protraction distance, prey is caught successfully in
80 100% of the presentations. If object are far away, the tongue projection is
inhibited. Only for prey slightly out of reach of the tongue, a larger percentage
of errors occurs a prediction for a behavior experiment.
The Simulander model demonstrates that a simple binary encoding scheme
can explain accurate object localization, and that the evaluation of information
contained in the neural activity is a straightforward process in a neural network.
15
Although the experimental setup employs a discrimination task, the model is concerned
with the localization accuracy of a single object the discrimination threshold was considered
to be equivalent to the resolution obtained. In a more recent approach (Eurich, 2003:L) the
theory of population coding is extended to include the simultaneous presentation of multiple
objects; this yields the possibility to study the discrimination experiments from a new point of
view.
16
A similar determination of sensory thresholds in rats could not be performed so far; a
quantitative comparison is therefore not possible.
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-61
The assumption of a binary code has proven successful in describing the local-
ization accuracy of a population of neurons with overlapping receptive fields. In
neurophysiological terms, the occurrence of a neural response in a binary scheme
can manifest itself as any spike activity (while the neuron is silent in its other
state) or as a significant enhancement of its firing rate (while there is only a small
background activity in its other state). For example, a typical neural response
consists of a phasic burst followed by a tonic response (spike frequency adap-
tation, e. g. Connors et al., 1982; Ahmed et al., 1998; Smith et al., 2001). The
investigations on binary coding suggest that the initial phasic response may not
only signal the presence of a stimulus but could as well encode some of its prop-
erties. In fact, Bethge et al. (2003) have shown that binary coding is superior to
rate coding if time is a critical constraint in the signal processing. According to
the idea of palimpsest coding, additional information about the stimulus may be
transmitted through the firing rate, spike timing, etc. during the phasic response
or the tonic component of the neural response.
where F is the mean peak firing rate, c X is the center of the tuning curve f (x),
and () is a function of a scalar variable with maxz (z) = 1. For a constant
density of tuning curves, (c) const, and uncorrelated neural activity (2.10)
the elements of the D D Fisher information matrix are given by
(
0 if i 6= j
Jij (x) = D2
(2.23)
K (F, , D) if i = j ,
where K (F, , D) is a function depending on the shape but not the width of
the (identical) tuning curves. The result suggests that for D 3 features broad
tuning yields a better encoding than small tuning.18
The issue of neural tuning, however, is much richer than the result (2.23)
obtained from the considerations in Zhang and Sejnowski (1999).
x2 x2
A B
x 2,s x 2,s
x1,s x1 x1,s x1
neurons and spikes. To assess tuning properties for this case, we define a minimal
square estimation error for feature i, averaged over a distribution (x) of stimuli:
Z Z
h2i,min i dxD (x) J(x)1 ii .
:= dx1 . . . (2.26)
Figure 2.16: Example for the stimulus-averaged squared minimal encoding er-
ror for Poissonion spiking and Gaussian tuning curves arranged on a reg-
ular 2-dimensional grid with spacing . Dashed curves: Numerical result
h21,min i/(2 F 1 1 ) as a function of 1 / for different values of 2 / (from
top to bottom): 2 / = 0.25, 0.5, 1, 2. Solid lines: analytical broad-tuning re-
sult, i. e. h21,min i/(2 F 1 1 ) = (1/J1 )/(2 F 1 1 ) from (2.25). The inset shows
Gaussian tuning curves of optimal width, opt 0.4 (adapted from Eurich and
Wilke, 2000:L).
Q
D
X := DK (F, , D)
j=d+1 j is an experimentally unknown constant. If the
neurons tuning curves have been measured in more than one dimension, X can be
eliminated by considering the tuning widths as a relative measure of information
content,
2i,min i2
Pd 2 = Pd 2
, (2.27)
j=1 j,min j=1 j
where i is one of the known dimensions 1, . . . , d. Equation (2.27) is independent
of the unknown (or experimentally ignored) stimulus dimensions d + 1, . . . , D,
which are to be held fixed during the measurement of the d tuning widths. On
the basis of the accessible tuning widths only, it gives a relative quantitative
measure on how accurately the individual stimulus dimensions are encoded by
the neural population.
Equation (2.27) states that if i2 is small compared to the sum of the squares
of all measured tuning widths, the population activity allows an accurate recon-
struction of the corresponding stimulus feature: the population contains much
I-66 CHAPTER 2. NEURAL CODING
information about xi . A large ratio, on the other hand, indicates that the popula-
tion response is unspecific with respect to feature xi . As an example, consider the
different pathways of signal processing which have been suggested for the visual
system (Livingstone and Hubel, 1988). The model states that information about
form, color, movement, and depth is in part processed separately in the visual
cortex. Based on the physiological properties of visual cortical neurons, (2.27)
yields a quantitative assessment of the specificity of their encoding with respect
to the abovementioned properties, i.e., our method provides a test criterion for
the validity of the pathway model.
This result generalizes the results obtained above. Figure 2.17 visualizes four
distributions P (1 , . . . , D ) for D = 2. These will be discussed in the following,
using the general case of D features. Distribution (a) is given by
D
Y
P (1 , . . . , D ) = (i );
i=1
all neurons have radially symmetric tuning curves of width . The average pop-
ulation Fisher information (2.29) yields the Fisher information (2.23) obtained
by Zhang and Sejnowski (1999), whereby the tuning width is given by .
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-67
Case (b) has different tuning widths for the different features:
D
Y
P (1 , . . . , D ) = (i i ), (2.30)
i=1
where i denotes the fixed width in dimension i. For i = j, the average population
Fisher information (2.29) reduces to (2.25) (Eurich and Wilke, 2000:L), whereby
the tuning widths are given by j (j = 1, . . . , D).
Case (c) employs the encoding of a stimulus with distributed tuning widths;
we consider the distribution
D
Y 1 bi bi
P (1 , . . . , D ) = i ( i ) ( i + ) i , (2.31)
b
i=1 i
2 2
where denotes the Heaviside step function. Equation (2.31) describes a uniform
distribution in a D-dimensional cuboid of size b1 , . . . , bD around ( 1 , . . . D ). A
straightforward calculation shows that in this case, the average population Fisher
information (2.29) for i = j becomes
( 2 " #)
QD 4
l 1 b i bi
hJii i = DK (F, , D) l=12 1+ +O . (2.32)
i 12 i i
A comparison with (2.25) yields the astonishing result that an increase in b i results
in an increase in the i-th diagonal element of the average population Fisher infor-
mation matrix and thus in an improvement in the encoding of the i-th stimulus
feature, while the encoding in dimensions j 6= i is not affected. Correspondingly,
I-68 CHAPTER 2. NEURAL CODING
Note that, for = 1, the population is uniform: all neurons have tuning width
for all stimulus dimensions. For 6= 1, the population is split up into D
subpopulations; in subpopulation i, i is different from the other tuning widths
j (j 6= i). The Fisher information for this population is given by
(D 1)2D/(D1) + 1
Jii . (2.34)
D2
It does not depend on i because of the symmetry in the subpopulations. The
-dependent factor that quantifies the change of encoding accuracy as compared
to the uniform population is plotted in Figure 2.18 for different values of D. It
turns out that the uniform case = 1 is a local minimum of Fisher information:
Uniform tuning properties yield the worst encoding performance within the model
class. Any other value of , i. e., any unsymmetrical receptive field shape for the
subpopulations, leads to a more precise stimulus encoding. The increase of Fisher
information for smaller reflects the better encoding with specialized subpopu-
lations for smaller i , as discussed above. In this limit, the i-th subpopulation
exclusively encodes feature i. For 1, on the other hand, the i-th subpopu-
lation encodes all features but xi . The Fisher information loss that results from
decreasing the corresponding tuning widths is compensated by the increase of i .
The fact that the approximation (2.34) diverges for 0 and for
is due to the idealization of assuming an infinite population size and an infinite
stimulus space. Under these assumptions, the tuning curves will be strongly de-
formed if approaches extreme values, but the number of tuning curves that
cover a given point in stimulus space always remains constant. In a finite neural
population, on the other hand, the coverage decreases for large or small for
two reasons. First, for some features the tuning curves become narrower, so that
neighboring neurons in these directions cease to encode the stimulus. Second,
in the other directions, where the tuning curves become broader, no new tuning
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-69
0
0 1 2 3 4 5
Subpopulation Asymmetry
curves reach the stimulus position since the corresponding tuning curve centers
would lie outside the boundaries of the population. Consequently, Fisher infor-
mation eventually decreases as 0 or if the number of neurons is
finite. The value of at which this effect sets in depends on the initial tuning
widths : The larger , the sooner a considerable fraction of receptive fields will
lie outside the stimulus space. A second effect that is expected is that values
for that are too large or too small will lead to unrepresented gaps in stimulus
space, thus yielding a breakdown of encoding quality by the mechanism described
above. In contrast to the first effect, this breakdown occurs sooner if is smaller,
since, at given , this implies a less dense coverage of stimulus space with tuning
curves. These arguments are illustrated by an example in Figure 2.19. The ini-
tial deviation from the analytical curve can be attributed to the first effect; it is
strongest for large . The drop of Fisher information of the finite population at
large/small is due to unrepresented areas in stimulus space: it is most severe
for the smallest value of . Thus, our analysis leads to the prediction that for a
finite-size neural population and a bounded stimulus space, there is an optimal
level of specialization in terms of tuning curve non-uniformity.
In this section, we study the effect of the noise model, i. e., the deviations from the
mean spike count in individual stimulus presentations, on the encoding accuracy
of a neural population.
I-70 CHAPTER 2. NEURAL CODING
15
0
0,1 1 10
Subpopulation Asymmetry
Several empirical studies (e. g., Dean, 1981; Lee et al., 1998; Gershon et al.,
1998; Maynard et al., 1999) have yielded the result that the logarithm of the
spike count variance is a linear function of the logarithm of the mean spike count.
Theoretically, additive noise (Abbott and Dayan, 1999; Yoon and Sompolinsky,
1999) and multiplicative noise (Abbott and Dayan, 1999) have been employed
next to Poissonian noise to describe neural spiking statistics. Both the empirical
results and the theoretical models are captured by the mean-variance relation of
fractional power law noise of the k-th neuron,
spike counts (Softky and Koch, 1992, 1993; Rieke et al., 1997; Gershon et al., 1998;
Lee et al., 1998). Hence, there are aspects to neuronal noise that are not captured
by a simple Poissonian spike count distribution. However, there are no clear ideas
on what effect it may have on the encoding properties of a neural population. The
following paragraphs discuss the dependence of Fisher information on the mean-
variance relation exponent with respect to two properties of neural firing: the
existence of a stimulus-independent background activity, and the peak activity F .
Whereas details of the computations can be found in Wilke and Eurich (2001:Lb)
and Wilke and Eurich (2001:La), we will focus on the results here.
Background noise. In this section, we assume that the spike count vector
n = (n1 , . . . , nN ) can be treated as a continuous random variable and specify the
probability density function p(n; x) to be
1 1 T 1
p(n|x) = p exp (n f ) Q (n f ) , (2.36)
(2)N det Q 2
where (k)2 is defined as in (2.24). The parameter measures the level of stimulus-
unrelated background activity at given maximum firing rate F ; it ranges from
zero-baseline firing ( = 0) to completely stimulus-independent firing ( = 1). As
an example, a Gaussian tuning curve is depicted in Figure 2.20 along with the
variance according to the mean-variance scaling of = 0; 1/2; 1.
For sufficient tuning curve overlap and a constant distribution of tuning
curves, the Fisher information for the model population is given by
Q !
D
i
J = 1 Pi=1 D 2
K (F, , D) . (2.39)
D i=1 i
Mean response
Additive (flat) noise
150 Proportional noise
Multiplicative noise
Response / Hz
100
50
0
4 3 2 1 0 1 2 3 4
Stimulus /
Figure 2.20: Example of a tuning curve and the variability of a model neuron.
Mean response (solid) +/ standard deviation. The mean, i. e. the tuning curve,
follows a Gaussian centered at the preferred stimulus with F = 100 Hz and a
baseline firing level of 10 Hz ( = 0.1). The shape of the response variance
depends on the noise model. Arrows indicate stimuli for which the neurons
Fisher information is maximal (from Wilke and Eurich, 2001:Lb).
Equation (2.41) shows that the -dependence for this kind of firing rate variance
is always proportional to (1)2 , regardless of the dimension D, see Figure 2.21.21
20
A qualitatively similar dependence on and D results for other tuning curve shapes,
e. g. cos2 -tuning.
21
A similar equation can be derived for correlated neural activity (Wilke and Eurich, 2001:La);
the dependence on and D remains the same for uniform noise correlation.
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-73
D=1
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
Baseline Activity Level
Further discussion of this issue can be found in Wilke and Eurich (2001:Lb, p.
164).
In Section 2.4.3.1 we discussed why neurons may specialize in a few stimulus
features rather than being sensitive for all kinds of stimuli. Next to the exponen-
tially growing number of neurons required to cover a high-dimensional stimulus
space and the positive effect of a fragmentation in subpopulations, the suscepti-
bility to background noise for non-additive noise could be another reason for this
type of neuronal organization.
|c(k) c(l) |
Qkl (x) = kl + (1 kl ) q + b exp afk (x) fl (x) (2.42)
L
|c(k) c(l) |
Qkl (x) = kl + (1 kl ) exp afk (x) fl (x) . (2.43)
L
Note that the free parameter allows the study of different variance-mean rela-
tions, just as in Section 2.4.3.2. = 0 corresponds to additive noise, = 1/2 to
Poisson-like noise, and = 1 to multiplicative noise. For D = 1, this situation
can be treated analytically (Wilke and Eurich, 2001:Lb, 2002).
An important point concerning limited-range correlations was recognized re-
cently by Yoon and Sompolinsky (1999): Adding neurons to a population with
given correlation range does not improve Fisher information beyond some limiting
value. In consequence, the encoding capacity of populations with limited-range
correlations appears to be severely limited. Our calculation shows, however, that
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-75
this conclusion is limited to the case of additive Gaussian noise (see Figure 2.22).
For noise models other than additive noise, Fisher information increases linearly
with the number of neurons (i. e., with the density of tuning curves), even in
the presence of limited-range correlations. Since the empirically found values for
are in the vicinity of 1/2, this result casts doubt on the conclusion that limited-
range correlations generally impose an upper bound on the encoding accuracy for
large populations.
To test this prediction, the population Fisher information was calculated for
a fixed positive uniform correlation level (q = 0.5) as a function of sign and
strength of an additional limited-range correlation contribution, as specified by
the parameter b in (2.42). The result is shown in the inset of Figure 2.23. For
positive short-range correlations, the encoding becomes worse, while it improves
for negative short-range correlations.22
The trend that uniform and limited-range correlations counteract holds for
a large range of values of the correlation length L. Figure 2.23 demonstrates
this result: The plot shows the level q of uniform correlations that is necessary
to compensate the negative effect of limited-range correlations of given range ,
where = exp(1/(L)). This quantity was determined by choosing a fixed
and increasing the level of uniform correlations q, while simultaneously decreasing
22
An analytical approximation is given in Wilke and Eurich (2001:Lb, p. 168); it is plotted
as the dashed line in the inset of Figure 2.23.
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-77
The diagonal elements are fixed at qii = 1 for i = 1, . . . , N , whereas the non-
diagonal elements qij for i < j are drawn from a Gaussian distribution with
mean q and standard deviation s, while the symmetry of Qij requires qji = qij .
Equation (2.44) corresponds to a variant of the Gaussian deformed ensemble
known from random matrix theory (Brody et al., 1981). The smallest eigenvalues
of the covariance matrix Q yield the largest contribution to Fisher information.
Random matrices have eigenvalues that are distributed around the eigenvalues of
the corresponding mean matrix. Thus, it is expected that eigenvalues smaller
than those of the corresponding covariance matrix without jitter occur in the
case of noisy correlations, leading to an increase of average Fisher information.
This is confirmed by expanding the inverse of the covariance matrix in terms
of s2 and neglecting terms of higher order. It turns out, that for D = 1, the
correction term is always positive (see Wilke and Eurich, 2001:Lb, p. 170 and
Appendix A.5). Hence, for a given mean correlation coefficient q, a population of
neurons with noisy correlation coefficients performs better on average than one
with uniform correlations.
This effect is demonstrated by a numerical simulation in Figure 2.24, where
the mean Fisher information of 105 covariance matrices is plotted as a function
of the standard deviation of the correlation coefficient. Figure 2.24 also shows
that the range of validity of the expansion is quite small. This is due tothe N -
dependence of the smallest eigenvalue, which is given by min = 1 q 2 N s in
this case (Wilke and Eurich, 2001:Lb, Appendix A.5). For N = 50 and q = 0.5,
this implies that the mean Fisher information must diverge around smax = (1
I-78 CHAPTER 2. NEURAL CODING
q)/ 4N 0.035, explaining why the quadratic approximation becomes invalid
as soon as s = 0.015. The numerical calculations indicate that the increase in
Fisher information is even more pronounced than predicted by the expansion
formula.
The result that correlations with jitter increase the Fisher information can be
interpreted as follows. A zero eigenvalue in the covariance matrix indicates the
existence of a linear combination of single-neuron activities that is completely
noiseless. This situation can of course not be achieved in realistic systems. Thus,
there must be limits to the choice of the covariance matrix in biological systems.
However, the form of these restrictions remains unclear, making a systematic op-
timization of the covariance matrix by theoretical analysis impossible. One may
therefore view the introduction of jitter in the correlation coefficients as a means
of randomly exploring the space of covariance matrices, which, according to the
above, yields a better average encoding accuracy than remaining on the subspace
of uniform correlation strength. The range of validity of this argument is limited
by the fact that the matrices eventually leave the admissible set.
Section 2.4.3 listed a multitude of parameters that influence the encoding accu-
racy of a neural population: the number of encoded features, tuning widths for
2.4. THE ENCODING ACCURACY OF A SINGLE OBJECT I-79
different features, the maximal firing rate, the noise model, background activity,
various kinds of noise correlations and their interaction, etc. Even subtle effects
like introducing a non-uniformity in the tuning widths or the correlation coef-
ficents may improve the sensory resolution. These results suggest that a large
number of empirical data are required to make reliable predictions for the signal
content of a population of neurons. Also, it is practically not feasible to argue
with optimal coding strategies as long as the biological constraints and the physi-
cal quantities that are (putatively) optimized are not completely identified. This
is even more true as long as the neural code is not identified: the current results
were obtained under the simple assumption of a rate coding.
f2(x1,x2)
f1(x)
x'' x2
x'
x' x'' x
x1
f2(x1,x2)
f1(x)
1.2
1
x2
0.8
c x
0.6
0.4
0.2
8 (c,c)
6 8
6
x2 4
(a) 2 2
4
x1 (b) x1
diagonal element of the inverse Fisher information matrix; 2 is the distance of
the point (x1 , x2 ) from the diagonal. Due to the symmetry, 21,min () is identical
to the minimal square error for x2 , 22,min (). The estimation error diverges as
Figure 2.27: Minimal square estimation error for stimulus x1 or x2 . Solid line:
F = 1; dotted line: F = 1.5. In both cases, k = 0.5, = 1, = 1, = 1; from
Eurich (2003:L).
20
1.5
(a) direction (b) direction
x1+x2 15
x2-x1
1/2 1/2
2 2
(r)
(r)
1 10
emin
emin
2
2
5
0.5 0
-4 -2 0 2 4 -4 -2 0 2 4
r r
Figure
2.28: Minimal square estimation error for (a) (x 1 + x 2 )/ 2 and (b) (x2
x1 )/ 2. Solid lines: F = 1; dotted lines: F = 1.5. Same parameters as in
Figure 2.27; from Eurich (2003:L).
20
(a) f2(x1,x2) (b) direction
1.2 15
x2-x1
1/2
1 2
(r)
0.8
0.6 10
emin
2
0.4
0.2
5
8
6 8
6
x2 4 4
x1
0
-4 -2 0 2 4
2 2
r
Figure 2.29: Neural encoding for one attended stimulus. (a) Tuning curve (2.45),
(2.46) for k = 0.75, i. e., stimulus x1 is attended. All other parameters asin
Figure 2.26a. (b) Minimal square estimation errors for the direction (x 2 x1 )/ 2
resulting from a diagonalized Fisher information matrix. Solid line: k = 0.5 as in
Figure 2.28b; dotted line: k = 0.75. F = 1, all other parameters as in Figure 2.27;
from Eurich (2003:L).
mation analysis reveals that the attended stimulus x1 yields a smaller minimal
square estimation error than it does in the non-attention case k = 0.5 whereas
the minimal square error for the unattended stimulus x2 is increased (data not
shown). Figure 2.29b shows the minimal square error for the difference of the
stimuli, (x2 x1 )/ 2. The minimal estimation error becomes larger as compared
to k = 0.5. This result can be interpreted as follows: Attending stimulus x 1
yields a better encoding of x1 but a worse encoding ofx2 . The latter results in
the larger estimation error for the difference (x2 x1 )/ 2 of the stimulus values.
I-84 CHAPTER 2. NEURAL CODING
Section 2.4 has shown that studying neural responses upon presentation of a
single object (comprising the case of multiple objects as described in the com-
pound space) yields far-reaching results for neural encoding strategies. Although
conceptually simple, it allows the identification of system parameters that are
important in the neural signal processing.
Chapter 3
Neural Dynamics
3.1 Introduction
This chapter deals with the dynamical systems approach to phenomena of the
nervous system. It uses a terminology different from the one in the functional
approach: instead of talking about encoding, decoding, representation, discrimi-
nation, information, object formation etc., the language switches to notions like
transients, steady states, limit cycles, stability properties, noise-induced tran-
sitions and the like. Accordingly, the mathematical methodology changes from
statistical estimation theory and information theory to dynamical systems theory
and differential equations.
Historically, the brain as a scientific object has been subject to a large number
of contradicting ideas, theories and fallacies, much more than any other organ.
For example, historical brain metaphors tended to reflect the respective world
view that pervaded the society. This is true not only for the middle ages but also
for the 20th century: In the 1950s many people considered the brain to operate
like a computer1 (see the overviews by Florey, 1996 and Gould, 1996).
Even today a fundamental understanding of brain processes und functions is
lacking. What is the reason for this situation? An explanation lies in the fact
that the brain is a complex system. Several factors contribute to this:
Brain processes take place on many time scales ranging from sub-millisecond
conformation changes in proteins of the cell membrane to weeks or months
in plastic changes in neural tissue, for example after the loss of a limb.
The different dynamical processes are not separated but are highly inter-
twined. For example, according to Hebbs postulate (Hebb, 1949), learning
as manifested in long-term metabolic or structural changes results from
1
Tthe program of Artificial Intelligence (AI) contained the idea of building expert systems
that incorporated a large body of knowledge on more or less specific topics. This knowledge
was accumulated in the form of data and rules that combined them. The general belief in the
AI community was that the brain has similar principles of symbolic knowledge organization.
I-85
I-86 CHAPTER 3. NEURAL DYNAMICS
-2 3
10 cortical areas 10
-3 2
10 macrocolumns 10 neuromodulation
LTP / LTD
-4 1
10 cortical neurons 10
Figure 3.1: Schematic overview of length scales (left) and time scales (right) in
the nervous system. The scale given on the left of each diagrams serves as a coarse
guide only: Structures and dynamics can vary greatly depending of the specific
system under consideration. LTP: Long-Term Potentiation, LTD: Long-Term
Depression.
3.1. INTRODUCTION I-87
Finally, processes in the nervous system are usually nonlinear, allowing for
analytical investigations only in rare cases. For example, the subthreshold
dynamics of integrate-and-fire neurons is linear, and a complete solution
theory is available (e. g., Tuckwell, 1988a, chapter 3). The firing thresh-
old, however, makes the system nonlinear and confines most investigations
to approximations or numerical methods. A rare (and therefore famous)
exception is the synchronization study by Mirollo and Strogatz (1990); typ-
ically, the existence and stability of steady states are proven in this article.
Even two-neuron systems modelled by integrate-and-fire neurons show a
rich dynamical behavior such as phase locking and multistability (Catsig-
eras and Budelli, 1992; Gomez and Budelli, 1996) and this behavior occurs
2
The Hodgkin-Huxley equations are an example also for dynamics on different time scales.
The dynamical variables can be separated in fast and slow ones which are subsequently merged
to obtain a simpler model, the so-called Fitzhugh-Nagumo neuron (Fitzhugh, 1961; Nagumo
et al., 1962).
I-88 CHAPTER 3. NEURAL DYNAMICS
In the light of the large range of time and length scales on which interactions
take place in the nervous system, the notion of neural dynamics as it is used here
includes not only the behavior of cells in a narrower sense but encompasses all
the different levels that elucidate phenomena in the nervous system.
models. But the dynamical system approach generally applies on all spatial and
temporal scales, for example in short-term synaptic dynamics (e. g., Tsodyks and
Markram, 1997) and Hebbian learning (Eurich et al., 1999:L). A brief overview
of the nonlinear dynamics approach for neural modeling is given by Arbib et al.
(1998, chapter 4).
A second, slightly different approach is located in statistical physics from a
topical and methodological point of view. It encompasses the study of systems
comprising a large number of interacting elements (particles, agents etc.) and
also strives for an understanding of general mechanisms. Typical phenomena
are self-organized criticality and stochastic resonance. They occur also in neural
networks.
In Section 3.2, a few topics in the field of neural dynamics will be highlighted.
Section 3.3 focuses on mechanisms in systems with delays and/or noise and inte-
grates my own research.
Models in discrete time. Among the simplest neuron models range those
used in neural networks that map an input to an output or whose dynamics is
defined on a discrete time axis. Figure 3.2 shows a schematic picture of such
x1
x2 w1
w2
w3 Figure 3.2: Model neuron for time-
x3 E f y
discrete networks. indicates a lin-
wn ear superposition of the inputs.
xn
a model. The neuron receives inputs x1 , . . . , xn from which the total excitation
e = w1 x1 + . . . + wn xn is computed. w1 , . . . , wn are synaptic weights, usually
modelled as real numbers (positive or negative for excitation or inhibition, re-
spectively). A transfer function f translates the excitation to the neural output
y. In the McCulloch-Pitts neuron (McCulloch and Pitts, 1943), neural outputs
y (and, consequently, also x1 , . . . , xn in a network) are binary; in this case, the
transfer function is a Heaviside function with some threshold excitation . In the
so-called analog neurons or sigmoidal neurons, y and x1 , . . . , xn take positive val-
ues, and the transfer function is some squashing function, for example the Fermi
function
1
f (e) = k(e)
(3.1)
1+e
with a gain parameter k. Instead of computing the excitation e by linear summa-
tion, multiplicative terms wj1 j2 ...jm xj1 xj2 . . . xjm may also be considered (so-called
sigma-pi units); cf. Brause (1991) for an introduction. Discrete-time models are
not used to study single-neuron properties but are preferably employed in the ma-
chine learning community and find applications in classical networks like Hopfield
networks and multilayer perceptrons (e. g., Hertz et al., 1991; Brause, 1991). In
such neural networks, however, the abovementioned problem of multi-dimensional
parameter spaces arises.
fields require a real time axis and the corresponding appropriate neuron models.
Such models are usually based on the nonlinear electric properties of cell mem-
branes; the dynamical variable for describing neural behavior is the membrane
potential (see Chapters 5 and 6 in Dayan and Abbott, 2001 for a systematic
textbook introduction to various neuron models in continuous time).
A basic distinction can be made between temporal and spatiotemporal neural
behavior. The former case concerns the modelling of a patch of membrane with-
out considering the spatial extent of the neural surface. The purely temporal
behavior also gives rise to simple neural models which neglect a spatial structure
(point neurons). In the latter case, the spread of excitation or inhibition in the
cell is studied in the form of passive conduction (in dendrites) or action poten-
tials (in dendrites and in the axon). The physiology and morphology of extended
cells is considered in so-called multi-compartment models. With respect to this no-
tion, purely temporal models are sometimes also referred to as single-compartment
models.
Models with purely temporal dynamics. In brief, the basic equation for
all single-compartment models is
dV Ie
cm = im + , (3.2)
dt A
where V is the membrane potential, cm is the specific membrane capacitance,
im is the membrane current per unit area, Ie is an electrode current, and A is
the total surface area of the cell (Dayan and Abbott, 2001). The different signs
on the right hand side of (3.2) are convention. According to the basic equation,
the membrance creates a capacitance due to the separation of charges carried
by ions like sodium and potassium. The dynamics of the membrane potential is
further determined by various membrane currents that can be modelled as Ohmic
currents or currents with voltage-dependent conductances.
The simplest model resulting from the model class defined by (3.2) was introduced
by Lapicque (1907); see Abbott (1999) for a historical account. It is now referred
to as the integrate-and-fire neuron. In this model, all voltage-dependent (active)
conductances are ignored such that the basic equation does not produce action
potentials but considers the subthreshold behavior only. Equation (3.2) results
in
dV V Ie (t)
+ = , (3.3)
dt C
where V is now the deviation of the membrane potential from the neurons resting
potential, C is the total membrane capacitance, and = rm cm = RC is the
membrane time constant, rm is the specific membrane resistance, and R is the
total membrane resistance. The subthreshold neuron behaves like an electric
circuit consisting of a resistor and a capacitor in parallel (Figure 3.3).
I-92 CHAPTER 3. NEURAL DYNAMICS
V (ts ) V (t+
s ) = 0, (3.4)
where V (t+s ) lim0,>0 V (ts + ). The firing threshold makes the neuron non-
linear, allowing for analytical results only in rare cases. For example, even a
simple oscillatory input Ie (t) = sin t yields spike events that cannot be written
down in closed form (Tuckwell, 1988a).
While the integrate-and-fire neuron (3.3,3.4) describes the simplest possible neu-
ron of the model class defined by (3.2), it can be equipped with additional features
to capture physiological properties in case they are important to account for a
system under study. Such features include the following.
is the spike response model (Gerstner and van Hemmen, 1993; Gerstner,
1995). The synaptic kernel yields the neural response to incoming spikes
and usually contains a transmission delay. Gerstner (1995) shows that the
integrate-and-fire neuron is a special case of the spike response model.
1
im = (V EL ) + g K n4 (V EK ) + g Na m3 h(V ENa ) (3.6)
rm
in the basic equation (3.2). The first term describes the leak current, the second
and third terms are the potassium and the sodium current, respectively. EL ,
EK , ENa denote reversal potentials, and g K , g Na are maximal conductances. The
voltage-dependence enters through the variables n, m, h which are called gating
variables. They describe the probability of a channel subunit to be in a conforma-
tion that corresponds to an open channel. Each gating variable satisfies a Master
equation with voltage-dependent parameters; the exponent with which it occurs
corresponds to the number of subunits in the respective channel. For details of
the Hodgkin-Huxley model, see Johnston and Wu (1995).
The Hodgkin-Huxley model produces action potentials and captures the phys-
iology of some types of neurons (for example, action potentials in the giant axon
of the squid for which the theory was developed). Although it describes only
the production of action potentials in the first place, it is also used as a neuron
model.
Further neuron models have been proposed. For example, Fitzhugh (1961);
Nagumo et al. (1962) introduced a simplification of the Hodgkin-Huxley neuron
by combining those dynamical variables that act on a similar time scale, resulting
in a two-dimensional model. Likewise, Abbott and Kepler (1990) perform a
stepwise reduction of the Hodgkin-Huxley dynamics, resulting in an integrate-
and-fire type model or in a binary neuron.
A more frequently employed model class uses the purely temporal dynamics
(3.2) exclusively in all compartments and connects them by longitudinal resis-
tances. An easy approach is the subthreshold integrate-and-fire dynamics (3.3).
One of the compartments (the somatic compartment) has a threshold condition
of the form (3.4) whereas the others are not active.
The number of compartments varies greatly according to the goal of the mod-
elling study. An important step is the use of two-compartment models (Rospars
and Lansky, 1993; Bressloff, 1995), because it is the simplest spatially extended
neuron and yields a behavior that is qualitatively different from the behavior of
an integrate-and-fire neuron. For example, an integrate-and-fire neuron has no
memory in the sense that a after a spike, the membrane potential is reset, and
information about the dynamics prior to the spike is lost. If a dendritic compart-
ment is attached, however, its membrane potential is not reset and reflects the
history of the neuron.
Some modelling studies strive for a detailed investigation of the signal pro-
cessing in individual neurons. A typical procedure is the exact reconstruction of
a living cell and a subsequent compartmentalization (in particular, of the den-
dritic tree), whereby each compartment is assigned parameters according to the
morphology and the physiology of the corresponding part of the cell. A com-
plete model may thus consist of thousands of compartments. Software packages
like GENESIS (Bower and Beeman, 1998, http://genesis.bbb.caltech.edu/
GENESIS/genesis.html) or NEURON (Hines and Carnevale, 1997, http://www.
neuron.yale.edu) support such an approach.
3.2.2 Networks
In neural network modelling, one may distinguish between dynamics in discrete
time and in continuous time. A second criterion is the modelling of individual
neurons versus the consideration of neural densities or neural fields. Traditional
neural network approaches (Hertz et al., 1991; Rojas, 1993) are mostly based on
single neurons in discrete time. Such models, however, suffer from the difficulties
described in Section 3.1: the large number of parameters does not allow for in-
ferences on biological network structures, and the networks are hard to analyze.
This is true also for new approaches where emphasis is laid on transient activi-
ties, so-called liquid state machines, (Maass et al., 2002) and echo state networks
(Jaeger, 2001a,b).
There are some exceptions to this rule. Hopfield networks (Hopfield, 1982,
1984) allow for an imprinting of stable fixed points of the dynamics which may
serve as memory patterns. Analyses are also possible in networks with a sim-
ple, uniform connectivity. Mirollo and Strogatz (1990) consider a network of
excitatory integrate-and-fire type oscillators whose time-continuous dynamics is
mapped to a time-discrete dynamics via a Poincare map. Their approach yields
an analytical proof that neurons show in-phase synchronization in finite time
I-96 CHAPTER 3. NEURAL DYNAMICS
for almost all initial conditions. Ernst et al. (1995, 1998); Timme et al. (2002)
extend the model with respect to inhibition and transmission delays. The re-
sulting network behavior is different from the case without delays: neurons form
clusters which are stable only for inhibitory interactions, whereas excitatory in-
teractions yield clusters which decompose and self-organize again with changing
members. More recently, Eurich et al. (2002:La) studied avalanches of spike ac-
tivity in uniform networks of integrate-and-fire neurons and were able to derive
a closed expression for avalanche size distributions in networks of arbitrary size;
see Section 3.3.2.1 for further details.
term dynamcis (cf. Figure 3.1). First, synapses are known to show a variability
on short time scales ranging from milliseconds to seconds (e. g., , Markram and
Tsodyks, 1996; see also Fisher et al., 1997; Zador and Dobrunz, 1997 for reviews).
Second, synaptic changes are often identified with learning 4 on time scales ranging
from minutes to hours, days, or longer.
The existence of dynamic synapses may have far-reaching consequences for the
dynamics and signal processing in neural populations. From a dynamical point
of view, it provides new degrees of freedom resulting in a rich dynamical behavior
(e. g., Tsodyks et al., 1998). From a signal processing point of view, synapses
may serve as a memory device reflecting past activity of presynaptic neurons
(Maass and Markram, 2002).
Long-term dynamics. Plasticity in the nervous system has long been associ-
ated with the dynamics of synapses. Hebb (1949) formulated the following famous
postulate about learning: When an axon of cell A is near enough to
excite a cell B and repeatedly or persistently takes part in firing
it, some growth process or metabolic change takes place in one or
both cells such that As efficiency, as one of the cells firing B,
is increased. Self-organized modifications of the synaptic site have since been
the main candidate for adaptivity in the nervous system, although Hebb also
considered other possible physiological realizations including axonal or dendritic
growth. The Hebbian postulate triggered a worldwide research program which
has been going for more than 50 years (Seung, 2000).
In the 1970s, the phenomena of Long-Term Potentiation (LTP) and Long-
Term Depression (LTD) were discovered, i. e., the increase (LTP) or decrease
(LTD) in the amplitude of postsynaptic potentials following a concurrent strong
stimulation of the presynaptic and the postsynaptic cell; see Brown et al. (1988,
1990); Bliss and Collingridge (1993) for reviews. Whereas these investigations
revealed only a coarse picture of the conditions under which plasticity can be
observed, the precise temporal mechanisms on a single-spike basis have only re-
cently been discovered (e. g., Bi and Poo, 1998; Zhang et al., 1998b; Egger et al.,
1999; Bi and Poo, 1999). Basically, a critical time window (now referred to as the
learning window ) of about 50 ms around postsynaptic spike events exists during
which incoming spikes result in a change of the average postsynaptic potential.
For example, in Xenopus tectum neurons (Zhang et al., 1998b), action potentials
arriving prior to a postsynaptic event yield an enhancement of postsynaptic ac-
tivity while action potentials arriving later trigger a long-term depression. The
temporal conditions for LTP and LTD, however, may differ depending on the
system under study; a review is given in Abbott and Nelson (2000). Engert and
Bonhoeffer (1999) show that LTP is accompanied by the formation of new spines
on the dendrites of postsynaptic neurons, thus establishing the tight correlation
between structure and dynamics in the nervous system. Finally, Rioult-Pedotti
et al. (2000) convincingly demonstrate in a combined behavioral and electro-
physiological experiment in rats that the occurrence of LTP is indeed a neural
correlate of learning a conjecture which had never been shown before.
In theoretical studies, synaptic dynamics in time-discrete neural network mod-
els prevailed for a long time. Most types of such neural networks have two kinds
3.3. MECHANISMS IN SYSTEMS WITH DELAYS AND NOISE I-99
of dynamics: a neural dynamics on a fast time scale, and learning that is,
changes in the synaptic weights on a slow time scale. Hebbian learning is one
type of unsupervised learning; a typical learning rule is given by
where wji (k + 1) and wji (k) denote the synaptic weight between presynaptic
neuron i and postsynaptic neuron j at times k + 1 and k, respectively; is
a small learning rate; xi is the input to neuron j from neuron i; and yi is the
output of neuron j. A typical problem of Hebbian learning with a rule like (3.8) is
that weights may diverge; constraints like clipping or normalization of a neurons
synaptic weights have to apply to prevent this (cf. Miller and MacKay, 1994 for
a corresponding discussion). A review on Hebbian learning rules in discrete time
is given by Brown et al. (1990).
Theoretical investigations on Hebbian learning in continuous time are based
on the new experimental findings on learning windows. Since the first such learn-
ing study motivated by the auditory system of barn owls (Gerstner et al., 1996),
numerous investigations have been published with applications in coincidence de-
tection, sequence learning, and path learning in navigation (e. g., Abbott and
Blum, 1996; Eurich et al., 1999:L; Kempter et al., 1999; Mehta et al., 2000). For
an example, see Section 3.3.1.3.
1 Synapses
0.3
(a)
V(t) [mV]
1 (b)
2 0
0 t [ms] 15
2 Dendritic Tree
3
3 Axon
Figure 3.5: Delays arising in a single neuron. At the synapses (1), it is in par-
ticular the postsynaptic potential (EPSP or IPSP) that takes time to evolve.
The graph shows two EPSPs as modelled by alpha functions; EPSP (a) has a
higher amplitude than EPSP (b) and it is also faster. Signal processing in the
dendritic tree (2) is also associated with delays which may differ depending on
the morphology of the neuron. In the colliculus, for example, the passive spread
of postsynaptic potentials takes 5 25 ms (A. Schierwagen, personal communi-
cation). Spikes backpropagating into the dendritic take 3 4 ms to reach the
most distal branches in cortical pyramidal cells (Stuart and Sakmann, 1994). In
the axon, conduction velocities range from 0.1 20 m/s in unmyelinated fibers
to 100 m/s in myelinated fibers (Reichert, 1990) depending, for example, on the
diameter of the fibre and properties of the myelin sheath. The photo (taken from
Hsiao et al. (1984)) shows the cross section of the optic nerve of a cat; it suggests
that conduction velocities may differ from fibre to fibre.
with delays may show a rich dynamical behavior ranging from steady state be-
I-102 CHAPTER 3. NEURAL DYNAMICS
havior and limit-cycle oscillations to chaotic behavior and phenomena that can
be ascribed to the presence of both noise and delays. For example, the synchro-
nization of spike activity in neural populations, as modelled by integrate-and-fire
neurons, is fundamentally different in systems with delay (Ernst et al., 1995,
1998) as compared to networks without delay (e. g., Mirollo and Strogatz, 1990):
without delays, excitatory connectiviy results in an in-phase synchronization of
all oscillators whereas in the presence of delays, clusters of synchronous neurons
emerge and decay after some time; see also Timme et al. (2002). In the follow-
ing paragraphs, a few topics in the context of interaction delays will be briefly
discussed.
The effects of noise and delay in human postural sway. As an examle for
the effect of a delay in a sensorimotor feedback system when it is combined with
noise, we consider the control of simple upright stance in human subjects. Data
are obtained either from a force platform that records the body center-of pressure
(Collins and De Luca, 1993, 1995a; Collins et al., 1995c,d) or from an ultrasonic
3.3. MECHANISMS IN SYSTEMS WITH DELAYS AND NOISE I-103
device that records the position of a microphone attached to the body (Franke,
2003). Two examples obtained by the latter method are shown in Figure 3.6.
15 10
10
5
x,y [mm]
y [mm]
0
0
5
5
PSfrag replacements
PSfrag replacements 10
10
x
y
15 15
15 10 5 0 5 10 15 0 20 40 60 80 100 120
x [mm] t [s]
15 15
10 10
5
5
x,y [mm]
y [mm]
0
0
5
5
PSfrag replacements 10
PSfrag replacements
10 15
x
y
15 20
15 10 5 0 5 10 15 0 20 40 60 80 100 120
x [mm] t [s]
ln K( t)
1
-1
-3 -2 -1 0 1 2 3
ln t
Figure 3.7: A two-point correlation function, K(t), for a healthy young subject.
Vertical lines segment K(t) into different scaling regions: (from left to right),
H 0.86, H 0.29, H 0 (Chow and Collins, 1995). Experimental data were
supplied by and published with permission of J. Collins. From Eurich and Milton
(1996:L).
R
{ j
Figure 3.8: The body (mass m,
height R of the center of mass) as
an inverted pendulum.
{
The model has three parameters, a (rescaled) delay , a (rescaled) restoring force
C and a (rescaled) noise intensity D. There are two thresholds at 1 and 1
6
The dynamics are not sensitive with respect to the exact mathematical form of the restoring
force.
3.3. MECHANISMS IN SYSTEMS WITH DELAYS AND NOISE I-105
beyond which a control force is exerted; in the central region, the body is subject
to gravitation and noise only.
In the noiseless case D = 0, solutions of (3.11) can be calculated analytically.
The bifurcation diagram of the system is shown in Figure 3.9. Depending on
1.0 E
O3
O2
0.5
c
a
O1
b
0.0
1 2 3 4
C
and C, the system shows three different types periodic orbits two of which(O2
and O3) comprise both thresholds at 1 and 1. O1 is an oscillation around a
threshold, i. e., two co-existing orbits around the left and right sensory detection
threshold exist in the system (see Figure 4 in Eurich and Milton (1996:L)). For
sufficiently long delays and/or small restoring forces C, the motion is unbounded
which corresponds to the loss of stable upright stance. For D 6= 0, physiolog-
ical parameters D, and C can be found numerically by fitting the two-point
correlation functions to real data. The following results have been obtained:
Parameter values for the delay and the detection threshold in the step-
wise feedback function as obtained from the model closely match empirical
values.
For a single subject, the delay and the threshold are constant, while single
trajectories differ only in the noise intensities. This can be expected from
the physiological interpretation of the parameters: the delay and the detec-
tion threshold are fixed in a subject (or may change on a longer time scale)
while noise intensity (resulting from fatigue, the consumption of coffee etc.)
may varies on short time scales.
I-106 CHAPTER 3. NEURAL DYNAMICS
In all cases that were tested, the system seems to be in the regime of
oscillation O1 of the bifurcation diagram.
Noise and delays in balancing tasks. Milton et al. (2000) and Cabrera and
Milton (2002) studied the balancing of sticks on the fingertip in human subjects.
The motion of the stick is typically composed of small-amplitude movements
(laminar phases) interrupted by larger-amplitude corrective movements. Both
the duration of the laminar phases and the power spectrum of deviations of the
stick from the vertical position hint at on-off intermittency as the underlying dy-
namical phenomenon (e. g., Platt et al., 1993; Heagy et al., 1994; Venkataramani
et al., 1996). A dynamical model with delay put forward in Cabrera and Milton
(2002) employs parametric noise (rather than additive noise as in the postural
sway) and locates the system close to a stability boundary in the corresponding
parameter space. The authors argue that fluctuations in the angle of the stick
resemble a random walk for which the mean value is approximately zero; i. e., the
stick is statistically stabilized.
Bormann et al. (in press) carry this line of research forward by introducing a
new experimental paradigm: the balancing of a target on a computer screen
by manual control of a computer mouse. This system is simpler than the stick
balancing because it operates in 2D. Furthermore, the artificial computer world
allows for an arbitrary manipulation of the dynamics of the target. For example,
additional delays can be introduced into the sensorimotor feedback loop to study
its effects on the dynamics. Finally, data are simpler to obtain just by recording
the mouse and target positions. A series of experiments suggests that stick bal-
ancing and mouse balancing are governed by the same statistics. A current issue
is the search for common motor control schemes and their underlying dynamical
mechanisms.
(e. g., Mackey and Glass, 1977; Mackey and an der Heiden, 1984; Milton et al.,
1990; Eurich and Milton, 1996:L; Bormann et al., in press). However, this case
3.3. MECHANISMS IN SYSTEMS WITH DELAYS AND NOISE I-107
may be oversimplified: current system changes may depend on the system be-
havior during a whole past time interval rather than on a single past instance
only. An important question is therefore if theoretical predictions change if delay
distributions are considered.
A straigtforward extension of (3.12) is a linear superposition of past instances,
Z
x(t) = f [x(t), x(t )( ) d ] , (3.13)
0
( ) = ( 0 ) , (3.14)
In (3.17), H is the Heaviside step function, and is a kernel function like in (3.13)
adapted to the distribution of fibre diameters. The rescaled activation threshold
( ) for the generation of an action potential with a rescaled delay time yields
an activity-dependent delay distribution in the sense that the actual minimal
delay depends on the systems activity in the past.
Analytical calculations and numerical computations suggest that the system
shows a bistability in the physiologically plausible parameter regime: two coex-
isting fixed points corresponding to firing frequencies in the range of 10 25 Hz
and 30 100 Hz. They might correspond to the and frequencies, respectively
(Traub et al., 1999). In comparison with a similar model by Mackey and an der
Heiden (1984) which employs neither a distribution of delay times nor a state-
dependence thereof higher order bifurcations and a transition to chaos get lost
in the physiological regime as the variance of the distribution is progressively in-
creased. That is, the dependence of the dynamics on a whole past interval yields
a simpler behavior than the case of a singular delay time, just like in the linear
case mentioned above (Bernard et al., 2001).
Smooting with delay distributions. Thiel et al. (2002); Thiel and Eurich
(2002) and Thiel et al. (2003:L) study the dynamical effects of the width of the
delay distribution more systematically in various nonlinear systems. As an exam-
ple, consider again the negative feedback system (3.16,3.17) but with a constant
activation threshold, ( ) 0 . Figure 3.10 illustrates that the dynamics of the
system becomes simpler as the length of the past interval that determines the
current behavior increases. Starting from apparently chaotic behavior, trajecto-
ries become periodic, whereby the period number decreases as a function of the
width of the rectangular distribution ( ). The scenario resembles an inverse
period-doubling cascade. This phenomenon is further investigated in Figure 3.11,
where the period number of oscillations is grey-scale coded as a function of and
the gain of the feedback term. It is clearly visible that complex oscillations (or
even chaotic behavior) is restricted to feedback with delay distributions having a
small variance.
A comparable behavior can be found in the Mackey-Glass system (Mackey
and Glass, 1977) with distributed delays (Thiel et al., 2003:L). Finally, a transi-
tion from oscillatory to fixed point behavior occurs in a model for single-species
dynamics in population ecology. The population density N satisfies the delay-
3.3. MECHANISMS IN SYSTEMS WITH DELAYS AND NOISE I-109
4 =0.0 2 2
3 1 1
v(t-1)
()
v(t)
2 0 0
1 -1 -1
0 -2 -2
4 =0.2 2 2
3 1 1
v(t-1)
()
v(t)
2 0 0
1 -1 -1
0 -2 -2
4 =0.4 2 2
3 1 1
v(t-1)
()
v(t)
2 0 0
1 -1 -1
0 -2 -2
4 =0.6 2 2
3 1 1
v(t-1)
()
v(t)
2 0 0
1 -1 -1
0 -2 -2
0 1 2 0 5 10 -2 -1 0 1 2
t v(t)
Figure 3.10: Distributions of delay times (left) used to compute the membrane
potential v as a function of time (center) in simulations of the hippocampus
model with distributed delays. Time is measured in units of the mean time lag
m = 1 of the recurrent signals, the interval shown corresponds to 1 s in real time.
Additionally, phase portraits v(t m ) as a function of v(t) are shown (right),
obtained by two dimensional embedding with delay coordinates. = 0.0 (top
row) denotes the case of a singular time lag m = 1. Transients are omitted. Bars
above the potential plots in the center column exemplify the integration interval
of past potentials for the given distribution, arrows mark the instant when this
integrated feedback acts on the change of v. From Thiel et al. (2003:L).
differential equation
Z
dN (t) 1
= rN (t) 1 N (t )( ) d (3.18)
dt K
0
1.0
0.8 76
0.6
0.4
0.2 1
0.0
0 50 100 150 200
The results obtained for the different feedback systems suggest that one may
obtain an inappropriate description of the dynamics if only mean values of inter-
action delays are considered. A diversity in time lags, as it is frequently observed
in nature, may have a major impact on the systems time evolution. In particular,
the irregular behavior observed in recurrent feedback models with delays could
be a model artifact resulting from oversimplified assumptions on the nature of
the feedback.
3.3. MECHANISMS IN SYSTEMS WITH DELAYS AND NOISE I-111
(a)6 0.004
2=0.0
0.003
4
N(t)
()
0.002
2
0.001
0 0.000
(b)6 0.004
2=0.5
0.003
4
N(t)
()
0.002
2
0.001
0 0.000
(c) 6 0.004
2=1.0
0.003
4
N(t)
()
0.002
2
0.001
0 0.000
0 20 40 60 80 100 0 2 4 6
m
t
Figure 3.12: Left: Time course of the population density N (t) of a single species
simulated by numerically solving (3.18) for 100 arbitrary time units. Right:
Distributions ( ) of delay times used for the simulations. The variance 2 of
the distributions increases from top to bottom while the mean delay m = 1.5 is
the same in all three cases (a): 2 = 0.0, corresponding to a singular delay time,
(b): 2 = 0.5; (c): 2 = 1.0. As 2 is increased, oscillations in population density
are reduced in amplitude compared to the case of a singular delay. Limit cycles
can even be replaced by an approach to a stable equilibrium if the variance is
sufficiently large. From Thiel et al. (2003:L).
most prominent example is the auditory system of barn owls (e. g., Knudsen
et al., 1977; Knudsen and Konishi, 1978; Konishi, 1986; Carr and Konishi, 1990;
Konishi, 1991; Wagner and Takahashi, 1992; Carr, 1993) where azimuthal object
localization has been shown to rely on the time (or phase) difference of sound
arriving at both ears. The mechanism consisting of signals that are phase-locked
to the input, a series of delay lines, and binaural neurons that realize coincidence
detectors was originally proposed by Jeffress (1948).7 Time-sensitive object
localization is supposed to play a role also in other species and other sensory
modalities; see Carr and Konishi (1990) for a review.
There exist more hints on adapted delays, in particular on equalized delays
because this special case of adaptation can be easily recognized as such. Stanford
(1987) measured the conduction time of X-cells in the cat retina. He noticed a
considerable variability in the intraretinal conduction times (due to the missing
myelination and the different distances between the somata and the blind spot)
but only small total conduction time differences: extraretinal signal conduction
to the LGN compensates for the intraretinal time differences. Innocenti et al.
(1994) analyzed axonal arborization in cat visual neurons with callosal axons. A
detailed reconstruction of the axons and a corresponding computational study
on conduction times suggests that the geometry of the fibres is adapted to syn-
chronously activating spatially separate target columns.
How does this adaptation evolve? The usual hypothesis is that unspecific tis-
sue (neurons, dendritic branches, synapses) grows and is subsequently selected by
Hebbian mechanisms; see for example Gerstner et al. (1996) for a model on the
auditory system of the barn owl. A different hypothesis is an adaptivity of the de-
lays themselves, for example through an adaptation of the overall neural (dendritc,
somatic, axonal, synaptic) morphology, through the positioning of synapses, or
even through an adaptive myelination (see Demerens et al., 1996; Stevens et al.,
1998, for the latter). A few models have been developed that employ the adap-
tation of time delays in machine learning (Day and Davenport, 1993; Baldi and
Atiya, 1994; Draye et al., 1996) and also in a biological context (Vibert et al.,
1994; Eurich et al., 1997a, 1998; Huning et al., 1998; Eurich et al., 1999:L; Steuber
and Willshaw, 1999; Eurich et al., 2000; Tversky and Miikulainen, 2002).
Eurich et al. (1999:L) consider self-organized delay adaptation according to
the scheme of spike-timing dependent plasticity (STDP) as it has recently been
discovered (e. g., Bi and Poo (1998); Zhang et al. (1998b); Bi and Poo (1999);
Egger et al. (1999); for modelling and reviews, see Gerstner et al. (1996); Song
et al. (2000); Abbott and Nelson (2000); Bi and Poo (2001)). The experimental
studies have shown that the strengthening or weakening of synapses depends on
the relative timing of pre- and postsynaptic action potential within a short time
interval of approximately 50 ms, the so-called learning window . In Eurich et al.
7
More recent investigations suggest that the auditory system of mammals does not fit this
model (McAlpine and Grothe, 2003).
3.3. MECHANISMS IN SYSTEMS WITH DELAYS AND NOISE I-113
1 2 3 N-2 N-1 N
1 2 3 2 1 x
W (x)
x
(a) (b)
Figure 3.13: (a) Schematic overview of the network under study. It is composed
of N presynaptic neurons and a single postsynaptic neuron. ti,k denotes the
kth spike time in the ith presynaptic neuron, i is the conduction delay, and i
the corresponing synaptic weight (i = 1, . . . , N ). (b) Examples for the window
functions W and W for delay selection and delay shift, respectively. From
Eurich et al. (1999:L).
An overview of the model is shown in Figure 3.13. The network under consid-
eration consists of N presynaptic neurons. The kth spike time of the ith neuron
is denoted as ti,k . After a conduction delay i in the ith fibre, a spike contributes
to the postsynaptic membrane potential with a synaptic weight i (i = 1, . . . , N ).
The postsynaptic neuron is modelled as an integrate-and-firen neuron (Tuckwell,
1988a, chapter 3). After each spike of the postsynaptic neuron at time ts , learn-
ing takes place. The delay shift rule reads i W (ti,k + i ts ) (i = 1, . . . , N ),
where W () represents the delay shift learning window, an example of which is
shown in Figure 3.13b. Numerical investigations show that this network self-
organizes in a way that spikes arrive simultaneously at the postsynaptic neuron,
thus enhancing the impact of single spikes. An example is shown in Figure 3.14.
To obtain analytical results on attractors of the delay dynamics and their sta-
bility, a continuous formulation of the problem is used. For simplicity, we assume
I-114 CHAPTER 3. NEURAL DYNAMICS
(a)
(b)
0.04
p()
0.02
0 1.3 1.4
1 1.1 1.2 1.5
0.1
(c)
p()
0.05
0
1.1 1.15 1.2 1.25 1.3
Figure 3.14: Delay shift dynamics with Poissonian input. (a) Schematic drawing
of the spiking behavior of three (out of a total of 400) presynaptic neurons. Action
potentials occur independently in all neurons according to a Poisson point process
with rate rb . During presentation of a stimulus (bar), this rate is simultaneously
increased to rp in all presynaptic neurons. (b) Randomized delay distribution
p( ) prior to learning and (c) after 1,700 learning steps. In this case, the delays
become very similar to each other due to the simultaneous shift in the rate; the
remaining variance of the distribution is due to the stochastic nature of the spikes.
Adapted from Eurich et al. (1999:L).
P
times of the postsynaptic neuron, P (, t) m ( m ), where m denotes
the m-th spike time. The firing of the postsynaptic neuron in turn acts on the
weights (, t) and the delays (, t) via one of the learning rules described in the
previous section. The dynamics of the input are governed by two simultaneous
equations: a balance equation for the input density,
J(, t) = (J(, t)v(, t)) + Q(, t) , (3.20)
t
and a continuiuty equation for (, t) indicating the conservation of the number
of neural connections,
(, t) = ((, t)v(, t)) . (3.21)
t
The drift velocity, v(, t), and the source term, Q(, t) are defined according to
Hebbian principles. Here we consider the special case of delay shift only; the case
of delay selection is described in Eurich et al. (1999:L).
During delay shift, the weights are not modified and the source term, Q(, t),
on the right hand side of (3.20) vanishes. The dynamics are governed by (3.21),
where the drift velocity, v = d /dt, of the delays realizes the Hebbian adaptation,
Z
v(, t) := W ( 0 )P ( 0 , t) d 0 , (3.22)
and denotes the learning rate. For the distribution of spike times we assume
a linear neural response, P (, t) = J(, t) it has been shown that adding a
small amount of noise to the input approximately linearizes the behavior (Yu
and Lewis, 1989), so this assumption is valid if there is some random background
activity. Figure 2 in Eurich et al. (1999:L) suggests that this can be true also
without additional noise.
In Eurich et al. (1999:L) is was shown that these delayPshift dynamics has
an equilibrium solution (, t) = ( 0 ) provided that n= W (nt) = 0
which is the case for antisymmetric window functions. The solutions form a
one-dimensional manifold described by a parameter 0 [0; T ] which R is a delay
Roffset0 common to all input neurons. The Liapunov functional L[] = (, t)(
( , t) 0 d 0 )2 d yields the result that the equilibrium solutions are marginally
stable in the 0 direction and stable in the other directions provided that W (x) >
0 for x < 0 and W (x) < 0 for x > 0. In other words, the self-organizing
process yields equalized delays as they are observed in biological neural systems.
Figure 3.15 illustrates this result.
Figure 3.15: Numerical iteration of (3.21) with W as in Figure 3.13b. (a) Initial
distribution of the density of neural connections with delay , (, 0) = 1 + ( ),
where ( ) is Gaussian white noise. (b)-(d) (, t) for t = 12.0, 18.0 and 20.0,
respectively. Two peaks transiently emerge and the merge into a single delta
peak which corresponds to the equilibrium solution of the distribution of delays.
T = 1.0, = 0.1, c = 0.2. From Eurich et al. (1999:L).
under the title of benefits of noise. This notion emphasizes the observation that
irregular dynamical behavior is not necessarily a nuisance in the sense that it
reduces signal transmission rates or interferes with some clear-cut dynamics by
blurring trajectories. For example, Silberberg et al. (in press) investigate the
response of neural populations to uncorrelated inputs and find both empirically
and theoretically that the variance of the input regarded as a signal enables
faithful transmission of graded signals with a temporal resolution that is much
higher than in the transmission of the mean input. A second example is the
observation by Yu and Lewis (1989) that the nonlinear process of spike generation
in neuron models is linearized by adding noise.
Two topics in statistical physics that are relevant also in neuroscience have
been discussed intensively in the literature for approximately two decades: Stochas-
tic resonance and self-organized criticality or, more generally, avalanches and the
occurrence of power-law behavior . The number of publications in each field is of
the order of 103 which makes even a coarse overview impossible in the context of
this thesis. We will therefore briefly discuss two examples.
it is the case in phase transitions (e. g., Bak et al., 1988; Kadanoff et al., 1989;
Grinstein et al., 1990; Pietronero et al., 1994; Sornette et al., 1995; Vespignani
et al., 1996; Vespignani and Zapperi, 1997; Dhar and Ramaswamy, 1989; Dhar,
1990; Hwa and Kardar, 1989; Tsuchiya and Katori, 2000). Analytical results
were derived for sandpile models (Dhar and Ramaswamy, 1989; Dhar, 1990), and
it was shown that the existence of a conservation law is a necessary prerequisite
to obtain SOC (Hwa and Kardar, 1989; Tsuchiya and Katori, 2000). A minimal
model exhibiting SOC is presented in Nagler et al. (1999).
A second class of models inspired by earthquake dynamics employs continu-
ous driving and nonconservative interaction between the elements of the system
(Feder and Feder, 1991; Olami et al., 1992). In the Olami-Feder-Christensen
(OFC) model (Olami et al., 1992), the amount of dissipation is controlled by a
parameter , and power-law behavior of avalanches occurs for a wide range of
values. Subsequent investigations emphasized the importance of boundary con-
ditions and tied the existence of the observed scaling behavior to synchronization
phenomena induced by spatial inhomogeneities (Christensen and Olami, 1992;
Socolar et al., 1993; Grassberger, 1994; Corral et al., 1995; Middleton and Tang,
1995). More specifically, Lise and Jensen (1996) introduced a random-neighbor
interaction in the OFC model to avoid the buildup of spatial correlations. Fur-
ther analysis indeed revealed that the random-neighbor OFC model does not
display SOC in the dissipative regime (Broker and Grassberger, 1997; Chabanol
and Hakim, 1997; Kinouchi et al., 1998). In these avalanche models with non-
conservative interaction, analytical results were obtained only for system size
N , i. e., in the Abelian case (Broker and Grassberger, 1995, 1997). The
model introduced by Eurich et al. (2002:La) circumvents the problem of system
boundaries, and also yields an analytical access for finite system sizes N .
=0.8000 =0.9900
(a) 0 (b) 0
p(x,N,), numerical p(x,N,), numerical
1 p(x,N,), analytical 1 p(x,N,), analytical
pd(x,N,), numerical pd(x,N,), numerical
2 2
3 3
4 4
5 5
6 6
7 7
4 3 2 1 0 4 3 2 1 0
=0.9990 =0.9999
(c) 0 (d) 2
p(x,N,), numerical
1 p(x,N,), analytical
pd(x,N,), numerical
3
2
3 4
4 5
5
6
6 p(x,N,), numerical
pd(x,N,), numerical
7 7
4 3 2 1 0 4 3 2 1 0
log x/N log10 x/N
10
curves look very similar for any other choice of N . Four qualitatively different
regimes can be distinguished which are termed subcritical, critical, supracritical,
and multi-peaked. For small values of , subcritical avalanche size distributions
exist which can be approximated by the general expression
c which we refer to as the critical case For c , the system has avalanche
distributions with an approximate power-law behavior with exponent 3/2 from
L = 1 almost up to the size of the system, where the usual exponential cutoff
is observed (Figure 3.16b). For finite N , c is in a regime where the system
is dissipative. Above the critical value c , avalanche size distributions become
nonmonotone (Figure 3.16c). Such supracritical curves have a minimum at some
intermediate avalanche size. Finally, multi-peaked avalanche size distributions
like the one shown in Figure 3.16d are characterized by multiple firing of single
neurons.
For the subcritical, critical and supracritical regimes, a closed expression for
p(L; N, ) is given by
L2 N 1 L1 N L1
p(L; N, ) = L 1L
L1 N N
N (1 )
for 1 L N . (3.24)
N (N 1)
This is obtained via a geometrical argument in the N -dimensional configuration
space of the system comprising the neurons membrane potentials; see Eurich
et al. (2002:La) for details.
60 60
#(unit)
200 0 200 100 0 100
40 40
20 20
60 60
#(unit)
100 0 100 50 0 50
40 40
20 20
Then some intermediate noise intensity exists which yields a resonance-like be-
havior; it is a matching of the time scales of the periodic forcing and the threshold
crossing due to the noise input (Gammaitoni et al., 1995, 1998). This noise level
allows for an optimal inference of properties of the coherent input from the su-
perthreshold signals alone. Bezrukov and Vodyanoy (1997) extend the notion of
stochastic resonance to non-dynamical systems even without response thresholds.
Stochastic resonance has been suggested to be a mechanism at work in a large
number of systems, for example the ice ages (Benzi et al., 1981), lasers (McNa-
mara et al., 1988), superconducting quantum interference devices (Rouse et al.,
1995), and physiological systems including somatosensory percepton (Collins
et al., 1996, 1997), human blood pressure regulation (Hidaka et al., 2000), human
muscle spindles (Cordo et al., 1996), mechanoreception in the crayfish (Douglass
et al., 1993) and electroreception in the paddlefish (Russell et al., 1999; Green-
wood et al., 2000).
Neurons as threshold elements have also attracted attention. Longtin et al.
(1991); Longtin (1993) and Longtin et al. (1994) compare the interspike inter-
val histogram of a sinusoidally stimulated auditory nerve from a cat with the
return-time distributions of a periodically driven bistable potential, which is a
toy model for stochastic resonance. Wiesenfeld et al. (1994) propose an approx-
imate theory for modeling neuron firing in the presence of noise and a periodic
stimulus; the system shows stochastic resonance and compares well with prop-
erties of a crayfish mechanoreceptor. Bulsara et al. (1996) use an approximate
image-source method to solve the Fokker-Planck equation in the presence of a
weak and slowly varying sinusoidal input of an integrate-and-fire neuron. The
same neuron model is treated in Plesser and Geisel (1999) in the context of
Markov processes to avoid the unrealistic assumption of a stimulus reset after
each spike.
Stochastic resonance occurs also in spatially extended systems and thus con-
tributes to the field of pattern formation in excitable media; see for example Gierer
and Meinhardt (1972); Winfree (1987); Murray (1993); Koch and Meinhardt
(1994) for overviews of wave propagation and pattern formation. In biological
tissue and networks, the propagation of large-scale activity such as spiral waves,
target waves and planar waves has been demonstrated in heart tissue (Nagai
et al., 2000), cortical spreading depression (Basarsky et al., 1998), LGN (Kim
et al., 1995) (the latter two in vitro), and in cultured networks of glial cells (Jung
et al., 1998). In addition, there is much evidence that cortical tissue is a medium
capable of pattern formation which is indicated by its distance-dependent con-
nectivity structure (e. g., Hellwig, 2000), by psychophysics (Wilson et al., 2001),
migraine aura dynamics (Dahlem and Muller, 2003) and modelling studies on
visual hallucinations (Ermentrout and Cowan, 1979; Bressloff et al., 2001). Re-
views on cortical waves and the modelling of neural networks as spatio-temporal
pattern-forming systems are given by Ermentrout and Kleinfeld (2001) and Er-
3.3. MECHANISMS IN SYSTEMS WITH DELAYS AND NOISE I-123
Wheras most studies consider dynamical noise, i. e., a noisy input to layered
neural networks, the existence of spiral waves in disordered neural networks has
also been demonstrated (Milton et al., 1993; Chu et al., 1994). For example,
Chu et al. (1994) study a lattice of integrate-and-fire neurons with a distance-
dependent connectivity. The actual weight between two neurons at a distance r
is determined by the number q of synaptic connections between them; q is drawn
from a Poisson distribution whose mean decays exponentially with r. Such net-
works can carry spiral waves.
Eurich and Schulzke (in press:L) consider the question if large-scale activity
spiral waves, target waves, planar waves in such disordered media spreads
most easily at some intermediate disorder level; this would correspond to a new
variant of stochastic resonance.
The model consists of a two-dimensional network of excitatory integrate-and-
fire neurons with a refractory period, whereby iterations are performed in discrete
time steps t. An interaction between units takes place when a neuron j firing
at time t gives an input current to another neurons i according to its connectivity
matrix Kji . Explicitely, the input of neuron i is given by
X
Ii (t + t) = I0 Kji , (3.25)
j
and the variance can therefore be varied independently of each other. To ensure
that the sum of all coupling weights is held constant on average, the means
of the distributions are given by the deterministic weights, which are distance-
dependent. The variance is then considered a free parameter and is interpreted
as a measure of the disorder introduced in the network. Scaled by the squared
mean it gives rise to the disorder parameter
2 1
1 = 2 = , (3.27)
x d
where the subscript indicates model one. For d 0 and x > 0, 1 is limited to
the range 0 1. As a consequence, connections between nearby neurons can
vary much stronger than between distant neurons, as the variance is proportional
to the squared mean for a fixed value of 1 . Model one therefore yields only weak
disorder in the connections. In order to obtain larger disorder that allows for
a b
c d
Figure 3.18: Evolution of a target wave in a 100 100 network (disorder model
two, 2 = 0.2, = 0.1). Snapshots are taken at the times of 1, 9, 17 and 25 ms,
respectively. Active units are marked white, refractory black and resting neuron
are grey. (a) shows the initial conditions, in (c,d) a central pot with units in the
three different states is visible, which is a consequence of disordered connectivity.
Adapted from Eurich and Schulzke (in press:L).
3.3. MECHANISMS IN SYSTEMS WITH DELAYS AND NOISE I-125
strong couplings also between distant neurons we define a second disorder model.
In this model, weights are shuffled: a certain number of pairs of coupling weights
from the connection matrices of all neurons are chosen by random. Subsequently,
the values of these pairs are interchanged. The sum over all coupling weights
remains constant, ensuring that eventually occuring activity waves do not re-
sult from stronger connectivity. For the second model the disorder parameter is
defined as
number of exchanged weights
2 = . (3.28)
total number of weights
An example of the dynamics is shown in Figure 3.18, demonstrating that a
network with disordered connectivity is able to support target waves.
The influence of disorder is quantified by the minimal global coupling strength
I0 in (3.25) henceforth called I0,min that is necessary for sustained large-scale
activity. The influence of disorder model one is shown in Figure 3.19a. I0,min
decreases monotonically with increasing disorder, which can be interpreted as
disorder-enhanced self-sustaining activity. For larger disorder, target or spiral
waves should break up. Indeed, this can be seen in the result for the second,
stronger disorder model in Figure 3.19b; for small values of the disorder parame-
ter (2 < 0.2) the same propagation-enhancing effect as in model one is obtained;
for higher values of 2 , however, disorder strongly disrupts the Euclidean topol-
ogy of the connections which results in hampered wave formation or propagation.
In between there is an optimal parameter value which is best suited for the self-
maintenance of neural activity. The minimum of I0,min as a function of 2 can
be interpreted as a stochastic resonance-like phenomenon induced by disordered
connectivity. In addition to the results presented here, a similar dependency of
plain wave propagation on structural disorder has been found, which leads to the
speculation that spread of activity in the visual cortex may be facilitated through
irregular connectivity.
1.1 1.1
a b
1 1
0,min
0,min
0.9 0.9
I
I
0.8 0.8
0.7 0.7
0 0.2 0.4 0.6 0.8 0 0.2 0.4 0.6 0.8
1 2
Figure 3.19: The minimal needed global coupling factor I0,min for activity main-
tenance as a function of the disorder parameter of disorder model one (a) and
disorder model two (b). Error bars represent one standard deviation. The ordi-
nate is normalized to I0,min in the case of no disorder. Note that the abscissae 1
and 2 are conceptually different variables with no direct relation except for their
range. (a) Result obtained by 15 realizations of random connectivity matrix sets
generated by disorder model one. The size of the network was 100 100 neurons,
= 0.1, absolute refractory period 2 ms. I0,min decreases monotonically with
1 . (b) Simulation with disorder model two, here obtained by 10 random matrix
set realizations. Network size was 200 200 neurons, other parameters are the
same as in (a). I0,min (2 ) exhibits a distinct minimum. Adapted from Eurich and
Schulzke (in press:L).
Chapter 4
According to the research plan mapped out in Chapter 1, the preceding two chap-
ters described the complementary approaches of neural coding and neural dynam-
ics, respectively. It is of course desirable to obtain an integrated understanding
of phenomena in the nervous system. A combined dynamical and functional ap-
proach is associated with questions like the following: Do dynamical mechanisms
that can be identified in the nervous system serve particular functions which are rele-
vant for the survival and reproduction of animals or humans? What measures can be
extracted from dynamical variables in single neurons or neural populations to explain
perception, action, or mental states? A few contributions exist that consider both
approaches.
Gerstner et al. (1996) and also Eurich et al. (1999:L, 2000) model the adap-
tation of time delays in neural populations via a Hebbian mechanism.
The direct functional relevance is the self-organized emergence of tapped
delay lines (for example, in the auditory system of birds) which transform
a temporal code into a spatial code, i. e. a neural map, for the localiza-
tion of objects. Related biological systems are echolocation in bats and the
two lateral line systems (electroreception and mechanoreception) in weakly
electric fish and in amphibians, respectively (Carr and Konishi, 1990; Suga,
1990; Carr and Friedman, 1999; Franosch et al., 2003).
Attempts have been made to relate the temporal activity in neural popu-
lations, synchronization (Eckhorn et al., 1988; Gray et al., 1989 and sub-
I-127
I-128 CHAPTER 4. TOWARDS A COMBINED APPROACH
20 ms
300 ms
Percept
A neural network model of the shine-through effect (Herzog et al., 2002; Ernst
et al., 2002; Herzog et al., 2003; Ernst et al., 2003; Herzog et al., 2003:L) quan-
titatively accounts for the visibility of the vernier for these and a number of
other stimulus configurations. A Wilson-Cowan type network employing partial
differential equations for the activities of an excitatory and an inhibitory layer of
neurons is chosen because of its relative simplicity: the model allows for the iden-
tification of neural mechanisms that yield the observed psychophyical properties
and is not meant to reproduce physiological details of a specific cortical area.
An overview of the model is given in Figure 4.2. The model employs the hor-
izontal axis x of the visual field only and neglects the vertical spatial direction
and the orientation tuning of cortical visual cells to simplify the analysis. The
network consists of a one-dimensional layer with excitatory and inhibitory neu-
ronal populations, mutually connected with coupling kernels W{e,i} , with typical
length scales {e,i} ,
!
0 2
1 (x x )
W{e,i} (x x0 ) = q exp 2
. (4.1)
2 2 2 {e,i}
{e,i}
I-130 CHAPTER 4. TOWARDS A COMBINED APPROACH
Excitatory We
Layer
Wi We
Wi
Inhibitory V
Layer
t
Stimulus
he(Je), h i(J i)
S(x,t)
se si
x
e i Je, Ji
with the spatio-temporal pattern S(x, t) modeling the stimulus sequences used in
the experiment yields the efferent input I(x, t) converging onto both populations,
I(x, t) = (S ? V ) (x, t).
Figure 4.3 shows simulations of the experimental conditions given in Figure 4.1.
The top row shows the spatio-temporal activity of the excitatory population in a
grey scale code. The spatial center (at x = 0) is of particular interest because it
is the location of the vernier. A comparison of Figures 4.3a and b clearly shows
that the neural activity corresponding to the vernier persists longer in the case
of the 25-bar grating than in the case of the 5-bar grating. The dynamics is the
result of two properties of the system: first, edges (like the edges of the mask and
the vernier itself) are enhanced, i. e., represented by large neural activity. Second,
there is an inhibitory interaction between such edge activities. In Figure 4.3b the
edges of the 5-bar grating are sufficiently close to the center position to quickly
suppress the activity of the vernier. In contrast, the edges of the 25-bar grating
are further away from the center and have less or no influcence which results in
a longer persistence of the center activity (Figure 4.3a).
The model results suggests that the visibility of the vernier as observed in psy-
chophysical experiments can be related to a dynamical property of the underlying
network, the duration T for which the neural activity is above some threshold
A0 (see Figure 4.3d where the center activity is plotted as a function of time).
This is consistent with the results of Figure 4.3c where the short duration of the
activity of the center population corresponds to a low visibility of the vernier for
the gap grating.
The suggested dynamical equivalent of vernier visibility can actually be used
as a quantitative measure. A further set of psychophysical experiments employing
parametrized stimuli is used to calibrate the model and quantitatively compare
model output and psychophysical thresholds in subjects. It turns out that the
duration T of neural activity is not sufficient to account for all data but that a
second measure is necessary in addition, the amplitude Am of neural activity at
the position of the vernier (see Figure 4.3d); details are described in Herzog et al.
(2003:L).
40 0.25
0.2
t [ms]
60
0.15
80
0.1
100
0.05
120 Ae(x, t)
2000 0 2000 2000 0 2000 2000 0 2000
x [arcsec] x [arcsec] x [arcsec]
(d) 0.3
A (0, t) [arb. units]
0.2
Am
0.1
A
0
e
T
0
0 20 40 60 80 100 120
t [ms]
Figure 4.3: Spatio-temporal activation levels Ae (x, t) of the excitatory layer for
(a) the 25 element, for (b) the 5 element, and for (c) the gap grating. Activation
levels are expressed on a grey scale (see right scale). Time t proceeds on the
vertical axis. The position x within the excitatory neuronal layer is depicted
at the horizontal axis. (d) Activation Ae (0, t) of the center neural population
at position x = 0, being directly stimulated by the preceding vernier and the
following central element of the various gratings. The solid line shows A e (0, t)
for the 25 element, the dashed-dotted line for the 5 element, and the dashed line
for the gap grating. While in the 25 element grating condition, Ae (0, t) remains
above the threshold A0 = 0.08 for a long period T , the inhibitory interactions in
the 5 element and the gap grating conditions drive Ae (0, t) very quickly below
A0 . The maximum activity is denoted with Am . Adapted from Herzog et al.
(2003:L).
A sufficiently simple model whose dynamics can be analyzed and from which
general mechanisms can be deduced;
A systematic investigation of the nervous system under this point of view is only
at its beginning. A particular interesting field of research is the consideration of
cognitive phenomena which may yield deeper insights into our position and our
actions in this world.
I-134 BIBLIOGRAPHY
Bibliography
L. F. Abbott, E. Farhi, and S. Gutmann. The path integral for dendritic trees.
Biol. Cybern., 66:4960, 1991.
E. D. A. Adrian. The Basis of Sensation, the Action of the Sense Organs. Norton,
New York, 1928.
I-135
I-136 BIBLIOGRAPHY
P. Baldi and A. F. Atiya. How delays affect neural dynamics and learning. IEEE
Trans. Neural Networks, 5:612621, 1994.
A. F. Dean. The variability of discharge of simple cells in the cat striate cortex.
Exp. Brain Res., 44:437440, 1981.
M. A. Dichter. Rat cortical neurons in cell culture: culture methods, cell mor-
phology, electrophysiology, and synapse formation. Brain Res., 149:279293,
1978.
R. Eckhorn and B. Popel. Responses of cat retinal ganglion cells to the random
motion of a spot stimulus. Vision Res., 21:435443, 1981.
C. W. Eurich and H. Mallot. Mathematics for the neural and cognitive sciences.
In IK 2001 Interdisziplinares Kolleg. arendtap, Bremen, 2001.
F. Gaillard. Binocularly driven neurons in the rostral part of the frog optic
tectum. J. Comp. Physiol. A, 157:4755, 1985.
J. Haag and A. Borst. Active membrane properties and signal encoding in graded
potential neurons. J. Neurosci., 18:79727986, 1998.
M. H. Herzog, U. Ernst, and C. W. Eurich. Ein Kamel ist keine Palme. Gehirn
und Geist, 1(1):1113, 2002.
E. Hida, K.-i. Naka, and K. Yokoyama. A new photographic method for mapping
spatiotemporal receptive field using television snow stimulation. J. Neurosci.
Meth., 8:225230, 1983.
C.-F. Hsiao, M. Watanabe, and Y. Fukuda. The relation between axon diameter
and axonal conduction velocity of Y, X and W cells in the cat retina. Brain
Res., 309:357361, 1984.
BIBLIOGRAPHY I-157
S. K. Itaya. Retinal efferents from the pretectal area in the rat. Brain Research,
201:436441, 1980.
H. Jaeger. The echo state approach to analysing and training recurrent neural
networks. GMD Report 148, Sankt Augustin, 2001a.
H. Jaeger. Short term memory in echo state networks. GMD Report 152, Sankt
Augustin, 2001b.
B. W. Knight. The relationship between the firing rate of a single neuron and
the level of activity in a population of neurons. J. gen. Physiol., 59:767778,
1972b.
L. Kruger. Photographic Atlas of the Rat Brain: The Cell and Fiber Architecture
Illustrated in Three Planes with Stereotaxic Coordinates. Cambridge University
Press, Cambridge, 1994.
A. Longtin and J. G. Milton. Insight into the transfer function, gain, and oscilla-
tion onset for the pupil light reflex using nonlinear delay-differential equations.
Biol. Cybern., 61:5158, 1989a.
A. Manwani and C. Koch. Detecting and estimating signals over noisy and
unreliable synapses: information-theoretic analysis. Neural Comp., 13:133,
2001.
H. Markram, Y. Wang, and M. Tsodyks. Differential signaling via the same axon
of neocortical pyramidal neurons. PNAS, 95:53235328, 1998.
F. Moss and K. Wiesenfeld. The benefits of background noise. Sci. Am., 273(2):
6669, 1995.
P. L. Nunez. Electric Fields of the Brain. Oxford University Press, Oxford, 1981.
M. I. Posner and C. D. Gilbert. Attention and primary visual cortex. PNAS, 96:
25852587, 1999.
A. Pouget, S. Deneve, and J.-C. Ducom. Narrow versus wide tuning curves:
whats best for a population code? Neural Comp., 11:8590, 1999.
P. Reinagel. How do visual neurons respond in the real world? Curr. Opin.
Neurobiol., 11:437442, 2001.
F. Rohrbein and C. Zetzsche. The statistics of natural scenes and Webers law.
In R. P. Wurtz and M. Lappe, editors, Dynamic Perception, pages 233238.
Akademische Verlagsgesellschaft, Berlin, 2002.
G. Roth. Fuhlen, Denken, Handeln Wie das Gehirn unser Verhalten steuert.
Suhrkamp, Frankfurt, 2001.
W. Singer and C. M. Gray. Visual feature integration and the temporal correlation
hypothesis. Annu. Rev. Neurosci., 18:555586, 1995.
S. Single and A. Borst. Dendritic integration and its role in computing image
velocity. Science, 281:18481850, 1998.
W. R. Softky and C. Koch. Cortical cells should fire regularly, but do not. Neural
Comp., 4:643646, 1992.
W. R. Softky and C. Koch. The highly irregular firing of cortical cells is incon-
sistent with temporal integration of random EPSPs. J. Neurosci., 13:334350,
1993.
N. Suga. Biosonar and neural computation in bats. Sci. Am., 262(6):3441, 1990.
R. Van Rullen and S. J. Thorpe. Rate coding versus temporal order coding: what
the retinal ganglion cells tell the visual cortex. Neural Comp., 13:12551283,
2001.
J. D. Victor. Temporal aspects of neural coding in the retina and lateral genicu-
late. Network: Comput. Neural Syst., 10:R1R66, 1999.
F. Worgotter and U. T. Eysel. Context, state and the receptive fields of striatal
cortex cells. TINS, 23:497503, 2000.
W. Yao, P. Yu, and C. Essex. Delayed stochastic differential model for quiet
standing. Phys. Rev. E, 63:021902, 2001.
II-1
Chapter 5
Journal Articles
1. C. W. Eurich, G. Roth, H. Schwegler and W. Wiggers, Simulander: a neural
network model for the orientation movement of salamanders, Journal of
Comparative Physiology A 176 (1995) 379389.
II-3
II-4 CHAPTER 5. LIST OF ORIGINAL PAPERS
The articles on binary coding and the Simulander model, which were part of my
dissertation, were included in the list because they fit very well in the development
of the field of neural coding as it was presented in Part I of this thesis.
Acknowledgments
First of all, I would like to thank my longstanding advisor, Prof. Dr. Helmut
Schwegler, for numerous discussions about physics and the rest of our constructed
world.
Many thanks also to the actual and former members of our institute, in par-
ticular to Prof. Dr. Klaus Pawelzik and Dr. Udo Ernst. Udo (Yahoudi) has
become a close friend. There are so many people to mention here that I would
surely forget about some. Therefore, instead, heres a pool of letters from which
everyone can assemble his or her name (an example is given by the underlined
letters):
AAaaaaaBBbbbbbCCcccccDDdddddEEeeeeeFFfffff
GGgggggHHhhhhhIIiiiiiJJjjjjjKKkkkkkLLlllllMMmm
mmmNNnnnnnOOoooooPPpppppQQqqqqqRRrrrrrSS
sssssTTtttttUUuuuuuVVvvvvvWWwwwwwXXxxxxx
YYyyyyyZZzzzzz`````
II-5