Vous êtes sur la page 1sur 485

List of Contributors

H. Ackermann, Department of General Neurology, Hertie Institute for Clinical Brain Research, University
of Tübingen, Osianderstr. 24, Tübingen, Germany
R. Adolphs, Division of the Humanities and Social Sciences, California Institute of Technology, Human-
ities and Social Sciences 228-77, 331B Baxter Hall, Pasadena, CA 91125, USA
K. Alter, School of Neurology, Neurobiology and Psychiatry, Newcastle upon Tyne, UK
S. Anders, Institute of Medical Psychology and Behavioral Neurobiology, University of Tübingen,
Gartenstr. 29, 72074 Tübingen, Germany
R. Assadollahi, Department of Psychology, University of Konstanz, P.O. Box D25, D-78457 Konstanz,
Germany
T. Bänziger, Swiss Center for Affective Sciences, University of Geneva, 7 rue des Battoirs, 1205 Geneva,
Switzerland
S. Baron-Cohen, Department of Psychiatry, Autism Research Centre, University of Cambridge, Douglas
House, 18B Trumpington Road, Cambridge CB2 2AH, UK
M.M. Bradley, Department of Psychology, Psychology Building Room 114, University of Florida, P.O.
Box 112250, Gainesville, FL 32611, USA
R. Cardinale, Department of Psychology, University of Bologna, Viale Berti Pichat, 5-40127 Bologna, Italy
M.A. Cato Jackson, Nemours Children’s Clinic, Neurology Division, 807 Children’s Way, Jacksonville, FL
32207, USA
B. Chakrabarti, Department of Psychiatry, Autism Research Centre, University of Cambridge, Douglas
House, 18B Trumpington Road, Cambridge CB2 2AH, UK
S.D. Chiller-Glaus, Department of Psychology, University of Zurich, Zurich, Switzerland
M. Codispoti, Department of Psychology, University of Bologna, Viale Berti Pichat 5, I-40127 Bologna, Italy
B. Crosson, Brain Rehabilitation Research Center, Malcolm Randall VA Medical Center, 1601 SW Archer
Rd, Gainesville, FL 32608-1197, USA
D.W. Cunningham, Department Bülthoff, Max Planck Institute for Biological Cybernetics, Spemann-
strasse 38, 72076 Tübingen, Germany
M. Davis, Behavioral Sciences and Psychology, School of Medicine, Center for Behavioral Neuroscience,
Emory University, 1639 Pierce Dr., Suite 4000, Atlanta, GA 30322, USA
A. De Cesarei, Department of Psychology, University of Bologna, Viale Berti Pichat, 5-40127 Bologna,
Italy
T. Demirakca, Central Institute of Mental Health, J5, Division of Neuroimaging, 68159 Mannheim, Ger-
many
S. Dietrich, Max Planck Institute for Human Cognitive and Brain Sciences, Stephanstrasse 1a, 04103
Leipzig, Germany
K. Döhnel, Department of Psychiatry, Psychotherapy and Psychosomatics, University of Regensburg,
Universitätstrasse 84, D-93053 Regensburg, Germany
G. Ende, Central Institute of Mental Health, J5, Division of Neuroimaging, 68159 Mannheim, Germany

v
vi

T. Ethofer, Section of Experimental MR of the CNS, Department of Neuroradiology, University of


Tübingen, Olfried-Müller Str 51, 72076 Tübingen, Germany
V. Ferrari, Department of Psychology, University of Bologna, Viale Berti Pichat, 5-40127 Bologna, Italy
I. Fischler, Department of Psychology, Psychology Building Room 114, University of Florida, P.O. Box
112250, Gainesville, FL 32611, USA
T. Flaisch, Department of Psychology, Institute of Psychology, University of Konstanz, Universitätstrasse
10, 78457 Konstanz, Germany
V. Gazzola, BCN Neuro-Imaging-Centre, University Medical Center Groningen, University of Groningen,
A. Deusinglaan 2, 9713 AW Groningen, The Netherlands
D. Grandjean, Swiss Center for Affective Sciences, University of Geneva, 7 rue des Battoirs, 1205 Geneva,
Switzerland
G. Hajak, Department of Psychiatry, Psychotherapy and Psychosomatics, University of Regensburg,
Universitätstrasse 84, D-93053 Regensburg, Germany
A. Hennenlotter, Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain
Sciences, Stephanstrasse 1a, D-04103 Leipzig, Germany
C. Herbert, Department of Psychology, University of Konstanz, P.O. Box D25, D-78457 Konstanz,
Germany
M. Junghöfer, Institute for Biosignalanalysis and Biomagnetism, University of Münster, Münster 48149,
Germany
A. Keil, Department of Psychology, ZPR Building 029, C 516, University of Konstanz, Box D23, D-78457
Konstanz, Germany
C. Keysers, BCN Neuro-Imaging-Centre, University Medical Center Groningen, University of Groningen,
A. Deusinglaan 2, 9713 AW Groningen, The Netherlands
J. Kissler, Department of Psychology, University of Konstanz, P.O. Box D25, D-78457 Konstanz,
Germany
S.A. Kotz, Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain
Sciences, Stephanstrasse 1a, 04103 Leipzig, Germany
B. Kreifelts, Department of Psychiatry, University of Tübingen, Osianderstr. 24, 72076 Tübingen,
Germany
K. Kucharska-Pietura, Whichurch Hospital, Cardiff and Value NHS Trust, Cardiff CF4 7XB, UK
P.J. Lang, NIMH Center for the Study of Emotion and Attention, Department of Clinical and Health
Psychology, University of Florida, 2800 SW Archer Road, Building 772, Gainesville, FL 32610, USA
S. Leiberg, Institute of Medical Psychology and Behavioral Neurobiology, University of Tübingen, MEG
Center, Otfried-Müller-Str 47, 72076 Tübingen, Germany
J. Meinhardt, Ludwig-Maximilian University, Munich, Germany
M. Meyer, Department of Neuropsychology, Institute for Psychology, University of Zurich, Zurich,
Switzerland
J.L. Müller, Department of Psychiatry, Psychotherapy and Psychosomatics, University of Regensburg,
Universitätstrasse 84, D-93053 Regensburg, Germany
S. Paulmann, Max Planck Institute for Human Cognitive and Brain Sciences, Stephanstrasse 1a, 04103
Leipzig, Germany
M.D. Pell, School of Communication Sciences and Disorders, McGill University, Neuropragmatics and
Emotion Lab, 1266 Avenue des Pins Ouest, Montŕeal, QCH3G 1A8, Canada
P. Peyk, Department of Psychology, University of Basel, Basel, Switzerland
H. Pihan, Department of Neurology, Schulthess Klinik, Lengghalde 2, 8008 Zurich, Switzerland
G. Pourtois, Neurology and Imaging of Cognition, Clinic of Neurology and Department of Neuroscience,
University Medical Centre, University of Geneva, Geneva, Switzerland and Swiss Center for Affective
Sciences, University of Geneva, Switzerland
vii

D. Sabatinelli, NIMH Center for the Study of Emotion and Attention, University of Florida, Building 772
SURGE, 2800 SW Archer Road, Gainesville, FL 32608, USA
K.R. Scherer, Swiss Center for Affective Sciences, University of Geneva, 7 rue des Battoirs, 1205 Geneva,
Switzerland
U. Schroeder, Klinik Holthausen, Am Hagen 20, D-45527 Hattingen, Germany
H.T. Schupp, Department of Psychology, Institute of Psychology, University of Konstanz, Universitätstr-
asse 10, 78457 Konstanz, Germany
A. Schwaninger, Department Bülthoff, Max Planck Institute for Biological Cybernetics, Spemannstrasse
38, first floor, 72076 Tübingen, Germany
J. Schwerdtner, Department of Psychiatry, Psychotherapy and Psychosomatics, University of Regensburg,
Universitätstrasse 84, D-93053 Regensburg, Germany
M. Sommer, Department of Psychiatry, Psychotherapy and Psychosomatics, University of Regensburg,
Universitätstrasse 84, D-93053 Regensburg, Germany
M. Spezio, Division of the Humanities and Social Sciences, California Institute of Technology, Humanities
and Social Sciences 228-77, Pasadena, CA 91125, USA
J. Stockburger, Department of Psychology, Institute of Psychology, University of Konstanz, Univer-
sitätstrasse 10, 78457 Konstanz, Germany
D.P. Szameitat, Max Planck Institute for Human Cognitive and Brain Sciences, Stephanstrasse 1a, 04103
Leipzig, Germany
H. Tost, Central Institute of Mental Health, J5, Division of Neuroimaging, 68159 Mannheim, Germany
P. Vuilleumier, Neurology and Imaging of Cognition, Clinic of Neurology and Department of Neuro-
science, University Medical Centre, University of Geneva, Geneva, Switzerland and Swiss Center for
Affective Sciences, University of Geneva, Geneva, Switzerland
C. Wallraven, Department Bülthoff, Max Planck Institute for Biological Cybernetics, Spemannstrasse 38,
72076 Tübingen, Germany
S. Wiens, Department of Psychology, Stockholm University, Frescati Hagväg, 10691 Stockholm, Sweden
D. Wildgruber, Department of Psychiatry, University of Tübingen, Osianderstr. 24, 72076 Tübingen,
Germany
Preface

Common sense and everyday experience imply that feelings, affects, and emotions dominate large parts of our
everyday lives and particularly our social interactions. Emotions seem to drive much of our behavior, albeit
apparently not always for the good, leading ancient philosophers to reason that we might be better off
without emotions and that, perhaps, the purpose of reason could be the control of these obscure corners of
our inner lives. In the 17th century, René Descartes proposed a dualism between body and mind which has
guided much of the Western philosophy and science until recently. Perhaps as a Cartesian heritage, affects,
emotions, and feelings were all viewed as not lending themselves easily to exact scientific study although
Descartes himself viewed emotion, unlike reason, as having a physiological base (in his Traité des passions de
l’âme, 1649, he even argued that the control of the physical expression of emotion would control the emotions
themselves, a view which has repercussions even today, but it is rarely traced back to Descartes).
Recent developments in the rapidly growing discipline of neuroscience, driven by the enormous meth-
odological and technological advances in neurophysiology, neuroimaging, and computational neurosci-
ences, are heralding a shift of paradigm. Emotions are no longer regarded as too elusive to be approached
with scientific methods, and some have proposed that human survival and success depend critically on the
functioning of the neural networks that are thought to underlie affect, emotions, and, perhaps, feelings.
But what are the results of the scientific study of emotion beyond such increasingly commonplace
statements? What are the data that allow us to draw such seemingly simple conclusions? Are these con-
clusions indeed simple? And, most importantly, what are the implications of emotion research on our
understanding of human social interaction and communication?
Researchers from all branches of behavioral and physiological sciences are trying to specify the mechanisms
that normally guide emotional processing and the malfunctions that may give rise to emotional disorders.
The present volume brings many of them together, spanning a wide spectrum of experimental approaches
in animals and humans; instrumentations from behavioral testing to neurophysiology and neuroimaging;
paradigms from passive, uninstructed stimulus perception to complex social interaction in different
processing domains. The common thrive behind all contributions is to elucidate emotion as a social
phenomenon that does not affect individuals in isolation, but is rapidly conveyed between individuals, some
of these processes being implicit and automatic, others being more explicit and conscious.
This volume is the result of a conference held in September 2004 at the Freudental Castle near Konstanz
in southwestern Germany. This conference and this volume would not have been possible without the
generous support from the Heidelberg Academy of Sciences and Humanities in Heidelberg and the Center
for Junior Researchers at the University of Konstanz, which is gratefully acknowledged.

May 2006
Silke Anders, Tübingen
Gabriele Ende, Mannheim
Markus Junghöfer, Münster
Johanna Kissler, Konstanz
Dirk Wildgruber, Tübingen

ix
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 1

Emotion, motivation, and the brain: Reflex


foundations in animal and human research

Peter J. Lang1, and Michael Davis2

1
NIMH Center for the Study of Emotion and Attention, Department of Clinical and Health Psychology, University of
Florida, FL 32610-0165, USA
2
Department of Psychiatry, Behavioral Sciences and Psychology, Emory University, Yerkes National Primate Center and
the Center for Behavioral Neuroscience, Atlanta, GA 30322, USA

Abstract: This review will focus on a motivational circuit in the brain, centered on the amygdala, that
underlies human emotion. This neural circuitry of appetitive/approach and defensive/avoidance was laid
down early in our evolutionary history in primitive cortex, sub-cortex, and mid-brain, to mediate behaviors
basic to the survival of individuals and the propagation of genes to coming generations. Thus, events
associated with appetitive rewards, or that threaten danger or pain, engage attention and prompt infor-
mation gathering more so than other input. Motive cues also occasion metabolic arousal, anticipatory
responses, and mobilize the organism to prepare for action. Findings are presented from research with
animals, elucidating these psychophysiological (e.g., cardiovascular, neuro-humoral) and behavioral (e.g.,
startle potentiation, ‘‘freezing’’) patterns in emotion, and defining their mediating brain circuits. Parallel
results are described from experiments with humans, showing similar activation patterns in brain and body
in response to emotion cues, co-varying with participants’ reports of affective valence and increasing
emotional arousal.

Keywords: fear; startle; amygdala; arousal; conditioning; valence

Introduction clear that motivated behavior, prompted by


appetitive or aversive cues, can be similar across
This essay presents a motivational framework, in- the mammalian phyla: When a rat first perceives a
tegrating animal and human data, with the aim of predatory cat, the rodent’s behavioral and phys-
explicating the biological foundation of human iological reactions are very much like those of a
emotion. The raw data of emotion are threefold: human who, for example, is abruptly aware of an
affective language, behavior, and physiology (Lang, intruder in her/his home. Both species initially
1985, 1994). The emphasis here is on the latter two ‘‘freeze,’’ and both show parallel changes in heart
data sources — considering emotion’s actions and rate and respiration. Both release similar chemicals
physiology to be founded on evolved neural circuits into their blood stream. In each case these events
that are shared by humans and other animals. prompt attention to the potential threat and a
We cannot know if animals experience emotion readying of the body for defensive action. Fur-
in the same way that humans do. However, it is thermore, if an animal or a human escapes the
danger, brain circuits change: The survivor will
Corresponding author. Tel.: +1-352 392-2439; Fax: +1-352 have learned to react to tell-tale-signs of the po-
392-6047; E-mail: plang@phhp.ufl.edu tential predator, and to be aroused and be more

DOI: 10.1016/S0079-6123(06)56001-7 3
4

wary. These adaptive patterns of behavior have Affective valence and arousal
been carefully preserved in evolution because they
are successful in promoting survival. Significantly, The motivated behavior of a simple organism such
it is in anticipation of, or subsequent to, these as the flatworm can be almost entirely characterized
survival situations that humans report the most by two survival movements: Direct approach to
intense emotional experiences. Such reports occur appetitive stimuli and withdrawal from aversive
when motivational systems (reflexive, goal-rele- stimuli (Schneirla, 1959). These are the only avail-
vant behaviors and their determining neural cir- able tools in achieving consummation or escape.
cuits) are activated, but action is delayed, or in the This modest repertoire is woefully insufficient, how-
aftermath of such actions, or when signs or signals ever, for more complex species that must implement
appear that recall previous encounters. Thus, many critical sub-goals and cope with a richly per-
much can be learned about the biological founda- ceived sensory environment. Humans have proved
tion of expressed emotion by studying how hu- to be successful survivors, seeming to surmount
mans and less complex animals confront appetitive niche limitations, adapting readily to a greater
or aversive events. variety of environments. Much of this human suc-
cess can be attributed to language — the ability to
Working on a definition of emotion communicate, to manipulate symbols in problem
solving, and to label and catalog our experience of
Emotion cannot be operationally defined by a single the world. Although language is not the major fo-
measure. Emotions involve multiple responses that cus of this discourse, it is a primary way through
are variously sequenced in time. Events that are which we infer emotional experience in others.
pleasant/appetitive or aversive/threatening initially Thus, the relationship between emotion’s language
engage heightened attention. They prompt informa- and motivationally determined behavior and phys-
tion gathering, and do so more readily than other iology merits a careful prefatory consideration.
less motivationally relevant cues. Motive cues also The language of emotion includes thousands of
occasion general metabolic arousal, anticipatory words with myriad shades of feeling, degrees of
responses that are oriented towards the engaging redundancy, and shared meaning. Faced with this
event, and neuromuscular mobilization of the body plethora, philosophers, psychologists, and psycho-
for action. These reflex reactions are differently linguists have all tried to condense the list into a
paced and form different patterns, reflecting a few primary emotions, or alternatively, to define
changing organismic state as punishment or reward dimensions of meaning that might underlay this
become more imminent, and specific goal-oriented vast vocabulary.
actions are deployed. Reports of emotional experi- The view that affects — subjective reports of ex-
ence occur in the context of these response events, or perienced emotion — might be organized under a
with their inhibition or delay, but correlations with limited number of overarching factors was proposed
any specific affective report are notoriously modest. by Wundt (1896) in the 19th century. Contemporary
In this chapter we will try to explicate what is studies of natural language categories (Shaver et al.,
special about emotional information processing in 1987; Ortony et al., 1988) suggest that emotional
the brain. We will propose that neural networks knowledge is hierarchically organized, and that the
underlying expressed emotion include direct con- superordinate division is between positivity (pleasant
nections to the brain’s primary motivational sys- states: love, joy) and negativity (unpleasant states:
tems, appetitive and defensive. These neural circuits anger, sadness, fear). Osgood and his associates (e.g.,
were laid down early in evolutionary history: In Osgood et al., 1957), using the semantic differential,
primitive cortex, sub-cortex, and mid-brain. They showed that emotional descriptors were distributed
determine the deployment of attentional resources, primarily along a bipolar dimension of affective
systemic mobilization, approach, and defensive valence — ranging from attraction and pleasure to
behaviors, and the formation of conditioned asso- aversion and displeasure. A dimension of activation
ciations fundamental to the survival of individuals. — from calm to aroused — also accounted for
5

substantial variance. Other investigators have drawn habituates (Sokolov, 1963). Cues to appetite or
similar conclusions from factor analysis of verbal aversion, however, lead to systemic adjustments
reports (e.g., Mehrabian and Russell, 1974; Tellegen, that facilitate sustained perceptual processing. For
1985), and even, of facial expressions (Schlosberg, example, an animal (reptile or mammal) orienting to
1952). No other factors have ever approached the appearance of a distant predator shows a pro-
the generality and significance of these two simple found deceleration in heart rate — ‘‘fear brad-
variables. ycardia’’ — not found in response to other events
We should not be too surprised, perhaps, to learn (Campbell et al., 1997). ‘‘Freezing’’ — a statue-like
that affective valence and arousal find a parallel in inhibition of movement — accompanies the change
motivational theories based on conditioning re- in heart rate, along with increased overall sensory
search with animals and humans. Konorski (1967), acuity. If the predator approaches (shows stalking
for example, founded a motivational typology of behavior), somatic and autonomic activation in-
unconditioned reflexes, keyed to the reflex’s survival creases progressively, culminating in defensive ac-
role. Exteroceptive reflexes were either preservative tion. Humans show similar attention and action
(e.g., ingestion, copulation, nurture of progeny) or readiness when confronted with motivational cues
protective (e.g., withdrawal from or rejection of (in life and in the laboratory) responding reflexively
noxious agents). He further suggested that affective even if stimuli are not actual events, but media rep-
states were consistent with this bi-phasic typology: resentations. In fact, stories, pictures, and films all
Preservative emotions include such affects as sexual prompt patterns of bodily change in observers that
passion, joy, and nurturance; fear and anger are co-vary with the rated affective valence (pleasant or
protective affects. Dickinson and Dearing (1979) unpleasant) and arousal (intensity) of their emo-
developed Konorski’s dichotomy into a theory of tional experience.
two opponent motivational systems, aversive and
attractive, each activated by a different, but equally
wide range of unconditioned stimuli, determining Emotional perception
perceptual-motor patterns and the course of learn-
ing. In this general view, affective valence is deter- In recent years, the psychophysiology of emotional
mined by the dominant motive system: The perception has been systematically studied, using a
appetitive system (preservative/attractive) prompts set of standard photographic picture stimuli, cali-
positive affect; the defense system (protective/avers- brated for affective response. There are currently
ive) is the source of negative affect. Affective arousal nearly 1000 pictures in the International Affective
reflects the ‘‘intensity’’ of motivational mobiliza- Picture System (IAPS — Lang et al., 1999) rated for
tion, appetitive or defensive, determined mainly by experienced pleasure and arousal by a large nor-
degree of survival need and the imminence or prob- mative subject sample. A representative distribution
ability of nociception or appetitive consummation. of IAPS pictures is presented in Fig. 1 (Bradley,
From this perspective, individual reported emotions 2000) located in a Cartesian space formed by inde-
would be based on differing, situation-determined pendent dimensions of rated pleasure and arousal.
action dispositions, as ‘‘fear’’ indicates a disposition Similar distributions have also been obtained for
to avoid or escape and ‘‘anger’’ is a disposition to collections of acoustic stimuli (International Affec-
attack. tive Digitized Sounds (IADS): Bradley et al., 1998a)
as well as verbal materials (Affective Norms for
English Words (ANEW): Bradley et al., 1998b).
Attention, perception, and emotion Studies of IAPS picture stimuli have uncovered
highly reliable patterns of physiological and behavi-
For both humans and other animals, the first reac- oral responses that vary consistently with the factor
tion to any cue is reflexive, directional orientation to structure uncovered in studies of emotional lan-
the stimulus (Pavlov, 1927). If the input is motiva- guage (see Fig. 1: Greenwald et al., 1989, 1998;
tionally irrelevant, this ‘‘orienting reflex’’ rapidly Bradley et al., 2003). Thus, when affective valence
6

Fig. 1. In the upper right of the figure, pictures from the International Affective Picture System (Lang et al., 1999) are plotted in a two-
dimensional (Cartesian) space. Picture location is defined by mean ratings of judged pleasure and emotional arousal as reported by a
normative sample. The vectors in the upper and lower portions of the affective space describe hypothesized increasing activation levels
in appetitive and defensive motivation that co-vary with reported arousal. The other three quadrants of the figure contain graphic
representations of the co-variation between physiological reactions to a sample of these picture stimuli and self-ratings of the emotional
experience of research participants. Pictures are rank ordered on the abscissa by either each participant’s ratings of affective valence or
their ratings of affective arousal. The mean change in corrugator muscle activity (top right) and heart rate (bottom right) are presented
across ranked ratings of affective valence (pleasure). Mean skin conductance responses and cortical event-related potentials (bottom
right) are plotted as a function of ranked affective arousal rating. The correlations are in all significant cases (po0.01).

ratings are ranked by picture from the most to the linearly as pictures are rated as more unpleasant;
least pleasant image, for each subject, facial muscle conversely, zygomatic (smile) muscle activity
activity during picture viewing shows a strong increases with judged pleasantness. Heart rate is
monotonic relationship with levels of affective va- also responsive to differences in affective valence:
lence: Corrugator (frown) muscle action increases Unpleasant pictures generally prompt marked
7

deceleration during viewing (recalling the ‘‘fear participants control viewing time, they look longer
bradycardia’’ seen in animals), more pronounced at emotionally arousing pictures, both pleasant
than that seen when subjects view pleasant pictures. and unpleasant, than at neutral pictures (Lang
Other physiological responses vary with changes et al., 1993). This latter relationship is not found,
in rated emotional arousal, rather than affective however, if pictures evoke very high levels of dis-
valence. Skin conductance — a good general index tress: When phobics view pictures specific to their
of autonomic activation — increments monoton- fear, viewing time is dramatically reduced (see
ically with increases in rated arousal, regardless of Hamm, et al., 1997). They also show heart rate
picture valence. Electroencephalographic measure- acceleration (rather than deceleration), consistent
ment (EEG) shows a distinct, voltage-positive cor- with a precipitous increase in defense motivation
tical response evoked directly by the picture and mobilization for active escape.
stimuli. This is also positively correlated with As the phobia data imply, relationships between
stimulus arousal (i.e., it is similarly enhanced for specific measures can vary widely for individuals
both pleasant and unpleasant arousing pictures: and to some extent between particular groups.
Cuthbert et al., 2000; Keil et al., 2002). These Gender differences, for example, are highly relia-
measures appear to index the intensity or activa- ble: Pleasantness ratings covary more closely with
tion level of the current motivational state, but are facial muscle activity in females than in males; on
silent about motivational direction (i.e., appetitive the other hand, skin conductance changes are
or defensive). more closely correlated with arousal ratings in
Behaviors elicited in the context of emotional males than in females (Lang et al., 1993).
picture perception (in reaction to secondary stim- The results of factor analyses of affect self-report,
uli) also covary with motivational engagement. physiological, and behavioral measures are pre-
When first exposed to a new picture, reaction time sented in Table 1. The data were obtained from
responses to probes are significantly slower for large groups of young, healthy participants. The
emotionally arousing than for affectively calm pic- obtained two-factor solution is clearly very strong:
tures (Bradley et al., 1992). These data suggest that Pleasantness ratings, heart rate, and facial muscles
new activating images may require more attentio- load on a first, affective valence factor; arousal and
nal resources at encoding. Furthermore, when interest ratings, viewing time, skin conductance, and

Table 1. Factor analyses of measures of emotional picture processing

Measure Factor 1 (Valence) Factor 2 (Arousal)

Sorted loadings of dependent measures on principal components (Lang et al., 1993)


Valence ratings 0.86 –0.00
Corrugator musclea –0.85 0.19
Heart rate 0.79 –0.14
Zygomatic musclea 0.58 0.29
Arousal ratings 0.15 0.83
Interest ratings 0.45 0.77
Viewing time –0.27 0.76
Skin conductance –0.37 0.74

Sorted loadings of dependents measures on principal components (Schupp et al., 1994)


Valence ratings 0.89 0.07
Corrugator musclea –0.83 –0.10
Heart rate 0.73 –0.02
Arousal ratings –0.11 0.89
Cortical slow wave –0.06 –0.79
Skin conductance 0.19 0.77
a
Bioelectric potentials from muscles that mediate facial expression.
8

cortical EEG load on a second, affective arousal Neural substrates of emotion: Attention, action and
factor. The cross-loadings for all measures are very the role of the amygdala
low. The data are consistent with the view that re-
ported affective experience is determined in signifi- Much recent research has shown that a brain area
cant part by the individual’s motivational state. That called the amygdala is a crucial structure in a neural
is, negative affective valence (unpleasant experience) network that mediates motivated attention and
is associated with activation of the defense system; preparation for action. In man, this almond-shaped
positive valence (pleasant feelings) is associated with structure lies deep in the brain’s temporal lobe. It is
activation of the appetitive system. Reports of composed of several different nuclei that serve
arousal are associated with both states, reflecting different network functions. The basolateral am-
an increase in incentive strength and organismic ygdala (Bla), which includes the lateral, basal and
mobilization. The motivational states elicited by basomedial nuclei, is of particular significance, as it
these affective cues (and the somatic, cortical, and receives information from the thalamus, hippocam-
autonomic substrates of their perception) appear to pus, and cerebral cortex (see McDonald, 1998), and
be fundamentally similar to those occurring when then projects (Fig. 2) to other amygdala nuclei, as
other complex animals ‘‘stop, look, and listen,’’ sift- well as to targets elsewhere in the brain relevant to
ing through the environmental buzz for cues of emotional memory and action. The Bla’s projec-
danger, social meaning, or incentives to appetite. tions to the central nucleus of the amygdala (CeA)

Fig. 2. Schematic diagram of the outputs of the basolateral nucleus of the amygdala to various target structures, and the subsequent
outputs and targets of the amygdala’s central nucleus (CeA) and the lateral basal nucleus of the stria terminalis (BNST: The
‘‘extended’’ amygdala). Known and possible functions of these connections are briefly described.
9

and the bed nucleus of the stria terminalis (BNST; vigilance and superior signal detection found in
or ‘‘extended amygdala’’) are relayed from these the attentional phase of emotional processing.
sites to specific hypothalamic and brainstem target As already noted, sensory orientation to threat in
areas that mediate most of the visceral and striate mammals and reptiles is accompanied by a pro-
muscle events that index emotional processing. found decrease in heart rate (‘‘fear bradycardia’’:
Their projections target lateral hypothalamus — a Campbell et al., 1997). Heart rate decrease is asso-
key center activating the sympathetic branch of the ciated with attention in humans (Graham and Clif-
autonomic nervous system in emotion (LeDoux ton, 1966) and furthermore, a greater deceleration
et al., 1988). In addition, direct projections from the is generally found in response to stimuli judged to
BNST go to the dorsal motor nucleus of the vagus, be unpleasant (Bradley, 2000; Lang et al., 1993).
the nucleus of the solitary tract, and the ventrolat- Several lines of research suggest that this cardiac
eral medulla. These brainstem nuclei are known to response can be mediated by the central nucleus of
regulate heart rate and blood pressure (Schwaber the amygdala. During Pavlovian aversive condi-
et al., 1982), and may thus modulate cardiovascular tioning in rabbits, one sees a rapid development of
responses in emotion. Projections to the par- conditioned bradycardia. Pascoe and Kapp (1985)
abrachial nucleus are likely to be involved in emo- found a high correlation (.71) between the firing
tion’s respiratory changes (with additional, perhaps frequency of individual neurons in the amygdala’s
indirect effects on the cardiovascular system), as central nucleus and extent of heart rate deceleration
electrical stimulation and lesions of this nucleus alter to a conditioned stimulus. Furthermore, the central
breathing patterns. Finally, indirect projections nucleus of the amygdala could have indirect, but
from the amygdala’s central nucleus to the para- widespread, effects on cortical activity — mediated
ventricular nucleus (via the BNST and preoptic by projections to cholinergic neurons that in turn
area) may mediate neuroendocrine responses that project to the cortex (Kapp et al., 1992). This path
are particularly prominent when emotional stimuli may account for changes in the EEG waveform,
are sustained. perhaps associated with enhanced sensory process-
ing, acquired during Pavlovian aversive condition-
ing at the same rate as conditioned bradycardia.
Attention, vigilance, and conditioned fear
Motor behavior
During emotional stimulation, projections from the
central nucleus or BNST to the ventral tegmental Emotion’s attentional phase is characterized by im-
area appear to mediate increases in dopamine meta- mobility, ‘‘freezing,’’ mediated in the rat by amygdala
bolites in the prefrontal cortex (Goldstein et al., (CeA) projections to the ventral periacqueductal
1996). Cells in the locus coeruleus, which release gray. In the action phase, projections to the dorsal
norepinephrine into the brain, are also activated, periacqueductal gray appear to mediate fight/flight
perhaps mediated by projections to its dendritic responses (Fanselow, 1991). As norepinephrine and
field, or indirectly, via projections to the par- serotonin facilitate excitation of motor neurons
agigantocellularis nucleus (Redmond, 1977; Aston- (McCall and Aghajanian, 1979; White and Neuman,
Jones et al., 1996). Furthermore, there are direct 1980), rapid defensive action could be mediated by
projections to the lateral dorsal tegmental nucleus lateral BNST activation of norepinephrine release in
and parabrachial nuclei. These latter nuclei have the locus coeruleus or via its projections to serotonin
cholinergic neurons that project to the thalamus and containing raphe neurons.
could mediate increased synaptic transmission of its
sensory relay neurons. The sensory thalamus is, of
course, a primary processor of environmental input. Amygdala stimulation
Thus, this sequence of projections, by augmenting
cholinergic activation and facilitating thalamic Electrical stimulation of the amygdala, or a
transmission, may contribute to the increased bnormal electrical activation via temporal-lobe
10

seizures, produces emotion-like behavioral and Amygdala lesions and drug infusion
autonomic changes (probably activating targets
seen in Fig. 4) that humans generally describe as Assuming the emotion circuit to be a ‘‘hard wired’’
an experience of fear or apprehension (Chapman set of connections, destruction of the amygdala is
et al., 1954; Gloor et al., 1981). In animals, elec- expected to disrupt or eliminate emotion’s sensory
trical or chemical stimulation of the amygdala can processing and motor output. Various investiga-
also produce prominent cardiovascular effects, tors have provided data in support of this hypoth-
and persistent stimulation may produce gastric esis, showing, for example, that lesions of the
ulceration, increases in blood levels of corti- amygdala block attentional responses to stimuli
sol and epinephrine, and sustained changes in paired with food (cf. Gallagher et al., 1990), and
respiration. in general, fail to benefit from procedures nor-
Stimulation of these same CeA sites can also mally facilitating attention to conditioned stimuli
produce responses associated with attention — (Holland and Gallagher, 1993a, b).
both bradycardia (Kapp et al., 1990) and low- Other research suggests that the central nucleus
voltage fast EEG activity in rabbits (Kapp et al., and the basolateral nucleus may make different
1994) and in rats (Dringenberg and Vanderwolf, contributions to the processing of emotional stim-
1996). Furthermore, depending on the state of uli: In odor-aversion learning, rats develop aver-
sleep, electrical stimulation of the amygdala in sions to novel odors associated with illness, but
some species activates cholinergic cells involved in only if the odor is part of a compound stimulus
arousal-like effects. Overall, the orienting reflex that includes a distinctive taste. Electrolytic (Berm-
has been described as the most common response udez-Rattoni et al., 1986) or chemical lesions
elicited by electrical stimulation of the amygdala (Hatfield et al., 1992) of the basolateral nucleus
(Ursin and Kaada, 1960; Applegate et al., 1983). — but not the CeA — block such odor-aversion
In many species, electrical or chemical stimula- learning, but do not impede taste-aversion learn-
tion of the amygdala prompts cessation of ongoing ing. Local infusion of N-methyl-D-aspartate
behavior, facilitating the sensory orienting critical (NMDA) antagonists into the Bla has a similar
to the attentional phase of emotion. It is associated selective effect (Hatfield and Gallagher, 1995).
with ‘‘freezing’’ in rats and by the cessation of Thus, it has been suggested that, whereas the cen-
operant bar pressing. Electrical stimulation also tral nucleus ‘‘regulates attentional processing of
activates facial motoneurons, eliciting jaw move- cues during associative conditioning’’ (Hatfield
ments, and may be the pathway mediating facial et al., 1996, p. 5265), the Bla is critically involved
expressed emotion. The broad effects of am- in ‘‘associative learning processes that give condi-
ygadala stimulation on the motor system include tioned stimuli access to the motivation value of
modulating brainstem reflexes, such as the mass- their associated unconditioned stimuli’’(Hatfield et
enteric, baroreceptor nictitating membrane, eye- al., 1996, p. 5264). The two nuclei appear to work
blink, and startle reflexes. in concert: The motivational significance of input
In summary, it is assumed that the above broad is mediated by amygdala’s basolateral nucleus and
network of amygdala connections shown by the the central nucleus maintains the relevant cue as a
stimulation data are already formed in an adult focus of attention.
organism, given that these effects are produced in
the absence of explicit, prior learning. This sug-
gests that the behavioral pattern evoked by emo- Conditioned emotional states
tional stimuli is, in significant part, ‘‘hard wired’’
during evolution. Thus, it is only necessary that an A large literature indicates that amygdala lesions
initially neutral stimulus activate the amygdala — block many measures used to assess conditioned
in association, for example, with an aversive event and unconditioned fear (cf. Davis, 2000), including
— for this formerly neutral cue to then produce changes in heart rate, blood pressure, ulcers, and
the full constellation of emotional effects. respiration; secretion of ACTH or corticosteroids
11

into the blood; and release of dopamine, nor- Bla, markedly disrupt this behavior (Parkinson
epinephrine, or serotonin in certain brain areas. et al., 2000a). This effect is apparently mediated by
They also block behavioral measures: Freezing, CeA’s projection to dopamine containing neurons
fear-potentiated startle, and vocalization, as well in the ventral tegmental area, releasing dopamine
as operant conflict test performance, conditioned in the nucleus accumbens (given that depletion of
emotional responses, and shock avoidance. Fur- dopamine in the nucleus accumbens core prevents
thermore, lesions of the amygdala cause a general acquisition of autoshaping; Parkinson et al., 1998).
taming effect in many species (Goddard, 1964) In contrast, approach behavior itself seems to be
perhaps analogous to the increase in trust found in mediated by the anterior cingulate cortex (Bussey
humans following surgical amygdala lesions (Ado- et al., 1997) via projections to the nucleus accum-
lphs et al., 1998). bens core (Parkinson et al., 2000b).
In another series of experiments, rats first trained
to associate an auditory conditioned stimulus (CS)
Effects of local drug infusion
with delivery of food were subsequently trained to
press a lever to obtain food. Later, presentation of
The intensity of fear is determined by the interplay
the original auditory CS increased lever pressing.
of a variety of brain chemicals, many acting di-
Lesions of the central nucleus of the amygdala (not
rectly in the amygdala. Fear is reduced when
the Bla) reduced this facilitatory effect — mediated,
GABA (a major inhibitory neurotransmitter), or
the authors speculated (Everitt et al., 2000), by
GABA agonists, benzodiazepines (e.g., valium) are
projections from the CeA to the mesolimbic dopa-
infused into the amygdala. Drugs that decrease
mine system. Opposite effects have been reported,
excitatory transmission in the amygdala, such as
however, in a similar paradigm measuring actual
glutamate antagonists, have similar actions.
consumption of food (Gallagher, 2000). In this
Local infusion studies have also evaluated chem-
case, projections from the posterior division of the
ical compounds that increase fear. These include
basolateral amygdala to the hypothalamus are
GABA antagonists, or peptides such as corticotro-
thought to be involved.
phin releasing hormone (CRH), cholecystokinin
(CCK), vasopressin, thyroid-releasing hormone
(TRH), and opiate antagonists. Extensive tables Extra-amygdalar projections of the basolateral
giving references, substances used, sites of admin- nucleus
istration (CeA, Bla, etc.), and the effects can be
found in Davis (2000). As illustrated in Fig. 2, connections between the
amygdala’s central nucleus and the BNST are crit-
ically involved in many of emotion’s autonomic
Approach behavior
and motor responses. However, projections from
the basolateral nucleus to other target areas in the
Research concerned with the amygdala’s role in
brain are also significant for emotional behavior.
conditioned appetitive approach (Everitt et al.,
2000) has emphasized a projection from the central
nucleus to dopamine neurons in the ventral teg- The ventral striatum pathway: Secondary
mental area that then projects to the nucleus reinforcement
accumbens.
In studies of ‘‘autoshaping,’’ a light (CS+) is The Bla projects directly to the nucleus accumbens
followed by food reward, delivered in various lo- in the ventral striatum (McDonald, 1991), in close
cations. Another stimulus (CS) is presented but apposition to dopamine terminals of A10 cell bod-
never followed by food. Under these conditions, ies in the ventral tegmental area (cf. Everitt and
rats learn to approach the CS+ light before going Robbins, 1992). The ventral striatum, Morgenson
to the food hopper to retrieve the food. Bilateral (1987) suggests, is where affective processes in the
lesions of the central nucleus, but not lesions of the limbic forebrain gain access to the subcortical part
12

of the motor system that results in appetitive ac- The dorsal striatum pathway
tions (cf. Morgenson, 1987).
Projections from the Bla to the nucleus accum- The amygdala modulates memory in a variety of
bens are critically involved in secondary reinforce- tasks such as inhibitory avoidance, motor or spatial
ment. In a relevant paradigm: A light is paired learning ( McGaugh et al., 1992, 1993; Packard et
with food, after which the experimental animals al., 1994; Cahill and McGaugh, 1998; Packard and
are presented with two levers. One lever turns on Teather, 1998). Thus, post-training intercaudate in-
the light, pressing the other does not. Normal rats jections and intra-hippocampal infusion of amphet-
press the ‘‘light’’ lever much more than the other amine have task-specific effects on memory —
lever. Hence, light is a secondary reinforcer, as it intercaudate injections enhances memory in a vis-
prompts new behavior via its prior association ible-platform water maze task, but has no effect
with food. Rats with Bla lesions fail to learn this when the platform is hidden (Packard et al., 1994;
discrimination, whereas rats with lesions of the Packard and Teather, 1998); the hippocampal in-
CeA do (Cador et al., 1989; Burns et al., 1993). fusion result is the opposite: Enhanced memory for
Connections between the Bla and the ventral stria- the hidden platform task and no advantage for
tum also are involved in conditioned place prefer- the visible platform. However, post-training intra-
ence (Everitt et al., 1991). amygdala injections of amphetamine enhance mem-
Nevertheless, other research suggests that the ory in both water maze tasks (Packard et al., 1994;
central nucleus of the amygdala also has an impor- Packard and Teather, 1998), suggesting that the
tant modulatory role in secondary reinforcement amygdala may have broad influence, modulating
paradigms. Drugs like amphetamine, that release both the hippocampal and caudate–putamen mem-
dopamine, increase lever pressing to cues previously ory systems.
associated with food. The effect occurs with local Perhaps similarly, lesions of the central nucleus
infusion of amphetamine into the nucleus accum- block freezing but do not escape to a tone previ-
bens (Taylor and Robbins, 1984), but is blocked ously paired with shock, whereas lesions of the
when 6-OHDA prompts local dopamine depletion basal nucleus of the basolateral complex have just
(Taylor and Robbins, 1986). However, 6-OHDA the opposite effect (Amorapanth et al., 2000).
does not block the conditioned reinforcement itself, However, lesions of the lateral nucleus block both
which is consistent with the idea that the reinforce- freezing and escape. Lesions of the Bla (but not the
ment signal comes from some other brain area, such CeA) also block avoidance of a bar associated with
as the Bla, that projects to the nucleus accumbens. shock (Killcross et al., 1997), suggesting that baso-
These results suggest that two relatively independ- lateral outputs to the dorsal or the ventral striatum
ent processes operate during conditioned reinforce- may be important in escape or avoidance learning.
ment. First, information from the amygdala
concerning the CS–US (US — unconditioned stim-
ulus) association is sent to the nucleus accumbens Projections to the cortex
to control instrumental behavior as a conditioned
reinforcer. Second, dopamine in the nucleus accum- Primate research shows that the basal nucleus of the
bens modulates this instrumental behavior. The amygdala projects to several areas in the inferior
central nucleus of the amygdala, via its projections temporal cortex, continuing into prestriate and stri-
to the mesolimbic dopamine system, seems to be ate areas of the occipital lobe (Amaral and Price,
critical for this invigorating or arousing effect of 1984; Iwai and Yukie, 1987). Furthermore, the lat-
dopamine. Thus, lesions of the central nucleus eral nucleus of the amygdala gets input from an
block the increase in bar pressing normally pro- adjacent site in the visual system (TEO), which in
duced by infusion of amphetamine into the nucleus turn receives hierarchical projections from the sev-
accumbens (Robledo et al., 1996), probably by pre- eral nuclei along the ventral visual stream, extend-
venting dopamine in the nucleus accumbens shell ing to the retinal mapping area of the calcarine
(Everitt et al., 2000). fissure. These projections could potentially ‘‘close
13

the loop’’ with the visual system (Amaral et al., cortex showed selectivity before the behavioral cri-
1992) representing an amygdala feedback circuit terion was reached and many fewer reversed their
that may be significant for the sustained perceptual selectivity during reversal training (Schoenbaum
evaluation seen in the early stages of emotional et al., 1999). These investigators suggest that cells
processing. Following Pavlovian conditioning, pres- in the Bla encode the associative significance of
entation of a conditioned stimulus appears to elicit cues, whereas cells in the orbitofrontal cortex are
some neural representation of the US with which it active when that information, relayed from the Bla,
was paired — as the sound of an electric can opener is required to guide motivated choice behavior,
might elicit a representation of food and signal ap- presumably via to both the motor cortex and to the
proach behavior in the family cat. On the basis of a dorsal striatum.
procedure called ‘‘US devaluation,’’ several studies The significance of Bla and frontal cortex in US
suggest that the basolateral amygdala — perhaps representation and in guiding motivated (and
via connections with cortical areas such as the per- choice) behavior is also supported by lesion stud-
irhinal cortex (cf. Gewirtz and Davis, 1998) — is ies with rhesus monkeys (Baxter et al., 2000). That
critical for retaining these US representations (e.g., is, when both areas were lesioned, monkeys con-
Hatfield et al., 1996). Second-order conditioning tinued to approach a cue associated with food on
also depends on a US representation elicited by a which they had recently been satiated; whereas,
CS. Again, lesions of the Bla, but not the CeA, control monkeys showed appropriate choice be-
block second-order conditioning (Everitt et al., havior and consistently switched to a new cue.
1989, 1991; Hatfield et al., 1996). This same effect
occurs with local infusions of NMDA antagonists
Studies of the amygdala in humans
into the basolateral nucleus of the amygdala
(Gewirtz and Davis, 1997).
In the past, the primary source of data on human
Converging evidence also now suggests that the
amygdala functioning was based on the behavior
connection between the basolateral nucleus and the
of patients with temporal-lobe lesions — either
prefrontal cortex is critically involved in the way in
from accident, disease, or surgical treatment.
which a representation of a US (e.g., very good,
Overall, these data suggest that the consequences
somewhat good, somewhat bad, very bad) guides
are emotional deficits in emotion perception (e.g.,
approach or avoidance behavior. Analogous to the
Adolphs et al., 1994; Adolphs and Tranel, 1999)
animal data, patients with lesions of the orbital
and expression (e.g., Lee et al., 1998). However,
regions of the prefrontal cortex frequently ignore
these findings are considerably less consistent and
important information that could usefully guide
specific than those available from studies of ex-
their actions and decision-making (Damasio, 1994;
perimentally lesioned animals. More recently,
Bechara et al., 1997; Anderson et al., 1999).
however, the advent of neural imaging techniques,
Studies using single-unit recording techniques in
Positron emission tomography (PET), and func-
rats indicate that cells in both the Bla and the
tional magnetic resonance imaging (fMRI), has
orbitofrontal cortex fire differentially to an odor,
opened a new, non-invasive window for the study
depending on whether the odor predicts a positive
of regional changes in the human brain.
(e.g., sucrose) or negative (e.g., quinine) US. These
differential responses emerge before the develop-
ment of consistent approach or avoidance behavior Brain imaging
elicited by that odor (Schoenbaum et al.,1998).
Many cells in the Bla reverse their firing pattern Functional MRI and PET do not directly measure
during reversal training (i.e., the cue that used to neural action, but rather, measure enhanced blood
predict sucrose now predicts quinine and vice versa flow in the capillaries of the cerebral parenchyma.
— Schoenbaum et al., 1999), although this has not This effect is, however, a reliable vascular sequel to
always been observed (e.g., Sanghera et al., 1979). regional neural firing. Thus, the method can be used
In contrast, many fewer cells in the orbitofrontal to assess functional anatomy mediating language,
14

reflexes (autonomic and somatic), and behavioral in activation with rated level of increased emo-
actions that are emotion’s output. tional arousal of picture content. Finally, both
As already noted, appetitive and threatening sites showed heightened activation when phobics
stimuli capture attention and appear to accentuate viewed pictures of feared stimuli, relative the re-
processing in primary sensory areas. Primate re- sponse of non-fearful participants (see Fig. 3).
search indicates, furthermore, that the amygdala There are several studies suggesting that pictures
projects to occipital and ventral temporal process- of emotional faces engage the amygdala, but the
ing areas of the visual system (Amaral et al., 1992). findings are difficult to interpret. That is, face
To evaluate emotional processing in the visual sys- stimuli generally do not arouse strong emotion (as
tem, Lane et al. (1997) using PET and Lang et al. defined by autonomic activation or verbal report),
(1998) using fMRI presented evocative picture and facial expressions vary in communicative clar-
stimuli (IAPS) to normal subjects and recorded ity (e.g., fear faces are less reliably distinguished
blood flow changes in the caudal cortex. Compared than angry faces). One view holds that the am-
to affectively neutral pictures, participants showed ygdala functions as an initial stimulus discrimina-
dramatically larger areas of activation (in lateral tor that screens for motivational significance, but
occipital, posterior parietal, and inferior temporal subsides when the stimulus is resolved (e.g.,
cortex) for pictures with emotionally evocative con- Whalen, 1998). Interestingly, in both non-human
tent. Subsequent fMRI research (Bradley et al., and human subjects, several amygdala-mediated
2003; Sabatinelli et al., 2004) has further shown that responses (Applegate et al., 1983; Whalen and
activation in these secondary-processing areas of Kapp, 1991) reach their peak early during condi-
the visual system progressively increases, covarying tioning and subside thereafter (Schneiderman,
monotonically, with the judged emotional arousal 1972; Masur et al., 1974; Weisz and McInerney,
of picture stimuli. The significance of heightened 1990; also see Kapp et al., 1990). Moreover, when
emotional/motivational intensity is also highlighted stimulus contingencies change (e.g., when a CS is
by PET research with phobic participants (Fredrik- suddenly not followed by shock at the beginning of
son et al., 1993, 1995), showing enhanced occipital extinction), there is a re-emergence of single-unit
activation when viewing pictures of specific fear activity in the lateral amygdala nucleus in rats
objects. (Quirk et al., 1995). Under analogous conditions,
Consistent with the animal experiments, ne- humans show a resurgence of amygdalar blood
uroimaging researchers have generally reported flow (LaBar et al., 1998). Considering the animal
amygdala activation when human participants research, one could speculate that blood flow oc-
process emotional stimuli. Thus, Pavlovian fear- curring ‘‘in the region of the amygdala’’ during
conditioned cues prompt increased fMRI am- face discrimination might indicate Bla activation,
ygdala signal (cf. Davis and Whalen, 2001). Irwin and not coincident activation of the CeA (with
et al. (1996) reported that fMRI signal intensity is implications for autonomic reflexes and an ‘‘expe-
also greater in the amygdala when subjects view rience’’ of emotion). Unfortunately, the spatial
graphic photographs of unpleasant material (e.g., resolution of our current imaging tools are not yet
mutilated human bodies), compared with neutral sufficient to reliably discriminate individual nuclei.
pictures. Comparable increases in amygdala activ-
ity to emotion-evoking visual input have also been
found with PET (Lane et al., 1997; Reiman et al., Emotional arousal: Physiology and cognition
1997); and Cahill et al. (1996) reported a relation-
ship between emotion-induced amygdala activa- From the perspective of the animal model pre-
tion and recall of the visual content. In more sented here, input to the amygdala’s Bla begins the
recent research, Sabatinelli et al. (2005) found that sequence of neural events that underlay emotion,
activation of the amygdala covaried closely with namely orienting and attention, approach, and
the activation in the inferotemporal, visual defensive behaviors such as avoidance. Bla outputs
processing area. Furthermore, both sites increased to the CeA and the extended amygdala appear to
15

Fig. 3. Effects of emotional intensity and fear relevance on amygdala and inferior temporal regions of interest (ROIs) as measured
with functional magnetic resonance imaging (fMRI: From Sabatinelli et al., 2005). Areas of functional activity during picture
processing (superimposed on the average brain) are present in the right panel. Average standardized blood oxygen level dependent
(BOLD) signal change in the control group (left panel, top) and the snake-fearful group (left, bottom) are presented for the amygdala
and inferotemporal cortex, as these signals varied over different picture content categories. Error bars represent Standard Errors of the
Mean. The picture types are ordered on the abscissas (left to right) as they increase in average affective arousal ratings (International
Affective Picture System standardization groups: Lang et al., 1999).

be critical in the increased processing of emotion- Bla to the ventral striatum as well as the orbito-
ally significant stimuli, be they pleasant or avers- frontal cortex are also likely contributors to the
ive. Outputs from the CeA and BNST in turn execution of approach and choice behavior.
mediate many of the autonomic and somatic com- The above circuitry constitutes a motivational
ponents of overt action. Direct output to the dor- system that is finely tuned to the probability that
sal striatum, or indirect output via the orbital events will require survival action, e.g., that a re-
frontal cortex, appear to be involved in the actual mote threat will become an imminent danger or
avoidance response. Furthermore, outputs from that a sexual provocation will likely to lead to
16

pleasant consummation. In animals, increasing of consummatory pleasures. We can consider this


imminence prompts a more general mobilization physiology in the context of conditioning proce-
of the organism, mediated by various neurotrans- dures, explaining links between behavior and simple
mitters such as acetylcholine, dopamine, nor- lights and tones, invoking concepts such as conti-
epinephrine, as well as many peptides such as guity, primary reinforcement, or generalization.
CRH. These substances act either within the am- However, we understand less how the neurophys-
ygdala or at various central target areas to facil- iology and chemistry of arousal connects to human
itate neural transmission, and they are associated experience and to the wider information processing
with increasing intensity of appetitive or defensive functions of the brain. Although it is possible
motivation (and roughly correlated with arousal that some emotional stimuli might be ‘‘hard wired’’
reports in humans). (Seligman, 1970; Öhman and Mineka, 2001) for
How might the above model relate to the expe- association to avoidance or approach, most affec-
rience of an emotion? Or from a neuro-chemical tive cues are indirectly learned, remote associations,
perspective: How does looking at a picture, think- complexly connected to memories of appetite or
ing pleasant or unpleasant thoughts, or remem- aversion, and reflect the unique reactional biogra-
bering a painful experience lead to a release of phies of individuals.
acetylcholine, dopamine, norepinephrine, or co- Cognitive psychologists (e.g., Anderson and
rticotropin releasing hormone? Bower, 1974; Kintsch, 1974) suggested that our
Considering the animal research, we know that knowledge about events is represented in networks
the CeA and the BNST have direct connections to of representations, linked by laws of association,
the neurons in the brainstem that release acetyl- and instantiated by stimuli that match elements in
choline, dopamine, norepinephrine and to neurons the network. Lang (1979, 1994) suggested that
in the basal forebrain that release ACH. Electrical emotional memories are networks of this type. For
stimulation of the amygdala has been shown to example, the fear network of a snake-fearful indi-
increase cell firing in many of these neuronal vidual would include information about the stim-
groups. Activation of cells in the brainstem that ulus (a coiled snake, but also the snake’s skin — as
release norepinephrine and epinephrine are espe- in a belt, a rustling sound), the stimulus context
cially important in modulating memory in many (swimming at the lake, alone in the woods), phys-
different brain areas via activation of nor- iology (changes in heart rate and the vasomotor
epinephrine receptors in the Bla (McGaugh, system, respiration, and behavior [running away]),
2004). Activation of acetylcholine receptors is also and related interpretive elaborations (snakes are
critical for memory formation (Gold, 2003; Power slimy; snakes can kill). When enough input cues
et al., 2003). In addition, cells in the lateral divi- match units in the network, activity in one unit is
sion of the amygdala’s central nucleus send CRH transmitted to adjacent units, and depending on
containing terminals to the BNST (Sakanaka et the strength of activation, the entire memory
al., 1986), where many of the actions of CRH may structure may be engaged. The probability of net-
actually be mediated (cf. Walker et al., 2003). work processing is likely to be increased with the
Thus, more arousing images and thoughts could number of units initially stimulated. The cognitive
activate more cells in amygdala that automatically network is presumed to overlay a neural network
lead to a release of these neurochemicals, helping — perhaps, an organization of Hebbian cell as-
to stamp in these images for later recall. semblies (Hebb, 1949). Of course, only a fraction
of its representational units would have higher
level, language representation that — passing
Cognitive networks through awareness — are the formative stuff of
affective reports.
We can roughly describe the neural structure and How do emotional networks differ from other
chemistry of emotional arousal in the context of knowledge structures in the brain? It is proposed
imminent nociceptive stimulation or in anticipation that emotional networks are singular because they
17

include associative connections to the primary positive potential, is systematically related to par-
motivational systems in the brain that are the fo- ticipants’ ratings of their experience of emotional
cus of this discourse. In brief, reciprocal projec- arousal (Cuthbert et al., 2000).
tions from cortex to the amygdala circuits engage Sabatinelli et al. (in press, Fig. 4) have shown,
the somatic and autonomic reflexes that evolved to furthermore, that this EEG response to emotional
ensure the survival of individuals and species. pictures is highly correlated with an increased
fMRI blood oxygen level dependent (BOLD) sig-
nal in visual cortex (extrastriate occipital, parietal,
Levels of activation
and inferior temporal), when this fMRI response is
prompted by the same picture stimuli. Thus, it is
It may be that stimuli evoke differences in physio-
suggested that both the BOLD response and the
logical arousal because networks activate different
late positive ERP results from activity in a com-
numbers of cells in the amygdala depending on the
mon neuronal pool in the posterior cortex that
associative history of those stimuli. In fact, most
determines the semantic meaning of visual input.
stimuli or situations that produce an emotional re-
In this case, the positive voltage ERP reflects vol-
action do so by virtue of prior conditioning. Mon-
ume conduction to the scalp of bio-electric changes
keys reared in the lab, where serpents are not
associated with actual neural-cell firing, while the
normally encountered, are generally not afraid of
delayed fMRI response is the replacement flow of
snakes, when compared to monkeys raised in the
oxygenated blood subsequent to metabolism in
wild (Mineka et al., 1984). A baby with its fingers in
this same cell complex. As already noted, emo-
a light socket does not feel afraid when the light
tional BOLD activation in sensory cortex (partic-
switch is turned on, whereas a child who was once
ularly, inferotemporal cortex) has been shown, in
shocked in a similar situation will be terrified. After
turn, to correlate highly with activation in the am-
this association is formed, putting a finger in a
ygdala (Sabatinelli et al., 2005). Taken together,
socket may be presumed to engage many cells in the
these findings suggest the late positive ERP is a
amygdala, leading to a large release of neurochem-
valid indicator of emotional network processing.
icals and strong activation of the defense motivation
From a cognitive perspective, emotional arousal
system. Similar but perhaps less robust amygdala
occurs when a stimulus activates a matching rep-
activations are assumed to occur whenever emotion
resentation in an emotional network. The immedi-
networks are activated — by distressing thoughts,
acy and intensity of the affective reaction depends
imagery, and when viewing emotionally evocative
on the general associative strength of the network,
media (e.g., pictures, written or spoken text). For
and specifically, on the strength of connections to
example, just telling people they might get a painful
the amygdala circuit. In humans, this net is broadly
shock is enough to increase blood flow in the am-
cast and affects can be prompted by representations
gydala (Phelps et al., 2001).
that are not readily discriminated linguistically (and
are therefore, outside awareness). Thus, many
EEG measures of network activation different stimuli, varying across individuals, can
prompt an amygdala-dependent release of neuro-
The EEG is recorded from electrodes on the surface chemicals, with a potentially widespread modula-
of the scalp that transmit the brain’s neural, bio- tion of sensory and motor systems (and in humans,
electric activity in real time. There is now consider- reports of increasing ‘‘arousal’’).
able evidence that emotional networks, activated by
picture stimuli, prompt distinct EEG, event-related
potential (ERP) waveforms. The most reliable ERP The startle reflex and emotional priming
is a slow positive-going voltage, typically beginning
400 ms after picture onset and continuing until off- From an evolutionary perspective, human emo-
set (Cacioppo et al., 1994; Keil et al., 2002; Schupp tions can be considered dispositions to action — as
et al., 2004). The amplitude of this waveform, or late the experience of anger is a throwback to attack,
18

Fig. 4. The event-related potential (ERP) scalp topography obtained from the electroencephalograph (EEG: upper left) and the fMRI
BOLD contrast overlay (lower left) are compared for pleasant, neutral, and unpleasant picture presentations. The ERP map represents
average microvolt change from 400 to 900 ms after picture onset, with red indicating postive and blue indicating negative voltage. The
BOLD contrast overlay, coronal slice y ¼ 68, represents a random effects analysis of picture elicited activity, peaking roughly 8 s after
picture onset, with red indicating more reliable increases in oxygenated blood flow and yellow indicating a threshold of po.000000001.
Time-course waveform of the ERP and the fMRI BOLD response prompted by pleasant, neutral, and unpleasant picture processing
are presented in the right panel. The ERP waveform represents the average of 27 centro-parietal sensors; the BOLD waveform
represents the average of activity in inferotemporal, lateral occipital, and parietal regions of interest. The arrow at time zero on the
abscissas indicates picture onset. The correlations between the late positive potential (400–900 s) and BOLD signal were 0.60 for the
lateral occipital cortex, 0.51 for the parietal cortex, and 0.52 for the inferotemporal cortex (all p values less than 0.001: Adapted from
Sabatinelli et al., in press).

or fear is an evolved disposition to flee. In humans, human subjects, sudden closure of the eyelids is
the ‘‘emotion’’ generally occurs in a context of in- one of the first, fastest (occurring within 25 to
hibition or delay, and the primed action may or 40 ms after startle stimulus onset), and most stable
may not ultimately occur. In contrast, affect is not elements in the reflex sequence. It is the preferred
easily inferred in more primitive animals, as the measure of startle in humans. In rats, whole body
overt response — attack or flight — is immediate startle is measured in specially designed cages.
and automatic. A few of these automatic survival Although, when startled mammals typically do
responses have been preserved in humans — no- not show actual escape, the amplitude of the star-
tably, the startle reflex. tle reflex is augmented when they are under threat
Any abrupt sensory event will prompt a startle of pain or predation. The first laboratory demon-
response, a chained series of rapid extensor-flexor stration of this was reported by Brown et al.
movements that cascade throughout the body (1951), who showed that the amplitude of the
(Landis and Hunt, 1939). This reaction is a defen- acoustically elicited startle reflex in rats was in-
sive reflex, facilitating escape in many species (e.g., creased when elicited in the presence of a light
the hermit crab; Elwood et al., 1998), and it may previously paired with footshock. This effect is
still serve a protective function in mammals (i.e., in considered an animal model of fear or anxiety be-
avoiding organ injury as in the eyeblink, or in re- cause drugs like diazepam, which reduce fear or
traction of the head and torso in the full body anxiety in humans, block the increase in startle in
startle reflex to avoid attack from above; Li and the presence of the conditioned light stimulus but
Yoemans, 1999). Abruptness is the key to startle do not affect the basic startle reflex itself (see Davis
elicitation: When effective, the rise time of the et al., 1993, for a review). In contrast, during an
eliciting stimulus is perceptually instantaneous. In appetitive state, the startle reflex appears to be
19

partially inhibited, i.e., startle amplitude is reduced When startle probes are administered while sub-
when elicited in the presence of a light previously jects view pictures that vary systematically in emo-
paired with food (Schmid et al., 1995) or reward- tional valence, results have consistently conformed
ing electrical brain stimulation (Yeomans et al., to the motivational priming hypothesis: The
2000). probe-startle reflex is inhibited when participants
These effects are very like what cognitive psy- view pleasant stimuli and is potentiated when pic-
chologists call ‘‘priming’’ in research on human as- tures were judged to be unpleasant (Vrana et al.,
sociative behavior. Cognitive priming occurs when 1988; Lang et al., 1990; Lang, 1995; see Bradley,
a prior stimulus raises the activation level of an 2000 for a review). These emotion-perceptual
associated stimulus–response event. For example, effects have also been reported in five-month old
the prime ‘‘bread’’ prompts a faster reaction time infants, viewing smiling, neutral, and angry faces
response to the word ‘‘butter.’’ States of the organ- (Balaban, 1995).
ism may also prime particular behavior. Thus, clin- Affective modulation of startle is observed for
ically depressed individuals respond to nearly all picture stimuli regardless of whether the startle
cues with associations that are affectively negative. probe is visual, acoustic, or tactile (e.g., Bradley
The potentiated startle observed in animal condi- et al., 1990; Hawk and Cook, 1997), suggesting
tioning can be understood as an instance of moti- that the effect is not modality specific. Further-
vational state priming. That is, the induced more, affective modulation is not confined to vis-
defensive state of the organism primes (augments) ual percepts: When the foreground stimuli consist
an independently instigated reflex that is connected of short, 6 s sound clips of various affective events
to the defense system, i.e., the startle response. (e.g., sounds of love making, babies crying, bombs
According to the motivational priming hypoth- bursting), and the startle probe is a visual light
esis (Lang et al., 1990, 1997), the defensive startle flash, the same affective modulation of the probe
reflex will be of significantly greater amplitude reflex is observed (Bradley et al., 1994). Research-
(and faster) when a prior stimulus has induced a ers have also found startle potentiation in subjects
consonant, defensive motivational state. Alterna- smelling unpleasant odors (Miltner et al., 1994;
tively, if the appetitive system has been activated, Ehrlichman et al., 1995).
as when a pleasant stimulus is perceived, the
defensive startle reflex will undergo a reciprocal
attenuation. Thus, the startle reflex can serve as Attention and arousal
a non-volitional, objective measure of affective
valence; and as will be seen, it is a method that Consistent with the motivational priming hypoth-
solidly links animal neuroscience research and esis, modulatory effects on the startle reflex appear
human psychophysiology. to increase with greater activation of a motive
system. However, the direction of effect is opposite
for appetitive and defensive activation: The most
Startle modulation in humans arousing pleasant pictures (e.g., romantic and erotic
couples) prompt the greatest startle inhibition (Cut-
Like rats, human subjects show elevated startle hbert et al., 1996). Conversely, when unpleasant
amplitude in the presence of cues previously paired arousing pictures are probed, the startle reflex is
with shock (Hamm et al., 1993; Lipp et al., 1994; strongly potentiated. Thus, when phobic individu-
Grillon and Davis, 1997) or simply when they are als look at pictures of their fear object (e.g., snakes
told they might receive a shock (Grillon et al., or spiders), they show dramatically augmented
1991, 1993). However, this potentiation phenom- startle potentiation compared to normal subjects
enon is not restricted to anticipated pain; rather, (Hamm et al., 1997; Sabatinelli et al., 2001).
the probe-startle reaction is modulated (up or Graham (1992) previously proposed that startle
down) in almost any emotion evoking, perceptual reflex inhibition was an index of attentional
context (Lang et al., 1990). engagement, and recent research supports the view
20

‘‘pre-pulse’’ inhibition may be greatest for emo-


tional arousing pictures (irrespective of valence).
The early startle inhibition could indicate initial
attention capture that is preferential for motiva-
tionally relevant input. In the subsequent seconds,
as the picture input is more fully processed, arous-
ing unpleasant pictures prompt the expected reflex
potentiation, while startle reflexes during pleasant
pictures remain inhibited. These data suggest that
relative reflex inhibition is the modal reaction
when attention is engaged, and that this reduced
response increases with more ‘‘interesting’’ and
more pleasantly exciting foregrounds. In this view,
potentiated startle is the exceptional response, oc-
curring only when attention is focused on very
upsetting or threatening stimulus contents. The
fact that inhibition of motor action is the early,
first reaction to a stimulus and that defensive po-
tentiation occurs later (and only to the most
arousing unpleasant stimuli) is consistent with the
defense motivational sequence processing from at-
tention to action, and has been discussed previ-
ously in a wider timeframe. That is, the first
reaction to threat is inevitably to stop, look, and
Fig. 5. Mean startle blink magnitude in response to probe noise listen (as prey does when first perceiving a pred-
stimuli while participants viewed specific unpleasant picture ator in the distance), and only after information is
contents. Picture contents are ordered on the abscissa by rated gathered, a strategy is devised, and the attack is
affective arousal from left to right (low to high arousal). The clearly imminent, are the acts of fight or flight de-
correlation between content arousal rank and blink magnitude
ployed.
was 0.86, po0.01.

that attention and emotional arousal are interact-


ing factors in probe-startle modulation. Thus, The amygdala and conditioned fear: Startle
probes presented during low-arousal, unpleasant modulation in the rat
stimuli — pictures of sad events, pollution, people
who are ill — actually prompt some reflex inhibi- Figure 6 shows pathways believed to mediate fear-
tion relative to less interesting, neutral stimuli. potentiated startle in rats during conditioned vis-
Startle magnitude increases, however, linearly as ual stimulation. Information from the retina
unpleasant pictures are reported to be more arous- projects to both the dorsal lateral geniculate nu-
ing — with threat to the viewer (a gunman point- cleus and lateral posterior nucleus of the thalamus.
ing his weapon at the viewer or animal attack) The dorsal lateral geniculate projects to the visual
prompting the greatest potentiation (Cuthbert et cortex whereas the lateral posterior nucleus
al., 1996; Bradley et al., 2000; see Fig. 5). projects directly to the Bla (and the perirhinal
There is also evidence that startle reflexes are cortex, Shi and Davis, 1996), which then projects
reduced for all picture contents, when attention is to the lateral and/or basolateral nucleus of the
initially engaged (i.e., the first few hundred milli- amygdala. Next, the basolateral nucleus projects
seconds after the onset of a stimulus; see Bradley to the central nucleus, which in turn has neural
et al., 2003), and furthermore, that this attentional projections to the startle pathway.
21

Electrolytic or chemical lesions of visual thala-


mus (Shi and Davis, 1996), perirhinal cortex
(Rosen et al., 1992; Campeau and Davis, 1995a)
or Bla (Sananes and Davis, 1992; Campeau and
Davis, 1995b) completely block the expression of
fear-potentiated startle using a visual CS. None of
these lesions affect startle amplitude itself. Lesions
(Hitchcock and Davis, 1986; Campeau and Davis,
1995b) or local infusion of glutamate antagonists
into the CeA (Kim et al., 1993; Walker and Davis,
1997) also block fear-potentiated startle.
Both conditioned fear and sensitization of star-
tle by footshocks appear to modulate startle at the
level of the nucleus reticularis pontis caudalis
(PnC) (Berg and Davis, 1985; Boulis and Davis,
1989; Krase et al., 1994). The CeA projects directly
to the nucleus reticularis pontis caudalis (Rosen et
al., 1991) and electrolytic lesions along this path-
way block the expression of fear-potentiated star-
tle (Hitchcock and Davis, 1991). However, an
obligatory synapse appears to exist in this pathway
because fiber-sparing chemical lesions in the deep
layers of superior colliculus or periacqueductal
gray also block fear-potentiated startle (Frankland
and Yeomans, 1995; Fendt et al., 1996), as does
local infusion of muscimol or glutamate antago-
Fig. 6. Schematic diagram of pathways believed to be involved nists into a part of the mesencephalic reticular
in the fear-potentiated startle effect in rats using a visual con-
ditioned stimulus. Visual information goes from the retina
formation (Meloni and Davis, 1999; Zhao and
through the lateral posterior nucleus of the thalamus either di- Davis, 2004).
rectly to the basolateral amygdala or indirectly through the
perirhinal cortex to the amygdala. Visual information also goes
from the retina to the dorsal lateral geniculate nucleus to the
visual cortex and then through the perirhinal cortex to the am-
Pleasure attenuated startle
ygdala. Shock information goes through a series of parallel
pathways to eventually reach the basolateral amygdala. The
basolateral nucleus of the amygdala projects to the central nu- As mentioned earlier, startle amplitude is de-
cleus which then projects directly to the startle pathway at the creased when elicited in the presence of a cue pre-
level of the nucleus reticularis pontis caudalis (PnC), as well as viously paired with food (Schmid et al., 1995).
indirectly via a synapse in the deep layers of the superior col-
Moreover, the nucleus accumbens is important
liculus and mesencephalic reticular formation (Deep SC/Me).
A pathway in the brain stem and spinal cord mediates the for this ‘‘pleasure-attenuated startle effect’’ be-
startle reflex itself. Afferents from the cochlea synapse onto a cause pretraining local infusion of the neuro-
small group of neurons embedding the auditory nerve called toxin 6-OHDA into the nucleus accumbens, which
cochlear root neurons (CRN) which send heavy projections to markedly reduces dopamine, blocks pleasure at-
the PnC. Axons from cells in the PnC form part of the reti-
tenuated startle (Koch et al., 1996). It is possible
culospinal tract that make both monosynaptic and polysynaptic
connections in the spinal cord onto motoneurons. that the connection between the Bla and the ven-
tral striatum may be involved in pleasure-attenu-
ated startle. Unilateral lesions of the Bla on one
side of the brain, and the nucleus accumbens ab-
lation on the other, would test this hypothesis.
22

Startle and amygdala in the human brain defense may be attack. When the threat stimulus is
distant, however, the rat freezes and orients to-
There is limited direct evidence that the amygdala’s ward the predator. The Blanchards noted further
role in determining startle reflex amplitude is sim- that increases in ‘‘the amplitude of the startle re-
ilar in rodents and humans. Bower et al. (1997) sponse to sudden stimuli accompany decreasing
have provided suggestive data: They studied epi- defensive distance’’ (p. S5).
leptic patients who were treated by resection of the Using concepts introduced by Timberlake and
temporal lobe, and found that greater post-surgery colleagues (Timberlake and Lucas, 1989; Timber-
amygdala volume was associated with larger base- lake, 1993), Fanselow (1991, 1994) has similarly
line startle reflexes. Funayama et al. (2001) also analyzed fear behavior, describing three stages of
looked at epilepsy patients with temporal pole re- increasing prey–predator proximity: Pre-encounter
section and found deficits in fear-potentiated star- defense, pre-emptive behavior that occurs in a for-
tle that were task and hemisphere specific. Patients aging area where predators were previously en-
with damage to the right temporal lobe did not countered; post-encounter defense, responses
potentiate to unpleasant pictures but did when prompted by the detection of a distant predator;
told they might be shocked. Patients with left tem- circa-strike defense, behaviors that occur in the
poral-lobe damage were the opposite and one pa- region of physical contact or its close imminence.
tient with bilateral damage did not potentiate in Behavior shifts from pre-emptive threat vigilance
either case. This suggests that many of the later- at pre-encounter to post-encounter freezing and
alized effects that have variously been reported orienting to a specific predator cue, to the circa-
may turn out to be task specific. An fMRI study of strike stage when the organism, beyond vigilance,
amygdala reactions to faces of different races was engages in vigorous defensive action.
also suggestive (Phelps et al., 2000): The researcher Mild amygdala stimulation (electrical) first stops
showed that a measure of implicit negative eval- on-going behavior (freezing), prompting brad-
uation and startle potentiation both co-varied with ycardia and EEG activation. As stimulation in-
increased left amygdala activation. There are, of creases, the animal becomes more active; at high
course, imaging studies showing amygdala activa- levels of stimulation, it vigorously attempts escape
tion in human conditioning studies similar to those from the source of stimulation. It appears that a
conducted with animals (e.g., Morris et al., 1996). predator at some distance prompts mild activation
However, startle is rarely included as an additional of the amygdala, functionally increasing attention.
dependent variable. As the predator comes closer, amygdala activation
increases, now mediating overt defensive behavior,
including escape. As Fanselow (1991, 1994) pro-
Emotional processing: From attention to action posed, the switch from an attentional mode to ac-
tive defense may involve a switch in activation from
When a wild rat sees a human at some distance ventral to dorsal periacqueductal gray. Bandler and
away, the rat freezes in a highly alert, attentive others (cf. Bandler and Shipley, 1994) have shown
posture. As the human (or other potential preda- that the ventral periacqueductal gray projects to
tor) gradually approaches, the rat suddenly darts both cardiovascular centers mediating bradycardia
away, if escape is possible. If escape is not possible and motor system inhibition. In contrast, the dorsal
and the human gets very close, the rat will attack periacqueductal gray projects to centers mediating
(Blanchard et al., 1986). Defensive behaviors in- tachycardia and active escape behavior. Assuming a
crease systematically with a reduction in distance low threshold for projected amygdala activation in
from predators and other dangerous or potentially ventral periacqueductal gray, we might expect this
painful stimuli (Blanchard and Blanchard, 1987): structure to mediate attentional responses, subse-
Given an available escape route, proximity is as- quent to the amygdala’s activation by a novel event.
sociated with an increased probability of active If the dorsal periacqueductal gray had a higher
flight. In the absence of an escape option, the best threshold for amygdala activation, however, then
23

its function would depend on greater activation in overt behaviors that may result with increasing im-
the amygdala (e.g., as with predator approach), minence (closer to the Circa-strike region) could
prompting an abrupt switch from passive vigilance look fearful, angry — or given overwhelming stress
to full-scale action. and no available coping behavior — and could lead
Lang et al. (1997) proposed an adaptation of the to autonomic and somatic collapse, hopelessness,
predator stage model for explicating human brain and depression.
and psychophysiological reactions to unpleasant The model’s assumed predator is, of course, en-
and threatening stimuli. They suggest that humans, gaged in parallel dance — first observing quietly
when first seated in the psychophysiological labora- the distant prey, stalking forward slowly, increas-
tory, are functionally similar to an animal at the pre- ingly mobilized for what becomes, at the Circa-
encounter stage, i.e., they are moderately alert in strike stage, a final charge. Overall, it is a parallel
an unfamiliar environment. For humans and other progression from attention to action, with a coin-
animals, presentation of an aversive stimulus (post- cident increment in appetitive arousal. While there
encounter) prompts focused attention, ‘‘freezing,’’ is currently little data on the predator’s anticipa-
apprehension, and arousal. Human participants tory pleasures — joy of the hunt, satisfaction in its
show this pattern when viewing threatening films consummation — the neurophysiology of the
or pictures, or attending to a previously neutral cue process is likely to be similar in psychophysiolo-
that augers a contingent aversive event, such as gy to what is observed in defense.
electric shock. During post-encounter, physiological
indices of attention at first increase with imminence
and severity of the threatened consequence — Emotion and the brain: Concluding thoughts
greater skin conductance, increased heart rate de-
celeration, and some inhibition of the probe-startle We have proposed here that basic emotions are
reflex. During this period, however, brain and body founded on brain circuitry in which the amygdala
are also mobilizing for possible action. One could is a central component. We conceive emotion, not
conjecture that amygdala transmissions to the cen- as a single reaction, but as the reflection of a mo-
tral gray are increasing with greater arousal (immi- tivational process. Emotions reflect sequenced, so-
nence of threat), ultimately switching the impact site matic and autonomic reflexes, organized by brain
from the ventral (freezing) to the dorsal (action) circuits that developed over evolutionary history
gray. The first motor consequence of this switch to insure the survival of individuals and the prop-
may be the change in direction of probe-startle agation of genes across generations. Thus, events
modulation, i.e., from an initial moderate reflex in- that are positive/appetitive or aversive/threatening
hibition to reflex potentiation. Startle potentiation have an initial advantage in capturing attention.
then progressively increases in magnitude with the Furthermore, these events prompt further infor-
aversiveness of stimuli (as probes occur more prox- mation gathering, evaluation, and action planning
imal to the Circa-strike stage in the animal model). — more so, according to their degree of survival
With a further increment in threat, the heart rate significance. Motive cues also occasion the release
response also reverses the direction of change from of neurochemicals, metabolic arousal, anticipatory
orienting to defense (Sokolov, 1963; Graham, 1979) responses that are oriented towards the engaging
— from a parasympathetically mediated fear brad- event, and a general mobilization of the organism
ycardia to action mobilization and sympathetically preparatory to action. Sometimes these motiva-
mediated cardiac acceleration. tional sequences play out in humans in the same
The biological model of emotion presented here stimulus-driven way that they do in less complex
suggests that, depending on level of stimulus aver- organisms. More often, given an elaborate brain,
sion (threat, apprehension), patterns of physiolog- equipped with language, interactive memory stor-
ical change (and in humans, reports of experienced age, and a vast behavioral repertoire, motivational
emotional arousal) will systematically vary with the reflexes may only be engaged in part, with overt
level of defense system activation. Furthermore, the action inhibited, delayed, or complexly adapted to
24

the context. Nevertheless, it is this reflex bedrock Aston-Jones, G., Rajkowski, J., Kubiak, P., Valentino, R.J.
that generally prompts humans to say that they are and Shipley, M.T. (1996) Role of the locus coeruleus in
emotional activation. Prog. Brain Res., 107: 379–402.
emotionally aroused — joyful, angry, fearful, anx-
Balaban, M. (1995) Affective influences on startle in five-
ious, or sad and hopeless. It is emotion’s reflex month-old infants: reactions to facial expressions of emotion.
automaticity, moreover, that may have much to do Child Dev., 66: 23–36.
with our sense of being less in control when emo- Bandler, R. and Shipley, M.T. (1994) Columnar organization in
tionally aroused, and instead, feel driven by un- the midbrain periaqueductal gray: modules for emotional
bidden impulses and helpless in emotion’s thrall. expression? Trends Neurosci., 17: 379–389.
Baxter, M.G., Parker, A., Lindner, C.C.C., Izquierdo, A.D.
and Murray, E.A. (2000) Control of response selection by
reinforcer value requires interaction of amygdala and orbital
prefrontal cortex. J. Neurosci., 20: 4311–4319.
Acknowledgments
Bechara, A., Damasio, H., Tranel, D. and Damasio, A.R.
(1997) Deciding advantageously before knowing the advan-
This work was supported in part by National In- tageous strategy. Science, 275: 1293–1294.
stitute of Mental Health Grants MH 47840, MH Berg, W.K. and Davis, M. (1985) Associative learning modifies
57250, MH 58922, MH 59906 and the Woodruff startle reflexes at the lateral lemniscus. Behav. Neurosci., 99:
191–199.
Foundation to MD, NSF Contract No. IBN987675,
Bermudez-Rattoni, F., Grijalva, C.V., Kiefer, S.W. and Garcia,
to Emory University, and P50 MH52384, MH J. (1986) Flavor-illness aversion: the role of the amygdala in
37757, and MH43975 to PJL. acquisition of taste-potentiated odor aversions. Physiol. Be-
hav., 38: 503–508.
Blanchard, R.J. and Blanchard, D.C. (1987) An ethoexperimen-
tal approach to the study of fear. Psychol. Rec., 37: 305–316.
References Blanchard, R.J., Flannelly, K.J. and Blanchard, D.C. (1986)
Defensive behavior of laboratory and wild Rattus nor-
Adolphs, R. and Tranel, D. (1999) Intact recognition of emo- vegicus. J. Comp. Psychol., 100: 101–107.
tional prosody following amygdala damage. Neuropsycholo- Boulis, N. and Davis, M. (1989) Footshock-induced sensitizat-
gia, 37: 1285–1292. ion of electrically elicited startle reflexes. Behav. Neurosci.,
Adolphs, R., Tranel, D. and Damasio, A.R. (1998) The human 103: 504–508.
amygdala in social judgment. Nature, 393: 470–474. Bower, D., Eckert, M., Gilmore, R., Leonard, C.M., Bauer, R.,
Adolphs, R., Tranel, D., Damasio, H. and Damasio, A.R. Roper, S., Bradley, M.M., Lang, P.J. and Barta, P. (1997)
(1994) Impaired recognition of emotion in facial expressions Startle eyeblink magnitude in humans depends on extent of
following bilateral damage to the human amygdala. Nature, right amygdala removal. Soc. Neurosci. Abstr., 23: 570.
372: 669–672. Bradley, M.M. (2000) Emotion and motivation. In: Cacioppo,
Amaral, D.G. and Price, J.L. (1984) Amygdalo-cortical pro- J.T., Tassinary, L.G. and Bernston, G. (Eds.), Handbook of
jections in the monkey (Macaca fascicularis). J. Comp. Ne- Psychophysiology. Cambridge University Press, New York,
urol., 230: 465–496. pp. 602–642.
Amaral, D.G., Price, J.L., Pitkanen, A. and Carmichael, S.T. Bradley, M.M., Codispoti, M., Cuthbert, B.N. and Lang, P.J.
(1992) Anatomical organization of the primate amygdaloid (2000) Emotion and picture perception: emotion and moti-
complex. In: Aggleton, J.P. (Ed.), The Amygdala: Neurobio- vation I: Defensive and appetitive reactions in picture
logical Aspects of Emotion, Memory and Mental Dysfunc- processing. Emotion, 1: 276–298.
tion. Wiley, New York, pp. 1–66. Bradley, M.M., Cuthbert, B.N. and Lang, P.J. (1998a) Inter-
Amorapanth, P., LeDoux, J.E. and Nader, K. (2000) Different national affective digitized sounds (IADS). The Center for
lateral amygdala outputs mediate reactions and actions elic- Research in Psychophysiology. University of Florida, Gain-
ited by a fear-arousing stimulus. Nat. Neurosci., 3: 74–79. esville, FL.
Anderson, S.W., Bechara, A., Damasio, H., Tranel, D. and Bradley, M.M., Greenwald, M.K., Petry, M. and Lang, P.J.
Damasio, A.R. (1999) Impairment of social and moral be- (1992) Remembering pictures: pleasure and arousal in mem-
havior related to early damage in human prefrontal cortex. ory. J. Exp. Psychol. Learn. Mem. Cogn., 18: 379–390.
Nat. Neurosci., 2: 1032–1037. Bradley, M.M., Lang, P.J. and Cuthbert, B.N. (1990) Startle
Anderson, J.R. and Bower, G.H. (1974) A propositional theory reflex modification: emotion or attention. Psychophysiology,
of recognition memory. Memory Cogn., 2: 406–412. 27: 513–523.
Applegate, C.D., Kapp, B.S., Underwood, M.D. and McNall, Bradley, M.M., Lang, P.J. and Cuthbert, B.N. (1998b) Affec-
C.L. (1983) Autonomic and somatomotor effects of am- tive norms for English words (ANEW). In: Technical manual
ygdala central n. stimulation in awake rabbits. Physiol. Be- and affective ratings. The Center for the Study of Emotion
hav., 31: 353–360. and Attention. University of Florida, Gainesville, FL.
25

Bradley, M.M., Sabatinelli, D., Lang, P.J., Fitzsimmons, J.R., Cuthbert, B.N., Bradley, M.M. and Lang, P.J. (1996) Probing
King, W. and Desai, P. (2003) Activation of the visual cortex in picture perception: activation and emotion. Psychophysiolo-
motivated attention. Behav. Neurosci., 117: 369–380. gy, 33: 103–111.
Bradley, M.M., Zack, J. and Lang, P.J. (1994) Cries, screams, Cuthbert, B.N., Schupp, H.T., Bradley, M.M., Birbaumer, N.
and shouts of joy: affective responses to environmental and Lang, P.J. (2000) Brain potentials in affective picture
sounds. Psychophysiology, 31(S29). processing: covariation with autonomic arousal and affective
Brown, J.S., Kalish, H.I. and Farber, I.E. (1951) Conditional report. Biol. Psychol., 52: 95–111.
fear as revealed by magnitude of startle response to an au- Damasio, A.R. (1994) Descartes’ Error. Grosset/Putnam, New
ditory stimulus. J. Exp. Psychol., 41: 317–328. York.
Burns, L.H., Robbins, T.W. and Everitt, B.J. (1993) Differen- Davis, M. (2000) The role of the amygdala in conditioned and
tial effects of excitotoxic lesions of the basolateral amygdala, unconditioned fear and anxiety. In: Aggleton, J.P. (Ed.) The
ventral subiculum and medial prefrontal cortex on respond- Amygdala, Vol. 2. Oxford University Press, Oxford, UK, pp.
ing with conditioned reinforcement and locomotor activity 213–287.
potentiated by intra-accumbens infusions of D-amphetamine. Davis, M., Falls, W.A., Campeau, S. and Kim, M. (1993) Fear-
Behav. Brain Res., 55: 167–183. potentiated startle: a neural and pharmacological analysis.
Bussey, T.J., Everitt, B.J. and Robbins, T.W. (1997) Dissocia- Behav. Brain Res., 58: 175–198.
ble effects of cingulate and medial frontal cortex lesions on Davis, M. and Whalen, P. (2001) The amygdala: vigilance and
stimulus–reward learning using a novel Pavlovian autoshap- emotion. Mol. Psychiatry, 6: 13–34.
ing procedure in rats: implications for the neurobiology of Dickinson, A. and Dearing, M.F. (1979) Appetitive-aversive
emotion. Behav. Neurosci., 111: 908–919. interactions and inhibitory processes. In: Dickinson, A. and
Cacioppo, J.T., Crites Jr., S.L., Gardner, W.L. and Berntson, Boakes, R.A. (Eds.), Mechanisms of Learning and Motiva-
G.G. (1994) Bioelectrical echoes from evaluative categoriza- tion. Erlbaum, Hillsdale, NJ, pp. 203–231.
tions: I. A late positive brain potential that varies as a func- Dringenberg, H.C. and Vanderwolf, C.H. (1996) Cholinergic
tion of trait negativity and extremity. J. Pers. Soc. Psychol., activation of the electrocorticogram: an amygdaloid activat-
67: 115–125. ing system. Exp. Brain Res., 108: 285–296.
Cador, M., Robbins, T.W. and Everitt, B.J. (1989) Involvement Ehrlichman, H., Brown, S., Zhu, J. and Warrenburg, S. (1995)
of the amygdala in stimulus–reward associations: interaction Startle reflex modulation during exposure to pleasant and
with the ventral striatum. Neuroscience, 30(1): 77–86. unpleasant odors. Psychophysiology, 32: 150–154.
Cahill, L., Haier, R.J., Fallon, J., Alkire, M.T., Tang, C., Kea- Elwood, R.W., Wood, K.E., Gallagher, M.B. and Dick, J.T.A.
tor, D., Wu, J. and McGaugh, J.L. (1996) Amygdala activity (1998) Probing motivational state during agonistic encoun-
at encoding correlated with long-term, free recall of emotional ters in animals. Nature, 39317: 66–68.
information. Proc. Natl. Acad. Sci. USA, 93: 8016–8021. Everitt, B.J., Cador, M. and Robbins, T.W. (1989) Interactions
Cahill, L. and McGaugh, J.L. (1998) Mechanisms of emotional between the amygdala and ventral striatum in stimulus–re-
arousal and lasting declarative memory. Trends Neurosci., ward associations: studies using a second-order schedule of
21: 294–299. sexual reinforcement. Neuroscience, 30: 63–75.
Campbell, B.A., Wood, G. and McBride, T. (1997) Origins of Everitt, B.J., Cardinal, R.N., Hall, J., Parkinson, J.A. and
orienting and defense responses: an evolutionary perspective. Robbins, T.W. (2000) Differential involvement of amygdala
In: Lang, P.J., Simmons, R.F. and Balaban, M.T. (Eds.), At- subsystems in appetitive conditioning and drug addiction. In:
tention and Orienting: Sensory and Motivational Processes. Aggleton, J.P. (Ed.) The Amygdala, Vol. 2. Oxford Univer-
Lawrence Erlbaum Associates, Hillsdale, NJ, pp. 41–67. sity Press, Oxford, UK, pp. 353–390.
Campeau, S. and Davis, M. (1995a) Involvement of subcortical Everitt, B.J., Morris, K.A., O’Brien, A. and Robbins,
and cortical afferents to the lateral nucleus of the amygdala T.W. (1991) The basolateral amygdala-ventral striatal
in fear conditioning measured with fear-potentiated startle in system and conditioned place preference: further evidence of
rats trained concurrently with auditory and visual condi- limbic-striatal interactions underlying reward-related proc-
tioned stimuli. J. Neurosci., 15: 2312–2327. esses. Neuroscience, 42(1): 1–18.
Campeau, S. and Davis, M. (1995b) Involvement of the central Everitt, B.J. and Robbins, T.V. (1992) Amygdala-ventral stria-
nucleus and basolateral complex of the amygdala in fear tal interactions and reward related processes. In: Aggleton,
conditioning measured with fear-potentiated startle in rats J.P. (Ed.), The Amygdala: Neurobiological Aspects of Emo-
trained concurrently with auditory and visual conditioned tion, Memory and Mental Dysfunction. Wiley-Liss, New
stimuli. J. Neurosci., 15: 2301–2311. York, pp. 401–429.
Chapman, W.P., Schroeder, H.R., Guyer, G., Brazier, M.A.B., Fanselow, M.S. (1991) The midbrain periaqueductal gray as
Fager, C., Poppen, J.L., Solomon, H.C. and Yakolev, P.I. a coordinator of action in response to fear and anxiety. In:
(1954) Physiological evidence concerning the importance of Depaulis, A. and Bandler, R. (Eds.), The Midbrain Per-
the amygdaloid nuclear region in the integration of circulat- iaqueductal Gray Matter: Functional, Anatomical and Neuro-
ing function and emotion in man. Science, 129: chemical Organization. Plenum Publishing Co., New York,
949–950. pp. 151–173.
26

Fanselow, M.S. (1994) Neural organization of the defensive (Eds.), Attention and Information Processing in Infants and
behavior system responsible for fear. Psychon. Bull. Rev., 1: Adults. Lawrence Erlbaum Associates, Hillsdale, NJ, pp. 3–29.
429–438. Graham, F.K. and Clifton, R.K. (1966) Heart rate change as a
Fendt, M., Koch, M. and Schnitzler, H.U. (1996) Lesions of the component of the orienting response. Psychol. Bull., 65:
central gray block conditioned fear as measured with the 305–320.
potentiated startle paradigm. Behav. Brain Res., 74: 127–134. Greenwald, M.K., Bradley, M.M., Cuthbert, B.N. and Lang,
Frankland, P.W. and Yeomans, J.S. (1995) Fear-potentiated P.J. (1998) Sensitization of the startle reflex in humans fol-
startle and electrically evoked startle mediated by synapses in lowing aversive electric shock exposure. Behav. Neurosci.,
rostrolateral midbrain. Behav. Neurosci., 109: 669–680. 112: 1069–1079.
Fredrikson, M., Wik, G., Annas, P., Ericson, K.A.J. and Stone- Greenwald, M.K., Cook, E.W.I. and Lang, P.J. (1989) Affec-
Elander, S. (1995) Functional neuroanatomy of visually elic- tive judgment and psychophysiological response: dimensional
ited simple phobic fear: additional data and theoretical anal- covariation in the evaluation of pictorial stimuli. J. Psycho-
ysis. Psychophysiology, 32: 43–48. physiol., 3: 51–64.
Fredrikson, M., Wik, G., Greitz, T., Stone-Elander, S., Ericson, Grillon, C., Ameli, R., Woods, S.W., Merikangas, K. and Da-
K.A.J. and Sedvall, G. (1993) Regional cerebral blood flow vis, M. (1991) Fear-potentiated startle in humans: effects of
during experimental phobic fear. Psychophysiology, 30: anticipatory anxiety on the acoustic blink reflex. Psycho-
126–130. physiology, 28: 588–595.
Funayama, E.S., Grillon, C.G., Davis, M. and Phelps, E.A. Grillon, C., Amelia, R., Merikangas, K., Woods, S.W. and
(2001) A double dissociation in the affective modulation of Davis, M. (1993) Measuring the time course of anticipatory
startle in humans: effects of unilateral temporal lobectomy. anxiety using the fear-potentiated startle reflex. Psychophys-
J. Cogn. Neurosci., 13: 721–729. iology, 30: 340–346.
Gallagher, M. (2000) The amygdala and associative learning. Grillon, C. and Davis, M. (1997) Fear-potentiated startle con-
In: Aggleton, J.P. (Ed.) The Amygdala, Vol. 2. Oxford Uni- ditioning in humans: effects of explicit and contextual cue
versity Press, Oxford, UK. conditioning following paired vs. unpaired training. Psycho-
Gallagher, M., Graham, P.W. and Holland, P.C. (1990) The physiology, 34: 451–458.
amygdala central nucleus and appetitive Pavlovian condi- Hamm, A.O., Cuthbert, B.N., Globisch, J. and Vaitl, D. (1997)
tioning: lesions impair one class of conditioned behavior. Fear and startle reflex: blink modulation and autonomic re-
J. Neurosci., 10: 1906–1911. sponse patterns in animal mutilation fearful subjects. Psy-
Gewirtz, J. and Davis, M. (1997) Second order fear condition- chophysiology, 34: 97–107.
ing prevented by blocking NMDA receptors in the amygdala. Hamm, A.O., Greenwald, M.K., Bradley, M.M. and Lang, P.J.
Nature, 388: 471–474. (1993) Emotional learning, hedonic changes, and the startle
Gewirtz, J.C. and Davis, M. (1998) Application of Pavlovian prove. J. Abnorm. Psychol., 102: 453–465.
higher-order conditioning to the analysis of the neural subst- Hatfield, T. and Gallagher, M. (1995) Taste-potentiated odor
rates of learning and memory. Neuropharmacology, 37: conditioning: impairment produced by infusion of an N-me-
453–460. thyl-D-aspartate antagonist into basolateral amygdala. Be-
Gloor, P., Olivier, A. and Quesney, L.F. (1981) The role of the hav. Neurosci., 109(4): 663–668.
amygdala in the expression of psychic phenomena in tempo- Hatfield, T., Graham, P.W. and Gallagher, M. (1992)
ral lobe seizures. In: Ben-Ari, Y. (Ed.), The Amygdaloid Taste-potentiated odor aversion: role of the amygdaloid
Complex. Elsevier/North-Holland, New York, pp. 489–507. basolateral complex and central nucleus. Behav. Neurosci.,
Goddard, G.V. (1964) Functions of the amygdala. Psychol. 106: 286–293.
Bull., 62: 89–109. Hatfield, T., Han, J.S., Conley, M., Gallagher, M. and Holland,
Gold, P.E. (2003) Acetylcholine modulation of neural systems P. (1996) Neurotoxic lesions of basolateral, but not central,
involved in learning and memory. Neurobiol. Learn. Mem., amygdala interfere with Pavlovian second-order conditioning
80: 194–210. and reinforcer devaluation effects. J. Neurosci., 16:
Goldstein, L.E., Rasmusson, A.M., Bunney, B.S. and Roth, 5256–5265.
R.H. (1996) Role of the amygdala in the coordination of Hawk, L.W. and Cook, E.W. (1997) Affective modulation of
behavioral, neuroendocrine, and prefrontal cortical mono- tactile startle. Psychophysiology, 34: 23–31.
amine responses to psychological stress in the rat. J. Neuro- Hebb, D.O. (1949) The Organization of Behavior. Wiley and
sci., 16(15): 4787–4798. Sons, New York.
Graham, F.K. (1979) Distinguishing among orienting, defense, Hitchcock, J.M. and Davis, M. (1986) Lesions of the amygdala,
and startle reflexes. In: Kimmel, H.D., van Olst, H. and but not of the cerebellum or red nucleus, block conditioned
Orelebeke, F. (Eds.), The Orienting Reflex in Humans. An fear as measured with the potentiated startle paradigm. Be-
International Conference Sponsored by the Scientific Affairs hav. Neurosci., 100: 11–22.
Division of the North Atlantic Treaty Organization. Law- Hitchcock, J.M. and Davis, M. (1991) The efferent pathway of
rence Erlbaum Associates, Hillsdale, NJ, pp. 137–167. the amygdala involved in conditioned fear as measured with
Graham, F.K. (1992) Attention: the heartbeat, the blink, and the fear-potentiated startle paradigm. Behav. Neurosci., 105:
the brain. In: Campbell, B.A., Hayne, H. and Richardson, R. 826–842.
27

Holland, P.C. and Gallagher, M. (1993a) The effects of am- Lane, R.D., Reiman, E.M., Bradley, M.M., Lang, P.J., Ahern,
ygdala central nucleus lesions on blocking and unblocking. G.L., Davidson, R.J. and Schwartz, G.E. (1997) Neuroana-
Behav. Neurosci., 107: 235–245. tomical correlates of pleasant and unpleasant emotion. Neuro-
Holland, P.C. and Gallagher, M. (1993b) Amygdala central psychologia, 35: 1437–1444.
nucleus lesions disrupt increments, but not decrement, in Lang, P.J. (1979) A bio-informational theory of emotional im-
conditioned stimulus processing. Behav. Neurosci., 107: agery. Psychophysiology, 16: 495–512.
246–253. Lang, P.J. (1985) The cognitive psychophysiology of emotion:
Irwin, W., Davidson, R.J., Lowe, M.J., Mock, B.J., Sorenson, fear and anxiety. In: Tuma, A.H. and Maser, J.D. (Eds.)
J.A. and Turski, P.A. (1996) Human amygdala activation Anxiety and the Anxiety Disorders, Vol. 3. Erlbaum, Hills-
detected with echo-planar functional magnetic resonance im- dale, NJ, pp. 131–170 No. 2, 3–62.
aging. NeuroReport, 7: 1765–1769. Lang, P.J. (1994) The motivational organization of emotion:
Iwai, E. and Yukie, M. (1987) Amygdalofugal and amygdale- affect-reflex connections. In: VanGoozen, S., Van De Poll,
petal connections with modality-specific visual cortical areas N.E. and Sargeant, J.A. (Eds.), Emotions: Essays on Emo-
in Macaques (Macaca fuscata, M. mulatta and M. fasci- tion Theory. Erlbaum, Hillsdale, NJ, pp. 61–93.
cularis). J. Comp. Neurol., 261: 362–387. Lang, P.J. (1995) The emotion probe. Am. Psychol., 50:
Kapp, B.S., Supple, W.F. and Whalen, P.J. (1994) Effects of 372–385.
electrical stimulation of the amygdaloid central nucleus on Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1990) Emotion,
neocortical arousal in the rabbit. Behav. Neurosci., 108: attention, and the startle reflex. Psychol. Rev., 97: 377–395.
81–93. Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1997) Moti-
Kapp, B.S., Whalen, P.J., Supple, W.F. and Pascoe, J.P. (1992) vated attention: affect, activation and action. In: Lang, P.J.,
Amygdaloid contributions to conditioned arousal and sensory Simons, R.F. and Balaban, M.F. (Eds.), Attention and Ori-
information processing. In: Aggleton, J.P. (Ed.), The Am- enting: Sensory and Motivational Processes. Lawrence Er-
ygdala: Neurobiological Aspects of Emotion, Memory, and lbaum Associates, Inc., NJ, pp. 97–135.
Mental Dysfunction. Wiley-Liss, New York, pp. 229–254. Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1999) Interna-
Kapp, B.S., Wilson, A., Pascoe, J.P., Supple, W.F. and Whalen, tional Affective Picture System (IAPS). The Center for Re-
P.J. (1990) A neuroanatomical systems analysis of search in Psychophysiology. University of Florida,
conditioned bradycardia in the rabbit. In: Gabriel, M. and Gainesville, FL.
Moore, J. (Eds.), Neurocomputation and Learning: Foun- Lang, P.J., Bradley, M.M., Fitzsimmons, J.R., Cuthbert, B.N.,
dations of Adaptive Networks. Bradford Books, New York, Scott, J.D., Moulder, B. and Nangia, V. (1998) Emotional
pp. 55–90. arousal and activation of the visual cortex: an fMRI analysis.
Keil, A., Bradley, M.M., Hauk, O., Rochstroh, B., Elbert, T. Psychophysiology, 35: 1–13.
and Lang, P.J. (2002) Large-scale neural correlates of affec- Lang, P.J., Greenwald, M.K., Bradley, M.M. and Hamm, A.O.
tive picture-processing. Psychophysiology, 39: 641–649. (1993) Looking at pictures: affective, facial, visceral and be-
Killcross, S., Robbins, T.W. and Everitt, B.J. (1997) Different havioral reactions. Psychophysiology, 30: 261–273.
types of fear-conditioned behavior mediated by separate nu- LeDoux, J.E., Iwata, J., Cicchetti, P. and Reis, D.J. (1988)
clei within amygdala. Nature, 388: 377–380. Different projections of the central amygdaloid nucleus me-
Kim, M., Campeau, S., Falls, W.A. and Davis, M. (1993) In- diate autonomic and behavioral correlates of conditioned
fusion of the non-NMDA receptor antagonist CNQX into fear. J. Neurosci., 8: 2517–2529.
the amygdala blocks the expression of fear-potentiated star- Lee, G.P., Bechara, A., Adolphs, R., Arena, J., Meador, K.J.,
tle. Behav. Neural Biol., 59: 5–8. Loring, D.W. and Smith, J.R. (1998) Clinical and physio-
Kintsch, W. (1974) The Representation of Meaning in Memory. logical effects of stereotaxic bilateral amygdalotomy for in-
Erlbaum, Hillsdale, NJ. tractable aggression. J. Neuropsychiatry Clin. Neurosci., 10:
Koch, M., Schmid, A. and Schnitzler, H.U. (1996) Pleasure- 413–420.
attenuation of startle is disrupted by lesions of the nucleus Li, L. and Yeomans, J.S. (1999) Summation between acoustic
accumbens. NeuroReport, 7: 1442–1446. and trigeminal stimuli evoking startle. Neuroscience, 90:
Konorski, J. (1967) Integrative Activity of the Brain: An Inter- 139–152.
disciplinary Approach. The University of Chicago Press, Lipp, O.V., Sheridan, J. and Siddle, D.A. (1994) Human
Chicago. blink startle during aversive and nonaversive Pavlovian con-
Krase, W., Koch, M. and Schnitzler, H.U. (1994) Substance P ditioning. J. Exp. Psychol. Anim. Behav. Process., 20:
is involved in the sensitization of the acoustic startle response 380–389.
by footshock in rats. Behav. Brain Res., 63: 81–88. Masur, J.D., Dienst, F.T. and O’Neal, E.C. (1974) The acqui-
LaBar, K.S., Gatenby, J.C., Gore, J.C., LeDoux, J.E. and sition of a Pavlovian conditioned response in septally dam-
Phelps, E.A. (1998) Human amygdala activation during con- aged rabbits: role of a competing response. Physiol. Psychol.,
ditioned fear acquisition and extinction: a mixed-trial fMRI 2: 133–136.
study. Neuron, 20: 937–945. McCall, R.B. and Aghajanian, G.K. (1979) Serotonergic facil-
Landis, C. and Hunt, W. (1939) The Startle Paradigm. Farrar itation of facial motoneuron excitation. Brain Res., 169:
and Rinehart, New York. 11–27.
28

McDonald, A.J. (1998) Cortical pathways to the mammalian Parkinson, J.A., Robbins, T.W. and Everitt, B.J. (2000a) Dis-
amygdala. Prog. Neurobiol., 55: 257–332. sociable roles of the central and basolateral amygdala in
McDonald, J. (1991) Topographic organization of amygdaloid appetitive emotional learning. Eur. J. Neurosci., 12: 405–413.
projections to the caudatoputamen, nucleus accumbens, and Parkinson, J.A., Willoughby, P.J., Robbins, T.W. and Everitt,
related striateal-like areas of the rat brain. Neuroscience, B.J. (2000b) Disconnection of the anterior cingulater cortex
44(1): 15–33. and nucleus accumbens core impairs Pavlovian approach
McGaugh, J.L. (2004) The amygdala modulates the consolida- behavior: further evidence for limbic cortico-ventral stria-
tion of memories of emotionally arousing experiences. Annu. topallidal systems. Behav. Neurosci., 114: 42–63.
Rev. Neurosci., 27: 1–28. Pascoe, J.P. and Kapp, B.S. (1985) Electrophysiological char-
McGaugh, J.L., Introini-Collison, I.B., Cahill, L., Castellano, acteristics of amygdaloid central nucleus neurons during
C., Dalmaz, C., Parent, M.B. and Williams, C.L. (1993) Ne- Pavlovian fear conditioning in the rabbit. Behav. Brain Res.,
uromodulatory systems and memory storage: role of the am- 16: 117–133.
ygdala. Behav. Brain Res., 58: 81–90. Pavlov, I.P. (1927) Conditioned Reflexes. Oxford University
McGaugh, J.L., Introini-Collison, I.B., Cahill, L., Kim, M. and Press, Oxford UK.
Liang, K.C. (1992) Involvement of the amygdala in neuromod- Phelps, E.A., O’Connor, K.J., Cunningham, W.A., Funayama,
ulatory influences on memory storage. In: Aggleton, J.P. (Ed.), E.S., Gatenby, J.C., Gore, J.C. and Banaji, M.R. (2000)
The Amygdala: Neurobiological Aspects of Emotion, Performance on indirect measures of race evaluation predicts
Memory, and Mental Dysfunction. Wiley-Liss, New York, amygdala activation. J. Cogn. Neurosci., 12: 729–738.
pp. 431–451. Phelps, E.A., O’Connor, K.J., Gatenby, J.C., Gore, J.C., Grill-
Mehrabian, A. and Russell, J.A. (1974) An Approach to En- on, C. and Davis, M. (2001) Activation of the left amygdala to
vironmental Psychology. MIT Press, Cambridge, MA. a cognitive representation of fear. Nat. Neurosci., 4: 437–441.
Meloni, E.G. and Davis, M. (1999) Muscimol in the deep layers Power, A.E., Vazdarjanova, A. and McGaugh, J.L. (2003)
of the superior colliculus/mesencephalic reticular formation Muscarinic cholinergic influences in memory consolidation.
blocks expression but not acquisition of fear-potentiated Neurobiol. Learn. Mem., 80: 178–193.
startle in rats. Behav. Neurosci., 113: 1152–1160. Quirk, G.J., Repa, J.C. and LeDoux, J.E. (1995) Fear condi-
Miltner, W., Matjak, M., Braun, C. and Diekmann, H. (1994) tioning enhances short-latency auditory responses of lateral
Emotional qualities of odors and their influence on the startle amygdala neurons: parallel recordings in the freely behaving
reflex in humans. Psychophysiology, 31: 107–110. rat. Neuron, 15: 1029–1039.
Mineka, S., Davidson, M., Cook, M. and Keir, R. (1984) Ob- Redmond Jr., D.E. (1977) Alteration in the function of the
servational conditioning of snake fear in rhesus monkeys. J. nucleus locus: a possible model for studies on anxiety. In:
Abnorm. Psychol., 93: 355–372. Hanin, I.E. and Usdin, E. (Eds.), Animal Models in Psychi-
Morgenson, G.M. (1987) Limbic-motor intergration. In: Ep- atry and Neurology. Pergamon Press, Oxford, UK, pp.
stein, A. and Morrison, A.R. (Eds.), Progress in Psychobi- 292–304.
ology and Physiological Psychology. Academic Press, New Reiman, E.M., Lane, R.D., Ahern, G.L., Schwartz, G.E., Dav-
York, pp. 117–170. idson, R.J., Friston, K.J., Yun, L. and Chen, K. (1997) Ne-
Morris, J.S., Frith, C.D., Perrett, D.I. and Rowland, D. (1996) uroanatomical correlates of externally and internally
A differential neural response in the human amygdala generated human emotion. Am. J. Psychiatry, 54: 918–925.
to fearful and happy facial expression. Nature, 383: Robledo, P., Robbins, T.W. and Everitt, B.J. (1996) Effects of
812–815. excitotoxic lesions of the central amygdaloid nucleus on the
Öhman, A. and Mineka, S. (2001) Fears, phobias, and prepar- potentiation of reward-related stimuli by intra-accumbens
edness: toward an evolved module of fear and fear learning. amphetamine. Behav. Neurosci., 110: 981–990.
Psychol. Rev., 108: 483–522. Rosen, J.B., Hitchcock, J.M., Miserendino, M.J.D., Falls,
Ortony, A., Clore, G.L. and Collins, A. (1988) The Cognitive W.A., Campeau, S. and Davis, M. (1992) Lesions of the
Structure of Emotions. Cambridge Press, Cambridge. perirhinal cortex but not of the frontal, medial prefrontal,
Osgood, C., Suci, G. and Tannenbaum, P. (1957) The Meas- visual, or insular cortex block fear-potentiated startle using a
urement of Meaning. University of Illinois, Urbana, IL. visual conditioned stimulus. J. Neurosci., 12: 4624–4633.
Packard, M.G., Cahill, L. and McGaugh, J.L. (1994) Amygdala Rosen, J.B., Hitchcock, J.M., Sananes, C.B., Miserendino,
modulation of hippocampal-dependent and caudate nucleus- M.J.D. and Davis, M. (1991) A direct projection from the
dependent memory processes. Proc. Natl. Acad. Sci. USA, central nucleus of the amygdala to the acoustic startle path-
91: 8477–8481. way: anterograde and retrograde tracing studies. Behav. Ne-
Packard, M.G. and Teather, L.A. (1998) Amygdala modulation urosci., 105: 817–825.
of multiple memory systems: hippocampus and caudate–put- Sabatinelli, D., Bradley, M.M., Cuthbert, B.N. and Lang, P.J.
amen. Neurobiol. Learn. Mem., 69: 163–203. (2001) Affective startle modulation in anticipation and perc-
Parkinson, J.A., Dally, J.W., Bamford, A., Fehrent, B., Rob- petion. Psychophysiology, 38: 719–722.
bins, T.W. and Everitt, B.J. (1998) Effects of 6-OHDA le- Sabatinelli, D., Flaisch, T., Bradley, M.M., Fitzsimmons, J.R.
sions of the rat nucleus accumbens on appetitive Pavlovian and Lang, P.J. (2004) Affective picture perception: gender
conditioning. J. Psychopharmacol., 12: A8. differences in visual cortex. Neuroreport, 15: 1109–1112.
29

Sabatinelli, D., Bradley, M.M., Fitzsimmons, J.R. and Lang, Taylor, J.R. and Robbins, T.W. (1984) Enhanced behavioral
P.J. (2005) Parallel amygdala and inferotemporal activation control by conditioned reinforcers following microinjections
reflect emotional intensity and fear relevance. NeuroImage, of D-amphetamine into the nucleus accumbens. Psychophar-
24: 1265–1270. macology, 84: 405–412.
Sabatinelli, D., Lang, P.J., Keil, A. and Bradley, M.M. (in Taylor, J.R. and Robbins, T.W. (1986) 6-hydroxydopamine le-
press). Emotional perception: Correlation of functional MRI sions of the nucleus accumbens, but not the caudate nucleus,
and event-related potentials. Cerebral cortex. attenuate enhanced responding with reward-related stimuli
Sakanaka, M., Shibasaki, T. and Lederis, K. (1986) Distribu- produced by intra-accumbens D-amphetamine. Psychphar-
tion and efferent projections of corticotropin-releasing fac- macology, 90: 390–397.
tor-like immunoreactivity in the rat amygdaloid complex. Tellegen, A. (1985) Structures of mood and personality and their
Brain Res., 382: 213–238. relevance to assessing anxiety, with an emphasis on self-report.
Sananes, C.B. and Davis, M. (1992) N-Methyl-D-Aspartate le- In: Tuma, A.H. and Maser, J.D. (Eds.), Anxiety and the Anx-
sions of the lateral and basolateral nuclei of the amygdala iety Disorders. Lawrence Erlbaum, Hillsdale, NJ, pp. 681–706.
block fear-potentiated startle and shock sensitization of star- Timberlake, W. (1993) Behavior systems and reinforcement: an
tle. Behav. Neurosci., 106: 72–80. intergrative approach. J.Exp. Anal. Behav., 60: 105–128.
Sanghera, M.K., Rolls, E.T. and Roper-Hall, A. (1979) Visual Timberlake, W. and Lucas, G.A. (1989) Behavior systems and
responses of neurons in the dorsolateral amygdala of the alert learning: from misbehavior to general principles. In: Klein,
monkey. Exp. Neurol., 63: 610–626. S.B. and Mowrer, R.R. (Eds.), Instrumental Conditioning
Schlosberg, J. (1952) The description of facial expression in Theory and the Impact of Biological Constraints on Learn-
terms of two dimensions. J. Exp. Psychol., 44: 229–237. ing. Erlbaum, Hillsdale, NJ.
Schmid, A., Koch, M. and Schnitzler, H.U. (1995) Conditioned Ursin, H. and Kaada, B.R. (1960) Functional localization
pleasure attenuates the startle response in rats. Neurobiol. within the amygdaloid complex in the cat. Electroencepha-
Learn. Mem., 64: 1–3. logr. Clin. Neurophysiol., 12: 109–122.
Schneiderman, N. (1972) Response system divergencies in Vrana, S.R., Spence, E.L. and Lang, P.J. (1988) The startle
aversive classical conditioning. In: Black, A.H. and Prokasy, probe response: a new measure of emotion? J. Abnorm. Psy-
W.F. (Eds.) Classical Conditioning, Vol. III: Current Re- chol., 97: 487–491.
search and Theory. Appleton-Century-Crofts, New York. Walker, D.L. and Davis, M. (1997) Double dissociation between
Schneirla, T. (1959) An evolutionary and developmental theory the involvement of the bed nucleus of the stria terminalis and
of biphasic processes underlying approach and withdrawal. the central nucleus of the amygdala in light-enhanced versus
In: Jones, M. (Ed.), Nebraska Symposium on Motivation. fear-potentiated startle. J. Neurosci., 17: 9375–9383.
University of Nebraska Press, Lincoln, pp. 1–42. Walker, D.L., Toufexis, D.J. and Davis, M. (2003) Role of the
Schoenbaum, G., Chiba, A.A. and Gallagher, M. (1998) Or- bed nucleus of the stria terminalis versus the amygdala in
bitofrontal cortex and basolateral amygdala encode expected fear, stress, and anxiety. Eur. J. Pharmacol., 463: 199–216.
outcomes during learning. Nat. Neurosci., 1: 155–159. Weisz, D.J. and McInerney, J. (1990) An associative process
Schoenbaum, G., Chiba, A.A. and Gallagher, M. (1999) Neural maintains reflex facilitation of the unconditioned nictitating
encoding in orbitofrontal cortex and basolateral amygdala membrane response during the early stages of training. Be-
during olfactory discrimination learning. J. Neurosci., 19: hav. Neurosci., 104: 21–27.
1876–1884. Whalen, P.J. (1998) Fear, vigilance, and ambiguity: initial ne-
Schupp, H.T., Cuthbert, B.N., Bradley, M.M., Hillman, C.H., uroimaging studies of the human amygdala. Curr. Directions
Hamm, A.O. and Lang, P.J. (2004) Brain processes in emo- Psychol. Sci., 7: 177–188.
tional perception: motivated attention. Cogn. Emotion, Whalen, P.J. and Kapp, B.S. (1991) Contributions of the am-
18(5): 593–611. ygdaloid central nucleus to the modulation of the nictitating
Schwaber, J.S., Kapp, B.S., Higgins, G.A. and Rapp, P.R. membrane reflex in the rabbit. Behav. Neurosci., 105: 141–153.
(1982) Amygdaloid basal forebrain direct connections with White, S.R. and Neuman, R.S. (1980) Facilitation of spinal
the nucleus of the solitary tract and the dorsal motor nucleus. motoneuron excitability by 5-hydroxytryptamine and nor-
J. Neurosci., 2: 1424–1438. adrenaline. Brain Res., 185: 1–9.
Seligman, M.E.P. (1970) On the generality of the laws of learn- Wundt, W. (1896) Gundriss der Psychologie. Entgelman, Leip-
ing. Psychol. Rev., 77: 406–418. zig, Germany.
Shaver, P., Schwartz, J., Kirson, D. and O’Connor, C. (1987) Yeomans, J.S., Steidle, S. and Li, L. (2000) Conditioned brain-
Emotion knowledge: further exploration of a prototype ap- stimulation reward attenuates the acoustic startle reflex in
proach. J. Pers. Soc. Psychol., 52: 1061–1086. rats. Soc. Neurosci. Abstr., 30: 1252.
Shi, C. and Davis, M. (1996) Anatomical tracing and lesion stud- Zhao, Z. and Davis, M. (2004) Fear-potentiated startle in rats
ies of visual pathways involved in fear conditioning measured is mediated by neurons in the deep layers of the superior
with fear potentiated startle. Soc. Neurosci. Abstr., 22: 1115. colliculus/deep mesencephalic nucleus of the rostral midbrain
Sokolov, E.N. (1963) Perception and the Conditioned Reflex. through the glutamate non-NMDA receptors. J. Neurosci.,
Pergamon Press, Oxford, UK. 24: 10326–10334.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 2

Emotion and attention: event-related brain potential


studies

Harald T. Schupp1,, Tobias Flaisch1, Jessica Stockburger1 and Markus Junghöfer2

1
Department of Psychology, University of Konstanz, Konstanz, Germany
2
Institute for Biomagnetism and Biosignalanalysis, Münster University Hospital, Münster, Germany

Abstract: Emotional pictures guide selective visual attention. A series of event-related brain potential
(ERP) studies is reviewed demonstrating the consistent and robust modulation of specific ERP components
by emotional images. Specifically, pictures depicting natural pleasant and unpleasant scenes are associated
with an increased early posterior negativity, late positive potential, and sustained positive slow wave
compared with neutral contents. These modulations are considered to index different stages of stimulus
processing including perceptual encoding, stimulus representation in working memory, and elaborate
stimulus evaluation. Furthermore, the review includes a discussion of studies exploring the interaction of
motivated attention with passive and active forms of attentional control. Recent research is reviewed
exploring the selective processing of emotional cues as a function of stimulus novelty, emotional prime
pictures, learned stimulus significance, and in the context of explicit attention tasks. It is concluded that
ERP measures are useful to assess the emotion–attention interface at the level of distinct processing stages.
Results are discussed within the context of two-stage models of stimulus perception brought out by studies
of attention, orienting, and learning.

Keywords: emotion; attention; EEG; ERP; EPN; LPP

Humans live in an environment presenting a seem- 2000b). In passive attention, the power to capture
ingly endless stream of stimuli. Only a subset of the attention derives from simple qualities of the
available information is consciously recognized, stimulus such as intensity, suddenness of onset, or
becomes the focus of sustained attention, and is novelty. In active attention, priority processing
subjected to controlled and elaborated processing. reflects the intentional effort to look for selected
Mechanisms of selective attention assure the prior- stimuli based on instructions, self-generated inten-
itized processing of some objects, events, or loca- tions, or associative learning. In addition, certain
tions and multiple avenues to command attention kinds of stimuli trigger selective attention due to
are indicated by distinguishing among active and their biological meaning. Organisms respond to
passive forms of attentional control1 (Öhman et al., environmental cues according to their emoti-
onal/motivational significance (Lang et al., 1997;
Öhman et al., 2000b). The attention capture of
Corresponding author. Tel.: +49-7531-882504; Fax: +49-
emotionally relevant stimuli has been dubbed
7531-882971; E-mail: Harald.Schupp@uni-konstanz.de ‘motivated attention’ referring to a natural state
1
Similar to the distinction between active and passive atten-
tion, other classification schemata distinguish between auto-
of selective attention, ‘‘y similar to that occurring
matic vs. controlled, bottom-up vs. top-down, reflexive vs. in an animal as it forages in a field, encounters
instructed, implicit vs. explicit attention. others, pursues prey or sexual partners, and tries

DOI: 10.1016/S0079-6123(06)56002-9 31
32

or void predators and comparable dangers’’ (Lang system determines foraging, ingestion, copulation,
et al., 1997, p. 97). An evolutionary perspective and nurture of progeny, and is accompanied
suggests that this form of attentional control by affectively pleasant states. The protective
is highly adaptive for survival giving primacy — defensive system coordinates withdrawal from
in terms of attentional selection and response — to and defense against nociceptive agents, and is
appetitively and aversively relevant events (Lang associated with the experience of unpleasant
et al., 1997). affects. The view that emotion is in part organ-
One aim of the present review is to present ized by underlying motivational factors is sup-
theory and data regarding the emotional guidance ported by research utilizing verbal reports, which
of selective attention in distinct processing stages. consistently demonstrates the primacy of the
The review will discuss the effects of motivated valence dimension. Furthermore, both motiva-
attention from the perspective of the biphasic tional subsystems can vary in terms of engage-
approach to emotion. Furthermore, we predomi- ment or activation reflecting the arousal level
nantly consider empirical evidence derived by which is reliably observed as second dimension in
event-related brain potential studies (ERPs).2 studies of natural language and verbal reports
ERP measures have the unique advantage of (Lang et al., 1990, 1997).
high temporal resolution, and it will be concluded Accordingly, emotions can be functionally con-
that they provide a useful tool to study the sidered as action dispositions preparing the organ-
emotional guidance of attention at the level of ism for either avoidance or approach-related
distinct processing stages. A second aim of the actions, interrupting ongoing behavior and mental
present review is to extend this perspective to processes (Frijda, 1986; Lang et al., 1997). Con-
recent approaches examining the emotion–atten- sistent with this view, a large number of studies
tion interface. Specifically, recent research is con- demonstrated reliable modulations in autonomic,
sidered examining effects of motivated attention in somatic, and reflex response measures while par-
relation to active and passive attention and learn- ticipants viewed pleasant, neutral, and unpleasant
ing. In the final section, the ERP indices of selective pictures (Bradley, 2000; Hamm et al., 2003).
attentional orienting to emotional cues are dis- Presumably, emotional modulations are not lim-
cussed from the perspective of two-stage models of ited to the behavioral output channels. They should
stimulus perception advanced in research of atten- also be evident during the preceding perceptual and
tion, orienting, and learning. evaluative processing of emotional cues. Specifi-
cally, efficient preparation and organization of
A biphasic view of emotion appropriate behavioral responses require a rapid
extraction of critical information from the environ-
Although emotional expressions, bodily changes, ment. In this respect, emotional cues direct
and reported feelings vary idiosyncratically ac- attentional resources (Öhman et al., 2000a). Fur-
cording to dispositional and situational factors, thermore, motivated attention reflects our evolu-
many theorists claim that the emotional or affect tionary heritage and is therefore most apparent for
system retains a much simpler biphasic organiza- stimuli with high evolutionary significance, that is,
tion of two distinct motivational subsystems prototypical stimuli related to threat and survival
(Schneirla, 1959; Konorski, 1967; Dickinson strongly engaging basic motivational systems
and Dearing, 1979; Lang et al., 1990, 1997; Lang, (Bradley et al., 2001).
1995; Cacioppo et al., 1999; Dickinson and
Balleine, 2002). The self-preservative appetitive Emotion and attention: event-related brain
2
potentials
Discussion of the interaction of emotion and attention based
on behavioral, autonomic, reflex, somatic, and neuroimaging
measures are provided in recent reviews (Lang et al., 1997;
ERP measures provide a unique window into the
Mogg and Bradley, 1999; Öhman et al., 2000a; Pessoa et al., brain’s processing of emotional cues assisting in
2002a; Compton, 2003; Vuilleumier, 2005). detailing information processing at the level of
33

distinct stages. Already, the operation of selective (L2-MNE; Hamalainen and Ilmoniemi, 1994;
processing triggered by reflexive or explicit atten- Hauk et al., 2002).
tion has been greatly facilitated by the use of ERP In the experiments summarized here, pleas-
measures. Specifically, an array of experimental ant, neutral, and unpleasant pictures from the In-
tasks including the attentional blink, refractory ternational Affective Picture System (IAPS; Lang
period, visual search, and spatial cuing paradigms et al., 2005) were presented in a passive task con-
revealed modulations of specific ERP components text in which subjects were instructed to simply
thought to reflect perceptual encoding, working view the pictures. Some of the reviewed studies
memory, and motor preparation processes (Luck used a 6 s picture presentation time and extended
and Hillyard, 2000; Luck et al., 2000). Similar to intertrial intervals, enabling the simultaneous as-
the domains of explicit and reflexive attention, it is sessment of autonomic and somatic measures of
expected that ERPs provide valuable information affect processing (Cuthbert et al., 2000; Schupp
regarding the emotional guidance of selective at- et al., 2004a). Other studies utilized the timing
tention. Three issues are of particular interest: (1) parameters of the modified oddball paradigm de-
Identifying distinct processing stages at which veloped by Cacioppo and colleagues (Cacioppo
emotional cues are selectively processed. (2) Deter- et al., 1993) with shorter presentation times (1.5 s)
mining whether the differential processing of emo- and faster presentation rates (Schupp et al., 2000a,
tional cues already affects obligatory processing 2003a, 2004b). In addition, more recent studies
stages. (3) Comparing the ERP signature of mo- using the rapid serial visual presentation (RSVP)
tivated and explicit attentional control. paradigm are described, in which the pictures were
ERPs provide a voltage measurement of neural presented briefly (333 ms) and as a continuous
activity that can be recorded noninvasively from stream (Junghöfer et al., 2001; Schupp et al.,
multiple scalp regions (Birbaumer et al., 1990). 2003b, 2006, submitted; Wieser et al., 2006;). Fur-
More specific, ERPs are considered to reflect thermore, while earlier studies used sparse sensor
summed postsynaptic potentials generated in the sampling, more recent research provided an im-
process of neural transmission and passively proved spatial sampling by using 128 or 256 sensor
conducted through the brain and skull to the skin recordings. Due to limitations in head coverage
surface where they contribute to the electroen- and the placements of sensors, reliable differenti-
cephalogram (EEG). Since ERPs are usually hid- ation of emotional and neutral contents within the
den in the larger background EEG activity, it is first 300 ms was brought out in comparatively re-
necessary to use multiple stimulus presentations cent research (Junghöfer et al., 2001; Schupp et al.,
and the calculation of stimulus-locked signal aver- 2003b). Finally, earlier research exploring emotion
aging to extract the ERP signal from the back- processing is acknowledged, complementing the
ground EEG activity. Biophysical considerations data reported in this review (Johnston et al., 1986;
suggest that large-amplitude ERP components re- Diedrich et al., 1997; Johnston and Oliver-Rod-
flect widespread, synchronous sources in cortical riguez, 1997; Palomba et al., 1997). Across studies,
regions (Lutzenberger et al., 1987). Brain activity the differential processing of emotional compared
locked to the processing of a stimulus becomes ap- to neutral pictures was consistently and reliably
parent as positive and negative deflections in the reflected by three ERP components, which are
ERP waveform. The amplitude and latency of spe- summarized in the following according to their
cific ERP components provide information regard- temporal occurrence in the processing stream.
ing the strength and time course of underlying
neural processes. Furthermore, given appropriate
spatial sampling (Tucker, 1993), the topography of Early posterior negativity
ERP components can be used to estimate the neural
generator sites by advanced analytic tools such as The first ERP component reflecting the differential
Current-Source-Density (CSD; Perrin, Bertrand processing of emotional compared to neutral stimuli
and Pernier, 1987) or L2-Minimum-Norm-Estimate is the early posterior negativity (EPN). The temporal
34

and spatial appearance of the early indicator of se- as a function of emotional arousal. Consistent with
lective processing of emotional pictures is illustrated this notion, the EPN co-varied with the arousal level
in Fig. 1 presenting data from research utilizing the of the emotional pictures. Specifically, highly arous-
RSVP paradigm (Junghöfer et al., 2001). A pro- ing picture contents of erotic scenes and mutilations
nounced ERP difference for the processing of elicited a more pronounced posterior negativity com-
emotionally arousing and neutral pictures devel- pared to less arousing categories of the same valence
oped around 150 ms which was maximally pro- (Junghöfer et al., 2001; Schupp et al., 2004b).
nounced around 250–300 ms. This differential Given that these studies used natural scenes
ERP appeared as negative deflection over temporo- varying widely in terms of physical characteristics,
occipital sensor sites and a corresponding polarity it is important to explore to what extent the EPN
reversal over fronto-central regions. Despite differ- effect is related to systematic differences in phys-
ences in the overall topography, research presenting ical characteristics of the stimulus materials. To
pictures discretely and with longer presentation times examine the effects of color of the stimulus mate-
(1.5 s and 1.5 s intertrial interval) also demonstrated a rials, Junghöfer and colleagues (2001) included a
more pronounced negative potential for emotional control condition presenting the same materials as
pictures in the same latency range over temporo- grayscaled images. An almost identical affect mod-
occipital sensors (Schupp et al., 2003a, 2004c). Ad- ulated early posterior negativity was observed as
ditional analyses served to test the prediction derived for the corresponding color images showing that
from the biphasic view of emotion that the differen- the early discrimination of emotional from neutral
tial processing of pleasant and unpleasant cues varies pictures did not depend on color. Additional

Fig. 1. Time course and topography of the early posterior negativity. Left side: Upper panel: Grand-averaged ERP waveforms at a
right occipital sensor while viewing high and low arousing (left side), and high, moderate, and low arousing pictures (right side),
respectively. Pictures were presented at 3 Hz. Lower panel: Grand-averaged ERP waveforms at the same right occipital sensor while
viewing high and low arousing pictures at 5 Hz. The relative difference component peaking around 64 ms reflects the EPN of the
preceding emotional picture. Right side: Top (upper panel) and back (lower panel) side views of the topographical distribution of the
difference waves of (high–low) arousing pictures for both the 3 Hz (left side) as well as the 5 Hz (right side) presentation rates. Scalp
topography is shown in terms of current source density distribution estimating difference activities in bilateral occipito-temporal and
right hemispheric parietal cortex areas.
35

control analyses determined that increased EPN functional magnetic resonance imaging (fMRI)
amplitudes to emotional pictures were not second- studies indicate increased activation of the extended
ary to differences in physical characteristics such visual cortex while viewing emotional pictures in
as brightness, measures of perceptual complexity, rapid serial visual presentations (Junghöfer et al.,
or spatial frequency. However, a recent study pro- 2005, 2006). Third, paying explicit attention to spe-
vided a more systematic investigation of the EPN cific stimulus features (such as color, orientation, or
component as a function of picture content (ob- shape) and higher-order categories (e.g., animal vs.
jects vs. people), emotional arousal (low vs. high), non-animal) is also reflected by a temporo-occipital
and picture type (single foreground item vs. scat- negativity with a similar latency (Thorpe et al.,
tered scenes and multiple items; Löw et al., 2005). 1996; Codispoti et al., in press).
Each of these variables affected the EPN compo- In sum, a negative potential over temporo-oc-
nent suggesting that the EPN is not specifically cipital sensor sites, developing around 150 ms post-
reflecting emotional arousal but is presumably stimulus and most pronounced around 200–300 ms,
more generally related to perceptual variables de- reliably reflects the differential processing of emo-
termining selective attention. For instance, greater tional compared to neutral visual stimuli. These
EPN amplitudes were obtained for people com- findings have been considered from the perspective
pared to objects, and pictures with a clear fore- of ‘natural selective attention,’ proposing that per-
ground object as opposed to more complex scenes. ceptual encoding is in part directed by underlying
Interestingly, emotional effects were observed motivational systems of avoidance and approach.3
independent from picture type and content. How-
ever, in this study, high compared to low arousing
pictures elicited augmented EPN amplitudes only Late positive potential
for pictures depicting people. Accordingly, to de-
termine the effects of emotional arousal, secondary As shown in Fig. 2, subsequent to the modulation
explanations based on picture type and content during perceptual encoding, it is consistently ob-
need to be experimentally controlled by selecting served that emotional (pleasant and unpleasant)
appropriate picture materials (Schupp et al., pictures elicit increased late positive potentials
2004a, b). From a theoretical perspective, it is un- (LPPs) over centro-parietal regions, most apparent
known to what extent variables such as salient around 400–600 ms poststimulus (also referred to
perceptual features, high evolutionary significance, as P3b; Palomba et al., 1997; Cuthbert et al., 2000;
and expert knowledge contribute to the emotional Schupp et al., 2000a, 2003a, 2004a, b; Keil et al.,
EPN modulation. Future research is needed to 2002; Amrhein et al., 2004). This LPP modulation
examine whether the EPN effect is limited to se- appears sizable and can be observed in nearly
lected classes of emotional stimuli. each individual when calculating simple differ-
It might be informative to briefly outline the rea- ence scores (emotional–neutral). In addition, the
sons that led to the decision to refer to this mod- LPP is also specifically enhanced for pictures that
ulation as early posterior negativity associated with are more emotionally intense (i.e., described by
emotion processing (rather than reduced positivity). viewers as more arousing, and showing a height-
First, estimates of the neural generators by current ened skin conductance response). Extending the
source density (cf., Fig. 1; Junghöfer et al., 2001) valence by arousal interaction, a recent study ex-
and minimum norm analysis (cf., Fig. 3; Schupp amined the LPP amplitude associated with the
et al., 2006) suggested that the difference in poten- processing of specific categories of human
tial between emotional and neutral pictures has its
3
sources over occipito-temporo-parietal sites. Sec- The present analyses might be extended to discrete states of
emotion, considered to reflect a subordinate level of emotional
ond, with the notion of a stronger positive potential
organization. For instance, focusing on the fear system, we re-
for neutral compared to emotional images, some cently obtained evidence that threatening faces are associated
readers might infer stronger neural activation for with increased EPN and LPP amplitudes compared to neutral
neutral materials. In fact, our accompanying and friendly faces (Schupp et al., 2004c).
36

Fig. 2. Time course and topography of the late positive potential. (A) Illustration of the experimental paradigm used. Pictures are
shown in blocks of six stimuli for 1.5 s each, separated by variable Intertrial Intervals (ITI) (1.5 s). (B) Left side: Grand-averaged ERP
waveforms of a right parietal sensor while viewing pleasant, neutral, and unpleasant pictures. Right side: Difference scalp potential
maps (emotional–neutral) reveal the topography of the LPP modulation in the time interval of 400–600 ms poststimulus. Illustrated is a
left and right-side view of the model head.

experience. Focusing on specific pleasant and un- correspondence between instruction to attend to
pleasant picture contents, it was found that picture specified stimuli and the LPP amplitude associated
contents of high evolutionary significance such as with target detection has been reported. Moreover,
pictures of erotica and sexual contents and con- dual-task studies revealed that the LPP amplitude
tents of threat and mutilations were associated to a primary counting task was reciprocally related
with enlarged LPP amplitudes compared to picture to the LPP amplitude observed for a competing
categories of the same valence but less evolution- secondary task (Donchin et al., 1986). These find-
ary significance (Schupp et al., 2004a). ings suggest that the LPP is a sensitive measure of
In the cognitive domain, it is a hallmark finding attention manipulations indicating the operation of
that increased LPP amplitudes are associated with a capacity-limited system. Functionally, the LPP
the meaning of task-relevant stimuli rather than just amplitude has been considered to reflect the repre-
simple physical stimulus characteristics (Johnson, sentation of stimuli in working memory (Donchin
1988; Picton, 1992). In addition, a close and Coles, 1988). Consistent with this view, research
37

with the attentional blink paradigm indicates that sustained perceptual operations and memory proc-
the LPP might reflect a gateway to conscious rec- esses (Ruchkin et al., 1988; Ritter and Ruchkin,
ognition (Luck et al., 2000; Kranczioch et al., 2003). 1992). Accordingly, it was suggested that the pos-
Considering these findings from cognitive research, itive slow wave reflects sustained attention to visual
it might be proposed that the increased LPP ampli- emotional stimuli. This hypothesis is supported by
tude elicited by emotional cues reflects their intrinsic research using a secondary acoustic probe pre-
significance similar to the distinct representation in sented during picture viewing. In these experi-
working memory of task-relevant stimuli. ments, the P3 component of the probe ERP is
An alternative interpretation might consider the reliably smaller when viewing pleasant or unpleas-
LPP findings from the perspective of internal stim- ant, compared to neutral pictures (Schupp et al.,
ulus probability rather than emotional relevance. 1997, 2004a; Cuthbert et al., 1998). This result
In fact, ample research documents that rare stimuli parallels findings from experiments on instructed
are associated with increased LPP amplitudes attention. When participants are told to attend to
(Johnson, 1988). However, the LPP modulations one stimulus and ignore others (Johnson, 1988), a
discussed here appear to be unrelated to the prob- smaller P3 response to a secondary probe is con-
ability of picture categories. When subjects had to sistently found during the attended stimulus. This
spontaneously categorize the pictures on the basis result is held to reflect reduced availability of at-
of valence, the categories of pleasant, neutral and tentional resources for the probe, assuming that the
unpleasant contents were equally probable. Alter- resource pool is limited and there is high allocation
natively, assuming subjects saw the pictures as to the primary stimulus (Donchin et al., 1986).
forming only two categories (emotional vs. neu-
tral), the emotional stimuli were twice as frequent Summary
as neutral contents, as a consequence resulting in
the opposite prediction as empirically observed: Summarizing these findings, a number of obser-
neutral pictures should have elicited larger LPP vations merit reemphasis:
amplitudes compared to emotional categories. To
(1) The emotional guidance of selective atten-
resolve this issue more conclusively, the effect of
tion is reflected by three distinct ERP com-
stimulus probability and emotional context was
ponents associated with different stages of
explored in a follow-up study using the modified
stimulus processing including perceptual en-
oddball paradigm. In this study, increased LPP
coding, stimulus representation in working
amplitudes to emotional compared to neutral pic-
memory, and elaborate stimulus evaluation.
tures were observed when these contents were in-
(2) ERP modulations induced by emotional
terspersed in blocks of either neutral, pleasant, or
cues were reliable and consistently observed
unpleasant context stimuli (Schupp et al., 2000b).
early in the processing stream (o300 ms).
Thus, these data strongly suggest that emotional
(3) Indices of selective attention observed for
arousal is considerably more potent in determining
motivated and instructed attention appear
LPP amplitude than local probability.
strikingly similar.

Positive slow wave Exploring the emotion–attention interface

Research using longer picture presentation times Building upon these findings, recent research was
revealed that the LPP is followed by an extended directed towards understanding effects of motivated
positive slow wave associated with the processing attention in interaction with active and passive
of emotional cues throughout the 6 s picture-view- forms of attentional control. One goal of these
ing period (Cuthbert et al., 1995, 2000). In the studies was to determine the boundary conditions
cognitive domain, positive slow waves have been of the differential emotion processing. Thus, inter-
found to be sensitive to manipulations requiring ference of the selective emotion processing was
38

determined as a function of stimulus novelty, com- neutral pictures. Interestingly, differential emotion
petition by primary cognitive tasks, and the process- processing did not vary as a function of stimulus
ing of emotional prime pictures. Another interest repetition and was similarly expressed across
was to determine effects of cooperation, that is, blocks of picture presentation (cf., Fig. 3). One
when selective attention is devoted to emotional might assume that presenting the pictures as con-
stimuli. Furthermore, learning and experience may tinuous stream may be particularly effective in
shape attentional orienting as investigated in se- triggering attentional orienting to each stimulus, in
lected subject populations (e.g., simple phobia, sub- effect preventing differential emotion processing
stance abuse). from habituating. To pursue this possibility, a fol-
low-up study presented the pictures for 120 ms,
Stimulus novelty and emotional perception preceding a blank period of 880 ms. In this in-
stance, three pictures displaying erotic, neutral, or
In passive attention, stimuli may capture attentio- mutilation scenes were repeated 700 times (Schupp
nal resources simply because they are novel. In the et al., unpublished data). Again, the erotic and
studies described above, no extensive stimulus rep- mutilation pictures elicited an increased EPN com-
etitions were implemented. Therefore, it remains pared to the neutral scene, revealing no habitua-
to be determined to what extent the observed tion as a function of excessive stimulus repetition.
emotional attention capture depends on stimulus This issue was examined in yet another study fo-
novelty. Widely studied in the context of the ori- cusing on the habituation of differential emo-
enting response, the repeated presentation of sen- tion processing indexed by the LPP amplitude
sory stimuli usually prompts habituation, that is, (Codispoti et al., 2006). Three pictures displaying
decrement across several response systems (re- pleasant, neutral, or unpleasant scenes were re-
viewed in Öhman et al., 2000b). Recent studies peated 60 times. Furthermore, to obtain auto-
have demonstrated the importance of stimulus nomic and somatic indices of emotion processing,
novelty for differential emotion processing meas- pictures were presented at slow rates. Results re-
uring various motor output responses. Specifically, vealed augmented LPP amplitudes to pleasant and
studying habituation in processing of emotional unpleasant compared to neutral pictures, which
stimuli, skin conductance responses, heart rate, was maintained across repeated presentations of
and corrugator muscle activity habituated rather the stimuli. These results concur with findings in
quickly. In contrast, the startle reflex magnitude the cognitive domain observing pronounced LPP
continued to be augmented for unpleasant com- amplitudes in visual search tasks promoting auto-
pared to pleasant picture contents across repeated matic target detection (Hoffman et al., 1983).
stimulus presentations (Bradley et al., 1993). Thus, These studies suggest that stimulus novelty is not
habituation effects differ among measures of mo- critical to observe differential processing of emo-
tor output and response preparation, and a series tional compared with neutral stimuli reflected by
of recent studies extended the database to ERP EPN and LPP component.
measures of stimulus perception and evaluation. It has been suggested that perceptual processing
Given the attention capture by emotional cues in the cortex is regulated by subcortical structures
may reflect a rather automatic phenomenon involved in appetitive or defensive responding
(Öhman et al., 2000a), stimulus novelty might be (Derryberry and Tucker, 1991; Davis and Whalen,
less critical during perceptual processing and stim- 2001; Junghöfer et al., 2005; Sabatinelli et al.,
ulus representation in working memory. 2005). The present findings appear inconsistent
To separate emotional meaning and stimulus with this hypothesis, considering that several stud-
novelty, the processing of erotica, neutral contents, ies observed relatively rapid habituation effects in
and mutilations was examined across 90 picture the amygdala to emotional facial expressions. In
repetitions (Schupp et al., 2006). Replicating pre- addition, habituation effects were also demon-
vious results, emotional stimuli were associated strated for the cingulate cortex, the hippocampus,
with increased EPN amplitudes compared to and the dorsolateral prefrontal cortex (Breiter
39

Fig. 3. Effects of emotion and stimulus repetition.(a) Grand-averaged ERP waveforms for a selected right occipital sensor as a
function of affect and first and last block of picture viewing. (b) Difference scalp potential maps (emotional–neutral) for the first and
last block of picture presentation separately. (c) Illustration of the statistical effects observed in repeated measure ANOVAs calculated
for each sensor and mean time interval. (d) Calculation of the L2-MNE for the scalp potential difference (emotional–neutral)
separately for first and last block of picture presentation. All maps display a right-side view.

et al., 1996; Wright et al., 2001; Phan et al., 2003). determined in future studies whether the amygdala
However, a number of critical issues and findings also reveals rapid habituation to these emotionally
need to be taken into account. First, as observed in evocative stimuli. Thus, more research is needed to
a classical conditioning fMRI study, habituation evaluate the hypothesis that the preferential
may vary for the various anatomical subregions of processing of emotional images in the extended
the amygdala serving distinct functions (Morris et visual cortex is secondary to the appraisal of
al., 2001). Second, a recent study observed that affective significance in the amygdala. Alterna-
patients with amygdala lesions did not show en- tively, as discussed in associative learning, limbic
hanced activity to fearful faces in the fusiform and structures such as the amygdala might modulate
occipital gyri, which was indeed found for a group the associative strength of cortical stimulus repre-
of healthy controls (Vuilleumier et al., 2004). Fi- sentations of emotionally significant materials
nally, sizeable BOLD activations in the amygdala (Büchel and Dolan, 2000).
have been observed for highly arous- Taken together, habituation studies have shed
ing emotional materials (Junghöfer et al., 2005; light on the role of stimulus novelty in emotion
Sabatinelli et al., 2005), and it remains to be picture processing. Across several studies, we
40

observed that the differential processing of emo- sites to fearful as compared to neutral faces spe-
tional compared to neutral stimuli indexed by the cifically when the faces were presented at attended
EPN and LPP components did not depend on locations while absent when presented at nonat-
stimulus novelty and were maintained across nu- tended locations. Taken together, these data sug-
merous stimulus repetitions. These findings sug- gest the interference of selective emotion
gest that detecting emotionally significant stimuli processing when attentional resources are directed
in the environment might be an obligatory task of to locations of explicitly task-relevant stimuli.
the organism, apparently not habituating as a Pursuing the interaction of emotion and atten-
function of passive exposure. tion, Schupp and colleagues (2003b) studied the
hypothesis of obligatory selective emotional
processing while subjects performed a feature-based
Emotion processing in the context of primary explicit attention task. Towards this end, task-ir-
attention tasks relevant pleasant, neutral, and unpleasant images
were presented while subjects had to detect target
Recent studies approached the interaction of emo- checkerboard images interspersed in the picture se-
tion and attention from the perspective of compe- quence. In order to increase perceptual load, stimuli
tition by concurrently presenting emotional stimuli were presented as a rapid continuous stream with
and nonemotional task-relevant stimuli (Vuilleu- individual presentation times of 333 ms. The find-
mier and Schwartz, 2001; Pessoa et al., 2002b; ings revealed increased EPN amplitudes to pleasant
Anderson et al., 2003). For instance, relying on a and unpleasant stimuli, particularly pronounced for
primary spatial attention task, Vuilleumier and stimuli of high evolutionary significance. Addition-
Schwartz (2001) demonstrated the selective acti- ally, behavioral and electrophysiological responses
vation of the amygdala and fusiform gyrus to to the task stimuli showed successful top-down at-
fearful faces independent of whether the faces were tentional control to nonemotional stimuli. This re-
presented at attended or unattended spatial loca- search illustrates the selective processing of
tions. In another study, Pessoa and colleagues emotional cues while these stimuli were irrelevant
(2002b) observed that selective emotion processing to the primary cognitive task.
depended on the availability of attentional re- A stronger test for the hypothesis of the auto-
sources. In their critical condition, subjects had to matic nature of selective emotion processing was
discriminate the orientation of eccentrically pre- implemented by a study presenting emotional
sented bars while maintaining fixation on centrally pictures and task-relevant information concur-
presented emotional or neutral faces. Contrasting rently. As shown in Fig. 4(A), target and distracter
with control conditions without additional task stimuli were created by grids of horizontal and
load, emotional faces did not elicit increased ac- vertical lines, which were presented overlaying the
tivation, neither in visual processing areas, nor in IAPS pictures. To systematically vary the task
the amygdala. These data were interpreted from load, the proportion of stimuli presenting the IAPS
the perspective of the biased competition model of pictures with or without the overlaid task-relevant
visual attention, assuming that emotional and grid-pattern, respectively, was changed in four ex-
task-relevant representations competed for proces- perimental conditions. Specifically, in separate
sing resources in the visual cortex (Desimone and blocks, task-relevant stimuli occurred with 10%,
Duncan, 1995; Pessoa et al., 2002a). Complemen- 50%, or 100% probability, respectively, while an
tary evidence was provided by recent ERP additional control condition presented no task
research. For instance, Pourtois and colleagues stimuli in order to replicate previous findings. The
(2004) observed that task-relevant bar stimuli cued order of the four experimental conditions was bal-
by fearful rather than neutral faces were associated anced across subjects. Results closely replicated
with increased P1 components over lateral occip- previous findings for the control condition with
ital leads. In addition, Holmes and Colleagues pleasant and unpleasant pictures eliciting a
(2003) observed a reduced N1 peak over frontal pronounced EPN amplitude relative to neutral
41

Fig. 4. Interaction of emotion and explicit attention. (A) Illustration of the used RSVP paradigm in two of four conditions. Upper
row: passive viewing condition with IAPS pictures. Lower row: task condition with overlaid grid patterns in each trial. In this example,
the fourth trial represents the rare target stimuli. (B) Upper panel: Grand-averaged ERP waveforms of pleasant, neutral, and
unpleasant pictures in the four experimental conditions for a right occipital sensor. Gray-shaded area is the analyzed EPN time interval
from 200–300 ms. Lower panel: Topographical difference maps for pleasant-neutral and unpleasant-neutral in the four conditions
projected on the back view of a model head.

cues. Furthermore, occasionally interspersed task selective emotion processing of pictures not pre-
stimuli (10% condition) did not interfere with the senting task-related information (Schupp et al.,
processing of emotional compared to neutral stim- 2003b). However, interference effects were pro-
uli. Again, an active task set did not interfere with nounced in those conditions with higher task load.
42

The differentiation between emotional and neutral the posterior negativity of subsequent target pic-
stimuli was greatly attenuated during the 50% and tures. While congruence in hedonic valence be-
100% conditions (cf., Fig. 4). As expected by an tween a prime and a target picture has no special
interference account, ERP indices of selective at- status in a motivated attention framework, such
tention to the task-related stimuli were obtained. effects might be predicted from the perspective of
Specifically, target compared to distracter stimuli affective priming. A typical finding in behavioral
revealed increased EPN and LPP amplitudes. priming tasks is that target words are evaluated
Taken together, consistent with the notion of faster when preceded by a short-duration (e.g.,
capacity limitations during perceptual encoding, 300 ms) prime of the same hedonic valence (for re-
these data suggest that explicit attention to task- views see Bargh, 1989; Klauer and Musch, 2003).
related stimuli interferes with selective processing Affective priming may reflect spreading activation
of emotional ‘background’ information. Interest- (Ferguson and Bargh, 2003) and this perspective
ingly, a similar pattern of results was obtained would predict an interaction of prime and target
when the visual oddball task was substituted by an picture category due to the facilitated processing of
auditory one (Schupp et al., unpublished data). target pictures preceded by prime pictures of the
The same ERP component, which was found to be same valence category.
insensitive to stimulus novelty, was modified by To examine this issue (Flaisch et al., submitted),
explicit attention control mechanisms suggesting subjects viewed a continuous stream of pleasant,
that the stimulus-driven attention capture interacts neutral, and unpleasant pictures, presented for
with the goal-driven control of attention. 335 ms each. In order to examine the effects of
emotional primes on target picture processing,
separate average waveforms were calculated for
Emotional perception as a function of prime stimuli nine experimental cells (three emotional picture
categories for the prime and target picture, respec-
Interference effects may not be limited to concur- tively). As expected, emotional target pictures
rently presented stimuli, but may extend in time. In were associated with a larger early posterior
the real world, stimuli often appear not in isolation negativity, compared to neutral ones. Moreover,
but follow each other, raising the question to what it was found that the magnitude of the EPN for a
extent the selective encoding of emotional pictures target picture varied systematically as a function of
varies as a function of the emotion category of a the preceding prime (cf., Fig. 5). When a prime
preceding ‘prime’ picture. For instance, is the picture was emotional (and itself elicited an en-
processing of an erotic scene facilitated when pre- hanced early posterior negativity), the EPN of the
ceded by another pleasant image rather than a subsequent target picture was reduced. This effect
scene of threat? Considered from the perspective of of an emotional prime image was identical regard-
motivated attention (Lang et al., 1997), processing less of whether the target picture was pleasant,
resources are automatically captured and sustained neutral, or unpleasant. Thus, the data revealed no
by emotional cues (Cuthbert et al., 2000; Keil et al., evidence for affective priming in the perceptual
2002). Accordingly, one might expect that an emo- domain. Rather, whether the hedonic valence of
tional prime picture drawing heavily from limited the prime was congruent or incongruent with the
processing resources is hampering the processing of following target, occipital negativity of the target
subsequently presented stimuli. Thus, the moti- picture was decreased if the prime picture was
vated attention hypothesis assumes that target pic- affectively engaging.
ture processing will vary as a function of processing The novel finding of this study is that current
resources devoted to both prime and target pic- target processing is affected not only by the emo-
tures. Specifically, emotional target pictures are as- tional content of the current image but also sys-
sociated with a larger EPN compared to neutral tematically varies with the emotional content of
images, and, emotional prime pictures, themselves the preceding prime picture. These findings imply
associated with an increased EPN, should reduce that the capture of processing resources extends in
43

Fig. 5. Effects on target picture processing as function of preceding prime pictures. (A) Illustration of the prime –target picture
combinations examined in this study. (B) Upper panel: ERP waveforms for a representative right-occipital sensor illustrating the main
effects of prime (right part) and target (left part) valence. The prime effect is illustrated by reduced posterior negativities following
emotional compared to neutral prime pictures averaged across pleasant, neutral, and unpleasant target pictures. Lower panel: Detailed
illustration of the prime effect on target processing by presenting waveforms separately for pleasant, neutral, and unpleasant target
pictures. (C) Scalp potential maps of the difference waves (pleasant–neutral) and (unpleasant-neutral) in the time interval of
248–288 ms post-target picture onset reveal the topographical distributions of the modulation of target picture processing as a function
of prime valence.

time and interferes with successively presented the pattern of results was observed here in the ab-
stimulus materials. These data are consistent with sence of any active task instruction, presumed to
the notion that successively presented pictures reflect the implicit capture of attention by emo-
are characterized by a distinct neural representa- tional cues (Öhman et al., 2000a).
tion, which may compete for a limited pool of
processing resources (Kastner and Ungerleider, Paying explicit attention to emotions
2000; Keysers and Perrett, 2002; Potter et al.,
2002). However, rather than being limited to con- Competition designs explore the boundary condi-
current stimulus processing, the present results tions of the differential processing of emotional
suggest that processing resources allocated to tar- cues. In real life, however, both forces assuring
get pictures systematically vary with the amount attentional orienting, namely motivated and ex-
of processing resources captured by preceding plicit attention, often cooperate. This raises the
emotional prime stimuli. Furthermore, previous question whether paying attention to emotional
studies assessed competition effects mainly by in- rather than neutral contents is even more effective
troducing an explicit attention task. In contrast, in drawing attentional resources. Thus, the
44

interplay of attention and emotion was examined effects is presumably secondary to changes in neural
under conditions when these two forces pull in the structures controlling the expression of selective
same direction (Schupp et al., submitted). attention. The structures believed to be involved
In this study, subjects viewed a rapid and in selective visual processing of target and emotional
continuous stream of pictures displaying erotic, task relevance are heavily interconnected, providing
neutral, and mutilation contents while the EEG was possible anatomical pathways to implement boosted
recorded with a 128 dense sensor array. In separate processing of emotional targets. For instance, heavy
runs, each of the three picture categories served as interconnections are described for the dorsolateral
target category in a counting task. Target stimuli prefrontal cortex, a key structure in the fronto-
were defined according to their emotional valence, parietal network organizing explicit attention
and therefore, both avenues to induce selective effects, and the ventromedial prefrontal cortex,
attention were assessed within the same experimen- part of the paralimbic and limbic network impli-
tal context. It is worth noting that detecting cated in emotional stimulus evaluation (Barbas,
emotional and neutral target stimuli elicited 1995; Davis and Whalen, 2001; Ghashghaei and
increased occipital negativity and late positive po- Barbas, 2002; Bar, 2003). Consistent with anatom-
tentials with similar topography and latency as in ical data, recent fMRI studies revealed the interplay
previous studies requiring subjects to detect animal among these structures when attention was directed
targets (Thorpe et al., 1996; Delorme et al., 2004; to the location of emotionally relevant stimuli
Codispoti et al., in press). As expected, selective or when target detection was compared to the
emotional processing closely replicated previous processing of emotional distracters (Armony and
findings (Cuthbert et al., 2000; Junghöfer et al., Dolan, 2002; Yamasaki et al., 2002). However, con-
2001; Keil et al., 2002; Schupp et al., 2003b, 2004b), sidering the interplay of attention and emotion from
and was associated with a similar cortical signature the perspective of specific subprocesses, a recent
as instructed attention. Establishing these cortical theory is relevant linking late positive potentials to
indices of selective attention due to implicit the activity in a specific neuromodulatory system,
emotional and explicit task-defined significance namely the ascending locus coeruleus–nor-
provides the foundation to meaningfully interpret epinephrine (LC–NE) system which is presumed
the observed interplay of attention and emotion. to increase the gain of cortical representations
Interestingly, overadditive effects of paying atten- (Nieuwenhuis et al., 2005). From this perspective,
tion to emotional rather than neutral contents were the overadditive effects of paying attention to
only observed at later stages of processing indexed emotional targets might be secondary to the engage-
by the LPP (c.f., Fig. 6). This effect appeared siz- ment of the neuromodulatory LC–NE system, a
able, the amplitude of the late positive potential hypothesis awaiting more direct evidence by future
almost doubled when attending to emotional rather research.
than neutral contents. In contrast, effects of Taken together, there is first evidence that
emotion and task relevance were independent from obligatory emotion processing interacts with
each other during earlier perceptual processes explicit attention processes. However, providing
indexed by the EPN. Thus, ERP indices tracking support for the claim of an interaction between
the unfolding of stimulus processing revealed a shift attention and emotion at the level of specific
in the interplay of directed attention and emotional subprocesses, the relationship between emotion
stimulus significance. Attending to emotional cues and attention varied across processing stages.
appeared particularly efficient to boost the LPP Explicit attention effects and emotional signifi-
component. This may point to a capacity-limited cance operated additively during early perceptual
serial processing system possibly involved with stimulus categorization while synergistic effects
short-term memory representations needed for of implicit emotion and explicit attention effects
focused attention and conscious recognition. were observed during the stage of higher-order
The shift in the interplay of target and emo- conceptual stimulus representation in working
tion relevance from independent to overadditive memory.
45

Fig. 6. Effects of cooperation between emotional and explicit attention. (A) Upper panel: Scalp difference maps (target–non-target
stimuli) for erotic, neutral, and mutilation images in the EPN time-window. Lower panel: Calculation of the L2-MNE for the scalp
potential difference illustrating emotion and explicit attention effects. (B) Upper panel: Scalp difference maps (target–non-target
stimuli) for erotic, neutral, and mutilation images in the LPP time-window. Lower panel: Calculation of the L2-MNE for the scalp
potential difference illustrating the interaction of emotion and attention. All maps display a right-side view.

Learned stimulus relevance and attentional orienting investigate subject populations responding highly
emotional to selected stimuli. Accordingly, the is-
The perspective that emotional cues capture at- sue is raised how individual experience and learn-
tentional resources provides a framework to ing shape the attentional orienting to these stimuli
46

in distinct processing stages. Previous studies al- current results appear not fully consistent and
ready provided evidence for a pronounced atten- await more conclusive research. In particular, fu-
tion capture by fear-relevant stimuli in specific ture studies using dense sensor ERP arrays may
phobia. For instance, subjects with small animal enable to provide a better assessment of the early
phobia display a pronounced defensive reactivity selective stimulus processing. Also, examining clin-
to fear relevant stimuli, that is, augmented defen- ically relevant stimuli as well as standard pleasant,
sive startle magnitude, as well as heightened auto- neutral, and unpleasant materials in a range of
nomic responsivity and increased activation in the disorder groups appears particularly informative.
amygdala and infero-temporal cortex (Hamm One promise of this endeavor is that ERP meas-
et al., 1997; Cuthbert et al., 2003; Sabatinelli urements may become a sensitive tool to evaluate
et al., 2005). Building upon these findings, we re- effects of behavior therapy complementing fMRI
cently used rapid serial visual presentations to ex- and startle probe measurements (Vrana et al.,
plore whether fear-related pictures are associated 1992; Straube et al., 2005).
with enhanced EPN amplitudes in snake or spider
phobics. Consistent with this hypothesis, the EPN Summary
was largest for snake and spider pictures in snake
and spider phobics, respectively. However, the Experimental approaches to explore the emo-
effect appeared more robust in snake phobics tion–attention interface further the understanding
(Schupp et al., unpublished data). Focusing on of the emotional capture of attentional resources
spider phobia, a follow-up study including a larger in distinct processing stages. One issue is to deter-
sample revealed only marginally increased EPN mine to what extent the EPN component reflects
amplitudes to fear relevant materials in spider an automatic phenomenon, that is, the uninten-
phobics compared to healthy controls. This study tional, unconscious, and effortless processing of
observed pronounced enlarged LPP and slow wave emotional cues. Suggestive of automaticity, the
amplitudes for the processing of fear-relevant EPN modulation to emotional cues is observed
stimuli in phobics compared to control partici- across repeated stimulus presentations of the same
pants (Michalowski et al., 2005). Similar results materials and occurs in the absence of an active
have also been reported by Miltner and colleagues task instruction. Thus, in the absence of interfer-
(2005). Yet another line of research examined this ence by concurrent stimulus processing, selective
issue focusing on addiction and the processing of emotional processing appears to be unintentional,
drug-associated cues such as heroin, alcohol, co- effortless, and presumably precedes conscious
caine, and tobacco (Franken et al., 2004). In- recognition. Conversely, the EPN is subject to in-
creased LPP and slow wave amplitudes have been terference by intentional goals held in mind and
observed in heroin-dependent compared to control the processing of immediately preceding emo-
individuals in drug-associated pictures depicting tional cues. Thus, concurrent and successively
cocaine and heroin (Franken et al., 2003, 2004). presented pictures are characterized by distinct
Taken together, evidence accumulates suggest- neural representations, which may compete for a
ing that ERP measures derived in research of limited pool of processing resources, attenuating
emotion processing in healthy individuals might or abolishing selective emotion processing. Taken
contribute to the understanding of clinically rele- together, albeit preliminary, results appear to
vant issues. Previous research provides ample ev- support component features rather than all-or-
idence that learning and experience triggers nothing concepts of automaticity (Bargh, 1989).
pronounced and fast physiological responses to Further research is needed to explore the bound-
clinically relevant stimuli (Globisch et al., 1999; ary conditions of the EPN modulation
Öhman et al., 2000b). Extending these data, dis- with regard to the various component features of
order-related stimuli appear to trigger exaggerated automaticity.
responding in specific ERP components partition- Another interesting issue is to delineate dissociat-
ing the attention capture in substages. However, ions among the various processing stages reflecting
47

the emotional attention capture. In passive viewing stimulus perception may provide a framework to
contexts, emotional EPN and LPP modulation gain more specific hypotheses regarding the func-
appear similar in many respects, that is, not subject tional significance of ERP indices of selective at-
to habituation and most pronounced for high- tention. Specifically, the EPN associated with the
arousing pleasant and unpleasant materials. How- perceptual encoding of emotional material may re-
ever, when paying attention to emotion, boosted flect the call for resources in the capacity-limited
processing effects were only observed for the LPP, second stage of processing. However, to become
suggesting that the relationship of emotion and consciously represented, stimuli need to have ac-
attention changes across time from independent cess to the second processing stage, which might be
(EPN) to synergistic effects (LPP). Thus, interac- indexed by the LPP component. Furthermore, the
tion effects of explicit attention and implicit emo- elaborate and sustained attentive processing of
tion processes vary for distinct processing stages. emotional stimuli is presumably reflected by the
Future studies are needed to extend these findings sustained positive slow waves. Thus, the call for
including subject populations (e.g., phobia, sub- processing resources triggered after initial stimulus
stance abuse) demonstrating exaggerated attentio- categorization assures that emotional stimuli have
nal orienting to specific stimuli. priority access to a capacity-limited stage required
for working memory consolidation and conscious
recognition. Considered from a functional and ev-
Selective attention to emotion: two-stage models of olutionary perspective, this priority processing of
stimulus perception emotional cues facilitates adaptive behavior, finally
promoting survival and reproductive success (Lang
It might be informative to consider the ERP find- et al., 1997; Öhman et al., 2000a).
ings reviewed above from the broader perspective The various two-stage models of stimulus per-
of two-stage models of stimulus perception. Inte- ception suggest the rapid perceptual categorization
grating research from cognitive psychology, the of stimuli. Consistent with this view, available data
orienting reflex, and associative learning (Öhman, suggest that the differential processing of emo-
1979, 1986), a two-stage model was proposed to tional cues is consequent upon access to seman-
explicate how emotional, novel, and task-relevant tic meaning. Indirect evidence was provided by
stimuli guide selective attention. The model pro- the similar latency of emotion discrimination and
posed a first large capacity perceptual scanning explicit attention effects obtained in simple
stage providing a more or less complete analysis of feature and higher-order object-based attention
sensory information. Upon detection of significant tasks (Smid et al., 1999; Potts and Tucker, 2001;
stimuli, the perceptual system may emit a call for Delorme et al., 2004; Codispoti et al., in press ).
processing resources in a capacity-limited second More direct evidence was provided in a recent
stage of stimulus processing acting as gateway study examining explicit attention to emotional
to focused attention and conscious recognition. cues (Schupp et al., submitted; see above). Find-
Empirical support for this view is brought out by ings revealed that explicit attention effects and se-
research with the attentional blink and rapid visual lective emotion processing appeared with similar
serial presentations. This research determined that latency suggesting that perceptual categorization
pictures can be recognized and comprehended at up to the level of semantic meaning was achieved.
rapid rates while memory probed immediately One advantage of considering the ERP findings
afterwards appears quite poor. Potter (Chun and in the context of two-stage models of stimulus per-
Potter, 1995; Potter, 1999) suggested that stimulus ception is that a common frame is provided for
recognition occurs rapidly within the first few hun- considering the effects of the various active and
dred milliseconds. However, for conscious stimulus passive forms of attentional orienting. Specifically,
recognition, this fleeting stage of processing needs the call for processing resources is thought to re-
to be followed by a second stage of consolidation flect short- and long-term memory mechanisms
in short-term memory. Two-stage models of (Öhman et al., 2000a). Emotional stimuli capture
48

attention because they include ‘tags’ on memory Acknowledgments


representations denoting significant environmental
elements in the environment. Active attention This work was supported by the German Research
effects are considered to reflect the expectancy of Foundation (DFG) Grants Schu 1074/7-3 and
certain objects; that is, the temporary activation of 1074/10-1. Address reprint requests to Harald T.
long-term memories. In contrast, passive attention Schupp at the Department of Psychology, Uni-
effects as observed by the classical orienting re- versity of Konstanz, Universitätsstr. 10, 78457
sponses are thought to reflect the failure to match Konstanz, Germany.
current stimuli to the content of the short-term
memory store. Considering the competition and
cooperation studies of the emotion–attention in- References
teraction suggests that the call for processing re-
Amrhein, C., Mühlberger, A., Pauli, P. and Wiedemann, G.
sources triggered by the various forms of
(2004) Modulation of event-related brain potentials during
attentional control are independent from each affective picture processing: a complement to startle reflex
other and may compete for processing resources. and skin conductance response? Int. J. Psychophysiol., 54:
Furthermore, consistent with the main theme 231–240.
of the present review to consider the emotional Anderson, A.K., Christoff, K., Panitz, D., De Rosa, E. and
guidance of attention in distinct processing stages, Gabrieli, J.D. (2003) Neural correlates of the automatic
processing of threat facial signals. J. Neurosci., 23: 5627–5633.
interaction patterns changed across processing Armony, J.L. and Dolan, R.J. (2002) Modulation of spatial
stages and showed synergistic effects in the second attention by fear-conditioned stimuli: an event-related fMRI
stage associated with stimulus representation in study. Neuropsychologia, 40: 817–826.
working memory. Clearly, these hypotheses are Bar, M. (2003) A cortical mechanism for triggering top-down
speculative and await future research. One prom- facilitation in visual object recognition. J. Cogn. Neurosci.,
15: 600–609.
ising avenue to pursue this view is to complement Barbas, H. (1995) Anatomic basis of cognitive-emotional in-
information regarding the temporal dynamics re- teractions in the primate prefrontal cortex. Neurosci.
vealed by ERPs with structural information ob- Biobehav. Rev., 19: 499–510.
tained with fMRI. Overall, reviewing the ERP Bargh, J.A. (1989) Conditional automaticity: varieties of auto-
matic influence in social perception and cognition. In: Ule-
findings from the perspective of two stage models
man, J.S. and Bargh, J.A. (Eds.), Unintended Thought.
specifies hypotheses regarding the functional sig- Guilford Press, New York, pp. 3–51.
nificance of distinct ERP components from a Birbaumer, N., Elbert, T., Canavan, A.G. and Rockstroh, B.
broader theoretical context and suggests multiple (1990) Slow potentials of the cerebral cortex and behavior.
avenues to explore the emotion–attention interac- Physiol. Rev., 70: 1–41.
tion in future research. Bradley, M.M. (2000) Emotion and motivation. In: Cacioppo,
J.T., Tassinary, L.G. and Berntson, G. (Eds.), Handbook of
Psychophysiology. Cambridge University Press, New York,
pp. 602–642.
Abbreviations Bradley, M.M., Codispoti, M., Cuthbert, B.N. and Lang, P.J.
(2001) Emotion and motivation I: defensive and appetitive
CSD current-source-density reactions in picture processing. Emotion, 1: 276–298.
Bradley, M.M., Lang, P.J. and Cuthbert, B.N. (1993) Emotion,
EEG electroencephalogram novelty, and the startle reflex: habituation in humans. Behav.
EPN early posterior negativity Neurosci., 107: 970–980.
ERP event-related potential Breiter, H.C., Etcoff, N.L., Whalen, P.J., Kennedy, W.A., Ra-
FMRI functional magnetic resonance uch, S.L., Buckner, R.L., et al. (1996) Response and habit-
uation of the human amygdala during visual processing of
imaging
facial expression. Neuron, 17: 875–887.
LPP late positive potential Büchel, C. and Dolan, R.J. (2000) Classical fear conditioning in
L2-MNE L2-minimum-norm-estimate functional neuroimaging. Curr. Opin. Neurobiol., 10:
RSVP rapid serial visual presentation 219–223.
49

Cacioppo, J.T., Crites Jr., S.L., Berntson, G.G. and Coles, Diedrich, O., Naumann, E., Maier, S.G.B. and Bartussek, D.
M.G. (1993) If attitudes affect how stimuli are processed (1997) A frontal positive slow wave in the ERP associated
should they not affect the event-related brain potential. Psy- with emotional slides. J. Psychophysiol., 11: 71–84.
chol. Sci., 4: 108–112. Donchin, E. and Coles, M.G. (1988) Is the P300 component a
Cacioppo, J.T., Gardner, W.L. and Berntson, G.G. (1999) The manifestation of context updating? Behav. Brain Sci., 11:
affect system has parallel and integrative processing compo- 357–427.
nents: form follows function. J. Pers. Soc. Psychol., 76: Donchin, E., Kramer, A.E. and Wickens, C. (1986) Applica-
839–855. tions of brain event-related potentials to problems in engi-
Chun, M.M. and Potter, M.C. (1995) A two-stage model for neering psychology. In: Coles, M.G.H., Donchin, E. and
multiple target detection in rapid serial visual presentation. Porges, S.W. (Eds.), Psychophysiology: Systems, Processes,
J. Exp. Psychol. Hum. Percept. Perform., 21: 109–127. and Applications. Guilford Press, New York, pp. 702–718.
Codispoti, M., Ferrari, V. and Bradley, M.M. (2006) Repetitive Ferguson, M.J. and Bargh, J.A. (2003) The constructive nature
picture processing: autonomic and cortical correlates. Brain of automatic evaluation. In: Musch, J. and Klauer, K.C.
Res., 1068: 213–220. (Eds.), Psychology of Evaluation: Affective Processes in
Codispoti, M., Ferrari, V., Junghöfer, M. & Schupp, H.T. Cognition and Emotion. Lawrence Erlbaum Associates,
(in press) The categorization of natural scenes: brain atten- Mahwah, NJ, pp. 169–188.
tion networks revealed by dense sensor ERPs. Neuroimage. Flaisch, T., Junghöfer, M., Bradley, M.M., Schupp, H.T. and
Compton, R.J. (2003) The interface between emotion and at- Lang, P.J. (submitted) Rapid picture processing: affective
tention: a review of evidence from psychology and neurosci- primes and targets.
ence. Behav. Cogn. Neurosci. Rev., 2: 115–129. Franken, I.H., Hulstijn, K.P., Stam, C.J., Hendriks, V.M. and
Cuthbert, B.N., Schupp, H.T., Bradley, M., McManis, M. and van den Brink, W. (2004) Two new neurophysiological in-
Lang, P.J. (1998) Probing affective pictures: attended startle dices of cocaine craving: evoked brain potentials and cue
and tone probes. Psychophysiology, 35: 344–347. modulated startle reflex. J. Psychopharmacol., 18: 544–552.
Cuthbert, B.N., Schupp, H.T., McManis, M.H., Hillman, C., Franken, I.H., Stam, C.J., Hendriks, V.M. and van den Brink,
Bradley, M.M. and Lang, P.J. (1995) Cortical slow waves: W. (2003) Neurophysiological evidence for abnormal cogni-
emotional perception and processing [Abstract]. Psychophys- tive processing of drug cues in heroin dependence. Psycho-
iology, 32: S27. pharmacology, 170: 205–212.
Cuthbert, B.N., Lang, P.J., Strauss, C., Drobes, D., Patrick, C.J. Frijda, N.H. (1986) The Emotions. Cambridge University
and Bradley, M.M. (2003) The psychophysiology of anxiety Press, Cambridge.
disorder: fear memory imagery. Psychophysiology, 40: Ghashghaei, H.T. and Barbas, H. (2002) Pathways for emotion:
407–422. interactions of prefrontal and anterior temporal pathways in
Cuthbert, B.N., Schupp, H.T., Bradley, M.M., Birbaumer, N. the amygdala of the rhesus monkey. Neuroscience, 115:
and Lang, P.J. (2000) Brain potentials in affective picture 1261–1279.
processing: covariation with autonomic arousal and affective Globisch, J., Hamm, A.O., Esteves, F. and Öhman, A. (1999)
report. Biol. Psychol., 52: 95–111. Fear appears fast: temporal course of startle reflex potentiat-
Davis, M. and Whalen, P.J. (2001) The amygdala: vigilance and ion in animal fearful subjects. Psychophysiology, 36: 66–75.
emotion. Mol. Psychiatry, 6: 13–34. Hamalainen, M.S. and Ilmoniemi, R.J. (1994) Interpreting
Delorme, A., Rousselet, G.A., Mace, M.J. and Fabre-Thorpe, magnetic fields of the brain: minimum norm estimates. Med.
M. (2004) Interaction of top-down and bottom-up processing Biol. Eng. Comput., 32: 35–42.
in the fast visual analysis of natural scenes. Brain. Res. Cogn. Hamm, A.O., Cuthbert, B.N., Globisch, J. and Vaitl, D. (1997)
Brain Res., 19: 103–113. Fear and the startle reflex: blink modulation and autonomic
Derryberry, D. and Tucker, D.M. (1991) The adaptive base response patterns in animal and mutilation fearful subjects.
of the neural hierarchy: elementary motivational controls Psychophysiology, 34: 97–107.
of network function. In: Dienstbier, A. (Ed.), Nebraska Hamm, A.O., Schupp, H.T. and Weike, A.I. (2003) Motiva-
Symposium on Motivation. University of Nebraska Press, tional organization of emotions: autonomic changes, cortical
Lincoln, NE, pp. 289–342. responses, and reflex modulation. In: Davidson, R.J., Scherer,
Desimone, R. and Duncan, J. (1995) Neural mechanisms of K. and Goldsmith, H.H. (Eds.), Handbook of Affective Sci-
selective visual attention. Annu. Rev. Neurosci., 18: 193–222. ences. Oxford University Press, Oxford, pp. 188–211.
Dickinson, A. and Balleine, B.W. (2002). The role of learning in Hauk, O., Keil, A., Elbert, T. and Müller, M.M. (2002) Com-
motivation. In: Gallistel C.R. (Ed.), Steven’s Handbook of parison of data transformation procedures to enhance top-
Experimental Psychology, Vol 3, Learning, Motivation and ographical accuracy in timeseries analysis of the human
Emotion, 3rd edn. Wiley, New York, pp. 497–533. EEG. J. Neurosci. Methods, 113: 111–122.
Dickinson, A. and Dearing, M.F. (1979) Appetitive-aversive Hoffman, J.E., Simons, R.F. and Houck, M. (1983) The effects
interactions and inhibitory processes. In: Dickinson, A. of automatic and controlled processing on the P300. Psy-
and Boakes, R.A. (Eds.), Mechanisms of Learning and chophysiology, 20: 625–632.
Motivation. Lawrence Erlbaum Associates, Hillsdale, NJ, Holmes, A., Vuilleumier, P. and Eimer, M. (2003) The process-
pp. 203–231. ing of emotional facial expression is gated by spatial
50

attention: evidence from event-related brain potentials. and monkey electrophysiology. In: Gazzaniga, M.S. (Ed.), The
Brain. Res. Cogn. Brain Res., 16: 174–184. Cognitive Neurosciences (2nd edn). MIT Press, Cambridge.
Johnson Jr., R. (1988) The amplitude of the P300 component of Luck, S.J., Woodman, G.F. and Vogel, E.K. (2000) Event-re-
the event-related potential: review and synthesis. In: Ackles, lated potential studies of attention. Trends Cogn. Sci., 4:
P.K., Jennings, J.R. and Coles, M.G.H. (Eds.) Advances in 432–440.
Psychophysiology, Vol. 3. JAI Press, Greenwich, pp. 69–138. Lutzenberger, W., Elbert, T. and Rockstroh, B. (1987) A brief
Johnston, V.S., Miller, D.R. and Burleson, M.H. (1986) Mul- tutorial on the implications of volume conduction for the
tiple P3 s to emotional stimuli and their theoretical signifi- interpretation of the EEG. J. Psychophysiol., 1: 81–90.
cance. Psychophysiology, 23: 684–694. Michalowski, J.M., Melzig, C.A., Schupp, H.T. and Hamm,
Johnston, V.S. and Oliver-Rodriguez, J.C. (1997) Facial beauty A.O. (2005) Cortical processing of emotional pictures in spi-
and the late positive component of event-related potentials. J. der phobic students [Abstract]. Psychophysiology, 42: S89.
Sex Res., 34: 188–198. Miltner, W.H., Trippe, R.H., Krieschel, S., Gutberlet, I., He-
Junghöfer, M., Bradley, M.M., Elbert, T.R. and Lang, P.J. cht, H. and Weiss, T. (2005) Event-related brain potentials
(2001) Fleeting images: a new look at early emotion discrim- and affective responses to threat in spider/snake-phobic and
ination. Psychophysiology, 38: 175–178. non-phobic subjects. Int. J. Psychophysiol., 57: 43–52.
Junghöfer, M., Sabatinelli, D., Bradley, M.M., Schupp, H.T., Mogg, K. and Bradley, B.P. (1999) Selective attention and
Elbert, T.R. and Lang, P.J. (2006) Fleeting images: rapid affect anxiety: a cognitive-motivational perspective. In: Dalgleish,
discrimination in the visual cortex. Neuroreport, 17: 225–229. T. and Power, M.J. (Eds.), Handbook of Cognition and
Junghöfer, M., Schupp, H.T., Stark, R. and Vaitl, D. (2005) Emotion. Wiley, Chichester, pp. 145–170.
Neuroimaging of emotion: empirical effects of proportional Morris, J.S., Büchel, C. and Dolan, R.J. (2001) Parallel neural
global signal scaling in fMRI data analysis. Neuroimage, 25: responses in amygdala subregions and sensory cortex during
520–526. implicit fear conditioning. Neuroimage, 13: 1044–1052.
Kastner, S. and Ungerleider, L.G. (2000) Mechanisms of visual Nieuwenhuis, S., Aston-Jones, G. and Cohen, J.D. (2005) De-
attention in the human cortex. Annu. Rev. Neurosci., 23: cision making, the P3, and the locus coeruleus-nor-
315–341. epinephrine system. Psychol. Bull., 131: 510–532.
Keil, A., Bradley, M.M., Hauk, O., Rockstroh, B., Elbert, T. Öhman, A. (1979) The orienting response, attention, and learn-
and Lang, P.J. (2002) Large-scale neural correlates of affec- ing: an information-processing perspective. In: Kimmel,
tive picture processing. Psychophysiology, 39: 641–649. H.D., van Olst, E.H. and Orlebek, J.F. (Eds.), The Orient-
Keysers, C. and Perrett, D.I. (2002) Visual masking and RSVP ing Reflex In Humans. Erlbaum, Hillsdale, NJ, pp. 443–471.
reveal neural competition. Trends Cogn. Sci., 6: 120–125. Öhman, A. (1986) Face the beast and fear the face: animal and
Klauer, K.C. and Musch, J. (2003) Affective priming: findings social fears as prototypes for evolutionary analyses of emo-
and theories. In: Musch, J. and Klauer, K.C. (Eds.), The tion. Psychophysiology, 23: 123–145.
Psychology of Evaluation: Affective Processes in Cognition Öhman, A., Flykt, A. and Lundqvist, D. (2000a) Unconscious
and Emotion. Erlbaum, Mahwah, NJ, pp. 7–49. emotion: evolutionary perspectives, psychophysiological data
Konorski, J. (1967) Integrative Activity of the Brain: An Inter- and neuropsychological mechanisms. In: Lane, R.D. and
disciplinary Approach. University of Chicago Press, Chicago. Nadel, L. (Eds.), Cognitive Neuroscience of Emotion. Ox-
Kranczioch, C., Debener, S. and Engel, A.K. (2003) Event- ford University Press, Oxford, pp. 296–327.
related potential correlates of the attentional blink phenom- Öhman, A., Hamm, A. and Hugdahl, K. (2000b) Cognition and
enon. Brain Res. Cogn. Brain Res., 17: 177–187. the autonomic nervous system: orienting, anticipation, and
Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1990) Emotion, conditioning. In: Cacioppo, J.T., Tassinary, L.G. and Berns-
attention, and the startle reflex. Psychol. Rev., 97: 377–395. ton, G.G. (Eds.), Handbook of Psychophysiology (2nd edn.).
Lang, P.J. (1995) The emotion probe: studies of motivation and Cambridge University Press, Cambridge, UK, pp. 533–575.
attention. Am. Psychol., 50: 371–385. Palomba, D., Angrilli, A. and Mini, A. (1997) Visual evoked
Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1997) Moti- potentials, heart rate responses and memory to emotional
vated attention: affect, activation, and action. In: Lang, P.J., pictorial stimuli. Int. J. Psychophysiol., 27: 55–67.
Simons, R.F. and Balaban, M. (Eds.), Attention and Emo- Pessoa, L., Kastner, S. and Ungerleider, L.G. (2002a) Attent-
tion: Sensory and Motivational Processes. Erlbaum, Mah- ional control of the processing of neutral and emotional
wah, NJ, pp. 97–135. stimuli. Brain Res. Cogn. Brain Res., 15: 31–45.
Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (2005). Interna- Pessoa, L., McKenna, M., Gutierrez, E. and Ungerleider, L.G.
tional affective picture system (IAPS): digitized photographs, (2002b) Neural processing of emotional faces requires atten-
instruction manual and affective ratings. Technical Report tion. Proc. Natl. Acad. Sci. USA, 99: 11458–11463.
A-6. University of Florida, Gainesville, FL. Phan, K.L., Liberzon, I., Welsh, R.C., Britton, J.C. and Taylor,
Löw, A., Lang, P.J. and Bradley, M.M. (2005) What pops out S.F. (2003) Habituation of rostral anterior cingulate cortex to
during rapid picture presentation? [Abstract]. Psychophysi- repeated emotionally salient pictures. Neuropsychopharma-
ology, 42: S81. cology, 28: 1344–1350.
Luck, S.J. and Hillyard, S.A. (2000) The operation of selective Picton, T.W. (1992) The P300 wave of the human event-related
attention at multiple stages of processing: evidence from human potential. J. Clin. Neurophysiol., 9: 456–479.
51

Potter, M.C. (1999) Understanding sentences and scenes: the Schupp, H.T., Öhman, A., Junghöfer, M., Weike, A.I., Stock-
role of conceptual short term memory. In: Coltheart, V. (Ed.), burger, J. and Hamm, A.O. (2004c) The facilitated processing
Fleeting Memories. MIT Press, Cambridge, MA, pp. 13–46. of threatening faces: an ERP analysis. Emotion, 4: 189–200.
Potter, M.C., Staub, A. and O’Connor, D.H. (2002) The time Schupp, H.T., Stockburger, J., Codispoti, M., Junghöfer, M.,
course of competition for attention: attention is initially labile. Weike, A.I. and Hamm, A.O. (2006) Stimulus novelty and
J. Exp. Psychol. Hum. Percept. Perform., 28: 1149–1162. emotion perception: the near absence of habituation in the
Potts, G.F. and Tucker, D.M. (2001) Frontal evaluation and visual cortex. Neuroreport, 17: 365–369.
posterior representation in target detection. Brain Res. Cogn. Schupp, H.T., Stockburger, J., Codispoti, M., Junghöfer, M.,
Brain Res., 11: 147–156. Weike, A.I. and Hamm, A.O. (submitted) Selective visual
Pourtois, G., Grandjean, D., Sander, D. and Vuilleumier, P. attention to emotion.
(2004) Electrophysiological correlates of rapid spatial orient- Schupp, H.T., Weike, A.I. and Hamm, A. (2000b) Affect and
ing towards fearful faces. Cereb. Cortex, 14: 619–633. evaluative context: high-density ERP recordings during pic-
Ritter, W. and Ruchkin, D.S. (1992) A review of event-related ture processing [Abstract]. Psychophysiology, 37: S88.
potential components discovered in the context of studying Smid, H.G., Jakob, A. and Heinze, H.J. (1999) An event-related
P3. Ann. NY. Acad. Sci., 658: 1–32. brain potential study of visual selective attention to conjunc-
Ruchkin, D.S., Johnson Jr., R., Mahaffey, D. and Sutton, S. tions of color and shape. Psychophysiology, 36: 264–279.
(1988) Toward a functional categorization of slow waves. Straube, T., Glauer, M., Dilger, S., Mentzel, H.J. and Miltner,
Psychophysiology, 25: 339–353. W.H. (2005) Effects of cognitive-behavioral therapy on brain
Sabatinelli, D., Bradley, M.M., Fitzsimmons, J.R. and Lang, activation in specific phobia. Neuroimage, 29: 125–135.
P.J. (2005) Parallel amygdala and inferotemporal activation Thorpe, S., Fize, D. and Marlot, C. (1996) Speed of processing
reflect emotional intensity and fear relevance. Neuroimage, in the human visual system. Nature, 381: 520–522.
24: 1265–1270. Tucker, D.M. (1993) Spatial sampling of head electrical fields:
Schneirla, T. (1959) An evolutionary and developmental theory the geodesic sensor net. Electroencephalogr. Clin. Ne-
of biphasic processes underlying approach and withdrawal. urophysiol., 87: 154–163.
In: Jones, M. (Ed.), Nebraska Symposium on Motivation. Vrana, S.R., Constantine, J.A. and Westman, J.S. (1992) Startle
University of Nebraska Press, Lincoln, pp. 1–42. reflex modification as an outcome measure in the treatment
Schupp, H.T., Cuthbert, B.N., Bradley, M.M., Birbaumer, N. of phobia: two case studies. Behav. Assess., 14: 279–291.
and Lang, P.J. (1997) Probe P3 and blinks: two measures of Vuilleumier, P. (2005) How brains beware: neural mechanisms
affective startle modulation. Psychophysiology, 34: 1–6. of emotional attention. Trends Cogn. Sci., 9: 585–594.
Schupp, H.T., Cuthbert, B.N., Bradley, M.M., Cacioppo, J.T., Vuilleumier, P., Richardson, M.P., Armony, J.L., Driver, J. and
Ito, T. and Lang, P.J. (2000a) Affective picture processing: Dolan, R.J. (2004) Distant influences of amygdala lesion on
the late positive potential is modulated by motivational rel- visual cortical activation during emotional face processing.
evance. Psychophysiology, 37: 257–261. Nat. Neurosci., 7: 1271–1278.
Schupp, H.T., Cuthbert, B.N., Bradley, M.M., Hillman, C.H., Vuilleumier, P. and Schwartz, S. (2001) Beware and be aware:
Hamm, A.O. and Lang, P.J. (2004a) Brain processes in emo- capture of spatial attention by fear-related stimuli in neglect.
tional perception: motivated attention. Cogn. Emotion, 18: Neuroreport, 12: 1119–1122.
593–611. Wieser, M.J., Mühlberger, A., Alpers, G.W., Macht, M., Ellg-
Schupp, H.T., Junghöfer, M., Weike, A.I. and Hamm, A.O. ring, H. and Pauli, P. (2006) Emotion processing in Parkin-
(2003a) Emotional facilitation of sensory processing in the son’s disease: dissociation between early neuronal processing
visual cortex. Psychol. Sci., 14: 7–13. and explicit ratings. Clin. Neurophysiol., 117: 94–102.
Schupp, H.T., Junghöfer, M., Weike, A.I. and Hamm, A.O. Wright, C.I., Fischer, H., Whalen, P.J., McInerney, S.C., Shin,
(2003b) Attention and emotion: an ERP analysis of facili- L.M. and Rauch, S.L. (2001) Differential prefrontal cortex
tated emotional stimulus processing. Neuroreport, 14: and amygdala habituation to repeatedly presented emotional
1107–1110. stimuli. Neuroreport, 12: 379–383.
Schupp, H.T., Junghöfer, M., Weike, A.I. and Hamm, A.O. Yamasaki, H., LaBar, K.S. and McCarthy, G. (2002) Disso-
(2004b) The selective processing of briefly presented affective ciable prefrontal brain systems for attention and emotion.
pictures: an ERP analysis. Psychophysiology, 41: 441–449. Proc. Natl. Acad. Sci. USA, 99: 11447–11451.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 3

Implicit and explicit categorization of natural scenes

Maurizio Codispoti, Vera Ferrari, Andrea De Cesarei and Rossella Cardinale

Department of Psychology, University of Bologna, Viale Berti Pichat, 5– 40127 Bologna, Italy

Abstract: Event-related potential (ERP) studies have consistently found that emotionally arousing (pleas-
ant and unpleasant) pictures elicit a larger late positive potential (LPP) than neutral pictures in a window
from 400 to 800 ms after picture onset. In addition, an early ERP component has been reported to vary with
emotional arousal in a window from about 150 to 300 ms with affective, compared to neutral stimuli,
prompting significantly less positivity over occipito-temporal sites. Similar early and late ERP components
have been found in explicit categorization tasks, suggesting that selective attention to target features results
in similar cortical changes. Several studies have shown that the affective modulation of the LPP persisted
even when the same pictures are repeated several times, when they are presented as distractors, or when
participants are engaged in a competing task. These results indicate that categorization of affective stimuli
is an obligatory process. On the other hand, perceptual factors (e.g., stimulus size) seem to affect the early
ERP component but not the affective modulation of the LPP. Although early and late ERP components
vary with stimulus relevance, given that they are differentially affected by stimulus and task manipulations,
they appear to index different facets of picture processing.

Keywords: emotion; attention; categorization; habituation; natural scenes

The ability to organize the world into meaningful recent studies on natural scene categorization and
and appropriate groupings is necessary for deter- motivated attention.
mining an organism’s behaviour in terms of ca-
pacity to achieve adaptive responses in different
situations. A number of studies using simple stim- Categorization of natural scenes
uli and explicit tasks suggest marked limits to the
capacity of visual attention. In contrast, we all Since the 1970s, Mary Potter and colleagues have
know that despite the perceptual complexity of crafted a careful series of rapid serial visual pres-
natural scenes, the processing of everyday scenes is entation (RSVP) studies showing that natural ob-
almost instantaneous and feels effortless. Several jects belonging to a target category may be
recent studies have confirmed this notion and have classified remarkably quickly (Potter, 1975,
experimentally demonstrated the impressive speed 1976). In these studies, a rapid sequence of unre-
with which natural scenes are categorized. More- lated pictures is presented and an immediate se-
over, it has been demonstrated that when natural mantic detection is required; for example,
scenes are motivationally relevant they rapidly participants had to respond when they saw a pic-
capture attentional resources leading to automatic ture of ‘‘a dog’’ (which they had never seen be-
categorization. This paper provides an overview of fore), presented within a sequence of 16 pictures.
The results of these studies suggested that picture
Corresponding author. Tel.: +39-51-2091836; Fax: +39-051- scenes are understood and become immune to vis-
243086; E-mail: maurizio.codispoti@unibo.it ual masking within about 100 ms.

DOI: 10.1016/S0079-6123(06)56003-0 53
54

Evidence from human event-related potentials complexity that characterize a target category.
(ERPs) also gives cause to expect that ‘‘early’’ These features of intermediate complexity might
processing occurs considerably quicker than is be used to discriminate between scenes that do or
commonly thought. Thorpe et al. (1996) investi- do not contain the target without necessarily fully
gated how long the visual system takes to perceive identifying it. If the target stimulus is still present
and understand a complex natural image through after this detection stage, focused attention binds
a paradigm in which subjects were asked to decide these features to form a more specific representa-
whether or not a briefly displayed colour picture tion of the target stimulus.
contains an animal. This study revealed an early Initially, the early differential ERP activity was
ERP difference between target (image with ani- considered to reflect anterior neural generator sites
mal) and distractor (image without animal) trials implicated in the inhibition of inappropriate be-
that started roughly 150 ms after stimulus onset. havioural responses during NoGo trials (Thorpe et
Similar findings have been reported when partic- al., 1996). Subsequent research has demonstrated
ipants had to categorize complex scenes on the that the early (150 ms) larger negativity elicited by
basis of the presence or absence of a clearly arti- targets compared to distractor stimuli is exten-
ficial category: means of transport (VanRullen and sively due to the involvement of occipito-temporal
Thorpe, 2001). regions in semantic processing of visual stimuli
Together with other recent findings (Li et al., (Delorme et al., 2004; Rousselet et al., 2004).
2002; Kirchner and Thorpe, 2005) the early differ- Similarly, in the same time window, Rudell de-
ential activity starting at 150 ms from stimulus scribed a posterior evoked potential named recog-
onset has been interpreted as evidence of a specific nition potential (RP) to refer to an electrical brain
mode of visual processing, termed ‘‘ultrarapid vis- response which occurs when a subject views a rec-
ual categorization’’ that relies on a parallel and ognizable image of words (Rudell, 1991; Rudell
automatic feedforward mechanism (Fabre-Thorpe and Hua, 1997) and pictures (Rudell, 1992). More
et al., 2001). recently, Hinojosa et al. (2000) showed that the
These findings challenge the conventional view neural generator of the RP is placed mainly in the
that focused attention is required for high-level infero-temporal cortex and they interpreted this
visual processing (Evans and Treisman, 2005). In- potential as related to the processing of semantic-
deed, several studies have shown that many visual conceptual aspects. The role of infero-temporal
tasks necessitate focused attention, including fea- neurons in object recognition has been known
ture-binding to form objects and change detection since the seminal work of Gross and co-workers
in complex scenes (Treisman and Gelade, 1980; (1973); furthermore, an important role has been
Wolfe and Bennett, 1997; O’Regan et al., 1999). recently ascribed to the prefrontal cortex which
Evans and Treisman (2005) have recently shown receives highly processed visual information from
evidence that when participants correctly detected the infero-temporal cortex and orchestrates vol-
that an animal target was present in a rapid serial untary, goal-directed behaviours (Riesenhuber and
visual sequence, they frequently failed to identify Poggio, 2002; Freedman et al., 2003; Miller et al.,
the specific exemplar suggesting that detection was 2003).
based on partial processing. Also, the authors The role of the occipito-temporal cortices in
showed that detection of the target was consider- natural scene categorization has been examined by
ably worse in sequences that also contained hu- Fize and co-workers (2000) in an event-related
mans as distractors, presumably because of shared fMRI study. Results seem to indicate a differential
features between animal and human scenes. Based activity in occipito-temporal cortical structures
on these findings, Evans and Treisman suggested elicited by natural target and distractor scenes
that natural scene categorization is characterized (Fize et al., 2000). Unfortunately, because of the
by two different stages of processing. A first stage low temporal resolution, fMRI cannot disentangle
involves a rapid and parallel detection of disjunc- differences in BOLD activation due to early
tive sets of unbound features of intermediate and late processing of target. However, selective
55

attention to a specific target stimulus is also re- negative deflection over temporo-occipital regions
flected at the post-perceptual level of stimulus (see Fig. 1). Furthermore, source estimation tech-
analysis. Specifically, the P3 component is a hall- niques (Current-Source-Density (CSD) and L2-
mark of selective attention. It is a positive-polarity Minimum-Norm-Estimate) suggested primary
ERP component, typically observed in a time win- sources of the early differential ERP activity in
dow between 300 and 700 ms post-stimulus and posterior, visual-associative, brain regions, and
reflects the amount of attention devoted to task- also a contribution of anterior sources to the early
relevant stimuli (Johnson, 1987; Kok, 2001). Fol- differential ERP. Also, in a time interval
lowing Kok (2001), the P3 is considered to reflect 300–600 ms after stimulus onset, target scenes
components of attention and working memory were associated with augmented late positive po-
with the ‘‘event categorization process’’ as a core tentials (LPPs) over centro-parietal sensor sites
element. Most of the previous ERP studies on the (see Fig. 1).
categorization of natural scenes used a go/no-go Together, these findings seem to indicate that
task making the late differential activity difficult to top-down influences early in processing (150 ms)
interpret because the target and non-target condi- may shape the activity in the ventral visual path-
tions were unbalanced in terms of motor activa- way during selective attention, facilitating catego-
tion (Rousselet et al., 2004; Macé et al., 2005). rization based upon target features in natural
In a recent study, we investigated cortical indi- scenes, and that later (300 ms), task relevant stim-
cators of selective attention underlying categori- uli determine larger allocation of both perceptual
zation based on target features in natural scenes and central resources compared with non-target
(Codispoti et al., 2006b). In particular, we were stimuli. Future studies should further clarify the
interested in whether the early differential ERP functional and neural meaning of these early and
activity exclusively reflects primarily neural gener- late ERPs components.
ators in the visual cortex responsible for the biased
processing of target stimuli or, additional sources
related to the generation of a biasing signal in Perception and categorization of emotional scenes
prefrontal areas. Furthermore, with regard to the
late differential ERP activity, we tested whether Typically, studies on categorization, object recog-
this component reflects target categorization and nition and selective attention employ explicit tasks
continued perceptual processing (Ritter and Ruch- where participants are asked to classify stimuli ac-
kin, 1992; Kok, 2001), which would suggest the cording to verbal instructions. A special type of
presence of neural generators in higher order vis- categorization is represented by the processing of
ual processing areas. emotional stimuli. In fact, to the extent that nat-
As in previous studies by Thorpe and associates ural images represent motivationally relevant cues,
(1996), participants had to categorize images ac- they are able to catch attentional resources and to
cording to whether or not they contained an an- be therefore quickly categorized (Lang et al.,
imal. However, our study did not require a go/no- 1997).
go response and instead subjects responded as to Emotional processing has been widely investi-
whether an animal was present in the image or not gated by presenting affective pictures that are
(De Cesarei et al., 2006) using a two-alternative effective cues in evoking a broad range of emo-
forced-choice task. Pictures were drawn from a tional reactions, varying in intensity, and involving
large database containing 1200 exemplars with the both pleasant and unpleasant affect (Lang et al.,
target and non-target pictures occurring with 1997). For instance, skin conductance responses
equal probability; each picture was shown for are larger when viewing emotionally arousing
24 ms and an inter-stimulus interval of 3–4 s. (pleasant or unpleasant) pictures, compared to
Replicating previous findings, the early differ- neutral pictures, whereas heart rate varies with
ential ERP activity appeared as a positive deflec- affective valence, with more deceleratory heart
tion over fronto-central sensor sites and as a rate responses elicited when viewing unpleasant,
56

Fig. 1. ERP waveforms showing early brain potentials over occipito-temporal (PO8) and frontal (F4) sites and late positive potentials
over a parietal site (Pz) during explicit categorization of natural scenes. Scalp potential maps also reveal the topography of the early
and late differential ERP activity (Target minus distractor).

relative to neutral pictures (Bradley et al., 2001). In picture stimuli (Lang et al., 1990). Specifically, the
addition, it has been shown that affective pictures startle reflex is larger when people view unpleasant
are effective cues in evoking not only autonomic rather than pleasant pictures (Vrana et al., 1988;
responses but also a broad range of neuroendo- Bradley et al., 2006). Consistent with the motiva-
crine changes (Codispoti et al., 2003). tional hypothesis, it has been shown that the
The emotional context, elicited by photographic strongest reports of emotional arousal, largest skin
slides, also affects the startle reflex. It is well es- conductance responses, and greatest modulation
tablished that the magnitude of the blink response of the startle reflex occurs when participants view
to a startling acoustic or visual probe varies ac- pictures depicting threats, mutilated bodies and
cording to the affective valence of the foreground erotica (Bradley et al., 2001, 2006).
57

In a number of recent experimental findings, (masked condition). The masking stimulus was a
ERPs were shown to be modulated by emotionally scrambled image. In the unmasked condition,
significant stimuli. In particular emotionally affective modulation in terms of subjective and
arousing (pleasant and unpleasant) pictures elicit cortical reactions was only slightly affected by the
larger LPPs then neutral stimuli in a window from pictures’ exposure time, while heart rate modula-
400 to 800 ms after picture onset (Cacioppo et al., tion appeared to rely on the presence of the stim-
1994; Cuthbert et al., 2000; Schupp et al., 2000) ulus. Moreover, in the masked condition,
and occipital and posterior parietal regions were subjective and cortical reactions were modulated
suggested to be a possible origin of the arousal- by the affective content of the stimulus even
modulated late positive wave (Keil et al., 2002). when pictures were presented very briefly
This effect has been linked to the concept of mo- (450 ms), whereas longer exposure times were
tivated attention, which proposes that motivation- needed to observe autonomic changes as a func-
ally significant stimuli are selectively processed tion of picture content. Taken together, these find-
because they naturally engage attentional re- ings indicate that stimulus categorization, reflected
sources (Lang et al., 1997). in cortical and subjective changes, is elicited even
Furthermore, recent studies also found an early when affective scenes are briefly presented. Differ-
selective processing of affective cues, in which entially, autonomic changes probably reflecting
emotional pictures prompt less positivity than stimulus intake and preparation for action are not
neutral pictures over occipito-temporal sites start- activated when the emotional impact of the stim-
ing at 150 ms after picture onset and lasting for ulus is reduced as a consequence of short picture
about 100 ms (Schupp et al., 2004). The affective exposure time.
modulation of this early time interval has been Brief presentation might be considered analo-
interpreted as indication that the emotional con- gous to a distant (rather than imminent) predator
tent of visual cues facilitates the sensory encoding or prey that should determine less intense appeti-
of these stimuli (Schupp et al., 2003). tive and defensive activation leading to less auto-
Moreover, affective modulation of early and nomic changes associated with preparation for
late ERP components do not rely on voluntary action (Lang et al., 1997). Consistent with the
evaluation of the hedonic content (Cuthbert et al., prediction of larger emotional response to proxi-
1995; Codispoti et al., 1998; Junghoefer et al., mal stimuli, a study comparing the reaction of
2001; Keil et al., 2002; Schupp et al., 2003). For snake-phobic participants to snakes presented at
instance, Cuthbert et al. (1995) compared LPP various distances (Teghtsoonian and Frost, 1982)
during passive viewing or when an evaluative rat- showed a linear increase of autonomic responses
ing task was required, and found similar modula- and self-reported fear as a function of distance.
tion, suggesting that affective evaluation, as Since distance and retinal size are strictly related
measured by the LPP, may be a relatively oblig- (Loftus and Harley, 2005), it can be expected that
atory process. changes in stimulus size determine arousal modu-
These cortical and autonomic changes during lations similarly to distance. Moreover, in an ev-
affective picture processing are obtained when olutionary framework, the physical size of an
participants view pictures for a sustained time pe- encountered object or organism may determine the
riod (e.g., 6 s). In a series of studies, we investigated motivational relevance for the observer. This pos-
whether brief presentations are also able to engage sibility is supported by the results of a recent study
the defensive and appetitive motivational systems (Reeves et al., 1999) which investigated autonomic
that mediate emotional responding (Codispoti response following arousing and non arousing
et al., 2001, 2002). In particular, pictures were stimuli presented in different sizes, suggesting a
presented briefly for different exposure times (from more pronounced emotional response for bigger
25 ms to 6 s) and in one condition were followed by stimuli.
a blank screen (unmasked condition), while in an- Recently, we assessed the possibility that chan-
other condition were followed by masking stimulus ges in stimulus size may influence the affective
58

modulation of early and late ERPs (De Cesarei and Explicit categorization and affective modulation:
Codispoti, 2006). Assuming that size reduction de- emotional scenes as distractors
termines lower relevance of the scene to the ob-
server, a reduction in affective modulation at both In everyday life, while monitoring the environ-
early and late stages of processing was expected ment, we evaluate the relevant stimuli, even when
for smaller compared to larger images. Alterna- not explicitly intending to do so. We make such
tively, since size reduction also results in decreased implicit categorization processes in order to un-
discriminability due to the loss of fine details in the derstand and encode the context. The process of
scene (Loftus and Harley, 2005), we expected that implicit categorization can be considered as the act
the earlier time window, reflecting stages of per- of responding differently to objects and events in
ceptual analysis, would have been more affected the environment that belong to separate classes or
by size reduction compared to the LPP, which categories, even when it is not explicitly required.
is thought to reflect processes initiated after stim- Recently, we examined whether selective atten-
ulus recognition. tion underlying categorization based on target fea-
While a decrease in the modulation of the early tures and motivated attention share similar
ERP time interval (150–300 ms) as a function of mechanisms and whether emotional modulation
stimulus content was observed following image of the LPP is reduced when affective stimuli are
size reduction, affective modulation of the LPP did presented as distractors during an explicit catego-
not change as a function of picture size, suggesting rization task (Cardinale et al., 2005). Participants
that at this stage a semantic, size-invariant repre- performed two categorization tasks in counterbal-
sentation of the stimulus has been attained (see anced order: in one condition participants were
Fig. 2). Similarly, other studies using explicit cat- asked to decide if a picture contained an animal or
egorization tasks have shown that when discrim- not, and in the other condition they had to decide
inability of the image is reduced, early modulation if the picture contained a human. The stimuli in-
in the 150–300 ms latency range is largely reduced volved three natural scene images: animal, non-
(Goffaux et al., 2003; Macè et al., 2005). living (natural landscape scenes and pictures of

Fig. 2. Early and late ERP affective modulation to pictures presented in different sizes, ranging from smallest (left end) to larger (right
end). Labels represent the ratio between the actual and the 100% stimulus size. Horizontal and vertical visual angles subtended by
images are 31  21 (S/8), 51  41 (S/4), 111  81 (S/2) and 211  161 (S/1).
59

Fig. 3. Scalp potential maps reveal similar topographies of the early and late differential ERP activity during an explicit categorization
task as well as during the perception of emotional non-target stimuli (animal target condition). Bilateral foci of occipito-temporal
negativity appeared for the differential ERP activity in the time window from 150 to 300 ms. Later in the classical P3 window
300–600 ms, target as well as emotional pictures elicit a larger positive potential compared to distractor and neutral images, which is
maximally pronounced over centro-parietal sensor sites. Illustrated is the right side view of the model head.

objects) and human (erotic couples, neutral peo- These findings seem to indicate not only that
ple, mutilated bodies). affective modulation of the LPP does not depend
Results seem to suggest that regardless of the on the evaluative nature of the task, but that is
cause of relevance of a stimulus, whether it be task present even when participants were asked to per-
relevant (Target) or inherently significant (emo- form a categorization task where pleasant, neutral
tional), similar cortical changes are involved (see and unpleasant images were distractors. Although
Fig. 3). Also, affective modulation of the LPP per- these studies suggest that emotional pictures are
sisted even when these stimuli (human pictures) automatically categorized and capture attentional
were distractors (animal target condition). Specifi- resources, it should be noted that in these studies
cally, when animal pictures were the target, they not only was the competing task undemanding,
elicited a larger LPP compared to neutral people and but the pictures had to be perceived and catego-
this positivity did not differ from the one determined rized in order to establish if they belonged to the
by emotional stimuli, suggesting that selective at- target category or not.
tention on target features and motivated attention Pessoa and co-workers (Pessoa and Ungerleider,
share similar neural mechanisms (see Fig. 3). 2004; Pessoa et al., 2005) have shown that the
Interestingly, in this study the task did not affect processing of emotional visual stimuli might be
ERP responses to animal pictures in the early time modulated by the availability of attentional re-
interval (150–300 ms), in which similar early brain sources. In a recent fMRI study, central faces with
potentials were observed across the two conditions neutral or fearful expression were presented cen-
(animal task and human task). This finding was trally along with peripheral bars. During the bar-
interpreted as a consequence of the similarities be- orientation task, subjects were asked to indicate
tween the two target categories in terms of features whether the orientation of the bars was the same
of intermediate complexity and is consistent with or not and the difficulty level of the task was ma-
recent evidence from Evans and Treisman (2005), nipulated according to the angular difference of
discussed above, showing that detection of target the bars, in order to investigate the effect of at-
in a RSVP study was considerably worse in se- tentional resources on emotionally relevant dis-
quences that also contained humans as distractors. tractor stimuli. The results showed that in the
60

amygdala, differential responses to fearful faces interval (150–300 ms) as a function of stimulus
compared to neutral faces were observed only content over occipito-temporal regions.
during low-load conditions. This findings seem to
be consistent with Lavie’s (1995; Lavie et al., 2004)
proposal that if the processing load of a target task Affective modulation and habituation
exhausts available capacity, stimuli irrelevant to
that task will not be processed. These previous findings seem to indicate that the
Similarly in a recent experiment, we manipu- affective modulation of the ERPs is an automatic
lated the perceptual load of a competing task to process that does not rely on voluntary evaluation
evaluate to what extent emotional stimuli are of the hedonic content of the stimulus.
automatically categorized as a function of availa- Along with this reasoning, several behavioural
ble resources. In this study, participants were pre- studies have found involuntary semantic process-
sented a picture (pleasant, neutral and unpleasant) ing of affective stimuli (Pratto and John, 1991;
in the left or right hemifield for 100 ms while Stenberg et al., 1998; McKenna and Sharma,
simultaneously performing a foveal task where 2004). Moreover, behavioural studies (Bargh et al.,
they were asked to detect an X or a Z presented 1996; Hermans et al., 2001) have shown that in a
alone (low perceptual load condition) or flanked priming paradigm, emotional pictures speed up the
by other letters (high load condition). Results response when prime and target share the same
indicated that early (150–300 ms) modulation of valence as compared to trials in which prime and
the ERP over occipito-temporal regions was target are of opposite valence. Similar findings
affected by the perceptual load of the foveal task. have been observed also in the absence of an
In fact, while emotional pictures elicited less explicitly evaluative context by having participants
positivity compared to neutral pictures in the low merely pronounce or make a lexical decision,
perceptual load condition, this differential activity rather then evaluate target words after briefly
disappeared in the high perceptual load condition. viewing picture primes (Wentura, 1998; Giner-So-
On the other hand, the late positive potential was rolla et al., 1999; for different findings see Storbeck
unaffected by the foveal task suggesting that the and Robinson, 2004).
LPP might index a post-perceptual stage of One important feature of an obligatory process
processing and that even in a highly demanding is resistance to habituation. There is considerable
condition emotional content of the picture is proc- evidence that prior exposure to a stimulus affects
essed by the brain. subsequent attentional processes and orienting re-
Interestingly, as discussed above, similar results sponse (Sokolov, 1963; Siddle, 1991; Bradley and
were found when the size of the images was ma- Lang, 2000), influences perceptual facilitation
nipulated (De Cesarei and Codispoti, 2006). In fact, associated with ‘‘neural suppression’’ (Tulving
picture size affected the modulation of the early and Schacter, 1990; Henson and Rugg, 2003),
ERP time interval (150–300 ms) as a function of and leads to changes in subjective ratings of pleas-
stimulus content, while affective modulation of the antness and arousal (Fechner, 1876; Zajonc, 1968;
LPP did not change as a function of picture size. Bornstein, 1989; Raymond et al., 2003; Codispoti
Taken together, these findings suggest that the et al., 2006a). Although a single repetition of a
affective modulation of the LPP, probably related stimulus facilitates subsequent recognition (e.g.,
to post-perceptual stages of processing, is not in- repetition priming), with increasing repetitions the
fluenced by perceptual factors, persists even when salience of the stimulus is reduced. Thus, habitu-
pictures are presented as distractors and attention ation can be defined as an unlearned behavioural
is manipulated by a demanding task (participants process that results in a diminution of response
were engaged in a demanding competing task). By (i.e., decreased response magnitude and/or
contrast, two different perceptual factors (the increases in response latency) to stimuli that are
perceptual load of a competing task and picture repeatedly presented (Harris, 1943; Thompson and
size) affect the modulation of the early ERP time Spencer, 1966).
61

Habituation of affective modulation has been


investigated for the first time by Bradley and col-
leagues (1993). In this experiment the same pleas-
ant, neutral and unpleasant picture stimuli were
repeatedly presented (12 repetitions for each pic-
ture). Startle reflex habituation was assessed and
compared with the habituation patterns of auto-
nomic responses (heart rate and skin conduct-
ance). Results indicated that whereas all responses
showed general habituation over trials, affective
modulation of the blink reflex was not affected by
picture repetition. As found in previous studies,
the blink response was potentiated when startle
probes were presented during processing of un-
pleasant pictures (relative to neutral stimuli) and
reduced when viewing pleasant pictures, and this
pattern persisted even after several repetitions of
the same stimulus.
Recently, we examined the affective modulation
of the late positive potential as it varied with stim-
ulus repetition (Codispoti et al., 2006a). Pleasant,
neutral and unpleasant pictures were presented up
to 60 times. If the LPP reflects automatic affective
evaluation, we expected no difference in the mod-
ulation of this component across stimulus repeti-
tions. We expected that the autonomic responses
(heart rate and skin conductance) would habituate
rapidly, providing evidence that the affective im-
pact of the stimulus was maximal early within the
habituation phase. Results showed that, although
the amplitude of the late positive potential during
picture viewing declined with stimulus repetition,
affective modulation remained intact. On the other
hand, autonomic responses habituated rapidly
with stimulus repetition (see Fig. 4). These find-
ings suggest that stimulus repetition does not Fig. 4. Late positive potential amplitude (Pz; 400–800 ms (A))
change the associative strength of connections to and skin conductance (B) changes elicited when viewing pleas-
subcortical motivational systems, as reflected in ant, neutral and unpleasant pictures during each of the three
the LPP, but does change the output to systems blocks within the habituation phase.
involved in orienting and action.
In this study, measurement of a relatively sparse sustained processing of each stimulus was the use
electrode array (3 sensors) did not provide an op- of a relatively long inter-picture interval (e.g.,
portunity to assess how picture repetition affects 10–20 s), which may have increased the ‘‘novelty’’
the early ERP component over frontal and occip- of each picture, despite its continued repetition.
ital sites, or whether picture repetition differen- Thus, we decided to further investigate affective
tially affects early and late ERP components that habituation in a new study, employing a dense
vary with picture emotionality. Also in this previ- sensor electrode array to assess early and late ERP
ous study, one factor that might have encouraged components as they varied with repetition of affec-
62

tive pictures and using a shorter inter-stimulus engaged in a competing task. This issue was fur-
interval (2–3 s) to reassess the effects of picture ther investigated in a recent experiment (Biondi et
repetition on the late positive potential (Ferrari al., 2005) where a picture flanked by two numbers
et al., 2005). This array allowed us to assess ERPs was presented for 150 ms and participants were
measured over both the occipital and frontal asked to decide whether the two numbers had the
cortex in order to assess the early ERP differences same parity, while ignoring the image. Results
reported in previous studies, as well as to assess suggest that reaction times were markedly longer
once more the late positive potential, which is for arousing pictures, compared with those which
maximal over centro-parietal sites. Consistent with were neutral, pictures, but after several repetitions
previous data (Codispoti et al., 2006a) emotionally of the same stimuli this effect vanished. Despite
arousing stimuli continued to prompt larger LPPs the fact that attention is occupied in a competing
than neutral pictures, regardless of repetition. task, affective modulation of the early and late
Nonetheless, the magnitude of the late positive ERP components persisted even after several
potential decreased somewhat for both pleasant repetitions of the same stimulus suggesting that
and unpleasant pictures following multiple repeti- categorization of unattended affective stimuli is an
tions. On the other hand, there were no effects of obligatory process that continues to occur when-
repetition on ERPs measured in the earlier time ever a sensory stimulus is presented.
window, suggesting that early effects of emotional
arousal may reflect a stimulus-driven process that
occurs automatically. Summary and future directions
Recently, Schupp and colleagues (2006) showed
similar findings using a rapid visual presentation Event-related potentials measured during picture
where pictures were presented for 330 ms without viewing vary with emotional arousal, with affec-
an inter-stimulus interval further confirming that tive (either pleasant or unpleasant), compared to
the early affective modulation of the ERP over neutral, pictures eliciting a larger LPP in the
occipito-temporal region is resistant to habituation 300–600 ms time interval over centro-parietal
even in a condition where pictures are presented regions. This component has been interpreted as
without an inter-picture interval. Interestingly, us- reflecting enhanced attention to motivationally
ing an explicit categorization task, Fabre-Thorpe relevant pictures. In addition, an early ERP com-
and co-workers showed that repetitive presenta- ponent was also reported to vary with emotional
tion of natural scenes (14 days over 3 week period) arousal in a window from about 150 to 300 ms,
did not facilitate information processing. That is, with affective, compared to neutral, stimuli,
the early (150 ms) differential ERP effect indicated prompting significantly less positivity over occi-
that the visual processing was just as fast with pito-temporal sites. Similarly, in explicit categori-
completely novel images as it was for images with zation tasks of complex natural scenes, the activity
which the subjects were highly familiar (Fabre- of target and non-target ERPs diverged sharply
Thorpe et al., 2001). around 150 ms after stimulus onset and this early
Since in our previous affective habituation stud- differential ERP activity (target minus non-target)
ies participants were only asked to look at the appeared as negative deflection over occipito-tem-
pictures, without an additional task, however, we poral regions. Furthermore, in a time interval
cannot effectively rule out the possibility that 300–600 ms after stimulus onset, target scenes were
affective modulation of the LPP results from a associated with augmented LPP over centro-pari-
process in which more attention is voluntarily etal sites suggesting that similar mechanism are
allocated to viewing affective pictures. A strategy involved in selective attention to target features
for determining whether affective modulation of and to motivationally relevant stimuli.
the LPP and its resistance to habituation is due to Several studies have shown that the affective
obligatory or voluntary processes would be to modulation of the LPP persisted even when the
present affective pictures when participants are same pictures are repeated several times, when
63

they are presented as distractors or when partic- Cardinale, R., Ferrari, V., De Cesarei, A., Biondi, S. and Codi-
ipants are engaged in a competing task, indicating spoti, M. (2005) Implicit and explicit categorization of nat-
ural scenes. Psychophysiology, 42: S41.
that categorization of affective stimuli is an oblig-
Codispoti, M., Bradley, M.M., Cuthbert, B.N., Montebarocci,
atory process. On the other hand, while the affec- O. and Lang, P.J. (1998) Stimulus complexity and affective
tive modulation of the LPP is not influenced by contents: startle reactivity over time. Psychophysiology, 35:
perceptual factors (e.g., stimulus size), these same S25.
factors strongly reduce the modulation of the early Codispoti, M., Bradley, M.M. and Lang, P.J. (2001) Affective
ERP time interval (150–300 ms) as a function of modulation for briefly presented pictures. Psychophysiology,
38: 474–478.
stimulus content. Although early and late ERP Codispoti, M., Mazzetti, M. and Bradley, M.M. (2002) Expo-
components vary with stimulus relevance, given sure time and affective modulation in picture perception.
that they are differentially affected by stimulus and Psychophysiology, 39: S62.
task manipulations, they appear to reflect different Codispoti, M., Gerra, G., Montebarocci, O., Zaimovic, A.,
Raggi, M.A. and Baldaro, B. (2003) Emotional perception
sensory and attentional processes. Future studies
and neuroendocrine changes. Psychophysiology, 40: 863–868.
should further investigate functional and neural Codispoti, M., Ferrari, V. and Bradley, M.M. (2006a) Repet-
mechanisms underlying natural scene categoriza- itive picture processing: autonomic and cortical correlates.
tion, and examine how and when motivational Brain Res., 1068: 213–220.
systems (e.g., appetitive and defensive) modulate Codispoti, M., Ferrari, V., Junghöfer, M. and Schupp, H.T.
the processing of stimulus features. Further work (2006b) The categorization of natural scenes: brain attention
networks revealed by dense sensor ERPs. Neuroimage, 31:
should also better clarify the nature of the percep- 881–890.
tual features that affect early stages of processing Cuthbert, B.N., Schupp, H.T., McManis, M., Hillman, C.,
in picture perception. Bradley, M.M. and Lang, P.J. (1995) Cortical slow waves:
emotional perception and processing. Psychophysiology, 32:
S26.
Cuthbert, B.N., Schupp, H.T., Bradley, M.M., Birbaumer, N.
References and Lang, P.J. (2000) Brain potentials in affective picture
processing: covariation with autonomic arousal and affective
Bargh, J.A., Chen, M. and Burrows, L. (1996) Automaticity of report. Biol. Psychol., 52: 95–111.
social behavior: direct effects of trait construct and stereotype De Cesarei, A., Codispoti, M., Schupp, H.T. and Stegagno, L.
activation on action. J. Pers. Soc. Psychol., 71: 230–244. (2006) Selectively attending to natural scenes after alcohol
Biondi, S., De Cesarei, A., Cardinale, R. and Codispoti, M. consumption: an ERP analysis. Biol. Psychol., 72: 35–45.
(2005) What is the fate of unattended emotional stimuli? De Cesarei, A. and Codispoti, M. (2006) When does size not
Cortical and behavioural correlates of affective habituation. matter? Effects of stimulus size on affective modulation.
Psychophysiology, 42: S36. Psychophysiology, 43: 207–215.
Bornstein, R.F. (1989) Exposure and affect: overview and meta- Delorme, A., Rousselet, G.A., Mace, M.J. and Fabre-Thorpe,
analysis of research, 1968–1987. Psychol. Bull., 106: 265–289. M. (2004) Interaction of top-down and bottom-up processing
Bradley, M.M. and Lang, P.J. (2000) Emotion and motivation. in the fast visual analysis of natural scenes. Brain Res. Cogn.
In: Cacioppo, J.T., Tassinary, L.G. and Berntson, G. (Eds.), Brain Res., 19: 103–113.
Handbook of Psychophysiology (3rd ed). Cambridge Uni- Evans, K.K. and Treisman, A. (2005) Perception of objects in
versity Press, New York. natural scenes: is it really attention free? J. Exp. Psychol.
Bradley, M.M., Lang, P.J. and Cuthbert, B.N. (1993) Emotion, Hum. Percept. Perform., 31: 1476–1492.
novelty, and the startle reflex: habituation in humans. Behav. Fabre-Thorpe, M., Delorme, A., Marlot, C. and Thorpe, S.J.
Neurosci., 107: 970–980. (2001) A limit to the speed of processing in ultra-rapid visual
Bradley, M.M., Codispoti, M., Cuthbert, B.N. and Lang, P.J. categorization of novel natural scenes. J. Cogn. Neurosci.,
(2001) Emotion and motivation I: defensive and appetitive 13: 171–180.
reactions in picture processing. Emotion, 1: 276–298. Fechner, G.T. (1876) Vorschule der Ästhetik. Breitkopf and
Bradley, M.M., Codispoti, M. and Lang, P.J. (2006) A multi- Härtel, Leipzig.
process account of startle modulation during affective per- Ferrari, V., Codispoti, M. and Bradley, M.M. (2005) Not just
ception. Psychophysiology, 43: 421–428. the same old thing: cortical and autonomic measures of
Cacioppo, J.T., Crites Jr., S.L., Gardner, W.L. and Bernston, affective habituation. Psychophysiology, 42: S55.
G.G. (1994) Bioelectrical echoes from evaluative categoriza- Fize, D., Boulanouar, K., Chatel, Y., Ranjeva, J.P., Fabre-
tions: I. A late positive brain potential that varies as a func- Thorpe, M. and Thorpe, S.J. (2000) Brain areas involved in
tion of trait negativity and extremity. J. Pers. Soc. Psychol., rapid categorization of natural images: an event-related
67: 115–125. fMRI study. Neuroimage, 11: 634–643.
64

Freedman, D.J., Riesenhuber, M., Poggio, T. and Miller, E.K. Loftus, G.R. and Harley, E.M. (2005) Why is it easier to iden-
(2003) A comparison of primate prefrontal and inferior tem- tify someone close than far away? Psychon. Bull. Rev., 12(1):
poral cortices during visual categorization. J. Neurosci., 23: 43–65.
5235–5246. Macé, M.J.M., Thorpe, S.J. and Fabre-Thorpe, M. (2005)
Giner-Sorolla, R., Garcia, M.T. and Bargh, J.A. (1999) The Rapid categorization of achromatic natural scenes: how ro-
automatic evaluation of pictures. Social Cogn., 17: 76–96. bust at very low contrasts? Eur. J. Neurosci., 21: 2007–2018.
Goffaux, V., Gauthier, I. and Rossion, B. (2003) Spatial scale McKenna, F.P. and Sharma, D. (2004) Reversing the emotional
contribution to early visual differences between face and ob- Stroop effect: the role of fast and slow components. J. Exp.
ject processing. Cogn. Brain Res., 16: 416–424. Psychol. Learn. Mem. Cogn., 30: 382–392.
Gross, C.G. (1973) Visual functions of the inferotemporal cor- Miller, E.K., Nieder, A., Freedman, D.J. and Wallis, J.D.
tex. In: Austrum, H., Jung, R., Loewenstein, W.R., Mackay, (2003) Neural correlates of categories and concepts. Curr.
D.M. and Teuber, H.L. (Eds.) Handbook of Sensory Phys- Opin. Neurobiol., 13: 198–203.
iology: Central Processing of Visual Information, Vol. 7. O’Regan, J.K., Rensink, R.A. and Clark, J.J. (1999) Change-
Springer, Berlin, pp. 451–482. blindness as a result of ‘mudsplashes’. Nature, 398: 6722–6734.
Harris, J.D. (1943) Habituatory response decrement in the in- Pessoa, L., Padmala, S. and Morland, T. (2005) Fate of unat-
tact organism. Psychol. Bull., 40: 385–422. tended fearful faces in the amygdala is determined by both
Henson, R.N. and Rugg, M.D. (2003) Neural response sup- attentional resources and cognitive modulation. Neuroimage,
pression, haemodynamic repetition effects, and behavioural 28: 249–255.
priming. Neuropsychologia, 41: 263–270. Pessoa, L. and Ungerleider, L.G. (2004) Neural correlates of
Hermans, D., Houwer, J. and Eelen, P. (2001) A time course change detection and change blindness in a working memory
analysis of the affective priming effect. Cogn. Emotion, 15: task. Cereb. Cortex, 14: 511–520.
143–165. Potter, M.C. (1975) Meaning in visual search. Science, 187:
Hinojosa, J.A., Martin-Loeches, M., Gomez-jarabo, G. and 965–966.
Rubia, F.J. (2000) Common basal extrastriate areas for the Potter, M.C. (1976) Short-term conceptual memory for pic-
semantic processing of words and pictures. Clin. Ne- tures. J. Exp. Psychol. Learn. Mem. Cogn., 2: 509–522.
urophysiol., 111: 552–560. Pratto, F. and John, O.P. (1991) Automatic vigilance: the at-
Johnson Jr., R. (1987) The amplitude of the P300 component of tention-grabbing power of negative social information. J.
the event-related potential: review and synthesis. In: Ackles, Pers. Soc. Psychol., 63: 380–391.
P.K., Jennings, J.R. and Coles, M.G.H. (Eds.) Advance in Raymond, J.E., Fenske, M.J. and Tavassoli, N.T. (2003) Se-
Psychophysiology, Vol. III. JAI Press, Greenwich, CT. lective attention determines emotional responses to novel
Junghoefer, M., Bradley, M.M., Elbert, T.R. and Lang, P.J. visual stimuli. Psychol. Sci., 14: 537–542.
(2001) Fleeting images: a new look at early emotion discrim- Reeves, B., Lang, A., Kim, E.Y. and Tatar, D. (1999) The
ination. Psychophysiology, 38: 175–178. effects of screen size and message on attention and arousal.
Keil, A., Bradley, M.M., Hauk, O., Rockstroh, B., Elbert, T. Media Pyschol., 1: 49–67.
and Lang, P.J. (2002) Large-scale neural correlates of affec- Riesenhuber, M. and Poggio, T. (2002) Neural mechanisms of
tive picture processing. Psychophysiology, 39: 641–649. object recognition. Curr. Opin. Neurobiol., 12: 162–168.
Kirchner, H. and Thorpe, S.J. (2005) Ultra-rapid object detec- Ritter, W. and Ruchkin, D. (1992) A review of event-related
tion with saccadic eye movements: visual processing speed potential components discovered in the context of studying
revisited. Vision Res., 46: 1762–1776. the P3. In: Friedman, D. and Bruder, G. (Eds.), Psycho-
Kok, A. (2001) On the utility of P3 amplitude as a measure of physiology and Experimental Psychopathology. The New-
processing capacity. Psychophysiology, 38: 557–577. York Academy of Sciences, NewYork, pp. 1–32.
Lang, P.J., Bradley, M.M. and Cuthbert, M.M. (1990) Emotion, Rousselet, G.A., Thorpe, S.J. and Fabre-Thorpe, M. (2004)
attention, and the startle reflex. Psychol. Rev., 97: 377–395. How parallel is visual processing in the ventral pathway?
Lang, P.J., Bradley, M.M. and Cuthbert, M.M. (1997) Moti- Trends Cogn. Sci., 8: 363–370.
vated attention: affect, activation and action. In: Lang, P.J., Rudell, A.P. and Hua, J. (1997) The recognition potential, word
Simons, R.F. and Balaban, M.T. (Eds.), Attention and Ori- difficulty, and individual reading ability: on using event-re-
enting: Sensory and Motivational Processes. Lawrence Er- lated potentials to study perception. J. Exp. Psychol. Hum.
lbaum Associates, Inc., Hillsdale, NJ, pp. 97–135. Percept. Perform., 23: 1170–1195.
Lavie, N. (1995) Perceptual load as a necessary condition for Rudell, A.P. (1991) The recognition potential contrasted with
selective attention. J. Exp. Psychol. Hum. Percept. Perform., the P300. Int. J. Neurosci., 60: 85–111.
21: 451–468. Rudell, A.P. (1992) Rapid stream stimulation and the recog-
Lavie, N., Hirst, A., Defockert, J.W. and Viding, E. (2004) nition potential. Electroencephalogr. Clin. Neurophysiol., 83:
Load theory of selective attention and cognitive control. J. 77–82.
Exp. Psychol. Gen., 133: 339–354. Schupp, H.T., Cuthbert, B.N., Bradley, M.M., Cacioppo, J.T.,
Li, F.F., VanRullen, R., Koch, C. and Perona, P. (2002) Rapid Ito, T. and Lang, P.J. (2000) Affective picture processing: the
natural scene categorization in the near absence of attention. late positive potential is modulated by motivational rele-
Proc. Natl. Acad. Sci. USA, 99: 9596–9601. vance. Psychophysiology, 37: 257–261.
65

Schupp, H.T., Junghofer, M., Weike, A.I. and Hamm, A.O. Thompson, R.F. and Spencer, W.A. (1966) Habituation: a
(2003) Attention and emotion: an ERP analysis of facilitated model phenomenon for the study of neuronal substrates of
emotional stimulus processing. Neuroreport, 14: 1107–1110. behavior. Psychol. Rev., 73: 16–43.
Schupp, H.T., Junghofer, M., Weike, A.I. and Hamm, A.O. Thorpe, S., Fize, D. and Marlot, C. (1996) Speed of processing
(2004) The selective processing of briefly presented affective in the human visual system. Nature, 381: 520–522.
pictures: an ERP analysis. Psychophysiology, 41: 441–449. Treisman, A. and Gelade, G. (1980) A feature integration the-
Schupp, H.T., Stockburger, J., Codispoti, M., Junghofer, M., ory of attention. Cogn. Psychol., 12: 97–136.
Weike, A.I. and Hamm, A.O. (2006) Stimulus novelty and Tulving, E. and Schacter, D.L. (1990) Priming and human
emotion perception: the near absence of habituation in the memory systems. Science, 247: 301–306.
visual cortex. Neuroreport, 17: 365–369. VanRullen, R. and Thorpe, S.J. (2001) The time course of vis-
Siddle, D.A. (1991) Orienting, habituation, and resource alloca- ual processing: from early perception to decision-making. J.
tion: an associative analysis. Psychophysiology, 28: 245–259. Cogn. Neurosci., 13: 454–461.
Sokolov, E.N. (1963) Perception and the Conditioned Reflex. Vrana, S., Spence, E.L. and Lang, P.J. (1988) The startle probe
Macmillan, New York. response: a new measure of emotion? J. Abnorm. Psychol.,
Stenberg, G., Wilking, S. and Dahl, M. (1998) Judging words at 97: 487–491.
face value: Interference in a word processing task reveals Wentura, D. (1998) Affektives Priming in der Wo-
automatic processing of affective facial expressions. Cogn. rtentscheidungsaufgabe: Evidenz fur postlexikalische Urteils-
Emotion, 12: 755–782. tendenzen [Affective priming in the lexical decision task:
Storbeck, J. and Robinson, M.D. (2004) When preferences need evidence for post-lexical judgment tendencies]. Sprache und
inferences: a direct comparison of the automaticity of cognitive Kognition, 17: 125–137.
versus affective priming. Pers. Soc. Psychol. Bull., 30: 81–93. Wolfe, J.M. and Bennett, S.C. (1997) Preattentive object files:
Teghtsoonian, R. and Frost, R.O. (1982) The effect of viewing shapeless bundles of basic features. Vision Res., 37: 25–43.
distance on fear of snakes. J. Behav. Ther. Exp. Psychiatry, Zajonc, R.B. (1968) Attitudinal effects of mere exposure, J.
13(3): 181–190. Pers. Soc. Psychol., 9: 1–27.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 4

Dynamics of emotional effects on spatial attention in


the human visual cortex

Gilles Pourtois1,2, and Patrik Vuilleumier1,2

1
Neurology & Imaging of Cognition, Clinic of Neurology, University Hospital & Department of Neurosciences,
University Medical Center, University of Geneva, Geneva, Switzerland
2
Swiss Center for Affective Sciences, University of Geneva, Geneva, Switzerland

Abstract: An efficient detection of threat is crucial for survival and requires an appropriate allocation of
attentional resources toward the location of potential danger. Recent neuroimaging studies have begun to
uncover the brain machinery underlying the reflexive prioritization of spatial attention to locations of
threat-related stimuli. Here, we review functional brain imaging experiments using event-related potentials
(ERPs) and functional magnetic resonance imaging (fMRI) in a dot-probe paradigm with emotional face
cues, in which we investigated the spatio-temporal dynamics of attentional orienting to a visual target when
the latter is preceded by either a fearful or happy face, at the same (valid) location or at a different (invalid)
location in visual periphery. ERP results indicate that fearful faces can bias spatial attention toward threat-
related location, and enhance the amplitude of the early exogenous visual P1 activity generated within the
extrastriate cortex in response to a target following a valid rather than invalid fearful face. Furthermore,
this gain control mechanism in extrastriate cortex (at 130–150 ms) is preceded by an earlier modulation of
activity in posterior parietal regions (at 40–80 ms) that may provide a critical source of top-down signals on
visual cortex. Happy faces produced no modulation of ERPs in extrastriate and parietal cortex. fMRI data
also show increased responses in the occipital visual cortex for valid relative to invalid targets following
fearful faces, but in addition reveal significant decreases in intraparietal cortex and increases in orbito-
frontal cortex when targets are preceded by an invalid fearful face, suggesting that negative emotional
stimuli may not only draw but also hold spatial attention more strongly than neutral or positive stimuli.
These data confirm that threat may act as a powerful exogenous cue and trigger reflexive shifts in spatial
attention toward its location, through a rapid temporal sequence of neural events in parietal and temporo-
occipital areas, with dissociable neural substrates for engagement benefits in attention affecting activity in
extrastriate occipital areas and increased disengagement costs affecting intraparietal cortex. These brain-
imaging results reveal how emotional signals related to threat can play an important role in modulating
spatial attention to afford flexible perception and action.

Introduction Accordingly, emotional processes may be inti-


mately linked to action tendencies (Lang, 1979;
Negative emotions such as fear or anger imply both Frijda, 1986) that can be mapped onto either a de-
the appraisal of the presence of some nuisance, and fensive motivational system (for negative emotions)
the elicitation of appropriate behavioral reactions. or an appetitive motivational system (for pleasant
emotions). Heightened vigilance and enhanced
Corresponding author. Tel.: +41-22-372-8478/379-5381; Fax: allocation of cognitive resources toward motiva-
+41-22-379-5402; E-mail: gilles.pourtois@medecine.unige.ch tionally relevant information in the environment,

DOI: 10.1016/S0079-6123(06)56004-2 67
68

particularly when potentially threatening, is cer- Navon and Margalit, 1983), subjects are required
tainly among the most crucial adaptive aspects of to detect a neutral visual target (i.e., a dot or
emotional processing (Frijda, 1986; Eysenck, 1992; probe) whose location is uncertain from trial to
Cacioppo and Gardner, 1999), presumably shaped trial but unpredictably preceded by another
during evolution across many species to promote emotionally significant stimulus (i.e., a cue). Crit-
adequate actions and behaviors in the service of ically, this emotional cue can appear at either
survival (Ohman and Mineka, 2001). In humans, the same location as the subsequent target (valid
abundant empirical evidence has converged from trial) or at another location in the visual field
different fields in clinical psychology and cognitive (invalid trials), such that if the emotional value of
sciences to indicate that processing of threat-related the cue can exert some capture on the spatial dis-
stimuli can exert strong influences on attentional tribution of attention, it will lead to a facilitation
mechanisms (Fox, 2002; Vuilleumier et al., 2004a), of target processing in the former (valid) situa-
which constitute one of the most central cognitive tion and/or to a distraction from target process-
abilities controlling perception and action. Fur- ing in the latter (invalid) situation (Mogg and
thermore, some of these interactions of emotions Bradley, 1999b). Accordingly, such effects on spa-
with attentional and cognitive functions are liable tial orienting have been documented by several
to important modulations by individual or contex- behavioral studies using a variety of threat-related
tual factors, including for instance anxiety, person- cues including pictures of aversive scenes (Mogg
ality traits, learning, or goals and values. Thus, and Bradley, 1998; Mogg et al., 2000), aversive
studies of emotion and attention interactions can words (MacLeod et al., 1986; Mogg et al., 1994),
provide many precious insights into the basic func- emotionally negative faces (Mogg and Bradley,
tional architecture of the human mind and more 1999b; Bradley et al., 2000), or aversely condi-
subtle individual differences. tioned stimuli (Armony and Dolan, 2002; Koster
Here, we review our recent brain imaging work et al., 2004b, 2005). Moreover, these effects have
using event-related potentials (ERPs) and func- often been found to vary as a function of individ-
tional resonance imaging (fMRI) that has at- ual anxiety levels, even in people without clinical
tempted to uncover the neuroanatomical substrates manifestations (Fox et al., 2001; Koster et al.,
and temporal dynamics underlying the regulation of 2004b). However, behavioral studies have provided
attentional resources by emotional processes. In somewhat conflicting results as to whether the
particular, we will review recent data concerning the emotional effects on spatial attention in the dot-
networks and time-course of neural activity in the probe task (and the influence of anxiety on these
healthy human brain that may control the spatial processes) might primarily result from a facilitation
distribution of visual attention in response to threat- of orienting to the targets following valid emo-
related cues in the peripheral visual field (Fox, 2002; tional cues (i.e., via shifting and engaging proc-
Vuilleumier, 2002). We used electrical and hemody- esses), or conversely from an interference on
namic brain-imaging approaches during a very sim- reorienting to the targets following invalid emo-
ple, well-standardized experimental paradigm (i.e., tional cues (i.e., on disengaging processes) (Fox
the dot-probe task, Bradley et al., 2000), allowing us et al., 2001; Koster et al., 2004a; Mathews et al.,
to explore the sequence of neural events at stake 2004; see Fig. 1B).
during the different processing stages necessary to In our studies, by combining ERPs and fMRI
orient spatial attention toward behaviorally relevant during the dot-probe task, we could distinguish
stimuli (e.g., shift, engage, and disengage Posner between brain responses to the cues themselves
et al., 1980), and thus enabling us to examine how and their potential effect on the subsequent target
emotional signals of threat might influence these as a function of cue validity, and thus show dis-
different attentional mechanisms. tinct emotional influences on orienting and reori-
In this classical dot-probe paradigm (Fig. 1A), enting stages in attention (Fox et al., 2001). While
derived from the spatial orienting task originally ERPs provide brain-imaging measures with a high
designed by Posner et al. (Posner et al., 1980; temporal resolution allowing the registration of
69

Fig. 1. (A) Example of a valid (left panel) and an invalid (right panel) trial used in classical dot-probe tasks (Mogg and Bradley,
1999b). The target (here a dot probe) can unpredictably appear at one of two possible locations, either replacing (after a short-time
interval) the location previously occupied by the emotional stimulus (the so-called valid, here a schematic angry face) or replacing
another neutral position (the so-called invalid, here a schematic neutral face). Subjects are required to indicate the position (left or
right) of the dot and to ignore the face display. (B) Typical behavioral results obtained in dot-probe tasks (see Bradley et al., 2000; Fox
et al., 2001), showing either a facilitation in RTs for valid-fear relative to invalid fear trials (with low-anxious participants, reflecting an
engagement benefit, left panel) or conversely an interference in RTs for invalid fear compared to valid fear trials (with high-anxious
participants, reflecting a disengagement cost, right panel). No such modulations are reported with positive/happy cues.

neural activity from the scalp on a millisecond temporal dynamics of the cascade of successive
time-scale, which is optimal to separate cue and stages implicated during spatial orienting. Taken
target processing, fMRI provides a measure with a together, our imaging data converge to suggest
much poorer temporal resolution of the order of (i) that the deployment of spatial attention toward
1–4 s but with excellent millimetric anatomical res- threat signals (here conveyed by fearful faces) can
olution. In addition, we also used modern source occur rapidly and produce an enhanced sen-
localization techniques allowing us to assess the sory response to targets within the extrastriate
70

visual cortex (at o200 ms); (ii) that this effect on attention might be more frequently found in
might result from a direct modulation of visual visual search tasks (Juth et al., 2005; Williams
cortex by emotional inputs which can facilitate et al., 2005) than in dot-probe orienting tasks,
spatial orienting toward the target, through mech- possibly reflecting some intrinsic differences in at-
anisms controlled by the posterior parietal cortex tentional processes engaged in these two situations.
and activated prior to the enhancement of target Orienting to dot probes typically involves a single
processing (at o100 ms); and (iii) that activation and rapid shift of attention toward brief stimuli
of these neural circuits by invalid emotional cues that are presented in peripheral visual field;
can produce a cost in disengaging attention toward whereas visual search requires serial shifts of at-
a target presented at another location, by reducing tention among stimuli that can be processed by
the activity of orienting mechanisms within poste- foveal vision and remain visible throughout the
rior parietal cortex, but at the same time inducing task, potentially recruiting other exploratory strat-
concomitant increases in the response of other egies, including eye movement control.
brain regions within the ventromedial prefrontal In the dot-probe task, emotional stimuli are
cortex. These data therefore indicate that neural generally irrelevant to the task, and subjects are
mechanisms responsible for orienting spatial at- instructed to respond only to the dot-probe target
tention toward threat-related stimuli may partly that can unpredictably appear at one of two
overlap with those controlling shifts of spatial possible locations (Fig. 1A), either replacing the
attention through nonemotional (neutral) cues, location previously occupied by the emotional
but also involve partly distinct mechanisms. We stimulus (the so-called valid) or replacing another
discuss these brain-imaging observations in a more neutral position (the so-called invalid). In the in-
general framework (Vuilleumier, 2005) proposing itial study using this paradigm (MacLeod et al.,
that emotion and attention may operate through 1986), words (one threatening, one neutral) were
parallel neural pathways that can exert additive presented at two separated spatial locations (in
modulatory influences on sensory processing in the the upper and lower visual field) and followed
visual cortex. after a short-time interval by a small dot probe at
the spatial location of one of the two words. Par-
ticipants were asked to respond as fast as possible
Emotion and spatial attention to the dot probe. Results indicated that people
with higher anxiety level directed their attention
A facilitation of spatial attentional orienting by more readily to the location of threat-related
threat signals has been documented in a variety of words, suggesting a facilitated detection of threat
behavioral tasks (for review see Vuilleumier, 2002), with anxiety (Eysenck, 1992). Subsequent variants
including not only covert orienting in dot-probe of this paradigm (Fox, 1993; Bradley et al., 1997)
tasks where emotional stimuli can act as exogenous have used different types of cues to examine emo-
spatial cues (MacLeod et al., 1986; Mogg and tional influences on spatial attention in both
Bradley, 1999b; Bradley et al., 2000; Fox et al., (subclinical) anxious and nonanxious individuals
2001), but also visual search tasks where emotion- (see Fox, 2002 for an overview). In many studies,
ally negative targets are detected faster than neutral stimuli have been presented laterally in either vis-
or positive targets (Hansen and Hansen, 1988; Fox ual hemifield (rather than vertically), but only a
et al., 2000; Eastwood et al., 2001; Ohman et al., few have found reliable differences between hemi-
2001). In many studies, these emotional effects on fields (Mogg and Bradley, 1999a; Hartikainen
attention were obtained with negative stimuli, such et al., 2000; Fox, 2002). Remarkably, similar
as fearful or angry faces, whereas positive stimuli emotional biases in spatial attention have been
produced less consistent effects, suggesting a spe- observed with different types of targets and
cial role of threat-related cues in neural systems tasks, including simple detection or more complex
governing emotion and attention interactions. discrimination tasks (see Mogg and Bradley,
Moreover, some effects of positive emotional cues 1999b; Bradley et al., 2000).
71

The dot-probe task can thus provide a useful have shown that their detection of stimuli in the
snapshot of attentional allocation toward emotional contralesional (neglected) visual field is better for
stimuli by computing the difference in reaction emotional than for neutral pictures (Vuilleumier
times and/or accuracy rate for dot-probe targets and Schwartz, 2001a,b; Vuilleumier et al., 2002;
appearing at invalid minus valid locations (Bradley Fox, 2002). Note that contralesional deficits in
et al., 2000). Note, however, that the ‘‘attentional these patients were reduced but not abolished for
bias’’ measure obtained by this invalid–valid differ- emotional stimuli, indicating that spatial attention
ence might in principle reflect either faster orienting was facilitated toward these contralesional events,
to targets at valid locations, or slower reorienting to yet still necessary and clearly compromised to
targets at invalid locations, or both (Fox et al., afford normal awareness. Importantly, these re-
2001; see Fig. 1B). In several studies (Fox et al., sults suggest that spatial attention can be captured
2001; Koster et al., 2004a; Mathews et al., 2004), by contralesional emotional stimuli despite a pro-
target detection was found to be slowed by threat found impairment in directing attention toward
cues more than by neutral or positive cues presented that side. This may already indicate at least a par-
at an invalid location, even though threat produced tial dissociation between brain mechanisms re-
no advantage over neutral or positive cues when the sponsible for controlling spatial attention in
targets were presented at valid locations, suggesting response to emotional vs. nonemotional cues,
that threat stimuli may not necessarily act by such that damage to spatial attention networks
attracting attention to their own location, but rather in fronto-parietal areas leading to neglect (see
by influencing the disengagement of attention from Driver and Vuilleumier, 2001; Kerkhoff, 2001;
current fixation. However, other findings demon- Corbetta and Shulman, 2002) may still leave intact
strate a substantial facilitation in visual processing some effects on attention that are triggered by
for targets presented at locations validly cued by an emotionally charged cues (through their valence
emotionally negative stimulus (Phelps et al., 2006). and/or arousal properties, see Lang et al., 1998;
Furthermore, the fact that such attentional biases Sabatinelli et al., 2005).
can arise in response to task-irrelevant emotional Emotional signals may also influence atten-
stimuli also suggests that these effects may operate tion during Stroop tasks (Pratto and John, 1991;
in a remarkably involuntary manner, even when Williams et al., 1996) and attentional blink exper-
potentially harmful for optimal performance (e.g., iments (Anderson, 2005). These nonspatial effects
on invalid trials). Some studies reported significant are beyond the scope of the present review (see
biases under conditions where emotional cues Vuilleumier et al., 2004a), but also indicate a pri-
(faces) were masked and rendered invisible (Mogg oritization of processing for emotionally relevant
et al., 1994, 1995; Mogg and Bradley, 1999a). In information.
our own experiments, similar effects were often
observed in subjects who failed to note any emo-
tional features in facial cues presented in their visual Time-course of spatial orienting to threat locations
periphery (Pourtois et al., 2004). Further behavioral
studies are needed to investigate whether some In a recent ERP study (Pourtois et al., 2004), we
changes in the strategic set to respond to peripheral used a modified version of the dot-probe task
events vs. to focus on central stimuli may affect the (adapted from Mogg et al., 1994) in normal (non-
degree of involuntary spatial orienting to emotional anxious) adult participants while we recorded
cues. However, some studies have reported signifi- high-density EEG (Fig. 2B) to track the time-
cant interference by peripheral emotional flankers in course of spatial orienting toward the location of
tasks where subjects always responded to central emotional stimuli. Cues were faces with fearful,
stimuli only (Fenske and Eastwood, 2003). happy, or neutral expression, appearing briefly
Similarly, neuropsychological studies in brain- prior to a single neutral target presented at the
damaged patients with impaired spatial attention same location as one of these faces (see Fig. 2A).
and hemi-neglect following parietal-lobe lesions On each trial, two faces were first shown together,
72

Fig. 2. (A) Dot-probe paradigm used in our event-related potential (ERP) and functional resonance imaging (fMRI) experiment,
showing the sequence of events within a trial. A bar-probe (target) could unpredictably appear in the upper visual field on either side,
following a pair of faces in which one had a neutral expression and one an emotional expression (fearful or happy). The face-target
interval was randomly jittered between 100 and 300 ms. At the time of target onset, the vertical or horizontal segment of the fixation
cross was slightly thickened to indicate the relevant orientation to be monitored for targets. Participants had to judge the orientation of
the peripheral target relative to the thick cross segment. Faces were by themselves always irrelevant to the task. (B) Spatial layout of the
62 sites used in our ERP experiment (Pourtois et al., 2004), positioned according to the extended international 10–20 System
(Oostenveld and Praamstra, 2001). Two positions over the occipital lobe are marked: a right occipito-temporal electrode (PO8 where
the P1 was maximal, in black) and an occipito-parietal electrode along the midline (POZ where the negative C1 to upper visual stimuli
was maximal, in gray). (C) Grand average ERPs time-locked to the onset of the face display (fearful and happy faces collapsed). A
clear C1, lateral occipital P1 and occipito-parietal N1 components were recorded (the C1 was larger for fearful compared to happy
faces, data not shown). (D) Grand average ERPs time-locked to the onset of the bar probe (all conditions collapsed). Again, a
conspicuous C1, lateral occipital P1 and occipito-parietal N1 components were recorded.

for a duration of 100 ms, one in the left visual field by one of the faces (Fig. 2A). All stimuli (faces
(LVF) and one in the right visual field (RVF), one and bar-probe) were presented in the upper visual
neutral and one with an emotional expression field to allow us to measure early retinotopic re-
(fearful or happy, Ekman and Friesen, 1976). The sponses in ERPs (Jeffreys and Axford, 1972; Clark
faces were then replaced by a small bar-probe (du- et al., 1995). Participants were asked to perform
ration of 150 ms), oriented either vertically or hor- a go/no-go matching task in which they had
izontally, appearing at the position just occupied to judge, on each trial, whether the orientation of
73

the bar-probe (in the LVR or RVF) matched that sensory gain or enhanced vigilance mechanisms
of the thicker line-segment within the fixation (Heinze et al., 1990; Luck, 1995; Hopfinger and
cross. The task was to press a button only when Mangun, 1998; Hillyard et al., 1998; Carrasco
the bar orientation was the same as the thicker line et al., 2000; Keil et al., 2005).
of the cross (infrequent go trials), but to withhold Behaviorally, our modified version of the dot-
responses otherwise (more frequent no-go trials). probe paradigm was successful to trigger ex-
This task ensured that participants maintained ogenous shifts of spatial attention toward the side
their gaze on the central fixation cross (as con- of the negative face cues (Mogg et al., 1994; Mogg
firmed subsequently by eye-tracking data, see be- and Bradley, 1999b; Bradley et al., 2000). Partic-
low) and that all visual inputs were indeed ipants showed a better discrimination of bar ori-
restricted to the upper (peripheral) visual field. entation when the latter appeared in valid rather
This also allowed us to record ERPs uncontami- than invalid locations, as demonstrated by higher
nated by any motor-related activity, since only d0 values from signal detection theory (Green and
ERPs for no-go trials were analyzed (Fig. 2C, D). Swets, 1966), and this spatial validity effect was
Critically, the bar-probe could appear either significantly greater for fearful than happy faces
on the side of the emotional face (valid condition) (see Mogg and Bradley, 1999b).
or on the side of the neutral face (invalid condi- Conventional analyses (Picton et al., 2000) on
tion), in an unpredictable (50% valid and 50% the exogenous visual ERPs confirmed that fearful
invalid) and randomized manner. However, faces faces (but not happy faces) significantly modulated
were entirely irrelevant to the participants’ task. the early sensory processing of bar-probes appear-
Thus, spatial validity of the target-bar was arbi- ing at the same location. The lateral occipital
trarily defined by the position of the preceding P1 component peaking at 135 ms post-stimulus
emotion face expression. Moreover, since partici- onset (Heinze et al., 1990; Luck et al., 1990) was
pants had to fixate the central cross, emotional significantly enhanced when the target-bar re-
cues appeared in a truly unattended location, placed a valid fearful face as compared with an
which allowed us to properly assess any spatial invalid neutral face (Fig. 3A–C), even though the
biases in distribution of attention to peripherally bars were always physically identical but differed
presented probes (see Fox et al., 2001). We used only due to the preceding emotional face. This
only short-time intervals between the face pair and effect on P1 was equally present for bar probes
the bar onset (100–300 ms, systematically ran- shown in the upper LVF or RVF, in agreement
domized) to tap exogenous mechanisms of spatial with our behavioral results that did not show any
orienting (Egeth and Yantis, 1997). hemispheric asymmetry in spatial orienting to
Our main question was whether sensory re- threat faces (but see Mogg and Bradley, 1999b;
sponses to the peripheral bar-probes would be en- Fox, 2002 for greater facilitation of RTs in the
hanced when replacing an emotional (valid) face, LVF in a similar paradigm). Source estimation
rather than a neutral (invalid) face, as predicted if methods (Pascual-Marqui et al., 1994) further
spatial attention was involuntarily oriented toward confirmed that the P1 component was generated
that particular location (see Bradley et al., 2000; in the extrastriate visual cortex (Fig. 3D), includ-
Armony and Dolan, 2002; Vuilleumier, 2002); and ing the middle occipital gyrus and inferior tempo-
whether such attentional bias would differ between ral gyrus (Pourtois et al., 2004). Such neural
negative (fearful) and positive (happy) emotional sources are consistent with previous ERP studies
cues. Our main comparison therefore concerned on P1 responses that were found to be modulated
the amplitude or latency of ERP generated by the by spatial attention in tasks using nonemotional
exact same bar-probe as a function on the different stimuli (Clark et al., 1995; Martinez et al., 1999;
emotional values of the preceding face context (Fig. Di Russo et al., 2002, 2003).
2D). We predicted that any spatial orienting of These results therefore suggest an amplification
attention should affect early visual processing of sensory responses to a neutral visual stimulus
stages activated by the target-bar, consistent with (bar-probes) taking place at early processing stages
74

Fig. 3. (A) Grand average waveforms in the fear condition (electrode PO8). The black vertical line bar indicates the onset of the bar
probe (target). The P1 (area highlighted by an orange shaded area) was larger for fear valid compared to fear invalid trials, although
the target stimulus was the exact same in these two conditions. (B) Mean global field power (GFP, see Lehmann and Skrandies, 1980)
recorded 130–140 ms post bar-probe onset (corresponding to a temporal window where P1 amplitude was maximal) across the different
experimental conditions. A signal increase is observed in the fear valid condition relative to the three other conditions (fear invalid,
happy valid, and happy invalid), as indicated by a significant validity  emotion interaction. (C) Voltage maps for the P1 in the fear
valid and fear invalid conditions (in the same 130–140 ms time interval following bar-probe onset) showing a more prominent P1 scalp
topography in the former than the latter condition but without any qualitative change in the dipolar configuration of this map across
conditions (amplitude modulation only). (D) Inverse solution by LAURA (Grave de Peralta Menendez et al., 2004) for the P1
revealing distributed brain sources in the extrastriate occipital cortex including the middle occipital gyrus (red line) and inferior
temporal gyrus (blue line).

within extrastriate visual cortex, induced by the trials, but with a longer SOA between the cue and
preceding emotional face presented at the same the target (600 ms) as compared with the present
location. A greater amplitude of P1 activity in ex- study. In our study, we found no effect on the
trastriate visual cortex in response to bar probes latency of ERPs (Pourtois et al., 2004). The effect
following a fearful face also extend the related of threat signals on visual processing is therefore
ERP findings by Stormark et al. (1995) who used strikingly similar to the effect previously obtained
emotion words (rather than facial expressions) and with explicit manipulations of spatial atten-
found enhanced P1 and P3 components for invalid tion (Heinze et al., 1990; Hillyard et al., 1998),
75

which is usually thought to operate by gain control Mangun, 1998) and is more sensitive to attentional
mechanisms imposed on visual pathways through manipulations requiring feature discrimination
top-down signals from parietal and frontal ar- rather than detection (Vogel and Luck, 2000).
eas. Such boosting of visual cortical processing Likewise, C1 component is thought to reflect in-
may thus provide a neural substrate for recent itial processing in primary visual cortex that is not
psychophysical findings showing that the pres- affected by spatial attention (Martinez et al., 1999;
ence of a fearful (as opposed to neutral) face can Di Russo et al., 2003), although primary visual
enhance contrast sensitivity and reduce detection cortex might be modulated at a later delay through
threshold for a following visual stimulus (Gabor feedback mechanisms from higher cortical areas
patch), an effect that is also magnified with tran- (Martinez et al., 1999; Noesselt et al., 2002). Thus,
sient covert attention (Phelps et al., 2006). By con- our results suggest that the spatial orienting in-
trast, we found that positive emotional cues duced by fearful faces may operate through a
conveyed by happy faces did not produce any modulation of early exogenous visual response in
effect on P1 responses to bar-probes appearing at extrastriate areas, and act on the same processing
valid vs. invalid locations (Fig. 3B). pathways as traditional influences of spatial atten-
Importantly, in a control ERP experiment in a tion controlled by fronto-parietal mechanisms.
different group of participants, we could further
ascertain that amplitude increases for P1 responses
to bar-probe following fearful vs. neutral or happy Early responses to emotional faces preceding visual
faces was truly driven by the preceding facial ex- targets
pression rather than any low-level pictorial differ-
ences in the facial cues. Thus, no modulation of P1 One limitation of behavioral measures in the dot-
was found when we used inverted (as opposed to probe task is that the attentional biases can only
upright) faces, which are known to impair the be measured at the time when the target-probe is
normal recognition of emotional expression in presented, while any effect triggered by emotional
faces (Searcy and Bartlett, 1996). Moreover, a cues themselves, prior to the subsequent attentio-
quantitative analysis of our face stimuli showed no nal consequence on target processing, cannot di-
significant difference in mean luminance, contrast, rectly be registered since behavioral measures
surface or central spatial frequency content bet- provide only a snapshot of attentional focus by
ween fearful, happy, and neutral faces (Pourtois comparing the different responses to targets
et al., 2004). (Mogg and Bradley, 1999b). In our ERP study,
Noteworthy, the effect of fearful faces on ERPs however, continuous EEG recordings could be
to subsequent target-bars was selective for the lat- obtained not only in response to the bar-probe (as
eral occipital P1 component, but did not affect described above) but also in response to the pre-
other exogenous visual components, such as the ceding face pair (Fig. 2C). This allowed us to as-
earlier C1 component arising from the primary sess whether fearful or happy faces in the initial
visual cortex (see Clark et al., 1995) or the subse- cue display actually produced any differential
quent N1 component presumably generated by brain response already prior to the onset of the
higher extrastriate areas within occipito-parietal target (bar-probe).
cortex (Vogel and Luck, 2000; Fig. 3A). Spatial Thus, although our main question concerned
validity effects produced by nonemotional exo- how the location of emotional faces in the cue dis-
genous cues (e.g., light flash or abrupt onset) play affected ERPs time-locked to the subsequent
have also been found to affect predominantly the bar-probe (see above), we also analyzed ERPs
lateral occipital P1 component on valid trials (see time-locked to the face pairs (Pourtois et al., 2004).
Clark and Hillyard, 1996; van der Lubbe and This analysis revealed a striking effect of emotion
Woestenburg, 1997; Hopfinger and Mangun, on very early visual responses to faces, affecting
1998), whereas the N1 component is classically the C1 component (Clark et al., 1995), which was
not affected by exogenous cues (Hopfinger and selectively enhanced by fearful expressions. Thus,
76

C1 had a significantly higher amplitude for dis- early latency (110 ms) of this emotion effect was
plays with a fearful face than a happy face, irre- puzzling, given that the authors only found a re-
spective of the visual hemifield in which the fearful liable difference between faces and scramble faces
face was presented, with mean latency of 90 ms (i.e., basic categorical effect) at a later latency,
post onset. The scalp topography of this effect around 165 ms post-stimulus onset, with sources
showed a negative polarity that corresponded to located more anteriorly within the fusiform gyrus.
the expected response to visual stimulations (bilat- According to Halgren et al. (2000), this early
eral face display) presented in the upper visual field differential response to emotional face expression
(Clark et al., 1995). A distributed source localiza- in early visual areas (V1) could serve to rapidly
tion estimate also confirmed that this C1 response decode these socially relevant stimuli in distant
was evoked by main generators in two regions of regions such as the amygdala, which begins to re-
the occipital visual cortex, including the cuneus spond to faces at 120 ms (Halgren et al., 1994)
and lingual gyrus, corresponding to early striate and which receives projections from early visual
cortex (see Pourtois et al., 2004). areas (Amaral et al., 2003). This could explain the
This early valence effect of face expressions on preserved emotional processing of faces sometimes
ERPs to the cue display was not observed for the reported in prosopagnosic patients with occipito-
subsequent P1 or N170 components elicited by temporal damage (Tranel et al., 1988). The nature
these faces. Moreover, this also allowed us to rule of this C1 effect remains unclear and needs further
out the possibility that the modulation of P1 am- replications; but it might also reflect a rapid mod-
plitude to targets could reflect a mere carry over ulation of primary visual cortex by reentrant feed-
effect, for example, due to some slow potential back from emotion-related areas such as the
wave following emotional faces that would persist amygdala (Anderson and Phelps, 2001; Amaral
and contaminate subsequent ERPs to bar-probes et al., 2003; Vuilleumier et al., 2004b), possibly
(see Walter et al., 1964) despite our carefully ran- activated at similar or even earlier latencies
domized SOAs between faces and targets. Fur- (Krolak-Salmon et al., 2004; Halgren et al., 1994;
thermore, analysis of low-level visual features in see Kawasaki et al., 2001 for early visual responses
faces did not suggest any difference that could to emotional stimuli in the ventral prefrontal cor-
have caused early C1 increases (see above), and tex), or alternatively reflect some other deeper
our control experiment with inverted faces did not sources activated at the same latency and contrib-
produce a similar effect. uting to a similar occipito-parietal scalp topogra-
Such early effects of emotional expression have phy as the classic C1 component.
not been previously reported for ERPs to fearful We further asked whether this early effect of
faces, which usually evoke later modulations in fearful faces on C1 responses might be functionally
the P1, N170, and/or subsequent components related to the subsequent enhancement of re-
(Pizzagalli et al., 1999, 2002; Krolak-Salmon et al., sponses to the bar-probes. To indirectly address
2001; Eimer and Holmes, 2002; Batty and Taylor, this issue, we examined whether there was any re-
2003; Ashley et al., 2004; see Vuilleumier and lationship between the amplitude of the C1 evoked
Pourtois, in press for a recent overview). However, by the face cues, and the magnitude of the validity
most studies on emotional faces have always pre- effect on amplitude of P1 evoked by the bar target
sented stimuli only centrally or aligned on the (Pourtois et al., 2004). Strikingly, we found a sig-
horizontal meridian, which may prevent a reliable nificant positive correlation between C1 time-
C1 component evoked by more peripheral stimu- locked to faces and enhancement of P1 time-
lation in upper or lower visual fields (see Jeffreys locked to subsequent bar-probe, selectively arising
and Axford, 1972; Clark et al., 1995). However, a in the fear condition, whereas there was no sig-
previous MEG study (Halgren et al., 2000) has nificant correlation in the happy condition. This
already reported a similar early visual effect for correlation suggests that, even though the time in-
centrally presented emotional faces (sad vs. happy terval between the two stimuli varied randomly,
faces) arising in the occipital striate cortex. The the larger the C1 response to a fearful face in the
77

peripheral visual field, the larger the subsequent Importantly, in these studies, the enhancement of
validity effect on the occipital P1 evoked by a P1 amplitude for attended relative to unattended
bar-probe appearing at the same location. These stimuli was not associated with any concomitant
data provide indirect support for the idea that di- change in latency, waveform, or scalp voltage to-
rect feedback from amygdala on early visual cor- pography of this component, suggesting that the
tex might induce a sustained boosting of sensory effect of spatial attention on target processing may
processing and attention to visual stimuli (Amaral primarily correspond to a gain control mechanism
et al., 2003; Vuilleumier et al., 2004b); and it raises arising in identical visual pathways (Luck, 1995;
further questions about whether such boosting can Hillyard et al., 1998; Carrasco et al., 2002). Ac-
be retinotopic or hemifield/hemispheric-specific. cording to this model, spatial attention is thought
In keeping with this idea, several previous fMRI to operate as an amplification of visual processing
studies have shown that fearful faces can induce a via top-down signals from fronto-parietal areas,
greater activation of face-sensitive regions in fusi- which are activated prior to target onset in the case
form cortex as compared with neutral stimuli of preparatory/endogenous attention or at an ear-
(Morris et al., 1998; Vuilleumier et al., 2001; lier latency post-stimulus onset in the case of re-
Armony and Dolan, 2002; Surguladze et al., 2003; flexive/exogenous attention. These fronto-parietal
see Sabatinelli et al., 2005 with emotional scenes), areas can then in turn enhance the neural re-
but similar emotional enhancements were also sponses in extrastriate cortex (see Kastner and
found in primary visual cortex (Vuilleumier et al., Ungerleider, 2000; Hopfinger et al., 2000; Corbetta
2001; Pessoa et al., 2002) and in more lateral oc- and Shulman, 2002).
cipital areas (Lane et al., 1998). These fusiform In a follow-up study using our dot-probe
and occipital increases to emotional faces are likely paradigm (Pourtois et al., 2005), we therefore,
to depend on direct feedback from the amygdala tested whether any differential neural activity (e.g.,
because they are abolished in patients with am- within the fronto-parietal network) might precede
ygdala lesions (Vuilleumier et al., 2004b) but per- the amplitude modulation of P1 responses to bar-
sist in patients with parietal lesions (Vuilleumier probes and thus correspond to a possible source of
et al., 2002). Moreover, such effects of emotion in attentional biases in spatial attention. To this
early visual cortex may interact with selective at- aim, we employed a different approach than our
tention and be further augmented when emo- previous waveform analysis and turned to topo-
tional stimuli appear in task-relevant location graphical segmentation methods (Lehmann and
(Vuilleumier et al., 2001; Pessoa et al., 2002). It Skrandies, 1980; Michel et al., 1999) that allowed
remains, however, unclear to what extent these in- the identification of subtle changes in the topo-
creases can only affect the processing of emotional graphic configuration of scalp EEG over time.
stimuli themselves, or also affect the processing of Such topographic changes can arise independently
neutral stimuli following emotional cues as in our of component waveforms and independently of
dot-probe paradigm. differences in field strength (i.e., the amplitude of
waveforms; see Lehmann and Skrandies, 1980). A
standard topographic analysis (Lehmann and
Cascade of neural events and source of sensory gain Skrandies, 1980; Michel et al., 1999) was per-
in extrastriate visual cortex formed on ERP data from the same dot-probe
task as used previously (Pourtois et al., 2004),
Our findings of increased P1 amplitude (with- which provides a spatio-temporal segmentation
out any change in the latency) for visual tar- of the successive field configurations activated
gets cued by fearful faces converge with similar during the sequence of stimulus processing (usu-
effects observed in electrophysiological studies of ally referred to as ‘‘microstate’’, Lehmann and
spatial attention using nonemotional cues (van der Skrandies, 1980). The rational of this approach is
Lubbe and Woestenburg, 1997; Hopfinger and to identify a series of statistically distinct topo-
Mangun, 1998; Hillyard and Anllo-Vento, 1998). graphic configurations (i.e., activity maps) over the
78

time-course of the evoked neural responses, re- preceding the topographical maps corresponding
flecting the succession of different functional states to P1. No differential activity was seen for valid
engaged by stimulus processing, with the underly- vs. invalid trials in the happy condition, although
ing assumption that topographic changes may de- the same map was also present prior to P1. The
note the activation of distinct neural sources neural sources estimated for this map were
(Brandeis and Lehmann, 1986). Given that only clearly distinct from extrastriate occipital sources
landscape or topographic differences are of inter- associated with P1, and involved cortical genera-
est in this spatio-temporal cluster analysis, dom- tors in posterior temporal and posterior parietal
inant maps are normalized to unitary strength regions instead (Fig. 4C). In other words, these
values by dividing the voltage at each electrode by data indicate that an early topographic microstate
a measure of the global field power (Lehmann and (at 40–80 ms post-target onset) was differen-
Skrandies, 1980). Here again, we focused on re- tially activated when targets appeared at the same
sponses evoked by the bar-probes, as a function of vs. different location as a fearful face; and that
the preceding emotional face context. this distinctive configuration of neural activity
This new analysis (Pourtois et al., 2005) preceded another subsequent microstate (at
first showed that EEG activity during the time- 120–160 ms) corresponding to P1, whose genera-
range of the P1 component evoked by bar-probes tors did not differ but whose amplitude was en-
(120–160 ms) did not exhibit any differences in hanced for valid vs. invalid targets following
topographic configuration across the different fearful cues (Pourtois et al., 2005). These results
emotion and validity conditions of preceding face are consistent with the idea that a first sweep
cues, whereas there was a significant increase in the of activity in posterior temporal and parietal re-
strength of this P1 topography (as indicated by a gions might take place rapidly after a visual target
higher global field power, Lehmann and Skrandies, onset (Bullier, 2001; Foxe and Simpson, 2002)
1980) for targets following a valid fearful face as and possibly provide the signal for subsequent
compared with an invalid fearful face (Fig. 3C). top-down control of target processing (Hopfinger
This pattern fully agrees with our hypothesis of a et al., 2000; Kastner and Ungerleider, 2000;
gain control mechanism enhancing visual target Bullier, 2001).
processing subsequent to orienting of spatial at- Remarkably, we also found that these two con-
tention toward threat-related cues (Pourtois et al., secutive neural events were positively correlated
2004), as typically found with attentional orienting (Pourtois et al., 2005; Fig. 4E), suggesting some
based on nonemotional cues (Hillyard et al., 1998). functional coupling between the early posterior
More importantly, this analysis revealed the parietal activity (40–80 ms) and subsequent P1
existence of an early (o100 ms post bar-probe activity (120–160 ms). The variance of the early
onset) and stable (40–80 ms) topographical map 40–80 ms map (indexing the strength of its expres-
that reliably distinguished valid from invalid target sion across subjects) showed a significant positive
trials during the fear condition (Fig. 4A, B), just linear correlation with the variance of the next

Fig. 4. (A) Grand average waveforms in the fear condition (electrode PO8) time-locked to the onset of the bar probe (valid and invalid
trials collapsed). Before the onset of the P1 (area highlighted by an orange shaded area), there was a significant topography difference
between valid and invalid trials in the fear condition, although no exogenous electric component was detectable at this specific lateral
occipito-temporal site (PO8) during this early time-period (40–80 ms post bar probe onset). (B) Voltage maps in the fear valid and fear
invalid conditions in the 40–80 ms post bar probe showing a significant modulation of the global scalp configuration (with no change in
amplitude). (C) Statistical parametric mapping provided by LAURA indicated that brain regions that were more activated by fear
valid than fear invalid trials in the 40–80 ms post bar-probe onset were mainly located in the left posterior parietal cortex (po0.001,
uncorrected). (D) Conversely, fear invalid trials evoked more activation than fear valid trials in medial frontal regions (corresponding
to rostral anterior cingulate cortex (ACC; po0.01, uncorrected) during the same time interval (40–80 ms post bar-probe onset). (E)
There was an enhanced positive correlation between this early scalp map (40–80 ms post stimulus onset) and the directly following P1
map in the fear valid condition (r ¼ 0.55, p ¼ 0.03) but a clear attenuation of this correlation in the fear invalid condition (r ¼ 0.004,
p ¼ 0.50), suggesting an enhanced coupling between these two successive functional microstates in the fear valid condition.
79
80

120–160 ms P1 map, which was present across benefit in covert orienting task (see Fox et al.,
all conditions but selectively enhanced for valid 2001; Koster et al., 2004b). Problems in disengag-
targets following fearful faces, and selectively sup- ing from threat signals might be particularly
pressed for invalid targets following fearful faces important in people with higher anxiety (even
(Fig. 4E). This enhanced coupling between parietal at subclinical level, see Fox et al., 2001; Mathews
and extrastriate activity might provide a plausible et al., 2004). Accordingly, increases in rostral
neural mechanism underlying the facilitation in ACC activity when ignoring emotional stimuli
orienting spatial attention toward targets appear- (Vuilleumier et al., 2001) have been found to be
ing at the location of threat-related cues (Fox, greater in anxious than nonanxious individuals
2002; Vuilleumier, 2002). If early activity in pos- (Bishop et al., 2004a).
terior parietal and temporal regions following tar- Taken together, our ERP data (Pourtois et al.,
get onset (o100 ms) is implicated in the generation 2004, 2005) reveal a precise cascade of neural
of top-down signals to influence ongoing visual events involved during spatial orienting to periph-
processing in occipito-temporal areas (120–200 ms) eral visual targets, with selective influences of
(Kastner and Ungerleider, 2000; Bullier, 2001), emotional cues conveyed by fearful faces. Targets
then these functional influences appear to be following valid fearful faces evoked an enhanced
enhanced for targets at valid conditions and dis- exogenous visual response in extrastriate cortex
rupted for targets at invalid locations following (presumably through gain control mechanism),
fearful faces, with no such effect for happy faces. preceded by specific enhancement of activity in
Conversely, our topographical analyses revealed posterior parietal cortex and posterior temporal
that ERPs to bar-probes invalidly cued by fearful regions; whereas the same targets appearing on the
faces were associated with a distinctive pattern of side opposite to a fearful face evoke no differential
activity at the same early latency (40–80 ms), re- visual responses but greater activation of medial
placing the posterior parietal and temporal acti- prefrontal regions. Collectively, these results have
vation related to spatial orienting (Fig. 4B, D). begun to provide novel insights into brain mech-
Neural sources for this distinct map were located anisms by which fear-related signals can mobilize
within ventromedial prefrontal areas, includ- processing resources (Ohman and Mineka, 2001)
ing the rostral anterior cingulate cortex (ACC) and trigger a prioritization of spatial attention
(Fig. 4D). Such activation in rostral ACC may be toward their location (Vuilleumier, 2002, 2005). In
consistent with a role in controlling attention in a follow-up fMRI study, we identified brain re-
conflict situations (MacDonald et al., 2000) and gions underlying this capture of spatial attention
error processing (Carter et al., 1999), particularly by threat-related stimuli.
based on affective or motivational signals (Bush
et al., 2000; Bishop et al., 2004a) or related to
breaches in implicit expectation generated by fMRI correlates for benefits and cost in spatial
invalid cues (Nobre et al., 1999). If spatial atten- attention produced by threat cues
tion was reflexively oriented toward the location of
the threat-relevant stimulus in face display, then Converging neurophysiological, neuropsychologi-
a target appearing on the opposite side might cal, and neuroimaging studies (see Driver and
require resolving a potential conflict between re- Vuilleumier, 2001; Kanwisher, 2001) have now
sponding to the task-relevant stimulus and disen- clearly established how nonemotional (neutral) ex-
gaging from the emotionally alerting stimulus. ogenous (e.g., light flash or abrupt onset/offset)
This selective invalidity effect of fearful faces in or endogenous/symbolic cues recruit specific
ERPs also converge with behavioral findings sug- brain regions associated with visual spatial atten-
gesting that negative stimuli may not only draw tion in a variety of paradigms (Gitelman et al.,
more easily, but also hold spatial attention more 1999; Corbetta et al., 2000; Hopfinger et al., 2000;
strongly than neutral stimuli, and thus lead to Woldorff et al., 2004). A distributed cortical net-
greater invalidity cost rather than greater validity work of dorsal regions in fronto-parietal cortex
81

directly participating to the control of spatial cues and were presented in either visual field in an
attention, including the intraparietal sulcus event-related manner, preceding the dot probe at
(IPS) and frontal eye field (FEF), is crucially in- either a valid or invalid location. Armony and
volved in the voluntary or endogenous control of Dolan (2002) reported an activation of fronto-
spatial attention (Mesulam, 1998; Kastner and parietal areas during shifts of attention toward a
Ungerleider, 2000), whereas a more ventral corti- unilateral dot probe when the aversively condi-
cal network in the ventrolateral prefrontal cortex tioned face stimulus (CS+) was briefly presented
(VLPFC) and temporoparietal junction (TPJ) con- at a different location in the visual field prior to the
tributes to the detection of unexpected, behavi- target onset (see Mogg et al., 1994; Bradley et al.,
orally relevant or salient stimuli, with the latter 2000), suggesting an involuntary capture of atten-
ventral system interacting with the dorsal system tion by the aversive face.
during involuntary or exogenous shifts in attention We recently conducted an event-related fMRI
(Downar et al., 2000; Corbetta and Shulman, study (Pourtois et al., 2006) using the same dot-
2002). Moreover, the more dorsal areas (FEF and probe paradigm as in our previous ERP work
IPS) may be rapidly activated following stimulus (Pourtois et al., 2004; Fig. 2A) to identify neural
onset (see Bullier, 2001; Foxe and Simpson, 2002) substrates responsible for a spatially selective mod-
to act as a regulatory source and directly contrib- ulation of attention when threat cues were conveyed
ute to the top-down selection of stimuli and by fearful expression in faces, rather than by prior
responses in distant brain areas located within explicit conditioning (unlike prior studies, see
the temporal, frontal, and occipital lobe. In this Fredrikson et al., 1995; Armony and Dolan,
perspective, this dorsal network plays a crucial 2002). Our aims were to determine how brain re-
role during the deployment of spatial attention sponses to a neutral visual target might be altered
and can bias activity in remote visual areas to when preceded by emotional signals at valid or
narrow sensory processing onto relevant visual invalid locations, as previously investigated by the
targets (Hopfinger et al., 2000). Posterior parietal modulation of visual ERP components recorded on
regions are then responsible for imposing feedback the scalp (Pourtois et al., 2004). However, given the
signals on sensory areas to enhance processing slow temporal resolution of fMRI (Bandettini,
of behaviorally relevant, attended events (Kastner 1999), it was not directly possible to separate neu-
et al., 1999). Moreover, when attention is oriented ral effects related to the processing of emotion in
reflexively toward behaviorally relevant or salient face pairs (cue) from those related to the processing
(but emotionally neutral) stimuli, the same dorsal of subsequent bar-probes (target). Moreover, in our
fronto-parietal network can be activated together ERP study, we used short temporal intervals be-
with the TPJ and VLPFC, to promote a shift tween faces and targets (100–300 ms) to assess
of processing resources toward the new events purely reflexive orienting mechanisms (Egeth and
(Downar et al., 2000; Corbetta and Shulman, Yantis, 1997), such that any fMRI activation in this
2002; Peelen et al., 2004). context could only correspond to a compound of
However, very few PET or fMRI studies have neural responses to both the face pair (cue) and bar-
investigated whether the same fronto-parietal probe (target). For this reason, to identify a true
areas are also differentially activated during shifts modulation of target processing, distinct from the
of spatial attention in response to emotionally effect of emotional faces by themselves, we designed
threatening stimuli. In a pioneer study, Fredrikson a slightly modified version of the dot-probe para-
et al. (1995) found increased activation in superior digm during fMRI and introduced ‘cue-only’ trials
parietal and frontal regions for fear-conditioned (face pairs with no subsequent target) that were
stimuli and suggested that such effects might re- unpredictably intermingled with the ‘‘cue-plus-tar-
flect increased attention toward these stimuli. get’’ trials (face pairs followed by a unilateral target
Another brain imaging study by Armony and at valid or invalid location, similarly to our ERP
Dolan (2002) was the first to use a dot-probe task study, see Pourtois et al., 2004). A similar approach
in which aversively conditioned faces served as has been used in neuroimaging studies of spatial
82

attention during typical, nonemotional version of regions, fMRI showed that occipital responses to
Posner orienting task (see Corbetta et al., 2000; visual targets were enhanced by their presentation
Shulman et al., 2002; Woldorff et al., 2004), allow- on the same (valid) rather than the opposite
ing a clear distinction of brain activity related to (invalid) side as a preceding fearful face, even
cueing effects from target processing. A second though the bars and their orientation were
change in this fMRI experiment as compared with otherwise identical across these two conditions,
our initial EPR study (Pourtois et al., 2004) con- and faces always task-irrelevant. The lateral occip-
cerned the task: in the fMRI experiment, subjects ital cortex is known to be critically involved in
were asked to respond on each ‘‘cue-plus-target’’ visual shape recognition (Grill-Spector et al., 2001;
trial and to judge whether the orientation of the Kourtzi and Kanwisher, 2001), and might therefore
bar-probe matched (50%) or did not match (50%) be enhanced by perceived threat signals (see also
the orientation of the thicker line-segment of the Lane et al., 1998). Note, however, that we found no
fixation cross on that particular trials (by pressing difference in these occipital responses as a function
two different keys; see Pourtois et al., 2006). Recall of the side of fearful faces, and no effect of emotion
that in our earlier ERP study, subjects were re- or validity in early visual areas within the lingual
quired to press a button only in a minority of trials gyrus that exhibited different retinotopic responses
where the peripheral bar’s orientation matched the to targets in LVF or RVF. These data do not seem
thick fixation-cross segment (see Pourtois et al., to support the idea that fearful stimuli may produce
2004). Thus, in our fMRI experiment, the dual task retinotopic increases in early visual areas.
requirement and the frequent withdrawal of re- More importantly, the comparison of event-
sponses did not provide a pure measure for spatially related fMRI responses to valid vs. invalid targets
selective orienting effects in valid trials, even though following fearful faces showed a unique pattern of
different detection or discrimination judgments spatially selective effects in the IPS in both hem-
have been reported to produce similar attentional ispheres — an area playing a major role in spatial
biases in emotional dot-probe tasks (Mogg and attention (Corbetta and Shulman, 2002). Right IPS
Bradley, 1999b). Otherwise, all stimuli and condi- was unresponsive to targets in the RVF after an
tions were similar in this fMRI study as in our invalid fearful face presented contralaterally on the
previous ERP study (with fearful and happy emo- left side, whereas the left IPS was conversely un-
tional expressions, unpredictably shown at valid responsive to targets in the LVF after an invalid
and invalid locations, equally probable and ran- fearful face in the contralateral RVF, even though
domized). both IPS could respond to targets on either side in
Critically, fMRI results in our modified dot- other conditions (see Fig. 6). IPS responses to
probe task disclosed a pattern of brain activation peripheral targets were selectively reduced when
clearly indicating that fearful faces again had a targets were presented in the ipsilateral visual field
unique impact on spatial orienting of attention after a fearful face in the contralateral (invalid)
toward the subsequent bar-probe target, whereas hemifield (Fig. 6). These data therefore suggest
happy faces did not produce similar effects. First, some suppression in the processing of ipsilateral
following cue displays with a fearful face, bar- targets when attention was focused on the contra-
probes appearing at valid locations were found to lateral side during invalid fear trials. By contrast,
produce an increased neural response in lateral oc- IPS was strongly activated by targets in the ipsi-
cipital cortex, as compared with bar-probes ap- lateral hemifield when preceded by a fearful face at
pearing at invalid locations (Fig. 5). There was the same (valid) location. In other words, for both
no difference in activation of the lateral occipital hemispheres, attentional processing in IPS was
cortex when comparing fearful and happy faces apparently restricted to contralateral targets fol-
on ‘‘cue-only’’ trials, indicating that this occipital lowing a fearful face on that same side, but more
activation reflected target processing, not face bilaterally responsive to targets on either side in the
processing. Thus, consistent with previous ERPs other conditions. These fear-dependent spatial
showing enhanced P1 response in extrastriate invalidity effects led to a significant emotion 
83

Fig. 5. (A) Right extrastriate cortex activation (random effect analysis carried out in SPM2, po0.005, uncorrected) obtained in the
fear valid4fear invalid contrast, irrespective of the spatial location of targets. This fMRI results is consistent with the ERP result
showing an early enhanced sensory processing in the extrastriate occipital cortex for bar probes in the fear valid condition. (B) Mean
beta signal estimates (proportional to percentage of signal change) for this right occipital region, across the different conditions of cue-
plus-target trials (* indicates significant comparison).

validity interaction in IPS and found symmetrically invalid fearful cues (see Kincade et al., 2005) rather
in both hemispheres. These spatially selective than a response to emotional faces alone. Again,
effects on target processing in IPS are highly con- we found no effect of happy faces (valid or invalid)
sistent with the idea that the presentation of a in IPS. Furthermore, eye-tracking data during
fearful face in visual periphery might lead to a fMRI allowed us to ensure that our subjects made
transient focusing of attentional resources medi- no systematic eye movement toward either the left
ated by contralateral parietal areas toward that or right upper visual field across the different stim-
side, with a relative suppression of responses to ulus conditions, but correctly maintained fixation
visual events occurring at other locations (Fig. 6). on the central cross, ruling out the possibility that
Moreover, no such effect was found in IPS when IPS activity may reflect different saccadic behavior
comparing contralateral and ipsilateral fearful during the task (Pourtois et al., 2006).
faces in ‘‘cue-only’’ trials, suggesting a spatially These brain-imaging data clearly show that
selective modulation of IPS responses to targets by threat-related cues may not only draw covert
84

Fig. 6. (A) Left IPS activation (random effect analysis carried out in SPM2, po0.001, uncorrected) obtained in the fear valid4fear
invalid contrast for bar probes presented in the LVF, overlaid on a single-subject T1-weighted anatomical template (normalized to
MNI space). (B) Mean beta signal estimates (proportional to percentage of signal change) for this left IPS region, across the different
conditions of cue-plus-target trials. (C) Right IPS activation (po0.001, uncorrected) obtained in the fear valid4fear invalid contrast
for bar probes presented in the RVF. (D) Mean beta signal estimates for this right IPS region, across the same conditions. In both
plots, the critical conditions compared in these contrasts are highlighted in gray (* indicates significant comparison).

attention more efficiently than neutral or positive This activation of orbitofrontal regions on trials
cues, and thus produce benefits in visual (i.e., occip- where a fearful face called for attention on one side
ital) responses to targets presented at the same/ and a visual target subsequently appeared on the
valid location; but may also hold attention more opposite side converges with a similar effect in our
strongly and produce greater cost on reorienting previous ERP study (Pourtois et al., 2005), in
(i.e., IPS) when targets are presented at a different/ which neural sources in ventro-medial prefrontal
invalid location (‘‘disengage’’ effect, see Fox et al., areas were found (Fig. 4D) in the early phase of
2001; Koster et al., 2004b for behavioral evidence). the orienting response to invalid bar-probes, again
In addition, we also found that bar-probes following fearful but not happy faces.
invalidly cued by fearful faces also increased acti-
vation in the left lateral orbitofrontal cortex, as
compared with validly cued trials. This may be fMRI responses to peripheral faces alone
consistent with a role of this region in regulating
the allocation of processing resources during All fear-selective effects in our fMRI study
breaches of expectation implicitly generated by described above were identified by examining
emotional cues (Nobre et al., 1999), or in the brain activations for ‘‘cue-plus-target’’ trials, but
presence of affective or motivational conflicts not ‘‘cue-only’’ trials, indicating that these likely
(Vuilleumier et al., 2001; Bishop et al., 2004a). resulted from a spatially selective modulation of
85

target processing by the position of the preceding including an habituation due to repetition of the
fearful faces, but not just emotional face process- same faces (Breiter et al., 1996; Phillips et al., 2001;
ing. However, as in our ERP study (Pourtois et al., Fischer et al., 2003) and the presence of one fearful
2004) in which we could examine ERPs time- or happy face in all bilateral cue displays (Pessoa
locked to the target onset, as well as ERPs time- et al., 2002; Zald, 2003), but amygdala responses
locked to the face cues, here we could test for to fearful faces is likely to play a crucial role in
fMRI responses not only during ‘‘cue-plus-target’’ eliciting attentional orienting behaviors toward
trials but also during ‘‘cue-only’’ trials to deter- threat (Holland and Gallagher, 1999).
mine any effect of peripheral faces alone. Our Taken together, these fMRI data have provided
analysis of ‘‘cue-only’’ trials showed that fearful us with a valuable refinement concerning a wide
but not happy faces produced an increased re- network of brain areas within extrastriate occipital
sponse of the right precuneus in medial occipito- cortex, superior and inferior parietal cortex, and
parietal cortex, regardless of the side of fearful ventromedial prefrontal regions, implicated in
faces but with a contralateral predominance. This attentional biases and spatially selective modula-
medial occipital response to fearful face might be tions of target processing induced by emotional
consistent with the previous EEG (Pourtois et al., cues such as fearful faces. More generally, our
2004) and MEG (Halgren et al., 2000) results brain-imaging results show how emotional stimuli
showing increased activity for occipital sources may act as exogenous cues on spatial attention by
arising at an early latency post face onset, corre- modulating activity in a network of brain areas that
sponding to C1 time-range (see above), and poten- partly overlap with cortical systems previously
tially reflecting some general alerting or arousal associated with the control of attention for non-
effect triggered by fearful faces (Lang et al., 1998; emotional stimuli (see Corbetta and Shulman,
Thiel et al., 2004). 2002), but also partly involve distinct neural system.
On the other hand, we found that peripheral
fearful faces in ‘‘cue-only’’ trials produce a selec-
tive activation of the inferior temporo-parieto- Role of anxiety in emotion–attention interactions
occipital junction on the side opposite to the fear-
ful faces (with no effect of happy faces). These Many behavioral studies (e.g., Fox et al., 2001;
increases might be consistent with a role of ventral Koster et al., 2004b) have shown that attentional
cortical regions within the attentional networks, biases induced by emotionally threatening cues in
more critically concerned with the detection of dot-probe tasks can be significantly exaggerated in
behaviorally relevant or unexpected stimuli, rather people with higher anxiety levels, even below clin-
than with top-down selection or focusing as con- ical levels. Whereas some findings suggest that
trolled by the more dorsal IPS (Corbetta and anxiety may facilitate orienting or engaging atten-
Shulman, 2002). However, unlike a study using tion toward emotional cues (Mogg and Bradley,
similar displays with bilateral faces presented in 1998), other findings indicate a greater difficulty to
peripheral visual field (Noesselt et al., 2005), we reorient away or disengage from threat cues (Fox
found no significant increases in fusiform cortex or et al., 2001; Koster et al., 2004b; Mathews et al.,
amygdala in the hemisphere contralateral to fear- 2004). However, the neural correlates of these
ful faces. Nevertheless, several previous findings effects of anxiety remain unclear. A number of re-
suggest that the amygdala (and fusiform) is con- cent fMRI studies have pointed to heightened re-
sistently activated by fearful faces (see Vuilleumier sponses to negative emotional stimuli in highly
et al., 2004a), and that such amygdala activation anxious subjects in various regions including the
might play an important role in triggering subse- amygdala (Bishop et al., 2004b; Etkin et al., 2004;
quent visual orienting of attention to threat loca- Sabatinelli et al., 2005) or ACC (Bishop et al.,
tions (Amaral et al., 2003; Vuilleumier, 2005). We 2004a). However, in our own studies, we have
believe that this lack of amygdala effects might found no reliable modulation of amygdala re-
result from a number of methodological factors, sponses to fearful faces in relation to general scores
86

of anxiety trait or state in similar tasks (Vuilleumier healthy (nonanxious) participants may orient cov-
et al., 2001, unpublished data), but we observed ertly and reflexively to the position briefly occupied
more consistent effects in ventromedial prefrontal by an irrelevant and nonpredictive fearful face in
regions (e.g., see Sander et al., 2005). the (upper) visual field, such that this will modify
Moreover, in the current series of ERPs and their behavioral performance and brain responses
fMRI studies using the dot-probe paradigm, we to a subsequent target appearing at the same
systematically recorded anxiety scores in our location. These results suggest that fearful faces
participants. But our preliminary analysis failed may act as powerful exogenous cues and produce a
to identify straightforward correlations between transient involuntary capture of spatial attention,
anxiety levels and ERPs indices related to P1 am- somehow similar to an abrupt onset or offset, or a
plitude or fMRI indices related to validity effects. sudden luminance change (Egeth and Yantis,
As a tentative explanation, this might be caused by 1997). By contrast, happy faces do not seem to
the weak variation and low score in state and trait produce similar effects, suggesting a special role of
anxiety levels (Spielberger, 1983) across our non threat cues in such interactions with mechanisms of
preselected participants. In any case, further covert spatial attention.
research is therefore needed to explore more sys- More specifically, brain imaging data obtained
tematically the anatomical substrate and time- by both ERPs (Stormark et al., 1995; Pourtois
course of emotional biases in perception and et al., 2004, 2005) and fMRI (Armony and Dolan,
attention, contrasting large but preselected (and 2002; Pourtois et al., 2006) during the dot-probe
representative) samples of anxious/fearful vs. non- task (Posner et al., 1980; Bradley et al., 2000) are
anxious participants (based on a careful selection now converging to delineate the precise brain
of more extreme anxiety scores in either the sub- pathways and spatio-temporal dynamics underly-
clinical or clinical range) and using well controlled ing emotional biases in spatial attention. We sug-
tasks that may provide more sensitive measures gest a current working model that implicates both
for different attentional subprocesses. direct effects of emotional signals from amygdala
on sensory processing and indirect effects on
attentional systems subserved by parietal cortical
Conclusions areas (Vuilleumier, 2005). Taken together, the
extant fMRI results (Vuilleumier et al., 2004b;
Selective attention is an essential cognitive mech- Noesselt et al., 2005; Pourtois et al., 2006) com-
anism governing the capacity-limited processing bined with ERPs results (Pourtois et al., 2004,
resources of our brains (Marois and Ivanoff, 2005) 2005) indicate that a fearful face cue in peripheral
and promoting an efficient selection of salient visual field may not only activate the amygdala but
or goal-related information in the environment also induce rapid feedback to visual cortex to
(Posner et al., 1980). Until recently, attention has enhance face-sensitive areas as well as earlier oc-
mainly been studied in conditions where such selec- cipital areas, possibly within less than 100 ms post-
tion was based on relatively simple bottom-up onset for the latter region (and probably within
sensory-driven mechanisms (e.g., pop-out) or 170–200 ms for the former). The effect of direct
higher-level top-down influences (Egeth and feedback to occipital areas may then outlast the
Yantis, 1997; Kastner and Ungerleider, 2000). presentation of the facial threat cue, to reduce
Here we have reviewed recent neuroimaging data sensory threshold in retinotopic or nonretinotopic
(ERP and fMRI) indicating that emotional values, regions of early occipital cortex during a brief
such as threat signals conveyed by fearful faces, can period (e.g., see Phelps et al., 2006), leading to a
also influence the spatial distribution of selective subsequent facilitation of spatial selection mecha-
attention (Mogg and Bradley, 1998; Vuilleumier nisms directing attention to the same location (or
et al., 2001, 2004a; Fox, 2002; Dolan and same side), and thus to an enhanced activation of
Vuilleumier, 2003). Using the classical dot-probe posterior parietal and posterior temporal regions
paradigm (Bradley et al., 2000), we could show that implicated in orienting to a subsequent visual
87

target (through processes presumably activated between reflexive emotional signals and goal-
at o80 ms post-target onset, see Bullier, 2001). driven attentional set.
This may in turn lead to enhanced sensory process- Collectively, our electrical and hemodynamic
ing of the target in the lateral extrastriate occipital brain-imaging results highlight the complexity of
cortex, and generate enhanced P1 responses to the spatio-temporal dynamics underlying the prior-
the target (at o130–150 ms post-target onset), in a itization of attention resources to threat. These data
similar manner as the result of preparatory base- converge with other evidence suggesting that some
line shifts of activity imposed by endogenous top- emotional influences originating from phylogenet-
down biases or by other exogenous signals in ically ancient systems in ‘‘limbic’’ brain regions can
attention (Hillyard et al., 1998; Kastner et al., act in parallel with top-down influences tradition-
1999; Super et al., 2003; Tallon-Baudry et al., ally associated with selective attention or executive
2005; Liu et al., 2005). This amplitude modula- functions of fronto-parietal cortical areas (LeDoux,
tion of early sensory responses to visual targets 1996; Ohman and Mineka, 2001; Vuilleumier,
following a valid emotional cue seems compatible 2005). In this perspective, emotion is not separated
with a gain control mechanism (Hillyard et al., from cognition (Zajonc, 1980) but plays a funda-
1998). The electrophysiological properties (i.e., la- mental role in regulating brain functions involved in
tency, polarity, topography and neural sources in perception, attention and adaptive behaviors.
the extrastriate cortex) are entirely consistent with
those previously reported by ERP studies of spa-
tial attention using nonemotional exogenous Acknowledgements
cues (e.g., Hopfinger and Mangun, 1998; Hillyard
and Anllo-Vento, 1998; Van der Lubbe and This work is supported by a grant from the Swiss
Woestenburg, 2000), suggesting a similar subst- National Science Fund to PV (grant # 632.065935)
rate and time-course but distinct sources for the and the Swiss National Centre for Competence in
top-down bias signal (Pourtois et al., 2004). Research in Affective Sciences (NCCR grant #
On the other hand, when a target follows a 51A240-104897).
threat cue at another location, posterior parietal
responses to the target are reduced in the ipsilat-
eral hemisphere, disclosing a spatially selective References
restriction of attention to the location invalidly
Amaral, D.G., Behniea, H. and Kelly, J.L. (2003) Topographic
cued by the preceding threat, which is consistent organization of projections from the amygdala to the visual
with behavioral observations that fearful stimuli cortex in the macaque monkey. Neuroscience, 118:
do not only draw spatial attention more readily 1099–1120.
than neutral or positive stimuli, but also hold at- Anderson, A.K. (2005) Affective influences on the attentional
dynamics supporting awareness. J. Exp. Psychol. Gen., 134:
tention more durably (Fox et al., 2001). In ERPs,
258–281.
this corresponds to a relative suppression of an Anderson, A.K. and Phelps, E.A. (2001) Lesions of the human
early microstate in posterior parietal and posterior amygdala impair enhanced perception of emotionally salient
temporal cortex (o80 ms post-onset) associated events. Nature, 411: 305–309.
with orienting attention to the target. These results Armony, J.L. and Dolan, R.J. (2002) Modulation of spatial
reveal dissociable neural substrates for the engage attention by fear-conditioned stimuli: an event-related fMRI
study. Neuropsychologia, 40: 817–826.
and disengage components of spatial attention Ashley, V., Vuilleumier, P. and Swick, D. (2004) Time course
with threat-related cues. Concomitantly, disengag- and specificity of event-related potentials to emotional
ing from an invalid fearful face to reorient toward expressions. Neuroreport, 15: 211–216.
a task-relevant target on the opposite side involves Bandettini, P.A. (1999) The temporal resolution of functional
MRI. In: Moonen, C.T.W. and Bandettini, P.A. (Eds.),
motivational processes in ventromedial frontal
Functional MRI. Springer, Berlin, pp. 205–220.
cortex and/or executive control processes in ante- Batty, M. and Taylor, M.J. (2003) Early processing of the six
rior ACC that may become activated within basic facial emotional expressions. Brain Res. Cogn. Brain
the same time-range to resolve any competition Res., 17: 613–620.
88

Bishop, S., Duncan, J. and Lawrence, A.D. (2004a) Prefrontal Dolan, R.J. and Vuilleumier, P. (2003) Amygdala automaticity
cortical function and anxiety: controlling attention to threat- in emotional processing. Ann. NY Acad. Sci., 985: 348–355.
related stimuli. Nat. Neurosci., 7: 184–188. Downar, J., Crawley, A.P., Mikulis, D.J. and Davis, K.D.
Bishop, S., Duncan, J. and Lawrence, A.D. (2004b) State anx- (2000) A multimodal cortical network for the detection
iety modulation of the amygdala response to unattended of changes in the sensory environment. Nat. Neurosci.,
threat-related stimuli. J. Neurosci., 24: 10364–10368. 3: 277–283.
Bradley, B.P., Mogg, K., Millar, N., Bonham Carter, C., Driver, J. and Vuilleumier, P. (2001) Perceptual awareness
Fergusson, E., Jenkins, J. and Parr, M. (1997) Attentional and its loss in unilateral neglect and extinction. Cognition,
biases for emotional faces. Cogn. Emotion, 11: 25–42. 79: 39–88.
Bradley, B.P., Mogg, K. and Millar, N.H. (2000) Covert and Eastwood, J.D., Smilek, D. and Merikle, P.M. (2001) Differ-
overt orienting of attention to emotional faces in anxiety. ential attentional guidance by unattended faces expressing
Cogn. Emotion, 14: 789–808. positive and negative emotion. Percept. Psychophys.,
Brandeis, D. and Lehmann, D. (1986) Event-related potentials 63: 1004–1013.
of the brain and cognitive processes: approaches and appli- Egeth, H.E. and Yantis, S. (1997) Visual attention: control,
cations. Neuropsychologia, 24: 151–168. representation, and time course. Annu. Rev. Psychol., 48:
Breiter, H.C., Etcoff, N.L., Whalen, P.J., Kennedy, W.A., 269–297.
Rauch, S.L., Buckner, R.L., Strauss, M.M., Hyman, S.E. Eimer, M. and Holmes, A. (2002) An ERP study on the time
and Rosen, B.R. (1996) Response and habituation of the course of emotional face processing. Neuroreport,
human amygdala during visual processing of facial expres- 13: 427–431.
sion. Neuron, 17: 875–887. Ekman, P. and Friesen, W.V. (1976) Pictures of Facial Affect.
Bullier, J. (2001) Integrated model of visual processing. Brain Consulting Psychologists Press, Palo-Alto.
Res. Brain Res. Rev., 36: 96–107. Etkin, A., Klemenhagen, K.C., Dudman, J.T., Rogan, M.T.,
Bush, G., Luu, P. and Posner, M.I. (2000) Cognitive and emo- Hen, R., Kandel, E.R. and Hirsch, J. (2004) Individual
tional influences in anterior cingulate cortex. Trends Cogn. differences in trait anxiety predict the response of the baso-
Sci., 4: 215–222. lateral amygdala to unconsciously processed fearful faces.
Cacioppo, J.T. and Gardner, W.L. (1999) Emotion. Ann. Rev. Neuron, 44: 1043–1055.
Psychol., 50: 191–214. Eysenck, M.W. (1992) Anxiety: The Cognitive Perspective.
Carrasco, M., Penpeci-Talgar, C. and Eckstein, M. (2000) Erlbaum Ltd, Hove, UK.
Spatial covert attention increases contrast sensitivity across Fenske, M.J. and Eastwood, J.D. (2003) Modulation of focused
the CSF: support for signal enhancement. Vision Res., 40: attention by faces expressing emotion: evidence from flanker
1203–1215. tasks. Emotion, 3: 327–343.
Carrasco, M., Williams, P.E. and Yeshurun, Y. (2002) Covert Fischer, H., Wright, C.I., Whalen, P.J., McInerney, S.C.,
attention increases spatial resolution with or without masks: Shin, L.M. and Rauch, S.L. (2003) Brain habituation during
support for signal enhancement. J. Vis., 2: 467–479. repeated exposure to fearful and neutral faces: a functional
Carter, C.S., Botvinick, M.M. and Cohen, J.D. (1999) The MRI study. Brain Res. Bull., 59: 387–392.
contribution of the anterior cingulate cortex to executive Fox, E. (1993) Allocation of Visual-Attention and Anxiety.
processes in cognition. Rev. Neurosci., 10: 49–57. Cogn. Emotion, 7: 207–215.
Clark, V.P., Fan, S. and Hillyard, S.A. (1995) Identification of Fox, E. (2002) Processing emotional facial expressions: the role
early visual evoked potential generators by retinotopic and of anxiety and awareness. Cogn. Affect Behav. Neurosci., 2:
topographic analyses. Hum. Brain Mapp., 2: 170–187. 52–63.
Clark, V.P. and Hillyard, S.A. (1996) Spatial selective attention Fox, E., Lester, V., Russo, R., Bowles, R.J., Pichler, A. and
affects early extrastriate but not striate components of the Dutton, K. (2000) Facial expressions of emotion: Are angry
visual evoked potential. J. Cogn. Neurosci., 8: 387–402. faces detected more efficiently? Cogn. Emotion, 14: 61–92.
Corbetta, M., Kincade, J.M., Ollinger, J.M., McAvoy, M.P. Fox, E., Russo, R., Bowles, R. and Dutton, K. (2001) Do
and Shulman, G.L. (2000) Voluntary orienting is dissociated threatening stimuli draw or hold visual attention in subclin-
from target detection in human posterior parietal cortex. ical anxiety? J. Exp. Psychol. Gen., 130: 681–700.
Nat. Neurosci., 3: 292–297. Foxe, J.J. and Simpson, G.V. (2002) Flow of activation from
Corbetta, M. and Shulman, G.L. (2002) Control of goal- V1 to frontal cortex in humans. A framework for defining
directed and stimulus-driven attention in the brain. Nat. Rev. ‘‘early’’ visual processing. Exp. Brain Res., 142: 139–150.
Neurosci., 3: 201–215. Fredrikson, M., Wik, G., Fischer, H. and Andersson, J. (1995)
Di Russo, F., Martinez, A. and Hillyard, S.A. (2003) Source Affective and attentive neural networks in humans: a PET
analysis of event-related cortical activity during visuo-spatial study of Pavlovian conditioning. Neuroreport, 7: 97–101.
attention. Cereb. Cortex, 13: 486–499. Frijda, N.H. (1986) The Emotions. Cambridge University
Di Russo, F., Martinez, A., Sereno, M.I., Pitzalis, S. and Press, New-York.
Hillyard, S.A. (2002) Cortical sources of the early compo- Gitelman, D.R., Nobre, A.C., Parrish, T.B., LaBar, K.S., Kim,
nents of the visual evoked potential. Hum. Brain Mapp., Y.H., Meyer, J.R. and Mesulam, M. (1999) A large-scale
15: 95–111. distributed network for covert spatial attention: further
89

anatomical delineation based on stringent behavioural and cortex during directed attention in the absence of visual
cognitive controls. Brain, 122(Pt 6): 1093–1106. stimulation. Neuron, 22: 751–761.
Grave de Peralta Menendez, R., Murray, M.M., Michel, C.M., Kastner, S. and Ungerleider, L.G. (2000) Mechanisms of visual
Martuzzi, R. and S.L., G.A. (2004) Electrical neuroimaging attention in the human cortex. Annu. Rev. Neurosci., 23:
based on biophysical constraints. Neuroimage, 21: 527–539. 315–341.
Green, D. and Swets, J. (1966) Signal Detection Theory and Kawasaki, H., Kaufman, O., Damasio, H., Damasio, A.R.,
Psychophysics. Robert E. Krieger Publishing Company, Granner, M., Bakken, H., Hori, T., Howard, M.A. and III
Huntington, NY. and Adolphs, R. (2001) Single-neuron responses to emotional
Grill-Spector, K., Kourtzi, Z. and Kanwisher, N. (2001) The visual stimuli recorded in human ventral prefrontal cortex.
lateral occipital complex and its role in object recognition. Nat. Neurosci., 4: 15–16.
Vision Res., 41: 1409–1422. Keil, A., Moratti, S., Sabatinelli, D., Bradley, M.M. and Lang,
Halgren, E., Baudena, P., Heit, G., Clarke, J.M., Marinkovic, P.J. (2005) Additive effects of emotional content and spatial
K. and Clarke, M. (1994) Spatio-temporal stages in face and selective attention on electrocortical facilitation. Cereb.
word processing. I. Depth-recorded potentials in the human Cortex, 15: 1187–1197.
occipital, temporal and parietal lobes [corrected]. J. Physiol. Kerkhoff, G. (2001) Spatial hemineglect in humans. Prog.
Paris, 88: 1–50. Neurobiol., 63: 1–27.
Halgren, E., Raij, T., Marinkovic, K., Jousmaki, V. and Hari, R. Kincade, J.M., Abrams, R.A., Astafiev, S.V., Shulman, G.L.
(2000) Cognitive response profile of the human fusiform face and Corbetta, M. (2005) An event-related functional mag-
area as determined by MEG. Cereb. Cortex, 10: 69–81. netic resonance imaging study of voluntary and stimulus-
Hansen, C.H. and Hansen, R.D. (1988) Finding the face in the driven orienting of attention. J. Neurosci., 25: 4593–4604.
crowd –– an anger superiority effect. J. Pers. Soc. Psychol., Koster, E.H.W., Crombez, G., Van Damme, S., Verschuere, B.
54: 917–924. and De Houwer, J. (2004a) Does imminent threat capture
Hartikainen, K.M., Ogawa, K.H. and Knight, R.T. (2000) and hold attention? Emotion, 4: 312–317.
Transient interference of right hemispheric function due to Koster, E.H.W., Crombez, G., Van Damme, S., Verschuere, B.
automatic emotional processing. Neuropsychologia, 38: and De Houwer, J. (2005) Signals for threat modulate at-
1576–1580. tentional capture and holding: fear-conditioning and extinc-
Heinze, H.J., Luck, S.J., Mangun, G.R. and Hillyard, S.A. tion during the exogenous cueing task. Cogn. Emotion,
(1990) Visual event-related potentials index focused attention 19: 771–780.
within bilateral stimulus arrays.1. Evidence for early selec- Koster, E.H.W., Crombez, G., Verschuere, B. and De Houwer,
tion. Electroen. Clin. Neuro., 75: 511–527. J. (2004b) Selective attention to threat in the dot probe par-
Hillyard, S.A. and Anllo-Vento, L. (1998) Event-related brain adigm: differentiation vigilance from difficulty to disengage.
potentials in the study of visual selective attention. Proc. Behav. Res. Ther., 42: 1083–1092.
Natl. Acad. Sci. USA, 95: 781–787. Kourtzi, Z. and Kanwisher, N. (2001) Representation of per-
Hillyard, S.A., Vogel, E.K. and Luck, S.J. (1998) Sensory gain ceived object shape by the human lateral occipital complex.
control (amplification) as a mechanism of selective attention: Science, 293: 1506–1509.
electrophysiological and neuroimaging evidence. Philos. Krolak-Salmon, P., Fischer, C., Vighetto, A. and Mauguiere, F.
Trans. R. Soc. Lond. B Biol. Sci., 353: 1257–1270. (2001) Processing of facial emotional expression: spatio-tem-
Holland, P.C. and Gallagher, M. (1999) Amygdala circuitry in poral data as assessed by scalp event-related potentials. Eur.
attentional and representational processes. Trends Cogn. J. Neurosci., 13: 987–994.
Sci., 3: 65–73. Krolak-Salmon, P., Henaff, M.A., Vighetto, A., Bertrand, O.
Hopfinger, J.B., Buonocore, M.H. and Mangun, G.R. (2000) and Mauguiere, F. (2004) Early amygdala reaction to fear
The neural mechanisms of top-down attentional control. spreading in occipital, temporal, and frontal cortex: a depth
Nat. Neurosci., 3: 284–291. electrode ERP study in human. Neuron, 42: 665–676.
Hopfinger, J.B. and Mangun, G.R. (1998) Reflexive attention Lane, R.D., Reiman, E.M., Axelrod, B., Yun, L.S., Holmes, A.
modulates processing of visual stimuli in human extrastriate and Schwartz, G.E. (1998) Neural correlates of levels of
cortex. Psychol. Sci., 9: 441–447. emotional awareness: evidence of an interaction between
Jeffreys, D.A. and Axford, J.G. (1972) Source locations emotion and attention in the anterior cingulate cortex. J.
of pattern-specific components of human visual evoked- Cogn. Neurosci., 10: 525–535.
potentials 1. Component of striate cortical origin. Exp. Brain Lang, P.J. (1979) A bio-informational theory of emotional im-
Res., 16: 1–21. agery. Psychophysiology, 16: 495–512.
Juth, P., Lundqvist, D., Karlsson, A. and Ohman, A. (2005) Lang, P.J., Bradley, M.M., Fitzsimmons, J.R., Cuthbert, B.N.,
Looking for foes and friends: Perceptual and emotional fac- Scott, J.D., Moulder, B. and Nangia, V. (1998) Emotional
tors when finding a face in the crowd. Emotion, 5: 379–395. arousal and activation of the visual cortex: An fMRI anal-
Kanwisher, N. (2001) Neural events and perceptual awareness. ysis. Psychophysiology, 35: 199–210.
Cognition, 79: 89–113. LeDoux, J. (1996) The Emotional Brain: The Mysterious
Kastner, S., Pinsk, M.A., De Weerd, P., Desimone, R. and Underpinnings of Emotional Life. Simon & Schuster,
Ungerleider, L.G. (1999) Increased activity in human visual New York.
90

Lehmann, D. and Skrandies, W. (1980) Reference-free identi- Navon, D. and Margalit, B. (1983) Allocation of attention ac-
fication of components of checkerboard-evoked multichannel cording to informativeness in visual recognition. Q. J. Exp.
potential fields. Electroencephalogr. Clin. Neurophysiol., 48: Psychol.-A, 35: 497–512.
609–621. Nobre, A.C., Coull, J.T., Frith, C.D. and Mesulam, M.M. (1999)
Liu, T., Pestilli, F. and Carrasco, M. (2005) Transient attention Orbitofrontal cortex is activated during breaches of expecta-
enhances perceptual performance and FMRI response in hu- tion in tasks of visual attention. Nat. Neurosci., 2: 11–12.
man visual cortex. Neuron, 45: 469–477. Noesselt, T., Driver, J., Heinze, H.J. and Dolan, R. (2005)
Luck, S.J. (1995) Multiple mechanisms of visual-spatial atten- Asymmetrical activation in the human brain during process-
tion: recent evidence from human electrophysiology. Behav. ing of fearful faces. Curr. Biol., 15: 424–429.
Brain Res., 71: 113–123. Noesselt, T., Hillyard, S.A., Woldorff, M.G., Schoenfeld, A.,
Luck, S.J., Heinze, H.J., Mangun, G.R. and Hillyard, S.A. Hagner, T., Jancke, L., Tempelmann, C., Hinrichs, H. and
(1990) Visual event-related potentials index focused attention Heinze, H.J. (2002) Delayed striate cortical activation during
within bilateral stimulus arrays. 2. Functional dissociation of spatial attention. Neuron, 35: 575–587.
P1 and N1 components. Electroen. Clin. Neuro., 75: 528–542. Ohman, A., Lundqvist, D. and Esteves, F. (2001) The face in
MacDonald, A.W., Cohen, J.D., Stenger, V.A. and Carter, C.S. the crowd revisited: a threat advantage with schematic stim-
(2000) Dissociating the role of the dorsolateral prefrontal and uli. J. Pers. Soc. Psychol., 80: 381–396.
anterior cingulate cortex in cognitive control. Science, 288: Ohman, A. and Mineka, S. (2001) Fears, phobias, and prepar-
1835–1838. edness: toward an evolved module of fear and fear learning.
MacLeod, C., Mathews, A. and Tata, P. (1986) Attentional bias Psychol. Rev., 108: 483–522.
in emotional disorders. J. Abnorm. Psychol., 95: 15–20. Oostenveld, R. and Praamstra, P. (2001) The five percent elec-
Marois, R. and Ivanoff, J. (2005) Capacity limits of informa- trode system for high-resolution EEG and ERP measure-
tion processing in the brain. Trends Cogn. Sci., 9: 296–305. ments. Clin. Neurophysiol., 112: 713–719.
Martinez, A., Anllo-Vento, L., Sereno, M.I., Frank, L.R., Pascual-Marqui, R.D., Michel, C.M. and Lehmann, D. (1994)
Buxton, R.B., Dubowitz, D.J., Wong, E.C., Hinrichs, H., Low resolution electromagnetic tomography: a new method
Heinze, H.J. and Hillyard, S.A. (1999) Involvement of striate for localizing electrical activity in the brain. Int. J. Psycho-
and extrastriate visual cortical areas in spatial attention. Nat. physiol., 18: 49–65.
Neurosci., 2: 364–369. Peelen, M.V., Heslenfeld, D.J. and Theeuwes, J. (2004) Endog-
Mathews, A., Yiend, J. and Lawrence, A.D. (2004) Individual enous and exogenous attention shifts are mediated by the
differences in the modulation of fear-related brain activation same large-scale neural network. Neuroimage, 22: 822–830.
by attentional control. J. Cogn. Neurosci., 16: 1683–1694. Pessoa, L., McKenna, M., Gutierrez, E. and Ungerleider, L.G.
Mesulam, M.M. (1998) From sensation to cognition. Brain, (2002) Neural processing of emotional faces requires atten-
121(Pt 6): 1013–1052. tion. Proc. Natl. Acad. Sci. USA, 99: 11458–11463.
Michel, C.M., Seeck, M. and Landis, T. (1999) Spatiotemporal Phelps, E.A., Ling, S. and Carrasco, M. (2006) Emotion facil-
dynamics of human cognition. News Physiol. Sci., 14: 206–214. itates perception and potentiates the perceptual benefit of
Mogg, K. and Bradley, B.P. (1998) A cognitive-motivational attention. Psychol. Sci., 17: 292–299.
analysis of anxiety. Behav. Res. Ther., 36: 809–848. Phillips, M.L., Medford, N., Young, A.W., Williams, L., Williams
Mogg, K. and Bradley, B.P. (1999a) Orienting of attention to , S.C., Bullmore, E.T., Gray, J.A. and Brammer, M.J. (2001)
threatening facial expressions presented under conditions of Time courses of left and right amygdalar responses to fearful
restricted awareness. Cogn. Emotion, 13: 713–740. facial expressions. Hum. Brain Mapp., 12: 193–202.
Mogg, K. and Bradley, B.P. (1999b) Some methodological is- Picton, T.W., Bentin, S., Berg, P., Donchin, E., Hillyard, S.A.,
sues in assessing attentional biases for threatening faces in Johnson, R., Miller, G.A., Ritter, W., Ruchkin, D.S., Rugg,
anxiety: a replication study using a modified version of the M.D. and Taylor, M.J. (2000) Guidelines for using human
probe detection task. Behav. Res. Ther., 37: 595–604. event-related potentials to study cognition: recording stand-
Mogg, K., Bradley, B.P. and Hallowell, N. (1994) Attentional ards and publication criteria. Psychophysiology, 37: 127–152.
bias to threat –– roles of trait anxiety, stressful events, and Pizzagalli, D., Regard, M. and Lehmann, D. (1999) Rapid
awareness. Q. J. Exp. Psychol.-A, 47: 841–864. emotional face processing in the human right and left brain
Mogg, K., Bradley, B.P. and Williams, R. (1995) Attentional hemispheres: an ERP study. Neuroreport, 10: 2691–2698.
bias in anxiety and depression –– the role of awareness. Brit. Pizzagalli, D.A., Lehmann, D., Hendrick, A.M., Regard, M.,
J. Clin. Psychol., 34: 17–36. Pascual-Marqui, R.D. and Davidson, R.J. (2002) Affective
Mogg, K., McNamara, J., Powys, M., Rawlinson, H., Seiffer, judgments of faces modulate early activity (approximately
A. and Bradley, B.P. (2000) Selective attention to threat: 160 ms) within the fusiform gyri. Neuroimage, 16: 663–677.
A test of two cognitive models of anxiety. Cogn. Emotion, Posner, M.I., Snyder, C.R.R. and Davidson, B.J. (1980)
14: 375–399. Attention and the detection of signals. J. Exp. Psychol.
Morris, J.S., Friston, K.J., Buchel, C., Frith, C.D., Young, Gen., 109: 160–174.
A.W., Calder, A.J. and Dolan, R.J. (1998) A neuromodula- Pourtois, G., Grandjean, D., Sander, D. and Vuilleumier,
tory role for the human amygdala in processing emotional P. (2004) Electrophysiological correlates of rapid spatial ori-
facial expressions. Brain, 121(Pt 1): 47–57. enting towards fearful faces. Cereb. Cortex, 14: 619–633.
91

Pourtois, G., Schwartz, S., Seghier, M.L., Lazeyras, F. and Vogel, E.K. and Luck, S.J. (2000) The visual N1 component as
Vuilleumier, P. (2006) Neural systems for orienting attention an index of a discrimination process. Psychophysiology, 37:
to the location of threat signals: an event-related fMRI study. 190–203.
Neuroimage, 31: 920–933. Vuilleumier, P. (2002) Facial expression and selective attention.
Pourtois, G., Thut, G., Grave de Peralta, R., Michel, C. and Curr. Opin. Psychiatr., 15: 291–300.
Vuilleumier, P. (2005) Two electrophysiological stages of Vuilleumier, P. (2005) How brains beware: neural mechanisms
spatial orienting towards fearful faces: early temporo-parietal of emotional attention. Trends Cogn. Sci., 9: 585–594.
activation preceding gain control in extrastriate visual cortex. Vuilleumier, P., Armony, J.L., Clarke, K., Husain, M., Driver,
Neuroimage, 26: 149–163. J. and Dolan, R.J. (2002) Neural response to emotional faces
Pratto, F. and John, O.P. (1991) Automatic vigilance – the with and without awareness: event-related fMRI in a parietal
attention-grabbing power of negative social information. patient with visual extinction and spatial neglect. Ne-
J. Pers. Soc. Psychol., 61: 380–391. uropsychologia, 40: 2156–2166.
Sabatinelli, D., Bradley, M.M., Fitzsimmons, J.R. and Lang, Vuilleumier, P., Armony, J.L. and Dolan, R.J. (2004a) Recip-
P.J. (2005) Parallel amygdala and inferotemporal activation rocal links between emotion and attention. In: Frackowiak,
reflect emotional intensity and fear relevance. Neuroimage, R.J.S. and Mazziotta, J. (Eds.), Human Brain Function.
24: 1265–1270. Elsevier, London, pp. 419–444.
Sander, D., Grandjean, D., Pourtois, G., Schwartz, S., Seghier, Vuilleumier, P., Armony, J.L., Driver, J. and Dolan, R.J. (2001)
M.L., Scherer, K.R. and Vuilleumier, P. (2005) Emotion and Effects of attention and emotion on face processing in the
attention interactions in social cognition: brain regions in- human brain: an event-related fMRI study. Neuron, 30:
volved in processing anger prosody. Neuroimage, 28: 848–858. 829–841.
Searcy, J.H. and Bartlett, J.C. (1996) Inversion and processing Vuilleumier, P. and Pourtois, G. (in press) Distributed and
of component and spatial-relational information in faces. interactive brain mechanisms during emotion face
J. Exp. Psychol. Hum., 22: 904–915. perception: evidence from functional neuroimaging. Ne-
Shulman, G.L., d’Avossa, G., Tansy, A.P. and Corbetta, M. uropsychologia.
(2002) Two attentional processes in the parietal lobe. Cereb. Vuilleumier, P., Richardson, M.P., Armony, J.L., Driver, J. and
Cortex, 12: 1124–1131. Dolan, R.J. (2004b) Distant influences of amygdala lesion on
Spielberger, C.D. (1983) Manual for the State-Trait Anxiety visual cortical activation during emotional face processing.
Inventory. Consulting Psychologists Press, Palo Alto, CA. Nat. Neurosci., 7: 1271–1278.
Stormark, K.M., Nordby, H. and Hugdahl, K. (1995) Attent- Vuilleumier, P. and Schwartz, S. (2001a) Beware and be aware:
ional shifts to emotionally charged cues — Behavioral and Capture of spatial attention by fear-related stimuli in neglect.
Erp data. Cogn. Emotion, 9: 507–523. Neuroreport, 12: 1119–1122.
Super, H., van der Togt, C., Spekreijse, H. and Lamme, V.A. Vuilleumier, P. and Schwartz, S. (2001b) Emotional facial ex-
(2003) Internal state of monkey primary visual cortex (V1) pressions capture attention. Neurology, 56: 153–158.
predicts figure-ground perception. J. Neurosci., 23: 3407–3414. Walter, W.G., Cooper, R., Aldridge, V.J., McCallum, W.C.
Surguladze, S.A., Brammer, M.J., Young, A.W., Andrew, C., and Winter, A.L. (1964) Contingent negative variation: An
Travis, M.J., Williams, S.C. and Phillips, M.L. (2003) A electric sign of sensorimotor association and expectancy in
preferential increase in the extrastriate response to signals of the human brain. Nature, 203: 380–384.
danger. Neuroimage, 19: 1317–1328. Williams, J.M.G., Mathews, A. and MacLeod, C. (1996) The
Tallon-Baudry, C., Bertrand, O., Henaff, M.A., Isnard, J. and emotional stroop task and psychopathology. Psychol. Bull.,
Fischer, C. (2005) Attention modulates gamma-band oscil- 120: 3–24.
lations differently in the human lateral occipital cortex and Williams, M.A., Moss, S.A., Bradshaw, J.L. and Mattingley,
fusiform gyrus. Cerebral Cortex, 15: 654–662. J.B. (2005) Look at me, I’m smiling: visual search for threat-
Thiel, C.M., Zilles, K. and Fink, G.R. (2004) Cerebral correlates ening and nonthreatening facial expressions. Vis. Cogn., 12:
of alerting, orienting and reorienting of visuospatial attention: 29–50.
an event-related fMRI study. Neuroimage, 21: 318–328. Woldorff, M.G., Hazlett, C.J., Fichtenholtz, H.M., Weissman,
Tranel, D., Damasio, A.R. and Damasio, H. (1988) Intact rec- D.H., Dale, A.M. and Song, A.W. (2004) Functional par-
ognition of facial expression, gender and age in patients with cellation of attentional control regions of the brain. J. Cogn.
impaired recognition of face identity. Neurology, 38: 690–696. Neurosci., 16: 149–165.
van der Lubbe, R.H. and Woestenburg, J.C. (1997) Modulation Zajonc, R.B. (1980) Feeling and thinking — preferences need
of early ERP components with peripheral precues: a trend no inferences. Am. Psychol., 35: 151–175.
analysis. Biol. Psychol., 45: 143–158. Zald, D.H. (2003) The human amygdala and the emotional
Van der Lubbe, R.H. and Woestenburg, J.C. (2000) Location evaluation of sensory stimuli. Brain Res. Brain Res. Rev.,
selection in the visual domain. Psychophysiology, 37: 662–676. 41: 88–123.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 5

The neural basis of narrative imagery:


emotion and action

Dean Sabatinelli1,, Peter J. Lang1, Margaret M. Bradley1 and Tobias Flaisch2

1
NIMH Center for the Study of Emotion and Attention, University of Florida, PO Box 100165 HSC, Gainesville,
FL 32608, USA
2
Department of Psychology, University of Konstanz, Universitätstrasse 10, 78457 Konstanz, Germany

Abstract: It has been proposed that narrative emotional imagery activates an associative network of stim-
ulus, semantic, and response (procedural) information. In previous research, predicted response compo-
nents have been demonstrated through psychophysiological methods in peripheral nervous system. Here we
investigate central nervous system concomitants of pleasant, neutral, and unpleasant narrative imagery
with functional magnetic resonance imaging. Subjects were presented with brief narrative scripts over
headphones, and then imagined themselves engaged in the described events. During script perception,
auditory association cortex showed enhanced activation during affectively arousing (pleasant and un-
pleasant), relative to neutral imagery. Structures involved in language processing (left middle frontal gyrus)
and spatial navigation (retrosplenium) were also active during script presentation. At the onset of narrative
imagery, supplementary motor area, lateral cerebellum, and left inferior frontal gyrus were initiated,
showing enhanced signal change during affectively arousing (pleasant and unpleasant), relative to neutral
scripts. These data are consistent with a bioinformational model of emotion that considers response mo-
bilization as the measurable output of narrative imagery.

Keywords: emotion; script imagery; fMRI

It has been proposed that narrative imagery is which narrative emotional imagery, relative to
characterized by a network assembly of stimulus, neutral imagery, is accompanied by heightened
response, and semantic or ‘meaning’ representa- physiological reactivity. This ‘efferent leakage’ can
tions (Lang, 1979, 1985, 1994). Considering emo- be demonstrated in cardiovascular, autonomic,
tion as disposition toward action (Frijda, 1986; and somato-motor systems (Lang, 1977; Miller et
Lang et al., 1994), emotional imagery may be con- al., 1987; Cook et al., 1988; Vrana and Lang, 1990;
sidered essentially equivalent to emotional re- McNeil et al., 1993; for a detailed review see Cut-
sponding, with overt motor aspects ‘gated out.’ hbert et al., 1991). While heightened peripheral
From this perspective, emotional engagement dur- reactivity during narrative emotional imagery has
ing narrative imagery is defined by the extent of been thoroughly demonstrated, central indices of
response mobilization measurable in central and response information processing have only begun
peripheral nervous system. This conception has to be explored.
been corroborated by experimental paradigms in Several neuroimaging studies of mental imagery
have examined simple sensorimotor tasks (as
Corresponding author. Tel.: +1-352-392-2439; opposed to dynamic, affect-relevant behavior).
Fax: +1-352-392-6047; E-mail: sabat@ufl.edu Focusing on the visual system, imagery paradigms

DOI: 10.1016/S0079-6123(06)56005-4 93
94

commonly present a stimulus and ask subjects to script contents, which may reflect a similar pattern
imagine the stimulus after it has been removed. of enhanced BOLD signal intensity during emo-
Visual cortical activation during imagery can be tional, relative to neutral, imagery script process-
seen in categorically specific regions of inferior ing, as is the case in the visual system during
temporal lobe (O’Craven and Kanwisher, 2000; emotional, relative to neutral picture processing
Mechelli et al., 2004) as well as functionally (Lang et al., 1998, Bradley et al., 2003; Sabatinelli
specific regions of occipital lobe (Kosslyn et al., et al., 2004, 2005).
1995; Thompson et al., 2001), in one report with As subjects vividly imagine the script narratives,
retinotopic consistency (Slotnick et al., 2005); we hypothesize that response-related neural cir-
however, the involvement of primary visual cor- cuitry will become active, and potentially show
tex in visual imagery is inconsistently reported enhanced activation during emotional, relative
(Goebel et al., 1998; Ishai et al., 2000; see Kosslyn to neutral contents. Past studies have included
and Thompson, 2003). Motor imagery studies only aversive and neutral imagery scripts; there-
have exploited similar, simple tasks, intermixing fore, differences identified may reflect differences
motor execution and motor imagery trials that re- in emotional intensity (which may be high in
port premotor and cerebellar activities during both appetitive and aversive contexts) or emotional
tasks (Hanakawa et al., 2003; Nair et al., 2003; valence. The inclusion of pleasant as well as un-
Solodkin et al., 2004). The involvement of primary pleasant imagery scripts will enable us to more
motor areas in motor imagery has been variously specifically investigate the role of emotional
reported as partial (Ehrsson et al., 2003) and arousal as well as emotional valence on brain
absent (Dechent et al., 2004). activation.
Mental imagery of diagnosis-relevant narratives
has been explored in a few clinical studies in an
attempt to understand the neural effects of therapy, Method
which often employ narrative imagery as a desen-
sitization method. These studies typically contrast Procedure
symptom-relevant (therefore aversive) imagery
with neutral imagery, and report increased Twenty-one introductory psychology students (11
activity in limbic and prefrontal cortex (Rauch et male) at the University of Florida participated in
al., 1996; Shin et al., 1997, 2004; Bystritsky et al., the experiment for course credit. All volunteers
2001; Lanius et al., 2003). The methodologies consented to participate after reading a description
used in the experiments were variants of narrative of the study, approved by the local human
imagery paradigms originated in earlier psycho- subjects’ review board. Before entering the scan-
physiological research (Lang, 1977). ner room, participants read all imagery scripts
Here we investigate directly the neural concom- and recorded ratings of pleasantness and arousal,
itants of pleasant, neutral, and unpleasant narra- using the Self Assessment Manikin (Lang, 1980).
tive imagery using functional MRI (fMRI) to The scripts consisted of 12 exemplars of pleasant
probe potential response-related activation that scene contents, 6 of neutral scene contents, and 12
has been associated with emotional imagery. We of unpleasant scene contents1 (see appendix for
adapted an established imagery paradigm to the scripts).
scanner environment, presenting brief story-scripts Prior to entering the bore of the Siemens 3 T
to a nonclinical sample of young, healthy partic- Allegra MR scanner, subjects were fitted with MR-
ipants, and asked them to vividly imagine them- compatible, shielded headphones, a button re-
selves engaged in the emotional and neutral scenes sponse paddle, and a patient-alarm squeezeball
described while blood oxygen level-dependent 1
Two additional specific fear-relevant script contents (dental
(BOLD) signal is recorded throughout the brain. and snake fear) were excluded from the current analyses. These
We expect to identify language processing struc- script contents were included in the design in order to allow for
tures involved in the perceptual intake of narrative comparisons with clinical populations.
95

that was positioned within easy reach. Padding Data analyses


and explicit verbal instruction were used to limit
head motion. Once comfortable, subjects were The functional time series was motion-corrected,
moved into the bore, and an 8 min 3D structural mean-intensity adjusted, linearly detrended, high
volume was collected. T1-weighted structural scan- pass filtered at 0.02 Hz, and spatially smoothed
ning involved the collection of 160 sagittal slices with a 5 mm spatial filter using BrainVoyager QX
with 1 mm isotropic voxels. (www.brainvoyager.com). All subjects’ preproc-
After a brief delay, the functional images were essed functional activity was coregistered to their
acquired. The functional prescription included fifty respective structural volumes, after transformation
2.5-mm-thick coronal slices (0.5 mm gap) and cov- into standardized Talairach coordinate space (Ta-
ered the entire cortex in most subjects. The session lairach and Tournoux, 1988). These standardized
included 580 images, with 3 s temporal resolution functional maps were then analyzed using ANOVA
and 16 ml voxel size (3 s TR, 35 ms TE, 160 mm to identify voxels whose intensity, after convolution
FOV, 64  64 matrix). Subjects were asked to close with a standard hemodynamic response function,
their eyes, and imagery trials were presented in a was associated with auditory script presentation as
continuous series (see Fig. 1). A trial began with an well as script imagery. A conservative false discov-
auditory script presentation for 12 s. To limit the ery rate (FDR) correction of po0.00001 revealed
masking effect of scanner noise, a thin sound-dead- clusters of activity associated with script presenta-
ening mat was attached to the inside of the bore, and tion and script imagery, from which regions of in-
the headphones were also shielded for sound. To be terest (ROI) were sampled in individual subjects
clearly audible over the background of scanner and broken down by script content.
noise, yet not uncomfortable to the listener, auditory ROIs were sampled from each subject’s ANOVA
scripts were presented at 98 dB. All subjects reported map of significant (po0.01) script- and imagery-
that the scripts were clear and decipherable. associated activity, within the area of interest, yet
Each auditory script was followed immediately allowing the center of the 100 ml ROI sample to be
by a tone signaling subjects to vividly imagine sensitive to the individual’s specific neuroanatomy.
yourself fully engaged in the scene described. A Peak signal in each ROI (average percent BOLD
second tone signaled the end of the 12 s imagery signal 3–9 s post script/imagery onset) for each
period, after which subjects listened to a series of subject was entered into an ANOVA including
nine numbers, one per second, and made a button- script content (erotica, pleasant arousing, neutral,
press when a prespecified number was presented. A contamination, and threat). This two-step analysis
delay of 3–9 s was followed by the next trial, in process enables conservative identification of active
pseudo-random order by content (not more than clusters in the average group activity map while
two trials of the same content in succession); the remaining sensitive to individual differences in
entire series lasted 29 min. Script presentation or- structural variation, which is typically lost after
der was unique for each subject. spatial standardization. ROI analyses also enable

Fig. 1. Trial structure. Narrative imagery scripts were presented over headphones as subjects lay in the magnet. Immediately after
script presentation, subjects imagined themselves as actively engaged in the scene described. After 12 s, a tone signaled subjects to listen
to a 9 s series of numbers over the headphones, and make a button-press when a predetermined number was presented. After a delay of
3–9 s, the next trial was presented. The entire series lasted approximately 29 min.
96

Fig. 2. Brain activation during narrative script presentation. Significant clusters of activity were evoked by auditory scripts in primary
(A) and secondary auditory cortex (B) as well as retrosplenium (C) and the left MFG (D). The minimum threshold of significance in
the random-effects analysis was po1  106. ROI analyses from individual subjects yielded each cluster’s time course of signal change,
with respect to script presentation (0–12 s). Blue waveforms represent signal change during pleasant scripts, green during neutral
scripts, and red during unpleasant scripts. Error bars represent standard errors of the mean.

the examination of event-related time courses of (see Fig. 2A, B, left panel). In addition, script
clusters of activity. presentation triggered activation in bilateral re-
trosplenial cortex (Fig. 2C, left) and left middle
Results frontal gyrus (MFG; Fig. 2D, left).2 ROI time
courses (waveform panels of Fig. 2) demonstrate
Script ratings that peak signal change was significantly affected
The narrative scripts were rated (1–9 scale) as in- by script content in secondary auditory cortex
tended, with neutral scripts (7.0, SD 0.6) falling be- (Fig. 2B, right panel), F(2,40) ¼ 21.75, po0.001,
tween pleasant (8.5, SD 0.3) and unpleasant (2.4, SD with greater signal change during pleasant
0.6) scripts in terms of valence, and ranking below and unpleasant, relative to neutral scripts (quad-
(3.0, SD 1.1) both pleasant (7.3, SD 1.3) and un- ratic trend F(1,20) ¼ 36.75, po0.001). No effects
pleasant scripts (6.7, SD 1.1) in terms of emotional or content interactions of hemisphere were found
arousal. Pleasant and unpleasant scripts were rated in tests of primary or secondary auditory BOLD
equivalently in emotional arousal (t ¼ 1.20 ns). signal.

2
Clusters of activation during script presentation that did not
BOLD signal during script-presentation
meet our conservative statistical threshold were seen in vent-
romedial prefrontal cortex and parahippocampal gyrus. During
The presentation of auditory scripts elicited script imagery, below-threshold clusters of activity were present
activity in primary and secondary auditory cortex in caudate, hippocampus, and posterior parietal lobe.
97

Script content also affected signal change in (Fig. 3C). ROI waveforms (right panels of
left MFG, F(2,38) ¼ 7.11, po0.01, again with Fig. 3A–C) show that the peak level of signal
greater signal change during pleasant and change in SMA was sensitive to imagery
unpleasant, relative to neutral scripts (quadratic content F(2,38) ¼ 3.94, po0.05, with greater
trend F(1,19) ¼ 8.77, po0.01). In retrosplenium, BOLD signal evoked during pleasant and
script valence modulated signal change as unpleasant, relative to neutral script imagery
well, F(2,34) ¼ 4.71, po0.05, with greater signal (quadratic trend F(1,19) ¼ 5.02, po0.05). Left
change during neutral relative to pleasant and un- IFG showed the same sensitivity to imagery con-
pleasant script presentations (quadratic trend tent F(2,32) ¼ 11.91, po0.001, with greater
F(1,17) ¼ 7.53, po0.05). signal change during pleasant and unpleasant
relative to neutral script imagery (quadratic
trend F(1,16) ¼ 26.50, po0.001). Signal change
BOLD signal during mental imagery in the right lateral cerebellum showed a marginal
sensitivity to imagery content F(2,34) ¼ 3.44,
The onset of the imagery period triggered an p ¼ 0.07, with a trend toward greater signal
increase in BOLD signal in supplementary change during pleasant and unpleasant, relative
motor area (SMA; Fig. 3A), left inferior frontal to neutral imagery (quadratic trend F(1,17) ¼
gyrus (IFG; Fig. 3B), and right lateral cerebellum 3.93, p ¼ 0.06).

Fig. 3. Brain activity during narrative imagery. Significant clusters of activity were evoked by narrative imagery in SMA (A), left IFG
(B), and right lateral cerebellum (C). The minimum threshold of significance in the random-effects analysis was po1  106. Region of
interest analyses from individual subjects yield each cluster’s time course of signal change, with respect to the imagery period (12–24 s).
Blue waveforms represent signal change during pleasant imagery, green during neutral imagery, and red during unpleasant imagery.
Error bars represent standard errors of the mean. In panel D, the relative coronal locations of the left MFG cluster active during script
presentation (Fig. 2D, presented here in blue), and the left IFG cluster active during script imagery (in red) are highlighted.
98

Interestingly, two of the three ROIs that showed the background of scanner noise, the presentation
signal increases at the onset of the imagery task of the auditory script led to sharp and widespread
(SMA and IFG) also showed significant effects of BOLD signal increases in primary and secondary
script content 3–9 s after the presentation of the auditory cortex (Fig. 2). As in the visual system
auditory script (see Fig. 3, right panels). These during picture perception (Lang et al., 1998; Brad-
signal increases were greater during pleasant and ley et al., 2003; Sabatinelli et al., 2004, 2005), the
unpleasant relative to neutral script presentation intensity of signal change in auditory association
(SMA, F(2,38) ¼ 6.71, po0.01, quadratic trend cortex was modulated by emotional intensity, i.e.,
F(1,19) ¼ 9.36, po0.01; IFG, F(2,32) ¼ 4.01, the arousing pleasant and unpleasant scripts evoked
po0.05, quadratic trend F(1,16) ¼ 4.85, po0.05). (equivalently) greater signal change than did the
The script content effect is not continuous from neutral scripts. Interestingly, the primary auditory
presentation to imagery, however, as neither SMA cortex did not show an effect of script emotionality
nor IFG shows content effects at the end of the at the peak of signal change, yet the descending leg
script presentation period, prior to the onset of the of signal change (the first 6 s after script offset, see
imagery period (SMA, F(2,38) ¼ 1.56 ns; IFG, Fig. 2) — did show a reliable emotional arousal
F(2,32) ¼ 2.94 ns). effect (F(2,40) ¼ 25.02, po0.001, quadratic trend
F(1,20) ¼ 47.64, po0.001).
Outside the auditory cortex, the activity initi-
Discussion ated during script presentation in retrosplenium
and left IFG tended to return to baseline
The act of imagining oneself engaged in a narrative much more slowly. This gradual decline in the
recruits areas of the brain involved in planning and BOLD signal may reflect the involvement of
executing action–supplementary motor area, pre- these structures in both the decoding and the im-
frontal cortex, and cerebellum. These effects are agination of the narrative script. It might also be a
consistent with fMRI studies of explicit motor im- result of the tendency for subjects to begin the
agery (Hanakawa et al., 2003; Nair et al., 2003; imagery task while the script is being presented.
Solodkin et al., 2004; Cunnington et al., 2005). In For the current dataset, it is defensible to suggest
this dataset, it is demonstrated that the intensity of that the retrosplenium and IFG are involved in
preparatory motor activation reflects in part the narrative script intake, and perhaps script imagery
nature of the narrative; scenes characterized by as well.
emotional events evoke stronger signal increases. Beyond the timing of signal change, it is clear
These data support a bioinformational account of that the activity in retrosplenial cortex shows a
emotional imagery, in which response components unique pattern with respect to the other ROIs — a
of associative networks are thought to be more reverse arousal effect, with greater signal change
strongly activated by emotional, than neutral nar- associated with neutral script processing relative to
rative imagery. Prior reports of enhanced physio- pleasant and unpleasant script processing. In ne-
logical mobilization observed in peripheral uroimaging and neuropsychological studies, the
nervous system during emotional imagery (Lang retrosplenium has been closely tied to environ-
et al., 1980; Miller et al., 1987; Cook et al., 1988; mental navigation and orienting in large-scale
Vrana and Lang, 1990; McNeil et al., 1993) can space (Aquirre et al., 1998; Grön et al., 2000; for
thus be associated with response-related central review see Maguire, 2001). We can speculate that
nervous system correlates. Emotional imagery ap- in the absence of an emotional focus, our subjects
pears to reflect neural disposition toward action were processing the spatial details of the imagined
(e.g., approach, withdrawal), which can be ob- local environment to a greater extent. Future stud-
served as enhanced preparatory motor activation. ies in which the degree of environmental detail is
Immediately prior to each imagery period, manipulated could address this issue.
subjects listened to a recorded speaker read the The narrative scripts used in the current study
narrative in an emotionally muted voice. Despite were constructed to evoke pleasant, neutral, and
99

unpleasant contexts, but were not explicitly con- and thalamic activity during trauma-relevant im-
trolled for the represented level of physical agery than traumatized subjects without PTSD,
activity. Considering that emotion itself can be and no amygdala activation was reported. The
defined as disposition toward action, the separa- variability in reports of limbic activity during
tion of emotion and activity may be impossible to traumatic imagery may in part reflect individual
achieve. However, it might be possible in future differences in dissociative symptoms within
work to manipulate the level of physical activity PTSD samples (Lanius et al., 2006). The lack of
represented within each script category — to in- limbic activation in the current study may in part
clude active, inactive, and neutral as well as emo- be a result of our conservative analyses, as a
tional narratives. In this way, the potentially more liberal statistical threshold during imagery
entangled roles of action and emotion may be (FDR po0.05) revealed subcortical activity in
investigated. caudate and hippocampus, but no clusters in or-
The distinction between subsections of left pre- bitofrontal cortex or amygdala. It may be that
frontal cortex during script presentation and script subcortical recruitment in emotional imagery is
imagery (Fig. 3D) is consistent with functional specially potentiated in clinical populations. An-
distinctions commonly identified in studies of pho- other possibility is that subcortical motivational
nological and semantic language processing. In the circuits are more readily engaged by perceptual,
current dataset, encoding of the narrative scripts rather than imagined stimuli.
evoked activity in a superior, posterior subsection In summary, these data demonstrate several ef-
of the left MFG that has been shown to be in- fects; that auditory cortex activation reflects the
volved in phonological processing. During script emotional intensity of narrative scripts during per-
imagery, activity increases were seen in a more in- ception. Despite the need to decipher language,
ferior and anterior subsection of the left IFG, a affective modulation of cortical response in sec-
region associated with semantic language process- ondary, and perhaps primary, auditory cortex is
ing (Wagner et al., 2001; Gold et al., 2005; Gitel- evident while the script is heard. The current de-
man et al., 2005; see Hagoort, 2005). Perhaps as sign preexposed subjects to the script stimuli prior
subjects listen to the script, content is input, but to the experimental session, thus some of the dy-
semantic elaboration is delayed until the entire text namic aspects of narrative presentation were re-
is understood and the imagery period is signaled. moved, potentially converting the onset of each
Investigations of imagery-induced clinical symp- script into a static cue for memory retrieval. Fu-
toms have primarily reported activity in limbic ture work in which preexposure is manipulated
and prefrontal structures. In panic disorder will provide more information regarding the de-
patients, relative to healthy controls, Bystritsky pendence of stimulus dynamics on this effect. In
et al. (2001) reported limbic activation during any case, this sensory cortical effect is consistent
anxiety-provoking narrative imagery, including with picture processing effects in visual system, in
inferior and orbitofrontal cortex, anterior and which lateral occipital and inferior temporal visual
posterior cingulate, and hippocampus. Shin et al. system show reliably greater activation during af-
(1997), using positron emission tomography (PET) fective arousing, relative to neutral picture
in combat veterans with and without posttrau- processing (Lang et al., 1998; Bradley et al.,
matic stress disorder (PTSD), identified stronger 2003; Junghöfer et al., 2006). Thus, there may be
activation during trauma-related imagery in or- a modality-nonspecific mechanism by which emo-
bitofrontal cortex, insula, anterior cingulate, and tionally arousing stimuli evoke enhanced cortical
amygdala. A more recent PET study (Shin et al., processing.
2004) suggested amygdala hyperactivity in PTSD The neural effects of narrative imagery appear
to be inversely related to hyporeactivity in medial to be essentially response-related. Activity in
frontal cortex. However, in a similar design, trau- premotor, cerebellum, and prefrontal cortex
matized subjects with diagnosed PTSD (Lanius et show clear time-locked increases at the onset of
al., 2003) showed relatively less anterior cingulate the imagery task. Considering that response
100

readiness is especially critical in emotionally 6. Music murmurs in the background.


charged situations, it is perhaps not surprising You’re together in the big bed, naked
that emotional imagery triggered greater activa- but apart, eyes locked. You feel fingers
tion in these structures. barely touching, gliding softly along your
thigh.
Abbreviations 7. A moan of pleasure. Your body responds
slowly at first, languorously, and then with a
BOLD blood oxygen level dependent more urgent rhythm. You feel gentle hands,
FDR false discovery rate a soft mouth, your back arches.
fMRI functional magnetic resonance imaging 8. It’s a beautiful day and you’re heading a
IFG inferior frontal gyrus new convertible to the beach. The CD player
MFG middle frontal gyrus is blasting, and you’re singing along at the
MR magnetic resonance top of your voice.
PET positron emission tomography 9. It’s the last few minutes of the big game and
PTSD posttraumatic stress disorder it’s close. The crowd explodes in a deafening
SMA supplementary motor area roar. You jump up, cheering. Your team has
come from behind to win.
10. You shiver as your bodies brush together.
Acknowledgments You reach out. You want to touch every-
where, kiss everywhere. You hear the words,
This work was supported by grants from the ‘I love you’.
National Institutes of Mental Health P50 11. The registered letter says ‘You have just won
MH072850-01. ten million dollars!’ It’s amazing — You
bought the winning ticket in the lottery. You
Appendix cry, scream, jump with joy!
12. You are both aroused, breathless. You fall
Pleasant scripts together on the couch. Kisses on your neck,
face — warm hands fumbling with clothing,
1. You are lying together, legs over legs, arms hearts pounding.
around bodies — kisses deep and sweet. In
love on a blanket, beneath a tree, on a warm
Neutral scripts
summer day.
2. You tense as the roller coaster reaches the 1. You run the comb through your hair,
crest. Then, you are all plunging down, straighten your collar, smooth out the shirt’s
screaming above the roar, together, laugh- wrinkles. Water is running in the sink. You
ing, and waving your arms. turn it off and leave.
3. As soon as you saw each other, the affair 2. You are relaxing on a lawn chair, looking
began. You remember beautiful eyes look- out into the garden. A child’s tricycle is
ing straight into yours — your heart in your abandoned on the grass. You hear the low
throat, at the first touch. buzz of a lawn mower in the distance.
4. The band is terrific. The room vibrates with 3. It’s good to be able to do nothing and just
sound and your skin tingles. You’re dancing stretch out on the couch. The television is on
together, moving effortlessly with the music. with the sound off. You can hear the low
You’re feeling great! rumble of traffic in the distance.
5. The mountain air is clear and cold. The sun 4. You unfold the map, spread it out on the
glistens on the powder as you head down the table, and with your finger trace a route south
slope in gliding turns, mastering the moun- toward the beach. You refold the map, pick
tain, moving with a sure, easy grace. up your bag, and leave.
101

5. It’s a quiet day without much to do. You’re with teeth bared, leaping, and snarling in a
sitting around your place, resting, reading, crazy rage.
and looking out the window — where leaves 10. You gag, seeing a roach moving slowly over
swirl gently in the wind. the surface of the pizza. You knock the pie
6. You are sitting at the kitchen table with yes- on the floor. Warm cheese spatters on your
terday’s newspaper in front of you. You push shoes.
back the chair when you hear the coffee 11. You bite hungrily into the hamburger, and
maker slow to a stop. abruptly catch the putrid smell of spoiled
meat. You spit out, and a greasy piece falls
down your chin onto your pants.
12. You flinch, at the screech of brakes; you
Unpleasant scripts
look up, and see the speeding car slam into
1. The garbage can is upset. Maggots crawl on your friend. Her leg is crushed, the artery
the rotted food spilling out on the floor, torn, and blood pumps on the road.
staining the carpet. Your throat tightens
with a wave of nausea, but you must clean
it up.
2. A night landing in high winds: your hands References
clutch the seat-arms in the swaying plane.
Bradley, M.M., Sabatinelli, D., Lang, P.J., Fitzsimmons, J.R.,
Stomach queasy. The engine coughs; stops;
King, W. and Desai, P. (2003) Activation of the visual
restarts with a strange whine. cortex in motivated attention. Behav. Neurosci., 117:
3. A vagrant, wino, approaches you, yellow 369–380.
teeth and scabs on his face, clothes smelling Bystritsky, A., Pontillo, D., Powers, M., Sabb, F.W., Craske,
of mold and urine. You cringe as his hand M.G. and Bookheimer, S.Y. (2001) Functional MRI
touches your sleeve. changes during panic anticipation and imagery exposure.
Neuroreport, 12: 3953–3957.
4. As you ease the car onto the wooden bridge, Cook III, E.W., Melamed, B.G., Cuthbert, B.N., McNeil, D.W.
it groans. In the headlights, a broken railing and Lang, P.J. (1988) Emotional imagery and the differen-
swings in the wind. A swift current rams tial diagnosis of anxiety. J. Consult. Clin. Psychol., 56:
against the pilings below. 734–740.
5. The bathroom is filthy, toilet overflowing Cunnington, R., Windischberger, C. and Moser, E. (2005) Pre-
movement activity of the pre-supplementary motor area and
onto the floor spreading toward your feet. the readiness for action: studies of time-resolved event-
The smell is overwhelming and you run for related functional MRI. Hum. Mov. Sci., 24: 644–656.
the door. Cuthbert, B.N., Vrana, S.R. and Bradley, M.M. (1991)
6. You’re alone in the alley in a bad part of the Imagery: function and physiology. In: Ackles, P.K., Jennings,
J.R. and Coles, M.G.H. (Eds.) Advances in Psychophysiol-
city. A street gang slowly surrounds you,
ogy, Vol. 4. JAI, Greenwich, CT, pp. 1–42.
knives out, laughing with menace. Your Dechent, P., Merboldt, K.D. and Frahm, J. (2004) Is the
heart pounds as they close in. human primary motor cortex involved in motor imagery?
7. It’s late at night in a poorly lit parking lot. Brain Res. Cogn. Brain Res., 19: 138–144.
You are tense, clutching the keys. Your car Ehrsson, H.H., Geyer, S. and Naito, E. (2003) Imagery of vol-
stands alone in the distance, when footsteps untary movement of fingers, toes, and tongue activates cor-
responding body-part-specific motor representations. J.
sound behind you. Neurophysiol., 98: 3304–3316.
8. You are leaving the concert. (When) A Frijda, N.H. (1986) The Emotions. Cambridge, New York.
drunk, smelling of smoke and alcohol, Gitelman, D.R., Nobre, A.C., Sonty, S., Parrish, T.B. and Me-
stumbles into you and throws up on your sulam, M.M. (2005) Language network specializations: an
analysis with parallel task designs and functional magnetic
jacket. You retch, as vomit drips onto your
resonance imaging. Neuroimage, 26: 975–985.
hand. Goebel, R., Khorram-Sefat, D., Muckli, L., Hacker, H. and
9. You jump back, muscles tense, as the large Singer, W. (1998) The constructive nature of vision: direct
dog strains against the chain, slobbering evidence from functional magnetic resonance imaging studies
102

of apparent motion and motion imagery. Eur. J. Neurosci., and Menon, R.S. (2003) Recall of emotional states in
10: 1563–1673. posttraumatic stress disorder: an fMRI investigation.
Gold, B.T., Balota, D.A., Kirchhoff, B.A. and Buckner, R.L. Biol. Psychiatry, 53: 204–210.
(2005) Common and dissociable activation patterns associ- Maguire, E.A. (2001) The retrosplenial contribution to human
ated with controlled semantic and phonological processing: navigation: a review of lesion and neuroimaging findings.
evidence from FMRI adaptation. Cereb. Cortex, 15: Scand. J. Psychol., 42: 225–238.
1438–1450. McNeil, D.W., Vrana, S.R., Melamed, B.G., Cuthbert, B.N.
Grön, G., Wunderlich, A.P., Spitzer, M., Tomczak, R. and and Lang, P.J. (1993) Emotional imagery in simple and social
Riepe, M.W. (2000) Brain activation during human naviga- phobia: fear versus anxiety. J. Abnorm. Psychol., 102:
tion: gender-different neural networks as substrate of 212–225.
performance. Nat. Neurosci., 3: 404–408. Mechelli, A., Price, C.J., Friston, K.J. and Ishai, A. (2004)
Hagoort, P. (2005) On Broca, brain, and binding: a new frame- Where bottom-up meets top-down: neuronal imteractions
work. Trends Cogn. Sci., 9: 416–423. during perception and imagery. Cereb. Cortex, 14:
Hanakawa, T., Immisch, I., Toma, K., Dimyan, M.A., Van 1256–1265.
Gelderen, P. and Hallett, M. (2003) Functional properties of Miller, G.A., Levin, D.N., Kozak, M.J., Cook III, E.W.,
brain areas associated with motor execution and imagery. J. McLean Jr., A. and Lang, P.J. (1987) Individual differences
Neurophysiol., 89: 989–1002. in imagery and the psychophysiology of emotion. Cogn.
Ishai, A., Ungerleider, L.G. and Haxby, J.V. (2000) Distributed Emotion, 1: 367–390.
neural systems for the generation of visual images. Neuron, Nair, D.G., Purcott, K.L., Fuchs, A., Steinberg, F. and Kelso,
28: 979–990. J.A. (2003) Cortical and cerebellar activity of the human
Junghöfer, M., Sabatinelli, D., Bradley, M.M., Schupp, H.T., brain during imagined and executed unimanual and biman-
Elbert, T.R. and Lang, P.J. (2006) Fleeting images: rapid ual action sequences: a functional MRI study. Brain Res.
affect discrimination in the visual cortex. Neuroreport, 17: Cogn. Brain Res., 15: 250–260.
225–229. O’Craven, K.M. and Kanwisher, N. (2000) Mental imagery of
Kosslyn, S.M. and Thompson, W.L. (2003) When is early visual faces and places activates corresponding stiimulus-specific
cortex activated during visual mental imagery? Psychol. Bull., brain regions. J. Cogn. Neurosci., 12: 1013–1023.
129: 723–746. Rauch, S.L., van der Kolkm, B.A., Fislerm, R.E., Alpertm,
Kosslyn, S.M., Thompson, W.L., Kim, I.J. and Alpert, N.M. N.M., Orrm, S.P., Savage, C.R., Fischman, A.J., Jenike,
(1995) Topographical representations of mental images in M.A. and Pitman, R.K. (1996) A symptom provocation
primary visual cortex. Science, 378: 496–498. study of posttraumatic stress disorder using positron
Lang, P.J. (1977) Imagery in therapy: an information process- emission tomography and script-driven imagery. Arch.
ing analysis of fear. Behav. Ther., 8: 862–886. Gen. Psychiatry, 53: 380–387.
Lang, P.J. (1979) A bio-informational theory of emotional Sabatinelli, D., Bradley, M.M., Fitzsimmons, J.R. and Lang,
imagery. Psychophysiology, 16: 495–512. P.J. (2005) Parallel amygdala and inferotemporal activation
Lang, P.J. (1980) Behavioral treatment and bio-behavioral as- reflect emotional intensity and fear relevance. Neuroimage,
sessment: computer applications. In: Sidowski, J.B., Johnson, 24: 1265–1270.
J.H. and Williams, T.A. (Eds.), Technology in Mental Health Sabatinelli, D., Flaisch, T., Bradley, M.M., Fitzsimmons, J.R.
Care Delivery Systems. Ablex Publishing, Norwood, NJ, pp. and Lang, P.J. (2004) Affective picture perception: gender
119–137. differences in visual cortex? Neuroreport, 15: 1109–1112.
Lang, P.J. (1985) The cognitive psychophysiology of emotion: Shin, L.M., Kosslyn, S.M., McNally, R.J., Alpert, N.M.,
fear and anxiety. In: Tuma, A.H. and Maser, J.D. (Eds.), Thompson, W.L., Rauch, S.L., Macklin, M.L. and Pitman,
Anxiety and the Anxiety Disorders. Erlbaum, Hillsdale, NJ, R.K. (1997) Visual imagery and perception in posttraumatic
pp. 131–170. stress disorder: a positron emission tomographic investiga-
Lang, P.J. (1994) The motivational organization of emotion: tion. Arch. Gen. Psychiatry, 54: 233–241.
affect-reflex connections. In: VanGoozen, S., Van de Poll, Shin, L.M., Orr, S.P., Carson, M.A., Rauch, S.L., Macklin,
N.E. and Sergeant, J.A. (Eds.), Emotions: Essays on Emo- M.L., Lasko, N.B., Peters, P.M., Metzger, L.J., Dougherty,
tion Theory. Erlbaum, Hillsdale, NJ, pp. 61–93. D.D., Cannistraro, P.A., Alpert, N.M., Fischman, A.J. and
Lang, P.J., Bradley, M.M., Fitzsimmons, J.R., Cuthbert, B.N., Pitman, R.K. (2004) Regional cerebral blood flow in the
Scott, J.D., Moulder, B. and Nangia, V. (1998) Emotional amygdala and medial prefrontal cortex during traumatic im-
arousal and activation of the visual cortex: an fMRI analysis. agery in male and female Vietnam veterans with PTSD. Arch.
Psychophysiology, 35: 199–210. Gen. Psychiatry, 61: 168–176.
Lanius, R.A., Bluhm, R., Lanius, U. and Pain, C. (2006) A Slotnick, S.D., Thompson, W.L. and Kosslyn, S.M. (2005)
review of neuroimaging studies in PTSD: heterogeneity of Visual mental imagery induces retinotopically organized ac-
response to symptom provocation. J. Psychiatr. Res., in tivation of early visual areas. Cereb. Cortex, 15: 1570–1583.
press. Solodkin, A., Hlustik, P., Chen, E.E. and Small, S.L.
Lanius, R.A., Williamson, P.C., Hopper, J., Densmore, M., (2004) Fine modulation in network activation during motor
Boksman, K., Gupta, M.A., Neufeld, R.W., Gati, J.S. execution and motor imagery. Cereb. Cortex, 14: 1246–1255.
103

Talairach, J. and Tournoux, P. (1988) Co-planar stereotaxic Vrana, S.R. and Lang, P.J. (1990) Fear imagery and the startle
atlas of the human brain. 3-dimensional proportional system: probe reflex. J. Abnorm. Psychol., 99: 189–197.
an approach to cerebral imaging. Thieme Medical Publishers, Wagner, A.D., Mari, L.A., Bjork, R.A. and Schacter, D.L.
Inc, New York. (2001) Prefrontal contributions to executive control: fMRI
Thompson, W.L., Kosslyn, S.M., Sukel, K.E. and Alpert, N.M. evidence for functional distinctions within lateral prefrontal
(2001) Mental imagery of high- and low-resolution gratings cortex. Neuroimage, 6: 1337–1347.
activates area 17. Neuroimage, 14: 454–464.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 6

Subliminal emotion perception in brain imaging:


findings, issues, and recommendations

Stefan Wiens1,2,

1
Department of Psychology, Stockholm University, Frescati Hagväg, 106 91 Stockholm, Sweden
2
Section of Psychology, Department of Clinical Neuroscience, Karolinska Institute, 171 76 Stockholm, Sweden

Abstract: Many theories of emotion propose that emotional input is processed preferentially due to its
relevance for the organism. Further, because consciousness has limited capacity, these considerations imply
that emotional input ought to be processed even if participants are perceptually unaware of the input
(subliminal perception). Although brain imaging has studied effects of unattended, suppressed (in binocular
rivalry), and visually masked emotional pictures, conclusions regarding subliminal perception have been
mixed. The reason is that subliminal perception demands a concept of an awareness threshold or limen, but
there is no agreement on how to define and measure this threshold. Although different threshold concepts
can be identified in psychophysics (signal detection theory), none maps directly onto perceptual awareness.
Whereas it may be tempting to equate unawareness with the complete absence of objective discrimination
ability (d0 ¼ 0), this approach is incompatible with lessons from blindsight and denies the subjective nature
of consciousness. This review argues that perceptual awareness is better viewed as a continuum of sensory
states than a binary state. When levels of awareness are characterized carefully in terms of objective
discrimination and subjective experience, findings can be informative regarding the relative independence of
effects from awareness and the potentially moderating role of awareness in processing emotional input.
Thus, because the issue of a threshold concept may never be resolved completely, the emphasis is to not
prove subliminal perception but to compare effects at various levels of awareness.

Keywords: consciousness; attention; emotion; brain imaging; subliminal perception; backward masking

Because humans are complex organisms, many evaluate a picture of another human face as threat-
processes need to occur automatically to permit ening, and to respond to it, we would first need to
proper functioning and survival. Although our own become consciously aware of the facial expression.
experience of consciousness accepts that conscious- The importance of perceptual awareness in re-
ness is clearly insufficient to mediate all of these sponding to emotional events has been challenged
processes (e.g., blood pressure adjustments during by evolutionary considerations and theories of
postural changes), it appears to us that conscious- emotion (Öhman, 1986; Robinson, 1998; LeDoux,
ness plays a critical role in important mental proc- 2000; Dolan and Vuilleumier, 2003; Öhman and
esses. For example, to evaluate an external event as Wiens, 2003). In particular, because of their rele-
good or bad, or for it to affect our behavior, we vance to organisms, threatening situations need to
would need to be consciously aware of it. That is, to be registered and handled swiftly. However, be-
cause consciousness is limited and slow (Shevrin
Corresponding author. Tel.: +46-8-163933; Fax: +46- and Dickman, 1980; Roser and Gazzaniga, 2004;
8-159342; E-mail: sws@psychology.su.se, stefan.wiens@ki.se Marois and Ivanoff, 2005), these considerations

DOI: 10.1016/S0079-6123(06)56006-6 105


106

suggest that emotional input needs to be processed Participants can be instructed to attend to par-
partly unconsciously to ensure survival. ticular visual input, and responses to unattended
Several research paradigms have been developed input can be studied. If participants attend to some
to study the role of perceptual awareness in process- pictures but fail to notice pictures that are outside
ing of emotional pictures. The most commonly used of their attentional focus, responses to the unat-
approach is the dissociation paradigm (Holender, tended pictures might qualify as subliminal per-
1986). The goal of this approach is to present emo- ception (Merikle and Joordens, 1997). Several
tional pictures of which people are not consciously brain imaging studies with unattended emotional
aware, and to study whether or not these emotional pictures suggest that unattended pictures are proc-
pictures have effects despite peoples’ unawareness. essed even though participants are unaware of the
If so, such findings would provide evidence that pictures (Vuilleumier et al., 2001; Anderson et al.,
emotional pictures are processed in the absence of 2003; Bishop et al., 2004; Williams et al., 2004). In
awareness. Stated differently, because emotional one study, participants were shown simultaneously
effects would be obtained after elimination of per- two houses and also two faces with either neutral
ceptual awareness, awareness may not be necessary or fearful expressions (Vuilleumier et al., 2001).
for their occurrence. Further, because people are Pictures of the same category were positioned
considered either aware or unaware, perception is left and right, or above and below fixation. For
treated as a dichotomous state. Therefore, research example, one house each was shown left and right
on subliminal (or implicit) perception studies the of fixation, and one fearful face each above and
degree to which visual input is processed below the below fixation. During different blocks of trials,
threshold (‘‘limen’’) of perceptual awareness. This participants were instructed to attend to either the
review summarizes main results from recent brain horizontal or vertical picture pairs, and determine
imaging findings on subliminal perception of emo- whether the two attended pictures were the same
tional visual input. However, even though there has or different. Results showed that the amygdala
been a surge of findings on this topic, no agreement responded more to fearful than neutral faces irre-
about the existence of subliminal perception has spective of whether or not the faces were attended.
been reached. This paper reviews the main issues In a separate behavioral study with a surprise
and presents alternative strategies for future re- awareness test, participants could not report facial
search. Although this review focuses on subliminal expression (fearful or neutral), gender, or identity
perception of emotional pictures, the issues and of the unattended face pair from the preceding
strategies are generally applicable to research on trial. These findings suggest that the amygdala
subliminal perception. differentiated between fearful and neutral expres-
sion even though participants were unaware of the
faces (as these were unattended). In another study
Findings (Anderson et al., 2003), pictures of places were
colored in green and superimposed with faces in
Perceptual awareness can be manipulated in a red, and participants were instructed to attend to
number of ways (Frith et al., 1999; Kim and Blake, the places or faces. When attending to the places,
2005). Although visual masking has been tradi- participants rated whether it was inside or outside,
tionally used in subliminal perception, alternative and when attending to faces whether it was male
approaches such as manipulations of attention or female. Results showed that the amygdala re-
might have comparable effects on awareness (Me- sponded to fearful faces irrespective of whether
rikle and Joordens, 1997). This review focuses on or not they were attended. In another study, pairs
brain imaging findings with emotional pictures of of emotional faces were shown superimposed on
which participants were unaware as a result of houses in the periphery. Again, amygdala activa-
manipulation of attention (i.e., unattended pic- tion was greater to unattended than attended fear-
tures), binocular rivalry (i.e., suppressed pictures), ful faces (Williams et al., 2004). Taken together,
and visual masking (i.e., masked pictures). these findings suggest that fearful faces elicit
107

amygdala activation even if participants are una- houses. During the experiment, participants per-
ware of them because they are unattended. formed a one-back task in which they had to report
However, this conclusion has been challenged by the repetition of the same image on consecutive
results from two studies. In one study, participants trials. Although participants detected repetition of
were shown a fearful, happy, or neutral face in the nonsuppressed pictures on most trials, they missed
center of the screen together with a small bar in the all repetitions of suppressed pictures. Nonetheless,
left and right periphery (Pessoa et al., 2002). During amygdala differentiated between suppressed fearful
different blocks, participants rated either the gender and neutral faces. Similarly, Pasley et al. (2004)
of the face or judged whether the two bars had the presented houses to one eye and suppressed fearful
same orientation. Results showed that amygdala faces or chairs to the other eye. After excluding the
differentiated among the expressions when partici- small number of trials in which participants de-
pants rated the faces but not when they performed tected the presence of suppressed faces or chairs,
the bar task. In a similar study, participants also results showed that amygdala responded more
judged bar orientation while faces were presented strongly to the suppressed fearful faces than the
centrally, but task difficulty was manipulated in chairs. Taken together, as suppressed fearful faces
three levels (Pessoa et al., 2005). Results showed that were processed even though participants reported
amygdala differentiated between unattended fearful that they were unaware of the faces, these findings
and neutral faces only when the bar task was simple. provide evidence for subliminal perception. How-
Based on these findings, Pessoa et al. concluded that ever, it has been argued that mixed states of per-
unattended pictures are not processed outside of ception are possible in binocular rivalry (Kim and
awareness if awareness is focused sufficiently. Al- Blake, 2005; Pessoa, 2005). This finding suggests
though the studies did not include a manipulation that the measures of awareness used in studies
check to determine that awareness of the faces was of binocular rivalry may not be sensitive enough
actually reduced by the bar task (cf. Williams et al., to assess participants’ awareness. For example, be-
2005), findings from these and other studies suggest cause participants performed a one-back task and
that whether or not unattended faces are processed monitored repetition of the clearly visible nonsup-
depends on how unaware participants are of them. pressed pictures, they might have been distracted
In addition, individual differences such as state anx- from reporting the suppressed pictures. That is,
iety may play a moderating role (Bishop et al., although aware of the suppressed pictures, they did
2004). Accordingly, it is a matter of debate if studies not report them. To conclude, as with studies of
of unattended faces provide evidence for subliminal unattended emotional pictures, there is a debate as
perception (Pessoa et al., 2005; Vuilleumier, 2005). to whether or not the findings provide unequivocal
Subliminal perception can also be studied in bin- evidence for subliminal perception.
ocular rivalry. In these studies, two pictures are Although manipulations of attention and binocu-
presented simultaneously in different colors (e.g., lar rivalry are important methods to study subliminal
red and green), and participants wear glasses with perception, visual masking has a long tradition in
differently colored lenses on each side (e.g., red on research on subliminal perception (Holender, 1986;
left and green on right). Under these conditions, Öhman and Soares, 1994). In visual masking, a tar-
participants are typically aware of only one picture get picture is shown briefly and typically followed by
at a time. For example, if a red house is presented another irrelevant picture (mask). When picture
to one eye and a green face to the other eye, par- parameters are adjusted carefully, people often re-
ticipants might be aware only of the house. Thus, port that they are not consciously aware of the
responses to the suppressed face can be studied. target pictures. Thus, subliminal perception of
Two studies support the idea that fearful faces masked target pictures can be studied. During the
are processed under conditions of binocular sup- last few years, there has been a surge of brain im-
pression (Pasley et al., 2004; Williams et al., 2005). aging studies using visual masking (Morris et al.,
Williams et al. (2005) showed brief pictures of fear- 1998; Whalen et al., 1998; Rauch et al., 2000; Sheline
ful, happy, and neutral faces simultaneously with et al., 2001; Critchley et al., 2002; Hendler et al.,
108

2003; Carlsson et al., 2004; Etkin et al., 2004; duration for regular display devices (Wiens et al.,
Killgore and Yurgelun-Todd, 2004; Phillips et al., 2004; Wiens and Öhman, 2005b). In one study
2004; Whalen et al., 2004; Liddell et al., 2005; Pessoa (Whalen et al., 1998), participants were shown
et al., 2006). However, because research has shown fearful and happy faces that were masked by neu-
that small changes in picture parameters can have tral faces. After the experiment, participants were
strong effects on perceptual awareness (Esteves and interviewed about their awareness and asked to
Öhman, 1993), it is important to use a setup that point out the faces that they had seen during the
controls picture durations reliably. If picture dura- experiment. Although participants did not report
tion cannot be held constant over repeated presen- seeing any emotional expressions and pointed only
tations, picture duration might vary and thus at neutral faces, results showed that amygdala re-
confound results (Wiens and Öhman, 2005b). For sponded more strongly to fearful than happy faces.
example, if a certain level of awareness is targeted Similarly, in another study (Morris et al., 1998),
but picture duration is unreliable over trials, it is participants were shown two angry faces and
difficult to maintain a particular level of awareness. either face was fear-conditioned (i.e., paired with
Similarly, if trials are to be sorted after the exper- a loud noise). Participants were instructed that
iment on the basis of individual responses, differ- after each picture they should respond ‘‘yes’’ or
ences in responding might be due to variable picture ‘‘no’’ depending on whether they detected either
duration rather than to differences in participants’ angry face. Although participants responded yes
processing of the pictures. Although traditional only when the pictures were nonmasked, amygdala
(bulky-type) cathode ray tube (CRT) monitors have differentiated between the two angry faces even
adequate reliability, they cannot be used in func- when masked. Findings from other studies are
tional magnetic resonance imaging due to magnetic broadly consistent with these results for faces
interference with the imaging process (Wiens et al., (Rauch et al., 2000; Sheline et al., 2001; Critchley
2004). Also, the reliability of recent (flat-display) et al., 2002; Etkin et al., 2004; Killgore and
technologies based on liquid crystal displays (LCD) Yurgelun-Todd, 2004; Liddell et al., 2005) and
and thin-film transistors (TFT) is often poorer than other emotional pictures (Hendler et al., 2003;
assumed by researchers and claimed by manufac- Carlsson et al., 2004). However, although findings
turers (Wiens et al., 2004). However, reliable picture in visual masking are consistent with subliminal
presentation is possible with a setup involving two perception, many studies have used only an indi-
data projectors and mechanical high-speed shutters rect awareness measure (preference measure, e.g.,
(Wiens and Öhman, 2005b). This setup permits Critchley et al., 2002), provided insufficient infor-
precise control of picture durations in milliseconds mation about whether and how awareness was
rather than refresh cycles. Because many studies do measured (Carlsson et al., 2004; Liddell et al.,
not describe their setup and do not provide convinc- 2005), or reported evidence for partial awareness
ing evidence that picture presentation was reliable, it (Rauch et al., 2000; Sheline et al., 2001; Killgore
cannot be ruled out that many brain imaging results and Yurgelun-Todd, 2004; Whalen et al., 2004).
from visual masking are confounded. Also, other studies have not found amygdala
Early brain imaging studies of subliminal perce- activation to masked fearful faces (Phillips et al.,
ption used masked facial expressions and reported 2004; Pessoa et al., 2006) and suggested a moder-
findings that supported subliminal perception ating role of individual differences (Etkin et al.,
(Morris et al., 1998; Whalen et al., 1998). In these 2004). In sum, it is a matter of debate (Pessoa,
and subsequent studies, emotional facial expres- 2005) whether visual masking provides evidence
sions have been mainly masked with neutral faces. for subliminal perception.
Although the use of faces as both targets and
masks might confound the results, faces continue Issues
to be used because it is easier to mask faces with
other faces (Costen et al., 1994); this procedure Despite these numerous approaches and findings,
is necessary with limitations in minimum picture researchers continue to disagree on whether these
109

data provide convincing evidence for subliminal Whereas subjective measures assess participants’
perception. To end this debate, researchers would self-reported (subjective) ability to discriminate the
have to concur on a concept of threshold (limen) stimuli, objective measures assess participants’
to determine whether emotional processing can actual (objective) ability to discriminate the stim-
occur below this threshold of awareness. However, uli. In this context, the terms subjective and ob-
there is no threshold concept that researchers jective refer to the content rather than the quality
agree upon. Indeed, research on subliminal per- of measurement. That is, if a subjective measure
ception has often blurred the distinction between requires participants to report about their aware-
concepts of measurement and threshold (Reingold ness by pressing buttons, it can be as reliable (i.e.,
and Merikle, 1990). That is, awareness is defined objective in a measurement sense) as an objective
only indirectly by the measure that is used. Also, measure of discrimination ability.
although awareness measures are often distin- Objective measures are often favored over sub-
guished in terms of objective and subjective meas- jective measures. The most important reason is that
ures, this distinction is vague and has unclear objective measures typically allow one to separate
references to contemporary psychophysics (Mac- discrimination ability from response criterion
millan, 1986; Macmillan and Creelman, 1991). The (Eriksen, 1960). That is, even though participants
remainder of this paper discusses pros and cons of may not differ in their actual awareness of the pic-
objective and subjective measures, and reviews tures, they might have different notions about their
potential candidates for threshold concepts that fit level of awareness. For example, some participants
with contemporary psychophysics, in particular might already report that they saw a face even if
signal detection theory. they noticed only of a pair of eyes, whereas others
Requirements for a valid measure of awareness might do so only if they could clearly identify eyes,
are that it should be exhaustive and exclusive nose, and mouth. If so, the latter participants would
(Reingold and Merikle, 1990; Merikle and Rein- perform less well on a subjective measure than the
gold, 1998; Wiens and Öhman, 2002). This means former participants, and the subjective measure
that it should capture all aspects of conscious would incorrectly suggest that the participants
processing (i.e., exhaustive) but no unconscious differed in awareness. Response criteria are affected
processes (i.e., exclusive). To illustrate, if a meas- by demand characteristics (Eriksen, 1960). Also,
ure is not sensitive enough to capture conscious participants tend to underestimate their perform-
processes completely (i.e., it is not exhaustive), any ance in difficult perception tasks (Björkman et al.,
emotional effects could be due to the conscious 1993). In support, when awareness is assessed both
processes that were not captured by the measure. objectively and subjectively, objective measures of-
Similarly, if a measure is too sensitive and captures ten show evidence of discrimination ability in the
unconscious as well as conscious processes (i.e., it absence of subjective awareness. Thus, subjective
is not exclusive), the apparent absence of uncon- measures of awareness tend to be less sensitive than
scious emotional effects could be due to this objective measures (Cheesman and Merikle, 1984,
mislabeling of unconscious as conscious processes. 1986). Further, because the experimenter has little
Numerous measures have been proposed as control over the participant’s criterion, ‘‘we are in
valid indexes of awareness. However, two forms of fact using as many different criteria of awareness as
awareness thresholds are often distinguished, we have experimental subjects’’ (Eriksen, 1960, pp.
namely objective and subjective thresholds (Chees- 292–293). In contrast, objective measures typically
man and Merikle, 1984, 1986). Cheesman and allow one to control for differences in response
Merikle (1984) defined the subjective threshold as criterion.
the ‘‘level at which subjects claim not to be able to However, this is true only if performance meas-
discriminate perceptual information at better than ures are used that allow one to separate discrimi-
at a chance level’’, and the objective threshold nation ability and response criterion. That is, even
as the ‘‘level at which perceptual information is if awareness is measured with an objective task, it is
actually discriminated at a chance level’’ (p. 391). possible that the performance measure might be
110

confounded by response bias. As such, it would not was absent. In this view, the absolute level of per-
be a pure (objective) index of participants’ discrim- formance is almost irrelevant, as long as mean
ination ability but might represent a measure of performance does not differ significantly from
participants’ subjective awareness. For example, chance. However, because unawareness is defined
although yes–no detection tasks are objective meas- on the basis of results from a statistical significance
ures, the commonly used measure of performance test, the outcome of this test depends greatly on
(percent correct) is affected by response bias as well statistical power. In fact, this definition of un-
as discrimination ability (Macmillan and Creelman, awareness based on statistical testing can result in
1991). That is, if participants are instructed to nonsensical conclusions (for review, see Wiens and
detect whether or not a target is shown, percent Öhman, 2005a). That is, if the absolute level of
correct is affected by their predisposition to respond performance is constant and only the number of
that a target was shown. Thus, a participant with a trials or participants changes, one would have to
particular discrimination ability might score any- conclude that an individual became suddenly
where from 50% to 95% correct on a yes–no aware when more trials were run, or that the
detection task (see Figure 1 in Azzopardi and whole group became suddenly aware when more
Cowey, 1997). participants were run. To illustrate, results from a
Although objective tasks can be used to assess statistical test (e.g., one-sample t test) depend on
discrimination ability, a potential problem with the number of observations and on the variability
objective measures is the possibly confounding effect among observations. Statistical power increases
from lack of motivation. When masking parameters with the number of observations and decreases
are chosen so that the masked pictures are barely with the variability among observations. Thus,
visible, participants might have no motivation when the number of observations is increased and
to perform a discrimination task (Duncan, 1985; everything else is held constant, a lower observed
Merikle, 1992) as it might be experienced as mean- significance value (p-value) is obtained. For exam-
ingless. Because participants would loose motiva- ple, assume that the observed mean performance is
tion, they might push buttons randomly and thus 60% and chance level is 50%. For the ease of
perform at chance. Hence, objective measures might argument, assume further that percent correct is
not necessarily index the objective threshold but the unbiased (see above). Then, at a constant mean
subjective threshold (i.e., participants’ self-reported performance of 60%, an individual might perform
unawareness). This discussion highlights the issue of significantly better than chance (by definition, be-
whether or not the use of objective measures guar- come aware) when the objective measure consists
antees that an objective threshold is assessed. Indeed, of 40 trials but not when it consists of 20 trials.
Merikle and Daneman (2000) suggested that factors Similarly, a group might perform significantly bet-
such as an insufficient number of trials on the ob- ter than chance (by definition, become aware)
jective measure as well as insufficient motivation of when the sample consists of 40 participants but not
the participants could explain why many reports of when it consists of 20 participants. Also, because
unconscious processes have been comparable for statistical power decreases with variability among
studies in which unawareness was assessed with an observations, heterogeneity among participants
objective measure (purportedly indexing the objec- might result in nonsignificance. However, these
tive threshold) or a subjective measure. data may not suggest that all participants were
Aside from these issues, a major problem is that unaware of the pictures, but may indicate that
unawareness is commonly defined in terms of there is substantial variation among participants.
statistical deviations from chance performance Thus, the mean may not be representative for the
instead of an absolute level of performance. For group. Although variability among participants
example, mean performance is measured for an can be evaluated in terms of confidence intervals
individual or a group, and if the mean is not signi- (Cumming and Finch, 2005) and correlations with
ficantly different from the level of performance variables of interest, this is not commonly dis-
expected by chance, it is concluded that awareness cussed in research. Further, because researchers
111

are mainly interested in retaining the null hypoth- Shanks (2002) argued, participants might be able
esis (i.e., participants do not perform better than to discriminate among the masked pictures with-
chance and are thus unaware), the commonly used out being able to verbally label the masked pic-
a ¼ 0.05 is probably too low. However, there is no tures. Because both measures assess different kinds
general agreement on which a (0.20, 0.10) to use. of discrimination abilities, it is unclear which
Another problem with objective and subjective measure indexes the supposed objective threshold.
measures is that they are often measured on differ- Similarly, if participants are shown masked spiders
ent time scales. Objective measures typically and snakes, but they report that they did not see
require participants to respond after each picture. any spiders or snakes, it might be concluded that
For example, on each trial, participants indicate if they were unaware (subjective threshold). How-
they saw a happy or a fearful face by pushing a ever, if participants also reported that they noticed
button. In contrast, subjective measures often re- that masked pictures were shown even though they
quire participants to respond only after a series of could not tell if the pictures were spiders or snakes,
pictures. For example, participants are interviewed these findings would suggest that participants had
at the end of the experiment if they saw any emo- some subjective awareness of the masked pictures.
tional facial expressions. This makes it difficult to Therefore, it is unclear if it can be concluded that
compare results from the two tasks, as any differ- participants were unaware at the subjective thresh-
ences might be partly due to contextual differences old. In fact, this decision might be particularly
during measurement (Lovibond and Shanks, difficult for tasks that combine subjective with
2002). That is, because in subjective measures, objective features. For example, Kunimoto et al.
participants are often asked about their integrated (2001) described a measure that combines (objec-
experience over trials, confounding effects of for- tive) discrimination ability with (subjective) confi-
getting seem more problematic for subjective dence.
measures than objective measures. Apparently, objective measures have the advan-
Yet another problem with both objective and tage of objectifying awareness by removing indi-
subjective measures is that because the definitions vidual differences in response criteria and assessing
are rather vague, it is unclear which specific meas- pure discrimination ability. However, this is their
ures ought to be used to capture these thresholds. greatest drawback, as they ignore the principally
For example, Cheesman and Merikle (1986) meas- subjective nature of awareness. That is, because
ured objective thresholds in terms of either per- awareness refers to phenomenological experience,
formance on a detection task (Experiment 2) or it may be more relevant to index what people
identification task (Experiment 3). That is, per- notice subjectively rather than what they can dis-
formance was indexed by participants’ ability to criminate objectively (Bowers, 1984; Wiens and
detect whether or not a word was presented (de- Öhman, 2002). In analogy, the phenomenological
tection task) or which word was presented (dis- experience of pain cannot be indexed in terms of
crimination task). However, because other whether people can discriminate stimuli objec-
research has used detection tasks as the objective tively but whether they experience them subjec-
measure and identification tasks as the measure of tively as painful (Wiens, 2006). Because awareness
unconscious processing (Snodgrass et al., 2004b), is a process that is closer to the process of ‘‘no-
it is unclear which measure is the correct one to ticing’’ than ‘‘discriminating,’’ a valid measure of
index a supposed objective threshold. Indeed, awareness ought to capture what participants
dissociations among measures might be expected notice rather than what they can discriminate.
by chance (Eriksen, 1960) or result from subtle Indeed, because research has shown that objective
differences in task requirements or performance measures often provide evidence for discrimina-
indexes (Fisk and Haase, 2005). To give a simple tion ability despite subjective unawareness, objec-
example, Öhman and Soares (1994) measured tive measures might fulfill validity requirements of
ability to classify masked spiders, snakes, flowers, exhaustiveness but not exclusiveness (Merikle and
and mushrooms. However, as Lovibond and Reingold, 1998). That is, objective measures might
112

capture not only conscious aspects but also un- or to indicate their responses with button presses.
conscious processing, and their apparent greater Also, although monkeys first had to be trained on
sensitivity might be due to this violation of exclu- the tasks, they showed similar dissociations in task
siveness. Hence, objective measures might not be performance when compared to human blindsight
valid indexes of awareness. patients.
Indeed, proponents of objective measures of These findings imply that monkeys have percep-
awareness would have to infer that performance tual awareness (Cowey, 2004). Critically, they also
better than chance demonstrates necessarily that challenge the notion of discrimination ability as a
people are aware of the pictures. This reasoning is valid index of awareness. That is, because objective
inconsistent with findings from studies of brain- measures (i.e., localization task) were used to index
damaged patients. In these studies, objective meas- performance outside of awareness, these data chal-
ures were used to demonstrate that people could lenge the notion that above-chance performance on
perform a task although they are unaware of the objective measures necessarily shows that people
target stimuli, as indexed by subjective measures. are aware of the target stimuli. However, because it
The most famous example is blindsight (Weiskrantz is possible that patients are not reporting accurately
et al., 1974; Weiskrantz, 1986; Cowey and Stoerig, about their awareness, potential alternative expla-
1995; Cowey, 2004). Blindsight is observed in nations must be considered. Regarding fellow
patients with damaged primary visual cortex (V1) humans, it may be tempting to trust them about
who report that they are completely unaware of the their self-report. In contrast, when considering data
stimuli in their damaged visual field; nonetheless, from animal research, it is easier to remain skeptical
they can discriminate among them (de Gelder et al., and to think of possible confounding variables. The
1999; Morris et al., 2001; for blindsight studies with role of potentially confounding variables has been
emotional stimuli and brain imaging, see Anders et studied extensively in monkey models of blindsight
al., 2004; Pegna et al., 2005). Blindsight is commonly (Cowey and Stoerig, 1995; Stoerig et al., 2002).
demonstrated by a dissociation of performance on The main challenge to blindsight is the argument
two visual tasks: a localization task (objective meas- that patients are perceptually aware of the visual
ure) and a classification task (subjective measure). In input in their damaged visual field, but because the
the localization task (objective), subjects focus on pictures are perceived as weaker than in the un-
the center of the screen. When they push a button to damaged visual field, patients report that they are
initiate a trial, a light flashes somewhere, and sub- unaware of them. That is, targets in the damaged
jects are instructed to push the screen at that posi- visual field are classified as blanks because they ap-
tion. Typical results are that subjects can localize the pear more similar to blanks than to targets in the
flashes accurately, even in their damaged visual field. undamaged visual field. In signal detection terms,
In the classification task (subjective), subjects have this argument can be conceptualized in terms of
an additional response option to indicate when an stimuli falling below the response criterion. Al-
empty trial (i.e., no light flash) was presented. Here, though signal detection will be explained below, this
typical results are that subjects report blanks only means that people use an arbitrary level of visibility
when light flashes are presented in their damaged at which they report whether or not they are aware
visual field. Hence, although subjects can localize the of a target. Accordingly, targets in the damaged
light flashes in their damaged visual field accurately visual field fall below this criterion and are thus
when forced to point at the location, they choose the reported as blanks. However, a series of experi-
blank button when given the option to do so. Such ments have addressed this potential confound (e.g.,
discrepant results between tasks have been obtained Stoerig et al., 2002). First, when target visibility was
for both humans and monkeys with similar lesions manipulated so that target contrast was near
in V1 (Cowey and Stoerig, 1995; Stoerig et al., 2002). threshold in the undamaged visual field and max-
In humans, similar performance on the subjective imal in the damaged field, subjects reported only a
classification task is obtained irrespective of whether small proportion of targets in the undamaged field
they are instructed to verbally report their awareness as targets. Still, subjects continued to report targets
113

in the damaged field as blanks. Importantly, these measurement are rather unclear. As illustrated
findings were obtained although performance on above, different measures are used interchangeably
the localization task was apparently better for tar- as indexes of the same process, and identical meas-
gets with maximal contrast in the damaged than ures are sometimes used as indexes of either con-
near-threshold contrast in the undamaged visual scious or unconscious processing. As a potential
field. These findings are inconsistent with the argu- solution, it has been advocated to use awareness
ment of a decreased visibility in the damaged field, measures based on signal detection analyses (e.g.,
as it would have been expected that with similarly Hannula et al., 2005). However, the theory of signal
poor levels of target visibility in undamaged and detection makes no reference to awareness (Mac-
damaged visual fields, differences in ratings would millan, 1986; Macmillan and Creelman, 1991). So,
disappear. Second, when number of trials was how can signal detection measures be used to index
greater for the damaged than undamaged visual something that it is not part of the theory of signal
field, subjects continued to classify targets in their detection?
damaged visual field as blanks. Importantly, even
though monkeys received rewards for reporting
Threshold concepts in signal detection theory
targets correctly and did not receive a reward when
they misclassified a target in the damaged field as a
Although signal detection theory (SDT) makes no
blank, they reported targets in the damaged field as
reference to perceptual awareness, several concepts
blanks and thus did not receive any rewards on the
can be construed as thresholds (Macmillan, 1986;
majority of trials. These findings argue against the
Macmillan and Creelman, 1991). These threshold
possibility that less visible targets in the damaged
concepts can then be evaluated for their usefulness
visual field were classified as blanks because they
in indexing the threshold idea in subliminal per-
occurred less often than targets in the undamaged
ception. From a psychophysics perspective, four
visual field or because they had different outcomes
thresholds might be distinguished: sensory, crite-
associated with them. Taken together, the most
rion, empirical, and energy threshold. For the ease
parsimonious explanation for these findings is that
of argument, the following discussion of these
greater-than-chance performance on a task (e.g.,
concepts is illustrated for a simple yes–no detec-
localization) does not demonstrate awareness per
tion task. For example, in a face detection task,
se. If this conclusion is rejected, then it indicates
face and no-face trials are presented and partici-
that blindsight does not exist, as greater-than-
pants decide after each trial whether or not they
chance performance would necessarily indicate that
detected a face (yes or no).
patients are aware of the target stimuli. In fact, this
reasoning makes it logically impossible to demon-
strate performance without awareness (Bowers, Sensory threshold
1984). Accordingly, because there is no a priori
reason that emotional effects might not be con- A sensory or observer threshold is a hypothetical
sidered indexes of perceptual awareness, any form threshold that is internal to the participant and
of discriminative responding could actually be determines if a stimulus is sensed or not. This sen-
viewed as evidence of awareness (Wiens and sory threshold cannot be mapped directly to par-
Öhman, 2002). However, in the absence of a con- ticipants’ overt responses. Thus, responding yes or
vincing alternative explanation for blindsight, les- no in a detection task does not correspond directly
sons from blindsight suggest that discrimination to sensory states that fall above or below the
ability per se is not indicative of awareness. sensory threshold, respectively. The concept of a
To conclude, the distinction between objective and sensory threshold is probably closest to the notion
subjective measures makes intuitive sense because it of a limen and the distinction between subliminal
reflects differences between actual discrimination and supraliminal. Although this concept was in-
performance and phenomenological aspects of cluded in many early theories of psychophysics,
noticing. However, underlying concepts and their there is little evidence that supports the idea of a
114

sensory threshold. In fact, SDT has been intro- choose an arbitrary level on the sensory continuum
duced as an alternative model that can account for as a cutoff score (criterion). Above this criterion,
many findings without resorting to the concept of they respond yes, and below this criterion, they re-
a sensory threshold (Macmillan and Creelman, spond no. Based on the criterion, it is possible to
1991). distinguish among hits (responding yes on signal tri-
SDT argues against a sensory threshold in favor als), false alarms (responding yes on no-signal trials),
of an internal continuum of sensory states (more misses (responding no on signal trials), and correct
generally called strength of evidence; Pastore et al., rejections (responding no on no-signal trials). Then,
2003). According to SDT, each presentation of a probabilities for hits and false alarms can be used to
stimulus (signal) occurs against a variable back- compute the actual distributions to determine the
ground of internal noise. In theory, the variability of relative location of signal and noise distributions.
this noise over trials can be captured by presenting This is commonly expressed in terms of d0 (d prime),
no-signal trials repeatedly, measuring the values of which is the distance between the means of the signal
these no-signal trials on the sensory continuum, and and noise distribution in z scores. Sensitivity (d0 ) is
forming a histogram of these values. This noise dis- high if signal and noise distributions are far apart
tribution characterizes the mean and variability of and low if the distributions overlap closely. Alterna-
the internal noise on the continuum of sensory tive indexes of sensitivity such as Br and A0 have
states. Due to continuous background noise, a signal been proposed. However, Br is based on a different
is then presented in the context of noise; that is, a threshold model, and A0 has been criticized for
signal is superimposed on the noise (i.e., signal plus alleged claims that it is nonparametric (Snodgrass
noise). As with no-signal trials, a hypothetical dis- and Corwin, 1988; Macmillan and Creelman, 1990;
tribution for signal (plus noise) trials can be plotted. Pastore et al., 2003). However, in practice, these in-
Theoretically, there is no point on the continuum dexes often give comparable results. Nonetheless, to
that allows one to determine unmistakably whether calculate z scores and thus d0 , hit and false alarm
or not a signal was presented. That is, a signal can rates must be greater than zero and less than 1. This
sometimes evoke a relatively weak internal response, is a potential problem when the perceptual input is
whereas the absence of a stimulus (no-signal trial) degraded, as participants might never report that
can even evoke a relatively strong internal response. they detected a signal. To permit calculation of d0 ,
Accordingly, it is possible to make only probability extreme scores are often dealt with by adding 0.5 in
statements. For example, if the point on the contin- the numerator and 1 in the denominator when cal-
uum is high relative to the mean of the noise dis- culating hit rates and false alarm rates (Snodgrass
tribution, it is likely that a signal was presented, and Corwin, 1988).
whereas if the point is low, it is likely that no signal A critical feature of SDT is that the observer’s
(noise) was presented. However, the more the signal placement of the criterion does not affect estimation
(plus noise) and noise distributions overlap, the more of discrimination ability, as the distance between
difficult it is to distinguish between both types of the signal and noise distributions is unaffected by
trials. If the distributions overlap perfectly, signal participants’ placement of their criterion. Different
and noise trials cannot be distinguished at all. indexes of criterion placement have been proposed
The locations of the noise and signal (plus noise) (Snodgrass and Corwin, 1988; Macmillan and
distributions cannot be measured directly. Instead, Creelman, 1991). The likelihood ratio b (beta) is
they need to be inferred. To do that, many signal the ratio of the heights of the signal and noise dis-
and no-signal trials need to be presented, and par- tributions, and the criterion C is the distance from
ticipants are asked to make a response on each trial. the intersection point of signal and noise distribu-
For example, in the face detection task, face and tions as a z score. A neutral criterion or absence of
no-face trials are shown and participants decide after a response bias is present if participants set their
each trial whether or not they detected a face (yes or criterion so that the probabilities for signal and
no). According to SDT, when observers are asked to noise are equally likely (i.e., where they cross). That
make overt responses in a detection task, they is, b ¼ 1 and C ¼ 0. If participants position the
115

criterion more towards the lower end of the sensory of awareness. Also, because a response is collected
continuum, then they exhibit a lax or liberal re- on every trial, it has an advantage over other sub-
sponse bias (as they are more willing to respond jective measures that are assessed only across a
yes); if they position it more to the higher end of the number of trials. But, a drawback is that it is un-
sensory continuum, then they exhibit a strict or clear how this measure can be averaged over trials.
conservative response bias (as they are less willing For example, if participants report awareness on
to respond yes). A liberal response bias results in 12% of the signal trials, it is unclear if they should
bo1 and Co0, whereas a conservative response be considered aware or unaware. Intuitively, one
bias results in b41 and C40. Although beta is might consider whether participants reported
used more commonly as an index of response bias, awareness of faces even when no faces were
a number of arguments favor C (Macmillan and shown. However, the false alarm rate cannot be
Creelman, 1990). Although the location of the cri- used to make inferences about participants’ sub-
terion is chosen arbitrarily by the observer, it is jective awareness, as this would assess merely their
often affected by the pay-off associated with differ- discrimination ability (as d0 is calculated from hit
ent response outcomes. For example, if there is a rates and false alarm rates).
big reward for detecting a signal, observers are Another approach to using the criterion as a
more willing to respond yes (lax response bias). In threshold might be to sort signal trials into de-
contrast, if there is a punishment for false alarms, tected and undetected trials (i.e., above and below
observers are less willing to respond yes (strict re- criterion). Then, effects of interest could be studied
sponse bias). for undetected signals, that is, signals of which
participants report to be unaware. This approach
has a long tradition in experimental psychology
Criterion threshold (for review, see Merikle et al., 2001). However,
SDT can account for seemingly surprising findings
The criterion itself might be viewed as a threshold that undetected stimuli can be discriminated. For
of awareness. For example, applied to the face example, undetected faces might be discriminated
detection task, participants could be instructed to in their facial expressions. The reason is that un-
respond yes only if they were consciously aware of detected signals do not necessarily indicate that
the faces. Thus, the placement of the criterion discrimination ability is absent (d0 ¼ 0) (Mac-
would correspond to a subjective measure of millan, 1986; Haase et al., 1999). Because the rel-
awareness. However, participants might have ative location of signal and noise distributions is
different notions about their awareness. In the unaffected by the location of the criterion, dis-
face detection task, some might respond yes crimination ability between signal and noise might
already if they noticed only eyes (lax response be quite high even if the signal is not detected on
bias) whereas others might respond yes only if they 95% of the signal trials (strict response bias).
noticed eyes, nose, and mouth (strict response Hence, there would be nothing mysterious if par-
bias). Indeed, in SDT the criterion is considered a ticipants can discriminate undetected faces in
pure index of response bias that says nothing terms of facial expressions. However, SDT would
about awareness. However, if participants receive predict that if detection ability were indeed absent
clear instructions about how they should place (d0 ¼ 0), participants should be unable to discrim-
their criterion (e.g., respond yes in the face detec- inate among signals, as the signal and noise dis-
tion task only if they can clearly see eyes, nose, and tributions would overlap perfectly. Although this
mouth), individual differences might be reduced. point is debated, evidence suggests that observed
Of course, this requires that the experimenter has a effects are small and might be due to slight differ-
clear and explicit definition of awareness (e.g., ences in task setup (Snodgrass, 2002; Haase
which experiences constitute awareness of a face). and Fisk, 2004; Holender and Duscherer, 2004;
But, if these instructions are clear, the criterion Reingold, 2004; Snodgrass et al., 2004b; Snodgrass
might be a useful measure of the subjective aspect et al., 2004a; Fisk and Haase, 2005).
116

Further, although effects below a subjective Energy threshold


threshold can be studied, this threshold seems
arbitrary. For example, instead of asking partici- The energy threshold is the level at which per-
pants to respond yes or no, they could rate level of formance level is null, that is, d0 ¼ 0. Thus, it can
awareness (visibility) on a continuous scale be conceptualized as a specific empirical threshold.
(Sergent and Dehaene, 2004). When subjective The energy threshold is the most common denom-
awareness is considered on a continuum, it comes inator among researchers in subliminal perception.
at no surprise that unreported (i.e., below thresh- That is, researchers agree that people are unaware
old) stimuli are processed. That is, participants may if their performance is null (d0 ¼ 0). However, the
not report pictures below a particular cutoff on a debate ensues as to whether performance above
continuous scale (e.g., 6 on a 10-point scale) but null reflects awareness. Some models propose that
still be able to discriminate among these pictures. subliminal perception ought to be studied at d0 ¼ 0
(Snodgrass, 2002; Haase and Fisk, 2004; Snod-
grass et al., 2004a, b; Fisk and Haase, 2005).
Empirical threshold However, this approach has methodological and
conceptual problems. A major problem is that it
An empirical threshold is defined arbitrarily as a attempts to prove the null hypothesis. This end-
particular level of behavioral performance. For eavor is generally known to be difficult if not
example, different empirical thresholds might cor- impossible. It requires thousands of trials to ob-
respond to various performance levels (e.g., d0 ¼ 1, tain a reliable estimate of d0 , and a lax significance
2). Because there is no theory of awareness that criterion (a ¼ 0.20) to guard against a type 2 error
equates particular empirical thresholds with aware- of retaining the null (i.e., participant is unaware)
ness, empirical thresholds appear to have limited even though the alternative hypothesis (i.e., par-
usefulness in indexing awareness and unawareness. ticipant is aware) is true. Also, because signal and
However, findings of qualitative differences would noise trials will be barely distinguishable, it is
provide some support for the validity of particular doubtful that participants will stay motivated
empirical thresholds. As discussed by Merikle and during this task. Therefore, it is likely that partic-
colleagues (e.g., Merikle et al., 2001), the distinction ipants will give up and start pushing buttons
between subliminal (unconscious) and supraliminal randomly. As a consequence, d0 ¼ 0 may not
(conscious) processes is supported if they have accurately reflect absence of discrimination ability
qualitatively different effects. In fact, the distinction but lack of motivation (Merikle and Daneman,
between subliminal and supraliminal perception 2000). Further, this approach concludes that per-
might be interesting only if they have effects that formance above null (d0 40) necessarily indicates
differ qualitatively rather than quantitatively. For perceptual awareness. However, it is not intuitive
example, Merikle and Cheesman (1987) found that to conclude that any deviation from 0, however
reaction times in the Stroop task were in opposite small, indicates awareness. Critically, as discussed
directions for masked (subliminal) and nonmasked above, it ignores the subjective nature of percep-
(supraliminal) words. Because awareness was tual awareness as well as lessons from blindsight.
indexed by whether or not participants reported In sum, central aspects of contemporary psy-
awareness of the masks, these findings support this chophysics, and in particular SDT, are that there
measure of awareness. However, although interest- is a continuum of sensory states (i.e., there is no
ing, qualitative differences have been reported only actual sensory threshold), and that the relationship
for a few experimental conditions (Merikle et al., between stimulus events and sensory states is pro-
2001). Also, this validation is rather indirect, as babilistic due to constant background noise in the
qualitative differences might indicate only that sensory system. It is therefore impossible to deduce
awareness is a marker or correlate of qualitative unequivocally from a given sensory activation
differences rather than a causal mechanism (Kuni- whether it resulted from a signal or noise trial.
moto et al., 2001; Wiens and Öhman, 2005a). Also, although a response (e.g., yes or no) is
117

obtained on each trial, it does not reflect awareness who are completely unable to discriminate visual
per se but the cutoff point on the continuum of input are unaware, researchers might have been
sensory states that participants choose arbitrarily tempted to use this definition of unawareness. In
to separate different response alternatives (e.g., yes SDT terms, this definition corresponds to the
and no). energy threshold or an empirical threshold that is
set at d0 ¼ 0. However, because this approach
attempts to prove the null, any null findings on the
Recommendations awareness task might be challenged on the
grounds of insufficient statistical power or con-
It is definitely an oversimplification to treat per- founding effects from lack of motivation (Merikle
ceptual awareness as a unitary concept. Perceptual and Daneman, 2000). In fact, the conclusion that
input often consists of various aspects, each with brain imaging studies of subliminal perception are
their own thresholds for awareness. For example, based on d0 ¼ 0 has been questioned (Hannula
a face comprises aspects such as features (eyes, et al., 2005; Pessoa, 2005). Therefore, if researchers
nose, and mouth), expression, gender, race, age intend to argue that d0 ¼ 0, they need to provide
and so on. Similarly, awareness of words might be convincing evidence that participants were actually
differentiated in terms of awareness of individual unable to discriminate the stimuli.
characters and of the whole words (Kouider and A possible reason why researchers tend not to be
Dupoux, 2004). Awareness is likely to differ for explicit about their definition of awareness may be
these various aspects and in their underlying that they actually feel uncomfortable about equat-
mechanisms (e.g., Stoerig, 1996). Therefore, ing unawareness with d0 ¼ 0. Indeed, if d0 40 is
future studies ought to include measures that cap- inevitably equated with awareness, then this ap-
ture the relevant stimulus dimension of interest. proach denies the subjective nature of awareness
Further, several masking studies have reported (Bowers, 1984). Also, it tends to make it logically
that awareness was not measured at all because impossible to demonstrate subliminal perception,
picture parameters were similar to other experi- and implies that findings from blindsight patients
ments (e.g., 30 ms SOA). However, reliability and are invalid (Wiens, 2006). Although many
luminance curves of picture parameters vary sub- researchers might agree with this conclusion, they
stantially for different display technologies (Wiens might be unsure about which measure to use to
et al., 2004; Wiens and Öhman, 2005b), and small capture subjective experience. Participants might
differences in picture parameters can have strong be instructed to make a yes–no decision about
effects on perceptual awareness (Esteves and their own awareness (i.e., placement of criterion).
Öhman, 1993). Therefore, it is recommended that If participants are not instructed where to place
reliability of display equipment is demonstrated this criterion, there will probably be as many defi-
and that participants’ awareness is measured ex- nitions of awareness as there are participants
plicitly rather than assumed. In fact, to rule out (Eriksen, 1960). However, if participants are pre-
potentially confounding effects from individual sented with an explicit definition of subjective
differences in awareness, individual performance awareness, participants will probably use this defi-
needs to be assessed. Also, if a particular level of nition accurately. Thus, subjective experience can
performance is targeted, specific stimulus param- be assessed objectively.
eters could be selected on the basis of an awareness Despite methodological difficulties, awareness
test prior to the actual experiment. This approach needs to be treated and assessed as a subjective
is recommended from a psychophysics approach, state. Indeed, shortcomings in dichotomizing per-
but a potential drawback is that participants might formance on subjective measures do not argue
habituate to the target pictures. against subjective measures in general but against
Most studies lack an explicit definition of aware- a conceptualization of awareness that assumes it to
ness (Reingold and Merikle, 1990). Because be a binary state. Hence, conclusions about aware-
nobody challenges the conclusion that participants ness may be more realistic and informative in
118

terms of relative awareness rather than as aware- effects from awareness or yield different effects
ness as present or absent. depending on awareness are informative. However,
Because awareness might be treated more accu- in order to characterize particular effects in terms of
rately as a continuum, a psychophysics approach awareness, it is necessary to document awareness
lends itself to study stimulus-response relationships carefully.
between awareness and effects of interest. First, To conclude, because the debate in defining and
participants might be asked to rate their perceptual measuring awareness is conceptual, results from
awareness on a continuous scale to determine if brain imaging cannot solve this issue. Nonetheless,
awareness changes gradually or dichotomously. by adopting an eclectic approach using subjective
For example, Sergent and Dehaene (2004) propose and objective measures, and treating awareness as
that participants experience the attentional blink as a continuum, brain imaging can provide inform-
dichotomous. In general, when series of pictures are ative insights on how the brain processes emo-
presented briefly and participants have to detect tional input at various levels of awareness. Thus,
two target pictures, participants often fail to detect past and future findings from brain imaging stud-
the second target if it follows about 200–500 ms ies should not be evaluated in terms of whether or
after the first target (i.e., attentional blink). In their not they demonstrate subliminal perception but
study, participants were instructed to rate visibility instead in terms of if and how effects differ at
of the second target on a 21-point scale. Results different levels of awareness.
showed a bimodal distribution of visibility ratings
for the second target. This bimodal distribution was
probably not due to response biases, as participants Acknowledgements
gave gradually higher visibility ratings for detected
targets when the duration of the target was length- Preparation of this article was funded by grants
ened. Second, a psychophysics approach allows from the Swedish Research Council.
studying how perceptual input is processed at
different levels of awareness. For example, in a fol-
low-up study on the attentional blink, Sergent et al. References
(2005) sorted trials based on visibility ratings to
study their neural correlates. This example illus- Anders, S., Birbaumer, N., Sadowski, B., Erb, M., Mader, I.,
trates that continuous measures of awareness can Grodd, W. and Lotze, M. (2004) Parietal somatosensory as-
sociation cortex mediates affective blindsight. Nat. Neurosci.,
be powerful tools to index different levels of aware- 7: 339–340.
ness and to study their neural correlates. Therefore, Anderson, A.K., Christoff, K., Panitz, D., De Rosa, E. and
if facial expressions compared to other pictures Gabrieli, J.D.E. (2003) Neural correlates of the automatic
were processed similarly at various levels of aware- processing of threat facial signals. J. Neurosci., 23:
5627–5633.
ness, such results would suggest that awareness
Azzopardi, P. and Cowey, A. (1997) Is blindsight like normal,
does not play a critical role in processing facial ex- near-threshold vision? Proc. Natl. Acad. Sci. USA, 94:
pressions. Alternatively, if emotional input can be 14190–14194.
shown to have (qualitatively) different effects at Bishop, S.J., Duncan, J. and Lawrence, A.D. (2004) State anx-
different levels of awareness, this would suggest iety modulation of the amygdala response to unattended
that awareness (as indexed by a particular measure) threat-related stimuli. J. Neurosci., 24: 10364–10368.
Björkman, M., Juslin, P. and Winman, A. (1993) Realism of
plays a moderating role in emotion face perception. confidence in sensory discrimination — the underconfidence
A similar approach might be useful for studying the phenomenon. Percept. Psychophys., 54: 75–81.
role of attention in processing of emotional input Bowers, K.S. (1984) On being unconsciously influenced and
(Pessoa et al., 2005; Vuilleumier, 2005). informed. In: Bowers, K.S. and Meichenbaum, D. (Eds.),
The Unconscious Reconsidered. Wiley, New York, pp.
So, even though findings for unattended, sup-
227–272.
pressed, and masked emotional pictures may not Carlsson, K., Petersson, K.M., Lundqvist, D., Karlsson, A., In-
permit absolute statements about subliminal percep- gvar, M. and Ohman, A. (2004) Fear and the amygdala: ma-
tion, findings that suggest a relative independence of nipulation of awareness generates differential cerebral
119

responses to phobic and fear-relevant (but nonfeared) stimuli. Hendler, T., Rotshtein, P., Yeshurun, Y., Weizmann, T., Kahn,
Emotion, 4: 340–353. I., Ben-Bashat, D., Malach, R. and Bleich, A. (2003) Sensing
Cheesman, J. and Merikle, P.M. (1984) Priming with and with- the invisible: differential sensitivity of visual cortex and am-
out awareness. Percept. Psychophys., 36: 387–395. ygdala to traumatic context. Neuroimage, 19: 587–600.
Cheesman, J. and Merikle, P.M. (1986) Distinguishing con- Holender, D. (1986) Semantic activation without conscious
scious from unconscious perceptual processes. Can. J. Psy- identification in dichotic-listening, parafoveal vision, and
chol., 40: 343–367. visual masking: a survey and appraisal. Behav. Brain Sci.,
Costen, N.P., Shepherd, J., Ellis, H. and Craw, I. (1994) 9: 1–23.
Masking of faces by facial and non-facial stimuli. Vis. Cogn., Holender, D. and Duscherer, K. (2004) Unconscious percep-
1: 227–251. tion: the need for a paradigm shift. Percept. Psychophys.,
Cowey, A. (2004) The 30th Sir Frederick Bartlett lecture: fact, 66: 872–881.
artefact, and myth about blindsight. Q. J. Exp. Psychol. A, Killgore, W.D. and Yurgelun-Todd, D.A. (2004) Activation of
57: 577–609. the amygdala and anterior cingulate during noncon-
Cowey, A. and Stoerig, P. (1995) Blindsight in monkeys. Na- scious processing of sad versus happy faces. Neuroimage,
ture, 373: 247–249. 21: 1215–1223.
Critchley, H.D., Mathias, C.J. and Dolan, R.J. (2002) Fear Kim, C.Y. and Blake, R. (2005) Psychophysical magic: render-
conditioning in humans: the influence of awareness and au- ing the visible ‘invisible’. Trends Cogn. Sci., 9: 381–388.
tonomic arousal on functional neuroanatomy. Neuron, Kouider, S. and Dupoux, E. (2004) Partial awareness creates
33: 653–663. the ‘‘illusion’’ of subliminal semantic priming. Psychol. Sci.,
Cumming, G. and Finch, S. (2005) Inference by eye — confi- 15: 75–81.
dence intervals and how to read pictures of data. Am. Psy- Kunimoto, C., Miller, J. and Pashler, H. (2001) Confidence and
chol., 60: 170–180. accuracy of near-threshold discrimination responses. Con-
de Gelder, B., Vroomen, J., Pourtois, G. and Weiskrantz, L. scious Cogn., 10: 294–340.
(1999) Non-conscious recognition of affect in the absence of LeDoux, J.E. (2000) Emotion circuits in the brain. Annu. Rev.
striate cortex. Neuroreport, 10: 3759–3763. Neurosci., 23: 155–184.
Dolan, R.J. and Vuilleumier, P. (2003) Amygdala automa- Liddell, B.J., Brown, K.J., Kemp, A.H., Barton, M.J., Das, P.,
ticity in emotional processing. Ann. N. Y. Acad. Sci., Peduto, A., Gordon, E. and Williams, L.M. (2005) A direct
985: 348–355. brainstem-amygdala-cortical ‘alarm’ system for subliminal
Duncan, J. (1985) 2 techniques for investigating perception signals of fear. Neuroimage, 24: 235–243.
without awareness. Percept. Psychophys., 38: 296–298. Lovibond, P.F. and Shanks, D.R. (2002) The role of awareness
Eriksen, C.W. (1960) Discrimination and learning without in Pavlovian conditioning: empirical evidence and theoretical
awareness: a methodological survey and evaluation. Psychol. implications. J. Exp. Psychol. Anim. Behav. Process.,
Rev., 67: 279–300. 28: 3–26.
Esteves, F. and Öhman, A. (1993) Masking the face: recogni- Macmillan, N.A. (1986) The psychophysics of subliminal per-
tion of emotional facial expressions as a function of the ception. Behav. Brain Sci., 9: 38–39.
parameters of backward masking. Scand. J. Psychol., Macmillan, N.A. and Creelman, C.D. (1990) Response bias:
34: 1–18. characteristics of detection theory, threshold theory, and
Etkin, A., Klemenhagen, K.C., Dudman, J.T., Rogan, M.T., nonparametric indexes. Psychol. Bull., 107: 401–413.
Hen, R., Kandel, E.R. and Hirsch, J. (2004) Individual Macmillan, N.A. and Creelman, C.D. (1991) Detection Theory:
differences in trait anxiety predict the response of the baso- A User’s Guide. Cambrigde University Press, New York.
lateral amygdala to unconsciously processed fearful faces. Marois, R. and Ivanoff, J. (2005) Capacity limits of informa-
Neuron, 44: 1043–1055. tion processing in the brain. Trends Cogn. Sci., 9: 296–305.
Fisk, G.D. and Haase, S.J. (2005) Unconscious perception or Merikle, P.M. (1992) Perception without awareness: critical
not? An evaluation of detection and discrimination as indi- issues. Am. Psychol., 47: 792–795.
cators of awareness. Am. J. Psychol., 118: 183–212. Merikle, P.M. and Cheesman, J. (1987) Current status of
Frith, C., Perry, R. and Lumer, E. (1999) The neural correlates research on subliminal perception. Adv. Consum. Res.,
of conscious experience: an experimental framework. Trends 14: 298–302.
Cogn. Sci., 3: 105–114. Merikle, P.M. and Daneman, M. (2000) Conscious vs. uncon-
Haase, S.J. and Fisk, G.D. (2004) Valid distinctions between scious perception. In: Gazzaniga, M.S. (Ed.), The New Cog-
conscious and unconscious perception? Percept. Psychophys., nitive Neurosciences. MIT Press, Cambridge, MA,
66: 868–871. pp. 1295–1303.
Haase, S.J., Theios, J. and Jenison, R. (1999) A signal detection Merikle, P.M. and Joordens, S. (1997) Parallels between per-
theory analysis of an unconscious perception effect. Percept. ception without attention and perception without awareness.
Psychophys., 61: 986–992. Conscious Cogn., 6: 219–236.
Hannula, D.E., Simons, D.J. and Cohen, N.J. (2005) Opinion Merikle, P.M. and Reingold, E.M. (1998) On demonstrating
— imaging implicit perception: promise and pitfalls. Nat. unconscious perception: comment on Draine and Greenwald
Rev. Neurosci., 6: 247–255. (1998). J. Exp. Psychol. Gen., 127: 304–310.
120

Merikle, P.M., Smilek, D. and Eastwood, J.D. (2001) Percep- Reingold, E.M. and Merikle, P.M. (1990) On the inter-relat-
tion without awareness: perspectives from cognitive psychol- edness of theory and measurement in the study of uncon-
ogy. Cognition, 79: 115–134. scious processes. Mind Language, 5: 9–28.
Morris, J.S., DeGelder, B., Weiskrantz, L. and Dolan, R.J. Robinson, M.D. (1998) Running from William James’ bear: a
(2001) Differential extrageniculostriate and amygdala re- review of preattentive mechanisms and their contributions to
sponses to presentation of emotional faces in a cortically emotional experience. Cogn. Emotion, 12: 667–696.
blind field. Brain, 124: 1241–1252. Roser, M. and Gazzaniga, M.S. (2004) Automatic brains —
Morris, J.S., Öhman, A. and Dolan, R.J. (1998) Conscious and interpretive minds. Curr. Dir. Psychol., 13: 56–59.
unconscious emotional learning in the human amygdala. Sergent, C., Baillet, S. and Dehaene, S. (2005) Timing of the
Nature, 393: 467–470. brain events underlying access to consciousness during the
Öhman, A. (1986) Face the beast and fear the face: animal and attentional blink. Nat. Neurosci., 8: 1391–1400.
social fears as prototypes for evolutionary analyses of emo- Sergent, C. and Dehaene, S. (2004) Is consciousness a gradual
tion. Psychophysiology, 23: 123–145. phenomenon? Evidence for an all-or-none bifurcation during
Öhman, A. and Soares, J.J.F. (1994) Unconscious anxiety: phobic the attentional blink. Psychol. Sci., 15: 720–728.
responses to masked stimuli. J. Abnorm. Psychol., 103: 231–240. Sheline, Y.I., Barch, D.M., Donnelly, J.M., Ollinger, J.M.,
Öhman, A. and Wiens, S. (2003) On the automaticity of au- Snyder, A.Z. and Mintun, M.A. (2001) Increased amygdala
tonomic responses in emotion: an evolutionary perspective. response to masked emotional faces in depressed subjects
In: Davidson, R.J., Scherer, K. and Goldsmith, H.H. (Eds.), resolves with antidepressant treatment: an fMRI study. Biol.
Handbook of Affective Sciences. Oxford University Press, Psychiatry, 50: 651–658.
New York, pp. 256–275. Shevrin, H. and Dickman, S. (1980) The psychological uncon-
Pasley, B.N., Mayes, L.C. and Schultz, R.T. (2004) Subcortical scious — a necessary assumption for all psychological theory.
discrimination of unperceived objects during binocular ri- Am. Psychol., 35: 421–434.
valry. Neuron, 42: 163–172. Snodgrass, J.G. and Corwin, J. (1988) Pragmatics of measuring
Pastore, R.E., Crawley, E.J., Berens, M.S. and Skelly, M.A. recognition memory: applications to dementia and amnesia.
(2003) ‘‘Nonparametric’’ A ‘ and other modem misconcep- J. Exp. Psychol. Gen., 117: 34–50.
tions about signal detection theory. Psychon. Bull. Rev., Snodgrass, M. (2002) Disambiguating conscious and uncon-
10: 556–569. scious influences: do exclusion paradigms demonstrate un-
Pegna, A.J., Khateb, A., Lazeyras, F. and Seghier, M.L. (2005) conscious perception? Am. J. Psychol., 115: 545–579.
Discriminating emotional faces without primary visual cor- Snodgrass, M., Bernat, E. and Shevrin, H. (2004a) Unconscious
tices involves the right amygdala. Nat. Neurosci., 8: 24. perception at the objective detection threshold exists. Percept.
Pessoa, L. (2005) To what extent are emotional visual stimuli Psychophys., 66: 888–895.
processed without attention and awareness? Curr. Opin. Snodgrass, M., Bernat, E. and Shevrin, H. (2004b) Unconscious
Neurobiol., 15: 188–196. perception: a model-based approach to method and evidence.
Pessoa, L., Japee, S., Sturman, D. and Ungerleider, L.G. (2006) Percept. Psychophys., 66: 846–867.
Target visibility and visual awareness modulate amygdala Stoerig, P. (1996) Varieties of vision: from blind responses to
responses to fearful faces. Cereb. Cortex, 16: 366–375. conscious recognition. Trends Neurosci., 19: 401–406.
Pessoa, L., McKenna, M., Gutierrez, E. and Ungerleider, L.G. Stoerig, P., Zontanou, A. and Cowey, A. (2002) Aware or un-
(2002) Neural processing of emotional faces requires atten- aware: assessment of cortical blindness in four men and a
tion. Proc. Natl. Acad. Sci. USA, 99: 11458–11463. monkey. Cereb. Cortex, 12: 565–574.
Pessoa, L., Padmala, S. and Morland, T. (2005) Fate of unat- Vuilleumier, P. (2005) How brains beware: neural mecha-
tended fearful faces in the amygdala is determined by both nisms of emotional attention. Trends Cogn. Sci., 9: 585–594.
attentional resources and cognitive modulation. Neuroimage, Vuilleumier, P., Armony, J.L., Driver, J. and Dolan, R.J. (2001)
28: 249–255. Effects of attention and emotion on face processing in the
Phillips, M.L., Williams, L.M., Heining, M., Herba, C.M., human brain: an event-related fMRI study. Neuron,
Russell, T., Andrew, C., Brammer, M.J., Williams, S.C.R., 30: 829–841.
Morgan, M., Young, A.W. and Gray, J.A. (2004) Differen- Weiskrantz, L. (1986) Blindsight. A Case Study and Implica-
tial neural responses to overt and covert presentations of tions. Oxford University Press, New York.
facial expressions of fear and disgust. Neuroimage, Weiskrantz, L., Warrington, E.K., Sanders, M.D. and Mar-
21: 1484–1496. shall, J. (1974) Visual capacity in hemianopic field following a
Rauch, S.L., Whalen, P.J., Shin, L.M., McInerney, S.C., restricted occipital ablation. Brain, 97: 709–728.
Macklin, M.L., Lasko, N.B., Orr, S.P. and Pitman, R.K. Whalen, P.J., Kagan, J., Cook, R.G., Davis, F.C., Kim, H.,
(2000) Exaggerated amygdala response to masked facial Polis, S., McLaren, D.G., Somerville, L.H., McLean,
stimuli in posttraumatic stress disorder: a functional MRI A.A., Maxwell, J.S. and Johnstone, T. (2004) Human am-
study. Biol. Psychiatry, 47: 769–776. ygdala responsivity to masked fearful eye whites. Science,
Reingold, E.M. (2004) Unconscious perception and the classic 306: 2061.
dissociation paradigm: a new angle? Percept. Psychophys., Whalen, P.J., Rauch, S.L., Etcoff, N.L., McInerney, S.C.,
66: 882–887. Lee, M.B. and Jenike, M.A. (1998) Masked presentations of
121

emotional facial expressions modulate amygdala activity In: Coan, J.A., Allen, J.J.B. (Eds.), The Handbook of Emotion
without explicit knowledge. J. Neurosci., 18: 411–418. Elicitation and Assessment. Series in Affective Sciences.
Wiens, S. (2006) Remain aware of awareness [Correspondence]. Oxford University Press, Oxford, in press.
Nat. Rev. Neurosci. Wiens, S. and Öhman, A. (2005b) Visual masking in magnetic
Wiens, S., Fransson, P., Dietrich, T., Lohmann, P., Ingvar, M. resonance imaging. Neuroimage, 27: 465–467.
and Öhman, A. (2004) Keeping it short: a comparison Williams, M.A., McGlone, F., Abbott, D.F. and Mattingley,
of methods for brief picture presentation. Psychol. Sci., J.B. (2005) Differential amygdala responses to happy and
15: 282–285. fearful facial expressions depend on selective attention.
Wiens, S. and Öhman, A. (2002) Unawareness is more than a Neuroimage, 24: 417–425.
chance event: comment on Lovibond and Shanks (2002). J. Williams, M.A., Morris, A.P., McGlone, F., Abbott, D.F. and
Exp. Psychol. Anim. Behav. Process, 28: 27–31. Mattingley, J.B. (2004) Amygdala responses to fearful
Wiens, S. and Öhman, A. (2005a). Probing unconscious and happy facial expressions under conditions of binocular
emotional processes: on becoming a successful masketeer. suppression. J. Neurosci., 24: 2898–2904.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 7

Neuroimaging methods in affective neuroscience:


Selected methodological issues

Markus Junghöfer1,, Peter Peyk2, Tobias Flaisch3 and Harald T. Schupp3

1
Institute for Biosignalanalysis and Biomagnetism, University of Münster, Münster, Germany
2
Department of Psychology, University of Basel, Basel, Switzerland
3
Department of Psychology, University of Konstanz, Konstanz, Germany

Abstract: A current goal of affective neuroscience is to reveal the relationship between emotion and dy-
namic brain activity in specific neural circuits. In humans, noninvasive neuroimaging measures are of
primary interest in this endeavor. However, methodological issues, unique to each neuroimaging method,
have important implications for the design of studies, interpretation of findings, and comparison across
studies. With regard to event-related brain potentials, we discuss the need for dense sensor arrays to achieve
reference-independent characterization of field potentials and improved estimate of cortical brain sources.
Furthermore, limitations and caveats regarding sparse sensor sampling are discussed. With regard to event-
related magnetic field (ERF) recordings, we outline a method to achieve magnetoencephalography (MEG)
sensor standardization, which improves effects’ sizes in typical neuroscientific investigations, avoids the
finding of ghost effects, and facilitates comparison of MEG waveforms across studies. Focusing on func-
tional magnetic resonance imaging (fMRI), we question the unjustified application of proportional global
signal scaling in emotion research, which can greatly distort statistical findings in key structures implicated
in emotional processing and possibly contributing to conflicting results in affective neuroscience fMRI
studies, in particular with respect to limbic and paralimbic structures. Finally, a distributed EEG/MEG
source analysis with statistical parametric mapping is outlined providing a common software platform for
hemodynamic and electromagnetic neuroimaging measures. Taken together, to achieve consistent and
replicable patterns of the relationship between emotion and neuroimaging measures, methodological as-
pects associated with the various neuroimaging techniques may be of similar importance as the definition of
emotional cues and task context used to study emotion.

Keywords: EEG; MEG; fMRI; average reference; sensor standardization; proportional global signal
scaling; SPM of EEG/MEG distributed source estimations

Neuroimaging methods have been increasingly Phelps, 2004). However, the inherent time lag of
used to explore the neural substrate of emotion. hemodynamic responses limits the temporal reso-
Over the last decade, a multitude of studies utilized lution of fMRI to reveal the dynamics of brain
functional magnetic resonance imaging (fMRI) to activity (Bandettini et al., 1992; Blamire et al.,
indirectly reveal brain activity by measuring 1992). Recordings of the brain’s magnetic and
blood-flow-dependent signal changes in magnetic electrical fields provide data with high temporal
resonance (Murphy et al., 2003; Phan et al., 2004; precision needed to determine the brain dynamics
of emotional processes. Availability of dense
Corresponding author. Tel.: +49-251-8-356-987; sensor electroencephalography (EEG; up to 256
E-mail: markus.junghofer@uni-muenster.de channels) and magnetoencephalography (MEG;

DOI: 10.1016/S0079-6123(06)56007-8 123


124

up to 275 channels) enable more adequate spatial draw inferences about underlying brain sources.
assessment of electromagnetic fields thereby im- One challenge to determine more precisely
proving the ability of these measures to uncover neural sources in ERP research is the adequate
brain sources. Thus, hemodynamic and electro- spatial assessment of brain field potentials.
magnetic neuroimaging measures provide comple- Technological advances enable the routine record-
mentary information regarding brain processes ing of dense sensor ERPs up to 256 channels
and progress may be expected in combining fMRI (Tucker et al., 2003), approaching the ideal of the
and EEG measures (Salek-Haddadi et al., 2003; reference-independent characterization of brain
Debener et al., 2005; Wang et al., 2006). field potentials.
Studying emotional perception from the perspec- The difficulty to achieve a reference-independent
tive of biphasic emotion theory, we utilized event- characterization of brain field potentials and the
related potentials (ERPs), event-related magnetic reference dependency of sparse sampling reflects the
field recordings (ERFs), and functional magnetic biophysics of EEG signal generation. It is widely
resonance imaging (fMRI) to reveal the brain dy- assumed that EEG represents signals meeting two
namics and neural structures of the processing of requirements: First, because individual neuron ac-
emotional visual cues (Schupp et al., this volume; tivity generates small field potentials, activity has to
Kissler et al., this volume; Lang et al., this volume; be synchronous in thousands of neurons to allow
Sabatinelli et al., this volume). In this research, for summation (Creutzfeld et al., 1966). Second,
methodological problems and challenges were en- activity has to occur in spatially aligned neurons
countered, unique to each neuroimaging method, assuring that summation is effective (rather than
with important implications for the interpretation cancellation as in closed fields; Lorente de No,
of data, design of analyses, and comparison across 1947; Coles et al., 1990). It is for this reason that
studies. Specifically, we first discuss reasons for the larger ERP components are considered to reflect
adequate spatial assessment in event-related brain predominantly excitatory postsynaptic potentials of
potential studies and highlight limitations associ- cortical pyramidal cells. Neuronal activity is vol-
ated with popular reference sites such as linked ume conducted through the scull to the scalp, where
mastoid or earlobes. Next, we provide reasons and it can be recorded through surface sensors (Caton,
ways to implement sensor standardization in MEG 1875; Berger, 1929). Measuring brain potentials a
research. In fMRI analyses, while proportional few millimeters away from the source, the measured
global signal scaling is routinely used, we report potential can be approximated by dipoles, with
findings that violations in assumptions of this cor- negligible contribution from multipole field poten-
rection can dramatically impact statistical findings. tials (Lutzenberger et al., 1987; Nunez, 1989). The
None of these problems is specific to emotion dipolar nature of potentials measured at the head
processing and, indeed, points of concern raised by surface can be described as the superposition of the
this review have been articulated repeatedly in cog- individual potentials of all active generators. Thus,
nitive neuroscience. However, we hope that illus- assuming a complete coverage of the head surface
tration of these issues with specific examples from (including the neck) and homogenous conductivity,
emotion research raises sensitivity to methodolog- the integral ERP activity would be zero. This logic
ical issues in affective neuroscience. Finally, we de- underlies the calculation of the so-called average
scribe a source spaced analysis of EEG/MEG reference (Offner, 1950; Bertrand, 1985). Thus,
data achieving electromagnetic data analyses with after recording, the average reference can be com-
the same SPM (Statistical Parametric Mapping) puted by subtracting the mean of all sensors from
routines as established for hemodynamic measures. each individual site. If the potentials at all body
surface points would be known, an average refer-
Event-related brain potentials and brain sources ence transformation would provide a real inactive
reference and thus complete reference independency
An enduring problem inherent to event-related of the EEG. However, EEG recordings according
potential (ERP) research is the limitation to to the traditional international 10/20 system use a
125

comparably small number of electrodes and even the fundamentally equivalent methods ‘‘Lapla-
the international 10/10 system provides limited cov- cian’’ or ‘‘Cortical Mapping’’ (Junghöfer et al.,
erage of the head surface. Similarly, even though 1997) are mathematically unique transformations
some researchers consider sensor arrays larger compensating for the strong spatial lowpass fil-
than 30 sensors as sufficient to calculate the aver- tering effect of the head as volume conductor —
age reference, the averaged potential across uncov- predominantly the ‘‘blurring’’ effect of the skull.
ered head surface areas may reveal residual mean These ‘‘deblurring’’ methods do not demand any
activity leading to an inaccuracy of the average a priori constraints or assumptions and are not
reference (Katznelson, 1981; Dien, 1998). Expanded affected by the ambiguity of the ‘‘inverse prob-
EEG sensor head coverage (Tucker, 1993; Gevins lem.’’ Thus, the CSD recommends itself as a ref-
et al., 1994) increasingly approximates the require- erence-independent method to uncover local
ments of the average reference and the residual cortical generator sources and simulation studies
average activity approaches the expected zero po- demonstrate that progressively more details about
tential. Simulation studies show that the step from a cortical potential distribution can be obtained as
10/20 system to a 128 whole head sensor recording spatial sampling is increased even beyond 128
approximately reduces the residual average refer- channels (Srinivasan et al., 1996). In CSD solu-
ence activity by 50%. The improved coverage of tions, a focal generator source is indicated by a
inferior frontal, temporal, and occipital regions of sink/source pattern of inward/outward flow of
256 sensor arrays provides a further substantial re- current, which on the other hand reveals a more
duction of the residual average reference activity complicated distribution of multiple inward and
(Junghöfer et al., 1999). outward currents in the case of activation of mul-
Residual average reference activity in regions tiple adjacent generator sources. For instance,
not covered by electrodes can be estimated by Junghöfer et al. (2001) used CSD to provide an
extrapolation of the measured potential distribu- increased spatial resolution of the early posterior
tion. The compensation for residual average negative (EPN) potential observed in emotion
reference activity can approach a so-called ‘‘infi- processing. Specifically, using the difference in
nite’’ reference (Junghöfer et al., 1999; Yao, evoked potentials of subjects viewing emotionally
2001). However, extrapolation of activity in arousing or neutral pictures, the CSD revealed
uncovered areas is not unique even if physiolog- bilateral symmetric sources in occipital areas ac-
ical constraints about reasonable field propaga- companied by right lateralized twin parietal sink
tion are taken into account and the extrapolation sources (see Schupp et al., this volume). However
accuracy shrinks dramatically with an increasing the goodness of ‘‘deblurring’’ methods heavily
integral of uncovered regions. For instance, in- depends on a sufficient spatial sampling ratio
ternational 10/20 system recordings are not (Srinivasan et al., 1996; Junghöfer et al., 1997) as
sufficient for any reasonable extrapolation. Sim- the sampling needs to meet the constraint of the
ulation studies demonstrate that a 128-sensor Nyquist sampling theorem in order to avoid ghost
system allows compensation of roughly a third effects consequent upon spatial aliasing. For scalp
while a 256-sensor system can approximate more potential interpolation, CSD computation and
than the half of the residual average reference ‘‘Cortical Mapping,’’ any choice of Green’s spline
activity (Junghöfer et al., 1999). functions could be used. However, optimized
A further mean to achieve a reference-free spline functions can be derived from additional
characterization of ERPs is to calculate the cur- information such as physiological conductivity
rent source density (CSD; Perrin et al., 1987, properties and estimated depth of generator struc-
1989; Gevins et al. 1991), the negative second tures as described in Junghöfer et al. (1997). A
spatial derivative of the scalp voltage distribution. further requirement for applying CSD is a high
As the derivative of a constant value is zero, the signal-to-noise ratio because the CSD technique
CSD extracts the globally constant effect of a profoundly emphasizes high spatial frequencies
reference. The calculation of the CSD as well as and spurious findings may emerge by noisy data.
126

The calculation of inverse distributed source es- these methods need to be considered providing
timations such as the minimum-norm-least-square limitations for the interpretation of the findings.
(MNLS; Hamalainen and Ilmoniemi, 1994) or Distributed source estimations considerably limit
low-resolution tomography (LORETA; Pascual- the detection of nearby focal sources and provide
Marqui et al., 1994) provides further methods to estimations of rather distributed neural generators.
achieve a reference-independent characterization Thus, the nonuniqueness of the inverse estimation
of ERP potentials. These models use a large requires criteria that do not automatically reveal
number of distributed test dipoles varying in the correct solution, and, in the absence of inde-
strength to represent the scalp measured field po- pendent further support, is probably best viewed
tentials. The MNLS, as a linear estimation tech- in relation to broader anatomical regions rather
nique, is based on the assumption that the then specific neural structures. While the spatial
measured scalp potential distribution (U) at each high pass filter characteristic of the CSD dimin-
point in time can be described as the product of a ishes the impact of extended potential distributions
so-called leadfield matrix (L), specifying each elec- of deeper neural activities and thus overestimates
trode sensitivity to each of the distributed sources superficial sources, the MNLS and LORETA tend
of the model head, and the generator activation to explain activities in deeper structures by widely
(G)-U ¼ LG. In order to estimate the generator distributed superficial activities, an effect which
distribution G ¼ LU the inverse of the leadfield can be compensated to some extent by depth
matrix L has to be multiplied with the measured weighting (Fuchs et al., 1994; Pascual-Marqui
scalp potential distribution. However, this matrix et al., 1994). Similar as CSD, inversely distributed
inversion is only defined if the number of columns source estimates demand a good signal-to-noise
(given by the number of sensors in U) and rows ratio because projecting electrophysiological data
(given by the number of sources in G) of L would from two-dimensional (2D) signal space into 3D
be identical and L would have maximal rank. With source space may lead to strongly magnified spa-
distributed source models, the number of sources tial variance.
by far exceeds the number of sensors, and thus L Taken together, dense sensor arrays provide
has to be replaced by the pseudoinverse leadfield multiple avenues to achieve reference-independent
matrix L+ leading to G ¼ L+U. In this case, the characterization of electrophysiological record-
inverse equation is underdetermined and the in- ings. Adequate spatial sampling is requested by
verse problem is ill posed, i.e., the scalp potential these techniques and this may heavily impact the
may be represented by an infinite number of so- outcome of their application. As each method in-
lutions that could produce the identical measured vokes unique assumptions, converging evidence
field potentials. Thus, selection of the ‘‘most real- across different analysis tools and complementary
istic’’ solution in distributed source models re- neuroimaging methods is particularly desirable.
quires further constraints or criteria. In addition to For instance, the early differential occipital nega-
representing the measured scalp field, the MNLS tivity elicited by emotional compared to neutral
estimate uses as an additional criteria that the pictures is revealed by brain maps based on the
pseudoinverse multiplication is characterized by average reference, CSD sink/source patterns in
minimizing the mean power (least square) of the occipito-temporo-parietal regions, and distributed
estimated current density of the sources. In con- posterior sources in minimum-norm solutions (cf.
trast, LORETA selects the distributed source so- Fig. 2; Junghöfer et al., 2001; Schupp et al., 2006).
lution that is maximally smooth assuming that the Furthermore, fMRI results and electromagnetic
higher spatial resolution of other solutions is recordings provide independent evidence for the
usually not justified for EEG data. While the increased activation of visual-associative struc-
most important advantage of distributed source tures by emotional cues amounting to converging
modeling is that these techniques do not depend evidence across several neuroimaging methods
on assumptions regarding location or number of and types of analyses (cf., Figs. 4 and 5; Jun-
brain generators, the assumptions introduced by ghöfer et al., 2005a,b, 2006).
127

Sparse sensor sampling and the reference issue expressions (de Gelder et al., 2002; Holmes et al.,
2003, 2005, Pourtois et al., 2004, 2005) or emo-
As discussed above, sparse sampling of field po- tional words (Chapman et al., 1980; Bernat et al.,
tentials would seriously violate the assumption 2001; Pauli et al., 2005). Other research utilized
underlying the calculation of the average reference dense sensor recordings providing an improved
(Katznelson, 1981; Dien, 1998). Sparse sampling description of the field potentials by calculation of
arrays provide a reference-dependent depiction the average reference in emotional picture (Jun-
of brain field potentials because EEG amplifier ghöfer et al., 2001; Keil et al., 2002, 2005; Schupp
systems have the requirement to obtain voltage et al., 2004a,b, 2006; Stolarova et al., 2006; Flaisch
recordings as difference between two locations on et al., 2005), emotional face (Batty and Taylor,
the head or body surface. In the past, many re- 2003; Schupp et al., 2003a,b; Meeren et al., 2005),
searchers hoped to minimize the recording of brain or emotional word processing (Skrandies et al.,
activity from one recording site by choosing po- 2003; Ortigue et al., 2004; Herbert et al., 2006;
sitions such as the linked earlobes, mastoids, or Kissler et al., this volume).
nose. The hope to approach monopolar recordings In the following, significant effects of the refer-
has been dubbed as ‘‘convenient myth’’ (Davidson ence choice (average reference vs. linked mastoids)
et al., 2000, p. 33) or ‘‘EEG/ERP folklore’’ are illustrated using data from a recent study in
(Nunez, 1990, p. 25). Considering that electrical which subjects viewed a continuous stream of
field potentials are volume conducted throughout emotionally arousing and neutral pictures, each
the head, there is no site on the head surface presented for 1 s (Junghöfer et al., 2003). Figure 1
showing a consistent zero activity across all pos- illustrates the time course of anterior, posterior,
sible brain sources. Specifically, depending on the and right inferior-lateral scalp potential activity
location of brain sources, electrodes considered to using mid-frontal, mid-occipital, and occipito-
represent an inactive reference may reveal sub- temporal electrode sites. On the basis of the
stantial field potentials varying dynamically across average reference, the occipital sensor (Fig. 1c)
time. To provide a potentially more intuitive anal- revealed a more negative potential for subjects
ogy, consider temperature measurements obtained viewing emotional compared to neutral cues. The
as difference between locations and compare relative negative difference component appeared
effects to absolute (true) values. On a sunny day, sizable, developed with the falling slope of the
the temperature may be 20 1C, 25 1C, and 30 1C in P100 and was maximally pronounced around
locations A, B, and C. Using A as reference, we 220 ms. Polarity reversal was observed over ante-
measure for B and C temperature of 5 1C and rior sensor sites as illustrated for a frontal sensor
10 1C, respectively. Using B as reference, we obtain (Fig. 1a). Specifically, emotional pictures were
–5 1C and +5 1C for locations A and C. Without associated with enhanced positivity compared
information about the reference temperature, ab- to neutral items. A much different pattern of re-
solute temperatures are not achievable. It is for sults emerged for the linked mastoid reference. Of
this reason that the choice of reference can have most relevance, the relative difference potential
rather dramatic effects on the appearance of ERP ‘‘Emotional minus Neutral’’ appears small at the
recordings and thus, interpretation of ERP find- occipital sensor (Fig. 1d). Furthermore, frontal
ings need to consider the choice of reference. Ap- positivity effects of the difference potential
plying these issues to the field of emotion research, are greatly amplified compared to the average
a large number of studies utilized visually pre- reference recordings (Fig. 1b). These differences in
sented stimuli and many of these studies relied the appearance of the effects of differential emo-
on the popular linked mastoid reference when tion processing are easy to explain: Mastoid sen-
studying the processing of emotional pictures sors are most sensitive to generator sources in
(Cuthbert et al., 2000; Schupp et al., 2000; Kemp occipito-temporal brain regions engaged during
et al., 2002; Pause et al., 2003; Amrhein et al., processing of visual cues. Consistent with this
2004; Carretie et al., 2004, 2005), emotional facial notion, pronounced ERP activity for the right
128

Fig. 1. While subjects viewed a continuous stream of emotionally arousing and neutral pictures, each presented for 1 s, ERPs were
measured with a 128-sensor whole head EEG system. On the basis of an average reference (left row), emotional pictures evoke a
relative negative ERP difference component (EPN) over occipital and occipito-temporal regions starting with the falling slope of the
P100 and finding its maximum around 225 ms at occipital (c) and roughly 250 ms at occipito-temporal (mastoid) sites (e). If referenced
to linked mastoids (right row) the strong negative difference component at the occipito-temporal mastoid sites is subtracted from all
other electrodes strongly diminishing the posterior negativity at occipital leads (d) and significantly increasing the corresponding
positive potential differences at frontal sites (b). The spreading brain activation from occipital to temporal areas leads to artificial
latency shifts between posterior negativities and frontal positivities if referenced to linked mastoids.

mastoid site is revealed with the average reference Brain potential maps (Fig. 2) further detail the
(similar left mastoid activity is omitted for brev- effects of linked mastoids on the appearance of the
ity). In contrast, the mean activity of both mas- ERP difference of emotional and neutral cues.
toids gets subtracted from all other sensors when Obviously, while the spatial characteristics of
using a linked mastoid reference, leading to a the potential distribution remain unaffected by
constant zero potential at the mastoids (Fig. 1e). choice of reference, the absolute magnitude of the
129

Fig. 2. Reference dependency of the EPN potential distribution at its maximum around 225 ms: The linked mastoid reference
artificially converts the posterior (a) and right hemispheric dominant (c) EPN into an anterior (b) positivity without clear laterality
effects (d). If based on the scalp potential distribution only, interpretations with respect to underlying generator structures are difficult.
Additional information from reference-independent analysis techniques like the CSD (e) or the minimum-norm-least-square (f) or
other neuroimaging methods such as fMRI (see Figs. 4 and 5) is needed.

potential characteristic depends entirely on the effects on the latency at which condition effects
chosen reference location. In effect, with a mastoid appear. For instance, on the basis of linked mas-
reference (Fig. 2b) the ERP correlate of differen- toid reference, latency differences between earlier
tial processing over distributed temporo-occipital negative posterior and later positive anterior po-
areas is strongly suppressed if compared to tential differences in visual processing have
the average reference (Fig. 2a), which in contrast been considered as evidence to reflect distinct gen-
reveals a pronounced differential occipital nega- erators in anterior/posterior brain regions (e.g.
tivity for emotional pictures compared to neutral Bentin et al., 1996; Eimer, 2000). Similarly, in the
pictures. example at hand, peak latency of emotional
In addition to such quite obvious spatial effects, discrimination appears also to be distinctly later
the choice of reference may also have more subtle for the fronto-central positivity (around 245 ms;
130

Fig. 1b) compared to the residual occipital nega- negative difference potential observed for the
tivity (around 220 ms; Fig. 1d) if applying a linked linked mastoid reference may also affect findings
mastoid reference. However, the different peak of hemispheric dominance.
latencies might be attributed to electrical activity
generated in occipito-temporal brain regions
Summary
picked up by the mastoid sensors. Possibly reflect-
ing the spread of brain activation from occipital to
ERP recordings are increasingly used to reveal the
occipito-temporal brain regions, peak latency of
brain dynamics in emotion processing. Dense sen-
emotional discrimination appears earlier at occip-
sor arrays provide multiple avenues to achieve the
ital sites (225 ms; Fig. 1c) and roughly 20 ms later
reference-independent characterization of field po-
at occipito-temporal electrodes such as the right
tentials and considerably improve the estimate of
mastoid (245 ms; Fig. 1e). Thus, neglecting the
cortical brain sources. Caveats specific to each of
reference issue and the principle that voltage
these analyses need to be considered as well as the
differences reflect the difference between bipolar
general principle that ERP field potentials reflect
sensor sites, the mastoid reference is suggestive of
the superimposition of all active brain sources.
anterior/posterior latency differences by pushing
However, limitations and shortcomings not
peaks at sites more posterior than the mastoids to
withstanding, inference of brain sources appears
earlier latencies and peaks at more anterior sites to
possible and reasonable in particular when con-
later latencies. In contrast, average reference mon-
verging evidence is provided by other neuroimag-
tage provides a rather different view suggesting
ing methods. The use of sparse sensor arrays
that mid-occipital and lateral occipital sensor sites
imposes limitations with regard to the assessment
have different peak latencies with regard to emo-
of surface potentials. Active brain sites cannot be
tion discrimination. It is important to note that
inferred and the reference issue needs to be con-
this example illustrates distinct latency effects for
sidered when interpreting data and comparing
different reference choices only and should of
across studies.
course not be taken as evidence against early
affect-driven difference activities in frontal brain
areas. Event-related magnetic fields: effects of sensor
In addition to latency effects and spatial effects standardization
in posterior–anterior direction, the choice of ref-
erence may also affect the appearance of hemi- The MEG detects weak magnetic fields generated
spheric differences in emotion processing. With by the flow of intracellular postsynaptic currents
respect to affective processing of visual stimuli, the (Williamson and Kaufman, 1981; Hari and Forss,
occipito-temporal areas of the right hemisphere 1999) of pyramidal cells, which constitute two
often showed stronger effects of motivated atten- thirds of the neurons of the cerebral cortex
tion for emotional pictures (Junghöfer et al., 2001) (Creutzfeldt, 1995). It has been estimated that a
and faces (Schupp et al., 2004b) while the left small area of about 40 mm2 including tens of
hemisphere showed dominance for the processing thousand synchronously active neurons can yield a
of emotional words (Kissler et al., this volme). As net dipole moment of about 10 nAm, which is
shown in Fig. 2c, the right hemispheric dominance strong enough to be detected extracranially by
for emotional picture processing can be readily MEG (Hamalainen, 1993). While the EEG meas-
observed in the posterior negativity with average ures the electric potential of the secondary cur-
reference recording. However, forcing effects rents, the MEG measures the additive overlay of
away from the occipito-temporal references, linked the weak magnetic fields of both primary and vol-
mastoids appear to sandwich difference effects ume currents (Sarvas, 1987). MEG is principally
from both sides to the frontal midline direction as sensitive to sources that are oriented tangentially
illustrated by the lateral symmetric anterior pos- to the skull, and much less sensitive to those ori-
itivity (Fig. 2d). Thus, the attenuated posterior ented radially. Hence, MEG is mainly constrained
131

to cortical areas that are bounded in the walls of One reason that MEG sensor standardization is
fissural cortex and the amplitude of the measured not routinely employed may be that MEG re-
MEG signal decreases rapidly as the source search often has the goal to use inverse modeling
depth increases. MEG measures provide a refer- to localize brain activity. The application of in-
ence-independent characterization of magnetic verse methods, like multiple equivalent current di-
fields. Furthermore, being differentially sensitive poles (ECD; Brazier, 1949; de Munck et al., 1988)
to tangential and radial generator orientations, or distributed source estimations like MNLS
MEG provides a different view on neural gener- (Hämäläinen and Ilmoniemi, 1994) or LORETA
ator sources compared to EEG. Combining (Pascual-Marqui et al., 1994), is based on individ-
EEG and MEG measures may therefore provide ual head-sensor configuration without necessity of
complementary evidence of neuronal emotion sensor standardization across sessions or individ-
processing. However, results of ERP and ERF uals. However, except for special cases where the
studies are difficult to compare because MEG number of underlying sources involved in neural
analysis is traditionally predominantly performed processing (ECDs) is fairly well known (Scherg
in the generator source space while ERP analyses and von Cramon, 1985) inverse source estimation
is most often based on measured sensor space is not unique, i.e., different generator distributions
subsequently extended to source space. The main can lead to identical magnetic field measures
reason for these different approaches is that a (Helmholtz, 1853). The nonuniqueness of the in-
standardized alignment of magnetic field sensors in verse estimation, as well as the substantial en-
MEG is almost unachievable as sensor positions largement of the signal space with transformation
are fixed within the MEG scanner and can thus from the 2D ‘‘sensor space’’ into the 3D ‘‘source
not be adjusted for individual head sizes and head space’’ adds significant spatial variance and might
shapes. In contrast, sensor positioning in EEG is eventually decrease statistical power compared to
usually standardized with respect to head land- the analysis in the sensor space. Consequently and
marks (nasion, inion, vertex, and mastoids). Thus, similar to the EEG, statistical analysis in MEG
standardized sensor positioning across different sensor space can be favorable if temporal aspects
subjects is almost impossible for the MEG and are the main focus of interest and regional local-
even exact repositioning of a subject in the MEG ization of neural activities would be sufficient.
scanner for a second measure is much harder to Moreover, due to (i) the better signal-to-noise ra-
achieve compared to the EEG. Consequently, to tio, (ii) the undisturbed DC coupling, (iii) the ref-
allow within subject comparisons across different erence independency, and (iv) the higher spatial
sessions, comparisons across different subjects, or resolution because of the less blurred topogra-
comparisons across different MEG systems (e.g. phies, MEG analysis in the sensor space may
magnetometer vs. gradiometer system), extrapola- still offer some advantages compared to the EEG
tion of the individually measured fields onto a sensor space analysis. Furthermore, procedures for
standard or target sensor configuration would be MEG sensor standardization have been already
necessary. Techniques for such MEG sensor validated on the basis of simulated data or phan-
standardization have been developed and recom- tom measurements (Knösche, 2002).
mended (Hämäläinen, 1992; Numminen et al., MEG sensor standardization is based on the
1995; Burghoff et al., 2000; Knösche, 2002) principle that a magnetic field distribution meas-
but, until now, did not find acceptance as stand- ured from some distance — the target sensor con-
ard application in MEG research. For instance, figuration — can be determined uniquely from a
neither MEG manufacturer data analysis software magnetic field distribution known at all sites of
(4D-Neuroimaging, VSM Medtech Ltd. or Elekta a closed surface enclosing all neural generators
Neuromag Oy) nor the most popular commercial (Yamashita, 1982; Gonzalez et al., 1991). However,
EEG/MEG analysis software programs (e.g., in reality the individual field is not known at all
BESAs or CURRYs) include MEG sensor points of a closed surface but only at some points
standardization techniques. (the magnetometer or gradiometer positions) of a
132

partial surface and thus, the accuracy of field ex- sensor extrapolations based on inverse MNLS es-
trapolation depends on an adequate spatial sam- timations, which has been suggested as sensor
pling (density of sensors), a sufficient coverage of standardization procedure in magnetocardiogra-
the magnetic fields generated by neural activity and phy (MCG) by Numminen et al. (1995) (see
adequate extrapolation functions. Demands for Knösche, 2002 for a similar procedure). In the al-
dense sensor coverage are even higher in MEG ternative stream, data analysis was achieved with-
compared to EEG because magnetic fields usually out compensation for sensor positioning similar to
reveal higher spatial frequencies compared to the previous MEG studies (Costa et al., 2003; Susac et
EEG, mainly a consequence of the low conductivity al., 2004). Furthermore, to compare statistical
of the cranial bone strongly affecting EEG record- power, the original sample (n ¼ 24) was reduced to
ings. However, modern whole head MEG systems either 18 or 12 subjects. Figure 3 illustrates the
provide an adequate spatial sampling, as well as statistical effects comparing CS+ and CS– (CS —
sufficient head coverage (Knösche, 2002). The ex- conditioned stimulus) processing with paired t-
trapolation of the 2D ‘‘sensor space’’ onto a stand- tests for the auditory N1 peak (95–125 ms) and the
ard sensor configuration (forward solution) poses original and reduced samples. Areas with statisti-
similar problems as the ‘‘inverse solution’’ in source cally significant (po0.01) enhanced outgoing mag-
modeling. In the same way as inverse methods netic fields for CS+ compared to CS stimulus
try to minimize ambiguities by application of phys- processing are marked by ‘‘+’’ signs and areas
iological constraints — e.g. information about rea- with significantly enhanced ingoing fields are indi-
sonable source locations in the gray matter (Nunez, cated by ‘‘‘‘signs, respectively. The statistical
1990) and orientations perpendicular to the gyri maps reflect a tangential N1 dipole field with neg-
(Lorente de No, 1938) — extrapolation functions ative ingoing fields over centro-parietal regions
should take into account corresponding physiolog- and positive outgoing magnetic fields over inferior
ical constraints with regard to reasonable magnetic fronto-temporal regions. Obviously, sensor stand-
field propagation. An optimal extrapolation should ardization of the same data is reflected by in-
thus estimate the ‘‘most reasonable’’ neural gener- creased statistically significant sensor areas to
ator distribution explaining the magnetic field to- detect classical conditioning effects of the N1 peak
pography measured at an individual sensor and the advantages of sensor standardization are
configuration and use forward modeling to reveal distinctly more pronounced for smaller subject
the corresponding magnetic field at the standard- samples. In addition, although superficially simi-
ized target positions. Similar to inverse modeling, lar, topographic differences emerge for the original
different techniques to estimate neural sources (e.g., sample in the topography of the effects. Sensor
MNLS, LORETA) provide somewhat differing in- standardization is superior to detect conditioning
verse solutions, extending in this case to differing effects over inferior temporal regions. Without
extrapolations in the standardized target space. sensor standardization, the ‘‘effective sensor cov-
However, the important point is that the extrapo- erage’’ is limited to areas covered by the majority
lated 2D magnetic field distributions based on of subjects.
differing inverse solutions reveal much less variance These data suggest that MEG sensor standard-
compared to the variance in the 3D source space. ization should be considered as standard routine in
Thus, estimation of the magnetic field in standard- MEG sensor space analyses. Statistical power and
ized sensor space may have increased statistical ‘‘effective sensor coverage’’ was increased in the
power compared to the analyses in ‘‘source space.’’ present example and there are reasons to suspect
To illustrate the effects of MEG analyses with that standardization effects were rather underesti-
and without sensor standardization, we used mated in the present study. Specifically, effects of
whole head MEG data (275 channels) obtained sensor standardization may become significantly
during a classical conditioning experiment (Jun- stronger with MEG recordings using less sensors
ghöfer et al., 2005a). Sensor standardization was and smaller coverage compared to the 275 whole
achieved by the application of inverse/forward head VSM Medtech Ltd. system. Advantages are
133

Fig. 3. Effects of sensor standardization for MEG sensor space data analysis. Statistical effects of paired t-tests comparing the
processing of affectively conditioned CS+ and CS tones with a 275-sensor whole head MEG are shown. Areas with statistically
significant (po0.01) enhanced outgoing (+) and ingoing () magnetic fields of the auditory N1 peak (95—125 ms) are projected onto a
model head. Sensor standardization leads to increased statistically significant sensor areas and widens the effective sensor coverage, in
particular over inferior temporal regions. The advantages of sensor standardization are distinctly more pronounced for smaller subject
samples.

not limited to the increased statistical power and about reasonable magnetic field propagation pro-
‘‘effective sensor coverage.’’ Sensor standardiza- vide a high goodness of extrapolation if they are
tion is a prerequisite for averaging across subjects based on adequate sensor density and sufficient
or sessions in EEG and MEG research to avoid head coverage. Sensor standardization signifi-
ghost effects as pure consequence of inconsistent cantly improves effect sizes in typical neuroscien-
sensor configuration. Furthermore, sensor stand- tific investigations, avoids the finding of ghost
ardization may facilitate comparison of MEG effects, and facilitates comparison of MEG wave-
waveforms across studies. forms across studies.

Summary Functional magnetic resonance imaging: effects of


proportional global signal scaling
As a consequence of both the nonuniqueness of
the inverse solution as well as the higher dimen- ERP and ERF measures provide excellent tempo-
sionality of the ‘‘source’’ compared to the ‘‘sensor ral resolution but are limited with respect to de-
space,’’ statistical analysis in MEG ‘‘sensor space’’ termining the exact location of brain generators.
can be favorable. However, sensor standardiza- In contrast, fMRI (Bandettini et al., 1992; Kwong
tion is a necessary prerequisite for such an anal- et al., 1992; Ogawa et al., 1992) is an important
ysis. Three-dimensional extrapolation functions method for determining the location of activated
taking into account physiological constraints brain regions during emotion processing with
134

much better spatial resolution. This method pro- comparing emotional and neutral pictures taken
vides an indirect assessment of brain activity by from the International Affective Picture System
measuring local concentration variations of oxy- (IAPS, Lang et al., 2005). One specific concern was
genated hemoglobin in the blood presumed to be to consider the effects of emotional intensity. As
mostly induced by synaptic events of neuronal ac- suspected, the strong and distributed BOLD acti-
tivity (Logothetis et al., 2001; Raichle, 2001). vations elicited by high-arousing emotional con-
However, utilizing this method in emotion re- tents dominated the global signal variance
search poses important challenges. Limbic and violating the orthogonality assumption of global
paralimbic target structures are notoriously diffi- signal and experimental condition for high arous-
cult to assess and often require specific protocols. ing emotional materials but not for low-arousing
For instance, sampling of the amygdala is im- emotional contents. In line with previous reports
proved by acquiring thin coronal slices that min- (cf. Aguirre et al., 1998), this violation of the or-
imize signal loss due to susceptibility artifacts thogonality assumption resulted in widely differing
(Ojemann et al., 1997; Merboldt et al., 2001). Sim- outcomes comparing two streams of fMRI anal-
ilarly, a number of recommendations for an opti- ysis with and without global proportional signal
mal assessment of blood oxygen level dependent scaling: As shown in Fig. 4, the unjustified appli-
(BOLD) activity in the orbitofrontal cortex also cation of proportional global signal scaling
suffering from strong susceptibility artifacts are (PGSS) leads to an attenuated effect of emotional
outlined by Kringelbach and Rolls (2004). In ad- activation in structures with a positive correlation
dition to such challenges in the assessment of neu- of local and global BOLD signal (‘‘activation’’;
ral target structures, methodological issues in data indicated by reddish colors). Omitting global pro-
analysis may be also of concern when studying portional signal scaling, structures associated with
emotional processes. A still unresolved contro- pronounced BOLD signal activations when con-
versy is the use of proportional global signal scal- trasting high-arousing and neutral picture condi-
ing (PGSS) of BOLD signals in fMRI analysis. tions were apparent in uni- and heteromodal
Global variations of BOLD-fMRI signal are sensory processing areas in the occipital, parietal,
changes common to the entire brain volume and temporal lobe. Invariably, although still sig-
and have been considered to reflect background nificant, application of PGSS was reflected by re-
activity rather than signal changes related to ex- duced effect sizes and cluster volumes. However,
perimental manipulations (Ramsay et al., 1993). both streams of analysis also differed qualitatively,
Consequently, global variations of the BOLD i.e., revealing significant findings in the analyses
signal are commonly considered as nuisance without PGSS, when focusing on structures with
effects, contributing unwanted sources of variance moderate effect sizes in the parieto-temporo-oc-
such as hardware scanner drifts, physiological cipital cortex as well as limbic and paralimbic
movements, or pulsations. While agreeing on the structures. Focusing on ‘‘deactivations,’’ the use of
fundamental necessity of global signal correction, global signal covariates augmented effects in struc-
discussion centered on the proper method of nor- tures with a negative correlation of emotional
malization, identifying global signal changes as ei- arousal and BOLD signal (indicated by bluish col-
ther ‘‘additive’’ or ‘‘multiplicative’’ compared to ors in Fig. 4a). Specifically, areas revealing signifi-
the regional effects of interest (Fox et al., 1988; cant deactivation in the analysis without PGSS,
Friston et al., 1990; Arndt et al., 1996). However, such as parietal, occipital, and temporal structures,
normalization of global signal changes might turn revealed enlarged effect sizes and cluster volumes
into a confound, i.e. significantly changing the in the PGSS analysis. Furthermore, the PGSS
outcome of the analysis when the global signal is analysis revealed significant deactivations not ob-
not orthogonal to the experimental paradigm served when omitting global signal covariates, es-
(Andersson, 1997). pecially left parietal and right hemispheric
We recently explored reservations regarding the subcortical structures. As can be seen in Fig. 4b,
use of PGSS in the domain of emotion research analysis with and without PGSS appeared similar
135

Fig. 4. Effects of MRI proportional global signal scaling (PGSS) in emotion research (random effect analysis; 21 subjects; k ¼ 20;
DF ¼ 20). Significant BOLD signal differences contrasting high-arousing emotional (a) or low-arousing emotional (b) with neutral
picture conditions are shown. Reddish colors indicate a stronger BOLD signal for the emotional stimuli compared to the neutral
stimuli; bluish colors indicate stronger activations for neutral conditions compared to emotional picture conditions. The unjustified
application of PGSS greatly distorts statistical findings in key structures implicated in emotional processing and might contribute to
conflicting results in affective neuroscience fMRI studies, in particular with respect to limbic and paralimbic structures.

when comparing low-arousing emotional and neu- emotion-driven activation of these structures
tral picture contents restricting the detrimental across studies (Davis and Whalen, 2001; Phan
effects of unjustified global scaling to emotional et al., 2002). The present data suggest that the un-
stimuli of high intensity. justified use of global signal scaling might contrib-
Emotional stimuli have been associated with ute to inconsistencies in findings. In contrast to the
activations in limbic and paralimbic structures analysis omitting global scaling, the PGSS-based
(Morris et al., 1998; Vuilleumier et al., 2001; Ha- analysis failed to reveal significant activations in
mann and Mao, 2002; Pessoa et al., 2002). How- limbic and paralimbic regions for high-arousing
ever, results were rather inconsistent regarding the emotional materials. It seems noteworthy that this
136

effect was apparent for the amygdala, a key struc- imaging (Friston et al., 1995). Please note that a
ture of emotional processing in current emotion similar approach will be implemented in future
theories (LeDoux et al., 2000; Lang et al., 2000; releases of SPM (www.fil.ion.ucl.ac.uk/spm). The
Lang and Davis, this volume). Thus, the unjusti- idea behind this approach is that the distributed
fied application of global signal covariates appears neural activity generating measurable scalp poten-
to be a strong confound with regard to limbic and tials or magnetic fields can be estimated by inverse
paralimbic structures, possibly contributing to in- distributed source methods like MNLS (Ham-
consistencies in the literature. It appears likely alainen and Ilmoniemi, 1984), LORETA (Pascual-
that the untested use of global signal covariates Marqui et al., 1994), sLORETA (Pascual-Marqui,
impedes the generalization across studies, subject 2002; Liu et al., 2005), or even Beamformer tech-
groups, and stimulus materials (Davis and Wha- niques (Van Veen et al., 1997; Robinson and Vrba,
len, 2001; Phan et al., 2002; Wager et al., 2003). 1999; Sekihara et al., 2001; Huang et al., 2004) and
Particularly troublesome errors of interpretation the resulting 3D volumes of estimated neural ac-
might arise when comparing results of experimen- tivity can be treated and analyzed in a similar
tal conditions with and without significant global fashion as fMRI volumes of BOLD activity. Thus,
signal correlation and thus, this information event-related single trial EEG or MEG data can be
should be provided in publications (Aguirre et averaged across all time samples in an interval
al., 1998). of interest, the underlying generator activity of
this time interval can be estimated using inverse
distributed source modeling, and these functional
Summary
volumes of brain activity can then be submitted to
the parametric analysis of SPM. The number of
Taken together, this study demonstrated that the
epochs in EEG or MEG studies does often out-
concerns and precautions questioning the standard
number the number of functional scans in fMRI
use of PGSS in the cognitive domain also apply to
examinations. However, the ratio is usually mod-
emotion research. The unjustified application of
est and state-of-the-art workstations can easily
PGSS in emotion research greatly distorts statis-
process this amount of data in reasonable time.
tical findings in key structures implicated in emo-
In order to demonstrate the practicability of this
tional processing and might contribute to
method, we will present a brief example where
conflicting results in affective neuroscience fMRI
MEG data of 16 subjects who viewed a rapid serial
studies, in particular with respect to limbic and
visual presentation (RSVP; 333 ms per picture; no
paralimbic structures. Reiterating Aguirre et al.
ISI) of 200 alternating aversive and neutral pic-
(1998), it is recommended to report the correlation
tures (IAPS; Lang et al., 1999) were reanalyzed
of global signal and experimental condition when
with this suggested method. As a first step, the
using PGSS and omit this confound in cases where
event-related fields were edited on the basis of a
the global signal and experimental condition show
method for statistical control of artifacts in high-
a significant relationship.
density EEG/MEG measures (Junghöfer et al.,
2000) which includes the extraction of globally
Distributed EEG/MEG source analysis with noise contaminated epochs and interpolation of
statistical parametric mapping regional contaminations within trials. In a second
step, the Tihkonov regularized MNLS with a reg-
A current challenge for progress to determine the ularization value of l ¼ 0.05 was calculated for the
neural underpinnings of emotion processing is to mean magnetic field in the EPN time interval
integrate ERP/ERF and fMRI measures. In the (150—230 ms; see Fig. 1) for each artifact cor-
present example, we provide an approach to anal- rected epoch. As recommended by Hauk et al.
yze ERP/ERF data (Junghöfer et al., 2005a) that (2002), a distributed source model with four con-
allows application of the statistical parametric centric shells and evenly distributed sources in
mapping toolbox of SPM used for functional both tangential directions (2  350, 2  197,
137

2  87, and 2  21 dipoles per shell, no radial di- early posterior negativity (EPN-M) time interval
rection in MEG) was used. The 3D MNLS vol- (150–230 ms) and, for comparison, the hemody-
umes of estimated neural source activity were than namic correlate (BOLD) of the corresponding con-
stored in a 3D voxel based file format (ANAL- trast (aversive vs. neutral picture processing;
YZE) with a field of view covering the whole in- Junghöfer et al., 2005b). Although the same stim-
verse model head (51  51  51 voxel) and an uli were used in both studies and picture presen-
isotropic voxel size of 4 mm (adapted to the small- tation rate was identical, the comparison of both
est spatial distance of neighboring sources). In the effects is somehow limited since different subjects
following step, the functional MNLS volumes have been investigated and even more important,
were spatially filtered by SPM (Friston et al., different paradigms have been applied — block
1995) with an isotropic 20 mm full width maxi- design in fMRI and alternating design in MEG.
mum Gaussian kernel. The motivation for this However, considering these methodological limita-
spatial lowpass filtering of the MNLS estimations tions, results demonstrate converging findings
is the same as in fMRI data smoothing: The ap- across both methods with respect to posterior
plication of a Gaussian random field theory (as brain regions. Both methods suggest that neural
applied in SPM) is based on smooth Gaussian activity is elicited as function of emotional arousal
fields and smoothing will normalize the error dis- in bilateral occipito-temporal and occipito-parietal
tribution and ensure the validity of statistical tests. areas, somewhat more pronounced on right hem-
With respect to across-subject random effects ispheric regions. In contrast, prefrontal activations
analysis, smoothing allows some compensation revealed by fMRI are not readily apparent in MEG
for functional and anatomical differences between results which might be explained by the fact that
subjects. Afterwards, the experimental conditions the chosen isotropic spherical head model is less
were analyzed using the statistical parametric appropriate for prefrontal areas revealing strong
mapping toolbox SPM. Within this procedure, conductivity inhomogenities and anisotropies
conditions were modeled with boxcar functions (Wolters et al., 2006). Of course, due to the lim-
and condition effects were estimated according ited depth resolution of the MEG and the MNLS
to the general linear model while the specific effect bias towards superficial sources, the MEG analysis
of interest (aversive versus neutral picture percep- did not reveal activations in subcortical structures
tion) was compared using a linear contrast. As like thalamus, hippocampus, or amygdale, which
second-level analysis, across-subject random thus appear in the fMRI analysis only. A strong
effects analysis was performed on this contrast advantage of the statistical parametric mapping of
of interest. With respect to the visualization of ERP/ERF data is the possible consideration of
the final calculated statistical parametric map, the additional covariates of interest like behavioral or
spherical model head was coregistered onto the autonomic measures.
Montreal Neurological Institute (MNI) standard
brain and functional results were superposed Summary
on it using the freely available software tool
MRIcro (www.sph.sc.edu/comd/rorden/mricro). Distributed EEG/MEG source analysis with sta-
In the posterior regions of interest, the goodness tistical parametric mapping is one approach that
of coregistration was high, but it was unsatisfying uses common software routines across different
in anterior regions. Thus, regional spheres, realis- neuroimaging measures. Applying the same anal-
tic boundary element (BE; Cuffin, 1996), or finite ysis routines across fMRI, MEG, and EEG may
element (FE; Weinstein et al., 2000) head models help to demonstrate the convergence across meas-
may be needed to explore activations in prefrontal ures. Furthermore, data processing of EEG/MEG
regions. measures using source localization procedures are
The results obtained with this analysis are illus- usually performed within subject-related coordi-
trated in Fig. 5, contrasting the differential process- nate systems. Even if source localization is per-
ing of aversive compared to neutral pictures in the formed in standardized coordinate systems, the
138

Fig. 5. Distributed EEG/MEG source analysis with statistical parametric mapping reveals convergent effects of increased blood flow
and increased MEG source activity in the EPN time interval (150–230 ms) in subjects observing emotionally stimulating and emo-
tionally neutral images.

interindividual variability of the ‘‘source space’’ our view, progress in the field of affective neuro-
can be considered to influence and bias results to science is facilitated by the improvement and
some extend. This is even more important if results standardization of data acquisition and analysis,
over subjects and/or different groups of subjects in particular regarding the goal to compare find-
are compared. The use of SPM together with ings across studies. To achieve consistent and rep-
source analysis procedures opens the gate to non- licable patterns of the relationship between
linear subject normalization methods — a widely emotion and neuroimaging measures, methodo-
used procedure in fMRI analysis — and therefore logical aspects associated with the various ne-
can be expected to significantly decrease spatial uroimaging techniques may be similarly important
variance across subjects. as the definition of emotional cues and task con-
text used to study emotion (Bradley, 2000).
Conclusion
Abbreviations
A current goal of affective neuroscience is to reveal
the relationship between emotion and dynamic BOLD blood oxygen level dependent
brain activity in specific neural circuits (LeDoux, CS conditioned stimulus
2000; Lang and Davis, this volume). In humans, CSD current source density
noninvasive neuroimaging measures are of pri- DC direct current
mary interest in this endeavor. The review raised ECD equivalent current dipoles
specific methodological issues, which we encoun- EEG electroencephalography
tered in our research, while neglecting many other EPN early posterior negativity
issues deserving similar consideration. However, in EPN-M early posterior negativity (meg)
139

ERF event-related magnetic field Blamire, A.M., Ogawa, S., Ugurbil, K., Rothman, D., McCar-
ERP event-related potential thy, G., Ellermann, J.M., Hyder, F., Rattner, Z. and Shul-
FMRI functional magnetic resonance man, R.G. (1992) Dynamic mapping of the human visual
imaging cortex by high-speed magnetic resonance imaging. Proc.
Natl. Acad. Sci. USA, 15; 89(22): 11069–11073.
IAPS international affective picture
Bernat, E., Bunce, S. and Shevrin, H. (2001) Event-related
system brain potentials differentiate positive and negative mood ad-
LORETA low-resolution tomography jectives during both supraliminal and subliminal visual
MCG magnetocardiography processing. Int. J. Psychophysiol., 42(1): 11–34.
MEG magnetoencephalography Bertrand, O., Perrin, F. and Pernier, J. (1985) A theoretical
MNLS minimum-norm-least-square justification of the average reference in topographic evoked
potential studies. Electroencephalogr. Clin. Neurophysiol.,
PGSS proportional global signal scaling 62(6): 462–464.
RSVP rapid serial visual presentation Bradley, M.M. (2000) Emotion and motivation. In: Cacioppo,
J.T., Tassinary, L.G. and Berntson, G. (Eds.), Handbook of
Psychophysiology. Cambridge University Press, New York,
pp. 602–642.
Acknowledgments Brazier, M.A.B. (1949) A study of the electric field at the surface
of the head. Electroenceph. Clin. Neurophysiol., 2: 38–52.
This work was supported by the German Research Burghoff, M., Nenonen, J., Trahms, L. and Katila, T. (2000)
Foundation (DFG) Grants FOR 348/3-3, 348/3-4 Conversion of magnetocardiographic recordings between
two different multichannel SQUID devices. IEEE Trans.
and by the Academy of Science, Heidelberg (WIN-
Biomed. Eng., 47(7): 869–875.
Project). Address reprint requests to Markus Jun- Carretie, L., Hinojosa, J.A., Martın-Loeches, M., Mercado, F.
ghöfer; Institute for Biosignalanalysis, University and Tapia, M. (2004) Automatic attention to emotional
of Münster, Malmedyweg 15, 48149 Münster; stimuli: neural correlates. Hum. Brain Mapp., 22: 290–299.
Germany. Carretie, L., Hinojosa, J.A., Mercado, F. and Tapia, M. (2005)
Cortical response to subjectively unconscious danger. Ne-
uroimage, 24(3): 615–623.
Caton, R. (1875) The electric currents of the brain. Br. Med. J.,
References 2: 278.
Chapman, R.M., McCrary, J.W., Chapman, J.A. and Martin,
Aguirre, G.K., Zarahn, E. and D’Esposito, M. (1998) The in- J.K. (1980) Behavioral and neural analyses of connotative
ferential impact of global signal covariates in functional ne- meaning: word classes and rating scales. Brain Lang., 11(2):
uroimaging analyses. Neuroimage, 8: 302–306. 319–339.
Amrhein, C., Muhlberger, A., Pauli, P. and Wiedemann, G. Coles, M., Gratton, G. and Fabiani, M. (1990) Event related
(2004) Modulation of event-related brain potentials during brain potentials. In: Cacioppo, J. and Tassinary, L. (Eds.),
affective picture processing: a complement to startle reflex Principles of Psychophysiology. Cambridge University Press,
and skin conductance response? Int. J. Psychophysiol., 54(3): Cambridge, New York, Port Chester, Melbourne, Sydney.
231–240. Costa, M., Braun, C. and Birbaumer, N. (2003) Gender differ-
Andersson, J.L. (1997) How to estimate global activity inde- ences in response to pictures of nudes: a magnetoencepha-
pendent of changes in local activity. Neuroimage, 6: 237–244. lographic study. Biol. Psychol., 63(2): 129–147.
Arndt, S., Cizadlo, T., O’Leary, D., Gold, S. and Andreasen, Creutzfeld, O., Watanabe, S. and Lux, H. (1966) Relations be-
N.C. (1996) Normalizing counts and cerebral blood flow in- tween EEG phenomena and potentials of single cortical cells.
tensity in functional imaging studies of the human brain. I. Evoked responses after thalamic and epicortical stimula-
Neuroimage, 3: 175–184. tions. Electroenceph. Clin. Neurophysiol., 20: 1–28.
Bandettini, P.A., Wong, E.C., Hinks, R.S., Tikofsky, R.S. and Cuffin, B.N. (1996) EEG localization accuracy improvements
Hyde, J.S. (1992) Time course EPI of human brain function using realistically shaped head models. IEEE Trans. Biomed.
during task activation. Magn. Reson. Med., 25: 390–397. Eng., 43(3): 299–303.
Batty, M. and Taylor, M.J. (2003) Early processing of the six Cuthbert, B.N., Schupp, H.T., Bradley, M.M., Birbaumer, N.
basic facial emotional expressions. Brain Res. Cogn. Brain and Lang, P.J. (2000) Brain potentials in affective picture
Res., 17(3): 613–620. processing: covariation with autonomic arousal and affective
Bentin, S., Allison, T., Puce, A., Perez, A. and McCarthy, G. report. Biol. Psychol., 52: 95–111.
(1996) Electrophysiological studies of face perception in Davidson, R.J., Jackson, D.C., Larson, C.L. (2000) Human elec-
humans. J. Cogn. Neurosci., 8: 551–565. troencephalography. In: Cacioppo, J.T., Tassinary, L.G. and
Berger, H. (1929) Über das Elektroenkephalogramm des Berntson, G.G. (Eds.), Handbook of psychophysiology. 2nd
Menschen. Arch. Psychiat, Nervenkr., 87: 527–570. ed., Cambridge University Press, New York, NY, USA. 27–52.
140

Davis, M. and Whalen, P.J. (2001) The amygdala: vigilance and Hamalainen, M.S. and Ilmoniemi, R.J. (1994) Interpreting
emotion. Mol. Psychiatr., 6: 13–34. magnetic fields of the brain: minimum norm estimates. Med.
de Gelder, B., Pourtois, G. and Weiskrantz, L. (2002) Fear Biol. Eng. Comput., 32: 35–42.
recognition in the voice is modulated by unconsciously rec- Hamann, S. and Mao, H. (2002) Positive and negative emo-
ognized facial expressions but not by unconsciously recog- tional verbal stimuli elicit activity in the left amygdala. Ne-
nized affective pictures. Proc. Natl. Acad. Sci. USA, 99(6): uroreport, 13: 15–19.
4121–4126. Hari, R. and Forss, N. (1999) Magnetoencephalography in the
de Munck, J.C., van Dijk, B.W. and Spekreijse, H. (1988) study of human somatosensory cortical processing. Philos.
Mathematical dipoles are adequate to describe realistic gen- Trans. R. Soc. Lond. B Biol. Sci., 354: 1145–1154.
erators of human brain activity. IEEE Trans. Biomed. Eng., Hauk, O., Keil, A., Elbert, T. and Müller, M.M. (2002) Com-
35(11): 960–966. parison of data transformation procedures to enhance top-
Debener, S., Ullsperger, M., Siegel, M., Fiehler, K., von Cra- ographical accuracy in time-series analysis of the human
mon, D.Y. and Engel, A.K. (2005) Trial-by-trial coupling of EEG. J. Neurosci. Meth., 113(2): 111–122.
concurrent electroencephalogram and functional magnetic Helmholtz, H. (1853) Ueber einige gesetze der vertheilung el-
resonance imaging identifies the dynamics of performance ektrischer strome in korperlichen leitern, mit anwendung
monitoring. J. Neurosci., 25(50): 1730–1737. auf die thierischelektrischen versuche. Ann. Phys. Chem., 89:
Dien, J. (1998) Issues in the application of the average refer- 211–233 353–377.
ence: review, critiques, and recommendations. Behav. Res. Herbert, C., Kissler, J., Junghöfer, M., Peyk, P. and Rockstroh,
Meth., Instrum. Comput., 30: 34–43. B. (2006) Processing of emotional adjectives: evidence from
Eimer, M. (2000) Effects of face inversion on the structural startle EMG and ERPs. Psychophysiology, 43(2): 197–206.
encoding and recognition of faces -evidence from event-re- Holmes, A., Vuilleumier, P. and Eimer, M. (2003) The process-
lated brain potentials. Cogn. Brain Res., 10: 145–158. ing of emotional facial expression is gated by spatial atten-
Flaisch, T., Junghöfer, M. and Schupp, H. (2005) Motivational tion: evidence from event-related brain potentials. Brain Res.
priming and emotional perception: an ERP-study. J. Psy- Cogn. Brain Res., 16(2): 174–184.
chophysiol., 19(2): 115–115. Holmes, A., Winston, J.S. and Eimer, M. (2005) The role of
Fox, P.T., Mintun, M.A., Reiman, E.M. and Raichle, M.E. spatial frequency information for ERP components sensitive
(1988) Enhanced detection of focal brain responses using in- to faces and emotional facial expression. Brain Res. Cogn.
tersubjects and change distribution analysis of subtracted Brain Res., 25(2): 508–520.
PET images. J. Cereb. Blood Flow Metab., 8: 642–653. Huang, M.X., Shih, J.J., Lee, R.R., Harrington, D.L., Thoma,
Friston, K.J., Frith, C.D., Liddle, P.F., Dolan, R.J., Lam- R.J., Weisend, M.P., Hanlon, F., Paulson, K.M., Li, T.,
mertsma, A.A. and Frackowiak, R.S. (1990) The relationship Martin, K., Millers, G.A. and Canive, J.M. (2004) Com-
between global and local changes in PET scans. J. Cereb. monalities and differences among vectorized beamformers
Blood Flow Metab., 10: 458–466. in electromagnetic source imaging. Brain Topogr., 16(3):
Friston, K.J., Holmes, A.P., Worsley, K.J., Poline, J.P., Frith, 139–158.
C.D. and Frackowiak, R.S. (1995) Statistical parametric Junghöfer, M., Bradley, M.M., Elbert, T.R. and Lang, P.J.
maps in functional imaging: a general linear approach human (2001) Fleeting images: a new look at early emotion discrim-
brain mapping. Hum. Brain Mapp., 2: 189–210. ination. Psychophysiology, Special Report, 38(2): 175–178.
Fuchs, M., Wagner, M. and Wischmann, H.A. (1994). Gener- Junghöfer, M., Elbert, T., Leiderer, L., Berg, P. and Rockstroh,
alized minimum norm least squares reconstruction algo- B. (1997) Mapping EEG-potentials on the surface of the
rithms. In: Skrandies W. (Ed.), ISBET Newsletter No. 5. pp. brain: a strategy for uncovering cortical sources. Brain Top-
8–11. ogr., 9(3): 203–217.
Gevins, A., Jian, L., Brickett, P., Reutter, B. and Desmond, J. Junghöfer, M., Elbert, T., Tucker, D. and Braun, C. (1999) The
(1991) Seeing through the skull: advanced EEGs accurately polar effect of average reference: a bias in estimating the head
measure cortical activity from the scalp. Brain Topogr., 4: surface integral in EEG recording. Clin. Neurophysiol.,
125–131. 110(6): 1149–1155.
Gevins, A., Le, J., Martin, N.K., Brickett, P., Desmond, J. and Junghöfer, M., Elbert, T., Tucker, D. and Rockstroh, B. (2000)
Reutter, B. (1994) High resolution EEG: 124-channel re- Statistical control of artifacts in dense array EEG/MEG
cording, spatial deblurring and MRI integration methods. studies. Psychophysiology, 37: 523–532.
Electroenceph. Clin. Neurophysiol., 90: 337–358. Junghöfer, M., Keil, A. and Peyk, P. (2003) An early posterior
Gonzalez, S., Grave de Peralta, R., Lfitkenhoner, B., Me- negative EEG difference component, mirroring facilitated
rminghaus, E. and Hoke, M. (1991) A strategy for the solu- processing of emotionally arousing material in the extended
lion of the inverse problem using simultaneous EEG and visual cortex, is almost independent of presentation rate.
MEG measurements. In: Hoke, M., Okada, Y., Erno, S.N. Psychophysiology, 40: S51–S51 Suppl. 1.
and Romani, G.L. (Eds.), Advances in Biomagnetism ‘91: Junghöfer, M., Peyk, P., Steinsträter, O., Schupp, H. and Pan-
Clinical Aspects. Elsevier, Amsterdam, pp. 99–100. tev, C. (2005a) ‘‘The sound of fear’’: MEG based evidence for
Hämäläinen, M.S. (1992) Magnetoencephalography: a tool for affective conditioning in the auditory pathway. Psychophys-
functional brain imaging. Brain Topogr., 5(2): 95–102. iology, 42, S5–S5 Suppl 1.
141

Junghöfer, M., Sabatinelli, D., Schupp, H.T., Elbert, T.R., Brad- Lorente de No, R. (1947) Analysis of the distribution of action
ley, M.M. and Lang, P.J. (2006) Fleeting images: rapid currents of nerve in volume conductors. Stud. Rockefeller
affect discrimination in the visual cortex. Neuroreport, 17(2): Inst. Med. Res., 132: 384–482.
225–229. Lorente de No, R. (1938) Cerebral cortex: architecture, intra-
Junghöfer, M., Schupp, H., Stark, R. and Vaitl, D. (2005b) cortical connections, motor projections. In: Fulton, J.F.
Neuroimaging of emotion: empirical effects of proportional (Ed.), Physiology of the Nervous System (Chap. 15). Oxford
global signal scaling in fMRI data analysis. Neuroimage, 25: University Press, Oxford, London.
520–526. Lutzenberger, W., Elbert, T. and Rockstroh, B. (1987) A brief
Katznelson, R.D. (1981) EEG recording, electrode placement, tutorial on the implications of volume conduction for the
and aspects of generator localization. In: Nunez, P.L. (Ed.), interpretation of the EEG. J. Psychophysiol., 1: 81–90.
Electric Fields of the Brain: The Neurophysics of EEG. Ox- Meeren, H.K., van Heijnsbergen, C.C. and de Gelder, B. (2005)
ford University Press, New York, pp. 76–213. Rapid perceptual integration of facial expression and emo-
Keil, A., Bradley, M.M., Hauk, O., Rockstroh, B., Elbert, T.R. tional body language. Proc. Natl. Acad. Sci. USA, 102(45):
and Lang, P.J. (2002) Large-scale neural correlates of affec- 16518–16523.
tive picture viewing. Psychophysiology, 39: 641–649. Merboldt, K.D., Fransson, P., Bruhn, H. and Frahm, J. (2001)
Keil, A., Moratti, S., Sabatinelli, D., Bradley, M.M. and Lang, Functional MRI of the human amygdala? Neuroimage,
P.J. (2005) Additive effects of emotional content and spatial 14(2): 253–257.
selective attention on electrocortical facilitation. Cereb. Cor- Morris, J.S., Ohman, A. and Dolan, R.J. (1998) Conscious and
tex, 15(8): 1187–1197. unconscious emotional learning in the human amygdala.
Kemp, A.H., Gray, M.A., Eide, P., Silberstein, R.B. and Nature, 393: 467–470.
Nathan, P.J. (2002) Steady-state visually evoked potential Murphy, F.C., Nimmo-Smith, I. and Lawrence, A.D. (2003)
topography during processing of emotional valence in Functional neuroanatomy of emotions: a meta-analysis.
healthy subjects. Neuroimage, 17(4): 1684–1692. Cogn. Affect. Behav. Neurosci., 3(3): 207–233.
Knösche, T.R. (2002) Transformation of whole-head MEG re- Numminen, J., Ahlfors, S., Ilmoniemi, R., Montonen, J. and
cordings between different sensor positions. Biomed. Tech. Nenonen, J. (1995) Transformation of multichannel magne-
(Berl.), 47(3): 59–62. tocardiographic signals to standard grid form. IEEE Trans.
Kringelbach, M.L. and Rolls, E.T. (2004) The functional ne- Biomed. Eng., 42(1): 72–78.
uroanatomy of the human orbitofrontal cortex: evidence Nunez, P.L. (1989) Estimation of large scale neocortical source
from neuroimaging and neuropsychology. Prog. Neurobiol., activity with EEG surface Laplacians. Brain Topogr., 2(1–2):
72(5): 341–372. 141–154.
Kwong, K.K., Belliveau, J.W., Chesler, D.A., Goldberg, I.E., Nunez, P.L. (1990) Localization of brain activity with electro-
Weiskoff, R.M., Poncelet, B.P., Kennedy, D.N., Hoppel, encephalography. In: Sato, S. (Ed.) Advances in Neurology,
B.E., Cohen, M.S., Turner, R., Cheng, H.M., Brady, T.J. and Magnetencephalography, Vol. 54. Raven Press, New York,
Rosen, B.R. (1992) Dynamic magnetic resonance imaging of pp. 39–65.
human brain activity during primary sensory stimulation. Offner, F.F. (1950) The EEG as a potential mapping: the value
Proc. Natl. Acad. Sci. USA, 89: 5675–5679. of the average monopolar reference. Electroenceph. Clin.
Lang, P.J., Bradley M.M. and Cuthbert, B.N. (1999). Interna- Neurophysiol., 2: 213–214.
tional Affective Picture System (IAPS): Technical Manual Ogawa, S., Tank, D.W., Menon, R., Ellermann, J.M., Kim, S.-
and Affective Ratings. The Center for Research in Psycho- G., Merkle, H. and Ugurbil, K. (1992) Intrinsic signal
physiology, Gainesville, FL. changes accompanying sensory stimulation: functional brain
Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (2005). Interna- mapping with magnetic resonance imaging. Proc. Natl. Acad.
tional affective picture system (IAPS): digitized photographs. Sci. USA, 89: 5951–5955.
Instruction Manual and Affective Ratings. Technical Report Ojemann, J.G., Akbudak, E., Snyder, A.Z., McKinstry, R.C.,
A-6, University of Florida, Gainesville, FL. Raichle, M.E. and Conturo, T.E. (1997) Anatomic localiza-
Lang, P.J., Davis, M. and Ohman, A. (2000) Fear and anxiety: tion and quantitative analysis of gradient refocused echo-pla-
animal models and human cognitive psychophysiology. J. nar fMRI susceptibility artifacts. Neuroimage, 6(3): 156–167.
Affect Disord., 61: 137–159. Ortigue, S., Michel, C.M., Murray, M.M., Mohr, C., Carbon-
LeDoux, J.E. (2000) Emotion circuits in the brain. Annu. Rev. nel, S. and Landis, T. (2004) Electrical neuroimaging reveals
Neurosci., 23: 155–184. early generator modulation to emotional words. Neuroim-
Liu, H., Schimpf, P.H., Dong, G., Gao, X., Yang, F. and age, 21(4): 1242–1251.
Gao, S. (2005) Standardized shrinking LORETA-FOCUSS Pascual-Marqui, R.D., Michel, C.M. and Lehmann, D. (1994)
(SSLOFO): a new algorithm for spatio-temporal EEG source Low resolution electromagnetic tomography: a new method
reconstruction. IEEE Trans. Biomed. Eng., 52(10): for localizing electrical activity in the brain. International
1681–1691. Journal of Psychophysiology, 18: 49–65.
Logothetis, N.K., Pauls, J., Augath, M., Trinath, T. and Oel- Pascual-Marqui, R.D. (2002) Standardized low resolution brain
termann, A. (2001) Neurophysiological investigation of the electromagnetic tomography (sLORETA): technical details.
basis of the fMRI signal. Nature, 412(6843): 150–157. Methods Find. Exp. Clin. Pharmacol., 24: 5–12.
142

Pauli, P., Amrhein, C., Muhlberger, A., Dengler, W. and late positive potential is modulated by motivational rele-
Wiedemann, G. (2005) Electrocortical evidence for an early vance. Psychophysiology, 37(2): 257–261.
abnormal processing of panic-related words in panic disorder Schupp, H.T., Stockburger, J., Codispoti, M., Junghöfer, M.,
patients. Int. J. Psychophysiol., 57(1): 33–41. Weike, A.I. and Hamm, A.O. (2006) Stimulus novelty and
Pause, B.M., Raack, N., Sojka, B., Goder, R., Aldenhoff, J.B. emotion perception: the near absence of habituation in the
and Ferstl, R. (2003) Convergent and divergent effects of odors visual cortex. Neuroreport, 17: 365–369.
and emotions in depression. Psychophysiology, 40(2): 209–225. Schupp, H., Junghöfer, M., Weike, A.I. and Hamm,
Perrin, F., Bertrand, O. and Pernier, J. (1987) Scalp current A.H. (2003a) Attention and emotion: an ERP analysis of fa-
density mapping: value and estimation from potential data. cilitated emotional stimulus processing. Neuroreport, 14:
IEEE Trans. Biomed. Eng., 34(4): 283–289. 1107–1110.
Perrin, F., Pernier, J., Bertrand, O. and EchaUier, J.F. (1989) Schupp, H., Junghöfer, M., Weike, A.I. and Hamm, A.H.
Spherical splines for potential and current density mapping. (2003b) Emotional facilitation of sensory processing in the
Electroenc. Clin. Neurophys., 72: 184–187. visual cortex. Psychol. Sci., 14: 7–13.
Pessoa, L., Kastner, S. and Ungerleider, L.G. (2002) Attentio- Schupp, H., Junghöfer, M., Weike, A.I. and Hamm, A.H.
nal control of the processing of neural and emotional stimuli. (2004a) The selective processing of briefly presented affective
Brain Res. Cogn. Brain Res., 15: 31–45. pictures: an ERP analysis. Psychophysiol., 41(3): 441–449.
Phan, K.L., Wager, T., Taylor, S.F. and Liberzon, I. (2002) Schupp, H., Öhman, A., Junghöfer, M., Weike, A. and Hamm,
Functional neuroanatomy of emotion: a meta-analysis of A. (2004b) Emotion guides attention: the selective encoding
emotion activation studies in PET and fMRI. Neuroimage, of threatening faces. Emotion, 4: 189–200.
16: 331–348. Sekihara, K., Nagarajan, S.S., Poeppel, D., Marantz, A. and
Phan, K.L., Wager, T.D., Taylor, S.F. and Liberzon, I. (2004) Miyashita, Y. (2001) Reconstructing spatio-temporal activ-
Functional neuroimaging studies of human emotions. CNS ities of neural sources using an MEG vector beamformer
Spectr., 9(4): 258–266. technique. IEEE Trans. Biomed. Eng., 48(7): 760–771.
Phelps, E.A. (2004) Human emotion and memory: interactions Skrandies, W. and Chiu, M.J. (2003) Dimensions of affective
of the amygdala and hippocampal complex. Curr. Opin. Ne- semantic meaning — behavioral and evoked potential
urobiol., 14(2): 198–202. correlates in Chinese subjects. Neurosci. Lett., 341(1):
Pourtois, G., Grandjean, D., Sander, D. and Vuilleumier, P. 45–48.
(2004) Electrophysiological correlates of rapid spatial orient- Srinivasan, R., Nunez, P.L., Tucker, D.M., Silberstein, R.B.
ing towards fearful faces. Cereb. Cortex, 14(6): 619–633. and Cadusch, P.J. (1996) Spatial sampling and filtering of
Pourtois, G., Thut, G., Grave de Peralta, R., Michel, C. and EEG with spline laplacians to estimate cortical potentials.
Vuilleumier, P. (2005) Two electrophysiological stages of Brain Topogr., 8(4): 355–366.
spatial orienting towards fearful faces: early temporo-parietal Stolarova, M., Keil, A. and Moratti, S. (2006) Modulation of
activation preceding gain control in extrastriate visual cortex. the C1 visual event-related component by conditioned stim-
Neuroimage, 26(1): 149–163. uli: evidence for sensory plasticity in early affective percep-
Raichle, M.E. (2001) Cognitive neuroscience. Bold insights. tion. Cereb. Cortex, 16(6): 876–887.
Nature, 412(6843): 128–130. Susac, A., Ilmoniemi, R.J., Pihko, E. and Supek, S. (2004) Ne-
Ramsay, S.C., Murphy, K., Shea, S.A., Friston, K.J., Lam- urodynamic studies on emotional and inverted faces in an
mertsma, A.A., Clark, J.C., Adams, L., Guz, A. and oddball paradigm. Brain Topogr., 16(4): 265–268.
Frackowiak, R.S. (1993) Changes in global cerebral blood Tucker, D.M. (1993) Spatial sampling of head electrical fields:
flow in humans: effect on regional cerebral blood flow during the geodesic sensor net. Electroenceph. Clin. Neurophysiol.,
a neural activation task. J. Physiol., 471: 521–534. 87: 154–163.
Robinson, S.E. and Vrba, J. (1999) Functional neuroimaging Tucker, D.M., Luu, P., Frishkoff, G., Quiring, J. and
by synthetic aperture magnetometry (SAM). In: Yoshimoto, Poulsen, C. (2003) Frontolimbic response to negative feed-
T., et al. (Eds.), Recent Advances in Biomagnetism. Tohoku back in clinical depression. J. Abnorm. Psychol., 112(4):
University Press, Sendai, pp. 302–305. 667–678.
Salek-Haddadi, A., Friston, K.J., Lemieux, L. and Fish, D.R. Van Veen, B.D., van Drongelen, W., Yuchtman, M. and
(2003) Studying spontaneous EEG activity with fMRI. Brain Suzuki, A. (1997) Localization of brain electrical activity via
Res. Brain Res. Rev., 43(1): 110–133. linearly constrained minimum variance spatial filtering. IEEE
Sarvas, J. (1987) Basic mathematical and electromagnetic con- Trans. Biomed. Eng., 44(9): 867–880.
cepts of the biomagnetic inverse problem. Phys. Med. Biol., Vuilleumier, P., Armony, J.L., Driver, J. and Dolan, R.J. (2001)
32(1): 11–22. Effects of attention and emotion on face processing in the
Scherg, M. and von Cramon, Y.D. (1985) Two bilateral human brain: an event-related fMRI study. Neuron, 30:
sources of the late AEP as identified by a spatio-temporal 829–841.
dipole model. Electroencephalogr. Clin. Neurophysiol., 62: Wager, T.D., Phan, K.L., Liberzon, I. and Taylor, S.F. (2003)
32–44. Valence, gender, and lateralization of functional brain anat-
Schupp, H.T., Cuthbert, B.N., Bradley, M.M., Cacioppo, J.T., omy in emotion: a meta-analysis of findings from neuroim-
Ito, T. and Lang, P.J. (2000) Affective picture processing: the aging. Neuroimage, 19: 513–531.
143

Wang, Z., Ives, J.R. and Mirsattari, S.M. (2006) Simultaneous current computation in a realistic head model: a simulation
electroencephalogram-functional magnetic resonance imag- and visualization study using high-resolution finite element
ing in neocortical epilepsies. Adv. Neurol., 97: 129–139. modeling. Neuroimage, 30(3): 813–826.
Weinstein, D., Zhukov, L. and Johnson, C. (2000) Lead-field Yamashita, Y. (1982) Theoretical studies on the inverse prob-
bases for electroencephalography source imaging. Ann. Bio- lem in electrocardiography and the uniqueness of the solu-
med. Eng., 28(9): 1059–1066. tion. IEEE Trans. Biomed. Eng., 29: 719–724.78.
Williamson, S.J. and Kaufman, L. (1981) Biomagnetism. J. Yao, D., Wang, L., Oostenveld, R., Nielsen, K.D., Arendt-
Magn. Magn. Mater., 22: 129–201. Nielsen, L. and Chen, A.C. (2005) A comparative study of
Wolters, C.H., Anwander, A., Tricoche, X., Weinstein, D., different references for EEG spectral mapping: the issue of
Koch, M.A. and Macleod, R.S. (2006) Influence of tissue the neutral reference and the use of the infinity reference.
conductivity anisotropy on EEG/MEG field and return Physiol. Meas., 26(3): 173–184.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 8

Emotional and semantic networks in visual word


processing: insights from ERP studies

Johanna Kissler, Ramin Assadollahi and Cornelia Herbert

Department of Psychology, University of Konstanz, P.O. Box D25, D-78457 Konstanz, Germany

Abstract: The event-related brain potential (ERP) literature concerning the impact of emotional content on
visual word processing is reviewed and related to general knowledge on semantics in word processing:
emotional connotation can enhance cortical responses at all stages of visual word processing following the
assembly of visual word form (up to 200 ms), such as semantic access (around 200 ms), allocation of
attentional resources (around 300 ms), contextual analysis (around 400 ms), and sustained processing and
memory encoding (around 500 ms). Even earlier effects have occasionally been reported with subliminal or
perceptual threshold presentation, particularly in clinical populations. Here, the underlying mechanisms are
likely to diverge from the ones operational in standard natural reading. The variability in timing of the
effects can be accounted for by dynamically changing lexical representations that can be activated as
required by the subjects’ motivational state, the task at hand, and additional contextual factors. Through-
out, subcortical structures such as the amygdala are likely to contribute these enhancements. Further
research will establish whether or when emotional arousal, valence, or additional emotional properties
drive the observed effects and how experimental factors interact with these. Meticulous control of other
word properties known to affect ERPs in visual word processing, such as word class, length, frequency, and
concreteness and the use of more standardized EEG procedures is vital. Mapping the interplay between
cortical and subcortical mechanisms that give rise to amplified cortical responses to emotional words will be
of highest priority for future research.

Keywords: emotion; semantics; word processing; event-related potentials; healthy volunteers; clinical
populations

Introduction and activity (Osgood et al., 1957), the first two ac-
counting for the majority of the variance. On the
Influential dimensional approaches to the study of semantic differential, a word’s evaluative connota-
emotion derive their basic dimensions from analy- tion is determined by ratings on a multitude of
ses of written language. Osgood and collaborators, seven-point scales, spanned by pairs of antonyms
using the ‘semantic differential’ technique, were such as hot–cold, soft–hard, happy–sad, etc. Factor
the first to empirically demonstrate that affective analyses of the judgments of many words on such
connotations of words are determined by three scales, given by large subject populations, reveal a
principal dimensions, namely evaluation, potency, three-dimensional evaluative space, whose structure
has been replicated many times and across different
cultures (Osgood et al., 1975). Figure 1 provides an
Corresponding author. Tel.: +49-7531-884616; illustration of the evaluative space determined by
Fax: +49-7531-4601; E-mail: johanna.kissler@uni-konstanz.de the semantic differential.

DOI: 10.1016/S0079-6123(06)56008-X 147


148

Fig. 1. A three-dimensional affective space of connotative meaning as postulated by Osgood (1957, 1975). The figure depicts the three
orthogonal bipolar dimensions, evaluation (E), activity (A) and potency (P), and gives examples of prototypical words for each
dimension and polarity. (Adapted from Chapman, 1979.)

Osgood’s principal dimensions are at the core of Fig. 2 for an illustration derived from the ratings of
other circumplex theories of affect (Lang, 1979; German word stimuli). The impact of valence and
Russel, 1980). For instance, Lang and colleagues arousal on various central nervous and peripheral
posit that human affective responses are determined physiological indicators of affective processing has
by the dimensions valence, arousal, and dominance; been repeatedly validated using picture and sound
again the first two having the largest impact, leading media (Lang et al., 1993; Bradley and Lang, 2000;
these authors to propose model of affect defined by Junghöfer et al., 2001; Keil et al., 2002).
the dimensions of arousal and valence. Within such a Word stimuli assessed for perceived arousal and
two-dimensional affective space, different classes of valence such as the ‘Affective Norms for English
stimuli such as pictures, sounds, and words cluster in Words’, ANEW (Bradley and Lang, 1998) have
a u-shaped manner. Highly arousing stimuli usually also been put to use in physiological research (see
receive valence ratings as either highly pleasant or e.g., Fischler and Bradley, this volume), although
highly unpleasant, low arousing material is generally so far the resulting evidence seems more restricted
regarded as more neutral with regard to valence (see than the one for pictorial material.
149

Fig. 2. Illustration of a two-dimensional affective space spanned by the dimensions arousal and valence. Examples of the relative
position of some German adjectives and nouns used in our studies are given in English translation. Along the x- and y-axis are depicted
the arousal and valence scales of the Self-Assessment Manikin (SAM, Bradley and Lang, 1994) used to rate individual emotional
responses to the words.

Emotions are generally viewed as culturally uni- cultural achievements, but during the acquisition
versal, largely innate evolutionary ‘old’ signaling of written language considerable regional special-
and activation systems, residing in ‘old’, subcorti- ization emerges in the human brain (Warrington
cal parts of the brain. They are designed to pro- and Shallice, 1979, 1980; Dehaene et al., 2005).
mote survival in critical situations, i.e., to signal Reading acts as a secondary process that utilizes
and activate fight, flight or feeding, attachment, the processing capabilities of the earlier acquired
and sexual behavior. auditory language system once the analysis of vis-
Reading and writing, by contrast, represent ual word form is completed (Perfetti, 1998; Everatt
comparatively recent developments in the history et al., 1999; Perfetti and Sandak, 2000). Language
of mankind, and in individual development these and emotion share a communicative function but
are acquired much later than oral language. Con- linguistic communicative functions are obviously
sequently, these skills are often regarded as not restricted to the communication of affect.
150

How do the ‘emotional brain’ and the ‘linguistic movement and the descriptive speech pattern, for
brain’ interact, when written words with emotional instance when a child observes or carries out a
connotations are encountered? Emotion theories gesture such as throwing and simultaneously hears
posit that linguistic expressions are stored within the caregiver say the respective word. Later, in the
semantic networks that encompass links to all as- acquisition of written language, this phonological
pects of their linguistic and pragmatic usages and code is mapped onto the visual word form (Perfetti
emotional connotations (Lang, 1979; Bower, 1981). and Sandak, 2000).
Thus, the word ‘gun’, for example, not only repre- For emotional concepts, Lang et al. (1993, 1994)
sents the object itself, but also includes links to its assume that not only associated semantic but also
operations, use, purposes, and their consequences physiological and motor response information is
as well as their emotional evaluation (Bower, 1981). coactivated in associative networks. Figure 3 illus-
A converging view is shared in neurolinguistics trates such a multilevel network representation of
(Pulvermüller, 1999) and cognitive semantics an emotional scene, encompassing a semantic code
(Barsalou, 1999): All information related to a word of the given situation as well as associated motor
is stored in a dynamic network. Recent evidence and physiological responses. Thus, the semantic
suggests that subnetworks1 representing different network that codes for ‘emotional semantics’
aspects of a word’s lexical representation can be could include the neuronal circuitry processing
separately and dynamically activated. For instance, the associated emotion (see also Cato and Cross-
differential neuromagnetic activations of semantic on, this volume, for a related suggestion).
subnetworks have recently been shown for sub- How does emotional content influence different
classes of plant or animal names (Assadollahi and stages of visual word processing? Here, the literature
Rockstroh, 2005). Moreover, biasing contextual is sparse. Event-related brain potentials (ERPs), the
constrains can affect the timing of access to the scalp recorded averaged synchronized activity of
dominant vs. subordinate meaning of homonyms several thousands cortical pyramidal cells, have suc-
(Sereno et al., 2003), challenging the modular view cessfully been used to delineate different stages of
that in word processing initially all lexical entries visual word processing (for reviews see e.g., Posner
have to be exhaustively accessed. et al., 1999; Tarkiainen et al., 1999). A closer look
Functional divisions of the semantic system also reveals that a considerable number of electro-
mirroring functional divisions in the organization physiological studies of emotional processing have
of the cortex have repeatedly been shown for verbs employed visually presented word-stimuli. Some,
denoting different types of actions (Pulvermüller et particularly early, studies have used the semantic
al., 2001b; Hauk et al., 2004). Investigating verbs differential as their theoretical vantage point. In fact,
pertaining to movements carried out with different in the wake of Osgood’s studies, the examination of
parts of the body, these authors demonstrate that ERP correlates of emotional semantics generated
the meaning of action words is reflected by the substantial research interest (Chapman et al., 1978,
correlated somatotopic activation of motor and 1980; Chapman, 1979; Skrandies, 1998; Skrandies
premotor cortex. These patterns of coactivations and Chiu, 2003). More recently, two-dimensional
presumably reflect individual learning history, valence  arousal models have been used as a
where the so-called referential meaning has been framework for research (see Fischler and Bradley,
acquired by repeated coactivation of the body this volume). However, many studies used words
with emotional connotations as experimentally con-
1
The terms sub-network or sub-representation as used here venient instances of a broader class of emotional
are not necessarily intended to imply a fixed hierarchical or- events (Anderson and Phelps, 2001; Dijksterhuis and
dering of the neural networks coding for different aspects of Aarts, 2003) or conversely, as a semantic class with-
semantics, although a certain degree of hierarchical ordering
out much reference to any particular theory of lan-
may indeed exist. Instead, sub-network or sub-representation
refers to the fact that different neural networks are likely to
guage and/or emotion (Begleiter and Platz, 1969). So
code for different aspects of a word’s meaning, such as ani- far, little systematic knowledge has been gathered on
macy, emotional connotation, grammatical gender, etc. the relationship between the emotion and language
151

Fig. 3. A network representation of a complex emotional scene (exam situation) illustrates how in dynamic emotional processing
perceptual, semantic and response systems are interactively linked. Activation on any level of this system can spread to other
subsystems. (Adapted after Lang, 1994.)

systems in visual word processing and possible im- to the available knowledge on ERP indices of
plications for the underlying neural implementation stages of visual word processing.
of meaning. First, we will address studies reporting early
ERP recordings have an excellent temporal res- effects of emotional content on ERP responses, oc-
olution, allowing for a fine-grained analysis of the curring within the first 300 ms after stimulus onset,
temporal sequence of different processing stages. separately for healthy volunteers and clinical pop-
Their spatial resolution is more restricted, and in- ulations; second, we will review effects of emotional
ferences from the spatial distribution of scalp content on the late ERPs to visual words; and third,
measured ERPs to their neural generators can only discuss how an interplay of subcortical and cortical
be made with caution. The number of electrodes mechanisms of emotion and word processing may
and the recording reference used influence the give rise to the observed effects. To facilitate com-
probability of finding effects, the generalizability parisons between studies, main methodological pa-
of these findings, and the accuracy of spatial lo- rameters and results concerning effects of emotional
calization (see also Junghöfer et al., this volume). content, as stated in the reviewed studies, are sum-
The present review will summarize and system- marized in two tables in the appendix. Table A1
atize existing studies on the role of emotion in describes studies with healthy volunteers; Table A2
visual word processing, both in healthy volunteers describes work with clinical populations. There, it
and in clinical populations and relate this evidence becomes immediately apparent that the studies
152

described vary considerably in theoretical approach recorded only from a single right occipital electrode
and methodology. (O2), but ERP differences due to emotional content
appeared in the ERP tracings already in the P1–N1
complex: twenty minimally above threshold pre-
Early effects — occurring within 300 ms after sented repetitions of each of two clearly negative
stimulus onset ‘four-letter –words’ led to larger responses than
twenty repetitions of each of the words ‘tile’ and
Healthy volunteers ‘page’. This pattern held for both a passive viewing
condition and a naming condition.
The extent to which early exogenous components of Several years later, Kostandov and Azurmanov
the human event-related potential are subject to (1977) contrasted ERP responses to both sublim-
modification by nonphysical stimulus characteristics inally and supraliminally presented ‘conflict’ and
is a matter of ongoing controversy. In visual word neutral words. In the subliminal condition, the
processing, a traditional view holds that within the earliest differences between ‘conflict’ and neutral
first 150–200 ms after a word has been presented, appeared around 200 ms after stimulus onset. Sub-
specific perceptual features of written words but no liminally presented ‘conflict words’ apparently re-
meaning-related attributes are extracted (Schendan lating to the subjects’ relationship conflicts caused
et al., 1998; Posner et al., 1999) and for many years by jealousy led to larger N200 responses than neu-
the N400 potential, a centro-parietal negativity aris- tral words. Supraliminally presented words led to
ing around 400 ms after stimulus onset has been later differentiations starting around 300 ms, in the
viewed as ‘the’ index of semantic processing (Kutas P3a window. Unfortunately, the description of the
and Federmeier, 2000). study is somewhat sparse on details regarding
Using other types of visually presented stimuli the materials used.
with emotional content, such as faces or pictures, Chapman et al. (1978) sought to electrophysio-
remarkable ERP differences between emotionally logically validate Osgood’s connotative dimen-
significant and neutral stimuli have been found sions of meaning (see Fig. 1), namely evaluation
within the first 300 ms after stimulus onset, some (E), potency (P), and activity (A). They recorded
even within the first 100 ms. from one midline electrode (CPz) referenced to
In his single subject study of ERP responses to linked mastoids while subjects had to name briefly
emotional words, Lifshitz (1966) failed to find a flashed (17 ms) words. Twenty words each from
visually impressive differentiation between emo- the extreme ends of the three connotative dimen-
tional and neutral words within the first 500 ms sions, E, P, and A, yielding six semantic classes
after word onset, although upon visual inspection (E+/-, P+/-, and A +/-) were used as stimuli.
the difference between the likewise presented erotic Random sequences of all these words were pre-
and neutral line drawings was sizeable. However, in sented 12–20 times to the subjects (until a set cri-
the meantime there is a considerable body of evi- terion of artifact-free trials was entered into the
dence indicating that even very early ERP responses average). Although the exact multivariate analysis
can diverge between emotional and neutral words. of the data in the original report is somewhat hard
Thus, the probably first quantitative study on the to reconstruct, in essence Chapman et al. (1978)
effects of emotional content on ERP indices of were able to statistically differentiate between all
word processing found differences between negative six connotative dimensions using ERP data from a
‘taboo words’ and neutral words already within single central channel. Whereas the different con-
the first 200 ms after word presentation (Begleiter notative dimensions are not mapped onto any
and Platz, 1969). Notably, this study was entitled: specific ERP components, in the original grand-
‘Cortical evoked potentials to semantic stimuli’, averages a second positive peak at around 260 ms
expressing a view of emotion as a vital part of after word onset is noticeable that clearly differ-
semantics. The technological standards at the entiates between the extreme ends (+ and -) of all
time were not very sophisticated and the authors three dimensions but not so much among E, P, and
153

A. In general, the material in this study was well (RSVP) design, i.e., a continuous stream of alter-
controlled for physical aspects but apparently not nating words without interstimulus interval. Skran-
for other linguistic attributes such as word fre- dies (1998) used a comparatively slow RSVP design
quency, word class, or concreteness. Replication presenting each word for 1 s. Ten stimuli per cate-
studies by the same group investigated the effects gory and polarity were selected and each stimulus
of explicit affective rating tasks on ERPs (Chap- was repeated 40 times, yielding 400 averages per
man, 1979) and directly compared the effects of category, thus optimizing the signal-to-noise ratio.
affective rating and word naming in one study, Subjects were instructed to visualize the words and
yielding similar results (Chapman et al., 1980). to remember them for a subsequent memory test in
Thus, a series of studies demonstrated reliable sta- order to ensure active engagement with the stimuli.
tistical differentiation between six connotative cat- EEG was recorded from 30 channels referenced to
egories within 500 ms after stimulus onset on the an average reference. Brain responses between the
basis of ERP data from a single channel. emotional word categories differed in six distinct
Extending their 1969 study across a larger selec- time windows, with respect to either their peak
tion of stimuli, Begleiter et al. (1979) recorded latency, associated global field power or the location
ERPs elicited by unpleasant, pleasant, and neutral of the centroids of the scalp distribution. Remark-
words, as 10 subjects had to either identify the last ably, most of the differences occurred within the
vowel in a presented word or give their personal first 300 ms after word presentation, starting with
affective evaluation of the word. Recordings were P1 at around 100 ms. These results were recently
made from three electrodes over each hemisphere, extended cross-culturally in a virtually identical
referenced to linked mastoids. However, results are study of affective meaning in Chinese, yielding a
reported only for two of these electrodes, namely P3 somewhat different but equally complex pattern of
and P4. Stimuli were the 62 most unpleasant, 62 differences depending on emotional word content,
most neutral, and 62 most pleasant five-letter words which were restricted to the first 300 ms after word
derived from a larger pool of words previously as- onset (Skrandies and Chiu, 2003). Together with
sessed with the semantic differential (Osgood et al., Begleiter and Platz (1969), Skrandies studies report
1957). Words were presented very briefly (20 ms). probably the earliest meaning-dependent differenti-
The words’ emotional meaning affected N1–P2 ation between words. Consequently, these studies
peak-to-peak amplitude during emotional evalua- are frequently cited in the visual word processing
tion. ERP responses to words evaluated as pleasant, literature, albeit without reference to the particular
neutral, or unpleasant could be distinguished sta- emotional semantic contents used.
tistically at both electrodes. Overall, effects of emo- Schapkin et al. (2000) recorded ERPs as subjects
tional content were somewhat more pronounced evaluated nouns as emotional or neutral. The stim-
over the left hemisphere and were restricted to the uli consisted of a total of 18 words, 6 pleasant, 6
affective evaluation condition. At the left hemi- unpleasant, and 6 neutral, which were matched
spheric electrode, ERPs were also generally larger across emotional categories for word length, fre-
when the words were shown in the emotional eval- quency, concreteness, and initial letters. Words
uation than in the letter-identification task. Unlike were presented peripherally, to the left and right
the previous studies, Begleiter et al. (1979) was the visual fields for 150 ms, while the EEG was re-
first to provide evidence for a major impact of task, corded from 14-linked mastoid referenced elec-
showing early ERP effects of emotional content trodes, eight of which were consistently analyzed.
only during active evaluation. Stimuli were repeated 32 times, 16 times in each
Skrandies (1998), studying ‘ERP correlates of se- visual field. The earliest effect of emotional signifi-
mantic meaning’, used an approach conceptually cance of the words on ERP responses was observed
similar to Chapman’s (1978, 1979, 1980). A pool of in the P2 window, peaking at 230 ms. At bilateral
60 nouns representing the bipolar extremes on Os- central sites, P2 responses to pleasant words were
good’s E, P, and A dimensions were selected and larger than responses to unpleasant and neutral
presented in a rapid serial visual presentation ones. While the P2 response was generally larger
154

over the left hemisphere, the emotion effect was not very briefly flashed (13 ms) stimuli. Subjects had to
lateralized. Similar effects of emotional content on indicate which of two simultaneously presented
ERP responses to words were also observed in later letter combinations in both visual fields consti-
time windows (see below). tuted an actual word. Stimuli were half neutral and
As already suggested by Kostandov and Arzuma- half emotional nouns of both pleasant and un-
nov (1977), the powerful effects of emotional con- pleasant valence. They were matched for word
notation on brain responses appear to extend even length and frequency and selected from a larger
below the limits of conscious perception. Bernat pool of words pre-rated on a bipolar seven-point
et al. (2001), recording from six-linked mastoid scale spanning the neutral-emotional continuum.
referenced scalp electrodes, report a differentiation Overall, emotional words presented in the right
between both subliminally and briefly (40 ms) pre- visual field were classified most accurately and
sented unpleasant and pleasant adjectives at left- fastest. However, the relative advantage for emo-
hemispheric electrode positions already in the P1 tional words was larger for words presented in the
and N1 time ranges, as participants simply main- left visual field. Using a source estimation ap-
tained fixation on a central cross and viewed the proach (LAURA) the authors identified a stable
computer screen without an explicit behavioral re- topographic pattern from the spatiotemporal dis-
sponse being required. In the P1 and N1 windows, tribution of the ERP data that accounted for the
unpleasant adjectives lead to larger ERP responses processing advantage of emotional words in the
in the left hemisphere. Overall larger responses to right visual field. This pattern emerged between
unpleasant as compared to pleasant adjectives were 100 and 140 ms after stimulus onset, i.e., mostly in
obtained in the subsequent P2, P3a, and LPC time the P1/N1 window. Curiously, it was localized to
ranges, the main effects of emotional content having primarily right-hemispheric extra-striate cortex.
earlier onsets in the subliminal condition. The affec- Surprisingly, no specific neurophysiological corre-
tive valence of the stimuli had been determined by late of the even more pronounced advantage for
assessing a larger pool of words on five bipolar emotional words in the left visual field was iden-
scales from the evaluative dimension of the semantic tified within the 250 ms after stimulus presentation
differential. Both the subsequent ERP study subjects that this study restricted its analysis to.
and an independent sample had repeatedly rated the Recent data from our own laboratory also pro-
stimuli. ERPs to 6 repetitions of the 10 most ex- duced evidence for early differences between cor-
tremely pleasant and unpleasant as well as to 12 tical responses to emotionally arousing (both
neutral words were recorded. Thus, study subjects pleasant and unpleasant) and neutral adjectives
had considerable experience with the words. Also, it and nouns. The word’s emotional content had
is unclear, whether the stimuli were assessed on been predetermined in a separate experiment, ob-
other potentially relevant emotional or linguistic di- taining valence and arousal ratings on two nine-
mensions such as arousal and dominance or word point rating scales (see Fig. 2) from 45 undergrad-
length, frequency, and abstractness. Still, these re- uate students. According to these ratings, highly
sults as well as data from Kostandov and Azurma- arousing pleasant and unpleasant and low arous-
nov (1977) or Silvert et al. (2004) and Naccache ing neutral words were selected. Different subsets
et al. (2005) in principle support the possibility of of these words were used in three studies where
measurable physiological responses to subliminally ERPs from 64 scalp sites were measured. Neutral
presented emotional words and add to the evidence and highly arousing pleasant and unpleasant
of emotional content-dependent P1 differences words matched for word length, frequency, and
(Begleiter, 1969; Skrandies, 1998; Skrandies and in one experiment also for concreteness were pre-
Chiu, 2003). sented in RSVP designs to subjects instructed to
Recently, Ortigue et al. (2004) also reported a read the words. Across three different stimulus
very early effect of word emotionality in a dense presentation durations (333, 666, 1000 ms) and re-
array ERP study recording from 123 scalp chan- gardless of word type (adjectives or nouns), a left-
nels. The task consisted of a lexical decision to hemispheric dominant occipitotemporal negativity
155

differentiated emotional (both pleasant and un- highly arousing pleasant and unpleasant as well as
pleasant) from neutral words. This negativity had neutral adjectives. The affective content of the
its maximum around 260 ms after stimulus onset. stimuli had been predetermined in a separate pop-
The influence of stimulus repetition was assessed, ulation using the above-described procedure.
but neither habituation nor sensitization was Words were presented for a relatively long period,
found for the emotional-neutral difference within namely 5 s. In this study the P2 component was the
the five repetitions used (Fig. 4 illustrates the effect first index of differential processing of emotional
for the 666 ms presentation rate). In one of the vs. neutral words. This P2 component primarily
studies we also manipulated task demands, in- responded to perceived stimulus intensity/arousal
structing subjects to attend to and count one of the and did not differentiate brain responses to pleas-
two word classes (adjective or noun). Interestingly, ant from those to unpleasant words. The same was
this manipulation did not affect the enhanced early true for the subsequent P3a component, but the
negativity to emotional words but had a significant picture changed for a later LPC component and
impact on the later positivity. the simultaneously recorded startle response that
Herbert et al. (2006) also recorded ERPs from were more pronounced for pleasant than for un-
64 average-reference linked scalp channels, as 26 pleasant and neutral words. This sequence of
subjects evaluated the emotional significance of effects is depicted in Fig. 5.

Fig. 4. Early arousal-driven enhancement of cortical responses to emotional words. Uninstructed reading of both pleasant and
unpleasant words in a rapid serial visual stimulation paradigm (RSVP, 666 ms stimulus duration) leads to larger occipitotemporal
negativities than reading of neutral words. The effect is illustrated at two occipital sensors (O9, O10) and the scalp topography of the
difference potential emotional–neutral words is depicted. Grand-averages from 16 subjects are shown.
156

Fig. 5. Difference maps of cortical activation for emotional minus neutral words in a covert evaluation task. Averaged activity in three
time windows is shown: P2 (180–280 ms), P3 (280–400 ms), and LPC (550–800 ms). For P2 and P3 both pleasant and unpleasant words
are associated with larger positivities than neutral ones. In the LPC window only processing of pleasant words diverges from neutral.
The time course of the activity is shown at electrode Pz. Grand-averages from 26 subjects are shown.

Clinical studies number of syllables, word frequency, and concrete-


ness but differed in emotional connotation. Non-
One of the first studies to use emotional words as a psychopathic subjects had faster reaction times and
tool to address processing biases (or a lack thereof) larger P2 responses to both pleasant and unpleasant
in clinical populations is Williamson et al. (1991) emotional words than to neutral ones. These differ-
who investigated behavioral and cortical responses ences induced by emotional arousal extended into
to pleasant, unpleasant, and neutral words in psy- the late positive component (LPC) time range in
chopathic and nonpsychopathic prisoners. Subjects controls but were completely absent in psychopaths.
had to detect words in a sequence consisting of Weinstein (1995) assessed correlates of enhanced
words and nonwords while their EEG was being re- processing of threatening and nonthreatening verbal
corded from five scalp positions referenced to linked information in university students with elevated or
mastoids. Stimuli were presented vertically for normal trait anxiety levels. Subjects read sentences
176 ms, separately to either visual field, and repeated with threatening and pleasant content that served as
three times. Stimuli had been matched for length, primes for subsequently presented threat related,
157

neutral, or pleasant words. ERPs in response to the words were likewise larger in the patients. In contrast
target words were assessed at Fz, Cz, and Pz, as to the first study, no differential effects of emotional
subjects had to decide whether the target word category were observed in subsequent time windows
contextually fit the previously shown sentence. (P3 and LPC). The ERP results are taken to reflect
Highly anxious subjects were reported to exhibit a heightened preconscious allocation of attention (N1)
larger frontal N1 in the threat-priming condition and and stimulus discrimination (N2) to disorder-related
an enhanced P400 (i.e., reduced N400, see below) in words in pain patients, but show no evidence of
a later time window. In retrospect, a number of further evaluative processing.
methodological problems seem to exist in this study In a similar vein, Pauli et al. (2005) studied cog-
or the data presentation may contain errors. For nitive biases in panic patients and healthy volunteers
instance, a large ERP offset at baseline is shown, the analyzing ERP responses from nine scalp channels
presented condition means occasionally do not to panic-related unpleasant and neutral words that
seem to correspond to the ERPs displayed and in- were presented, in separate runs, at individually de-
formation on linguistic properties of the stimuli termined perceptual thresholds and for 1000 ms.
used is missing. However, taken at face value, the Early ERPs differentiated panic patients from com-
results indicate heightened selective attention to (N1) parison subjects. At threshold presentation, patients
and facilitated semantic integration of (N400/P400) showed two enhanced early frontal positivities in re-
threat-related information in students with elevated sponse to panic words, one between 100 and 200 ms
trait anxiety levels. (P2) and the other (P3a) between 200 and 400 ms
A similar pattern of early ERP differences was post-stimulus onset. Both early effects were absent at
also found in a study investigating the processing the longer exposure duration and did not occur at all
of pain-, body-related, and neutral adjectives in in the control group. Interestingly, subsequent pos-
healthy volunteers and prechronic pain patients. itivities between 400 and 600 ms as well as between
Knost et al. (1997) report enhanced N1 responses 600 and 1000 ms differentiated between panic words
to the pain-related stimuli at a left frontal sensor from neutral words in both groups and for both
(F3) in the patient group. ERPs had been recorded presentation durations. This pattern of data resem-
from 11 mastoid linked scalp positions while sub- bles the results by Knost et al. (1997).
jects had to name words that were presented at the Kissler and colleagues (in preparation) assessed
individually determined perceptual threshold. In processing biases in depressed patients and com-
both groups, pain- and body-related words pro- parison subjects using the above-described RSVP
duced larger positivities than neutral ones in a paradigm, recording from 256 scalp electrodes and
later time window (600–800 ms) and were also as- comparing the amplitude and scalp distribution of
sociated with larger startle eye-blink responses on the previously described early negativity to pleas-
separately administered startle probe trials. The ant, unpleasant, and neutral adjectives matched
authors interpret their findings as evidence for for length and frequency. Around 250 ms after
preconsciously heightened attention to unpleasant, word onset (see Fig. 3), comparison subjects dis-
pain-related stimuli in the patients. played the above-described left-hemispheric dom-
These results were paralleled in an analogous inant enhanced negativity for emotional words,
study with chronic pain patients (Flor et al., 1997). pleasant and unpleasant alike. Depressed patients,
Chronic pain patients had a larger frontal N1 re- by contrast, exhibited this enhanced negativity
sponse to pain-related words than comparison sub- solely in response to the unpleasant words and
jects. Additionally, a general hemispheric asymmetry only in the right hemisphere.
emerged: N1 responses were larger over the right for
pain-related words and over the left side of the head Comparing early emotional and early semantic
for neutral words. The enhanced responses for pain- processing
related words in the patient group were also visible
in a centro-parietally maximal N2 component. In the In sum, numerous studies have found early
P2 window, right hemispheric responses to pain (o300 ms) amplifications of ERPs in response to
158

words with emotional content compared with neu- processing of the emotion dimension is required
tral words. The occurrence of such effects is re- by the task (Begleiter et al., 1979).
markable since controlled conscious processing Directly comparing the impact of pleasant vs.
has been suggested to arise only with the P3/N400 unpleasant word content yields mixed results, with
components (Halgren and Marinkovic, 1995), im- some studies finding larger early effects for pleas-
plying that emotion can affect preconscious stages ant words (Schapkin et al., 2000) and others larger
of word processing. effects of unpleasant ones (Bernat et al., 2001). As
Such effects appear to be more pronounced in mentioned above, the subjects’ clinical or motiva-
various clinical populations, reflecting heightened tional status may bias their cortical responses in
sensitivity and orienting to unpleasant, disorder- either direction. Also, task characteristics as well
related material (Weinstein, 1995; Flor et al., 1997; as the timing of stimulus presentation may have an
Knost et al., 1997; Pauli et al., 2005). Other additional impact but so far the influence of these
patients groups, by contrast, seem to selectively parameters is not well understood.
lack processing advantages for emotional words, Of note, some of the described effects occurred
psychopaths showing no cortical differentiation even before 200 ms after word onset, in a time range
between emotional and neutral words (Williamson in which from a traditional theoretical standpoint
et al., 1991) and depressed patients showing pref- meaning-related processing differences would not
erential processing of unpleasant but not pleasant be expected (Schendan et al., 1998; Cohen et al.,
words (Kissler et al., in preparation). At any rate, 2000).
processing biases in a number of clinical pop- These very early effects of emotional content on
ulations are reflected in their patterns of early ERP indices of visual word processing are rather
responses to pleasant, unpleasant, and neutral heterogeneous with regard to timing, locus, and
words. direction. Some of the inconsistencies are probably
A debated issue in emotion research pertains to related to differences in instrumentation and re-
whether, when, and how cortical responses differ cording methodology, number of electrodes, and
as a function of arousal, valence, and additional choice of reference electrode(s) representing but
factors such as dominance, or complex interac- the most obvious differences.
tions of these. Here, the data must remain some- The described studies also differ vastly in the
what inconclusive, as the studies discussed differed way emotional content of the stimulus material
vastly on the dimensions included and assessment was assessed as well as in the extent to which other,
methods used. Studies that assessed their materials nonemotional, linguistic factors such as word
with the semantic differential found that brain re- class, length, and frequency or concreteness were
sponses differentiate between all dimensions and controlled. Nevertheless, the bulk of the evidence
polarities within the first 300 ms after stimulus on- suggests that, indeed, under certain circumstances
set (Chapman et al., 1978, 1980; Skrandies, 1998; the emotional connotation of words can affect
Skrandies and Chiu, 2003). However, the arising even the earliest stages of preconscious sensory
pattern of results is so complex that it is hard to processing.
gauge the effect of each individual dimension on Thus, the challenge is to specify under which cir-
brain responses. The vast majority of studies re- cumstances such emotional modulation of earliest
port generally larger ERP responses to emotional processing may occur and what the underlying
than to neutral words, with some studies reporting mechanisms are. Two experimental factors arise
these effects even in the absence of a task that from the reviewed studies that may contribute to
would explicitly require processing of emotional the emergence of very early emotion effects. First,
content or other types of semantic access (Begleiter very brief stimulus presentation, near or even below
and Platz, 1969; Bernat et al., 2001; Kissler et al., the perceptual threshold (Begleiter and Platz, 1969;
submitted manuscript). However, occasionally, Kostandov and Arzumanov, 1977; Chapman et al.,
early emotion effects in word processing were 1978, 1980; Flor et al., 1997; Knost et al., 1997;
found restricted to situations where explicit Bernat et al., 2001; Ortigue et al., 2004; Pauli et al.,
159

2005) and second, repeated presentation of com- experiments and functional neuroimaging indeed
paratively small stimulus sets (Begleiter and Platz, reveals the existence of such subcortical ‘short-cut’
1969; Chapman et al., 1978, 1980; Skrandies, 1998; routes in emotional processing, particularly of
Skrandies and Chiu, 2003; Ortigue et al., 2004). fear-relevant stimuli (Davis, 1992; LeDoux, 1995;
None of the cited studies have explicitly assessed Morris et al., 1999). Direct pathways from the su-
the effect of stimulus repetition on the latency of perior colliculi and the thalamus to the amygdala
emotional-neutral ERP differences. Our own stud- and the cortex allow for the automatic processing
ies of repetition effects on negative difference waves of relevant stimuli outside of conscious awareness,
distinguishing emotional from neutral content preparing rapid behavioral responses. In humans,
around 250 ms after stimulus onset show no evi- this subcortical pathway has been mapped by
dence of change within five repetitions. However, functional neuroimaging during fear conditioning
some of the cited studies used by far more than five of subliminally presented faces (Morris et al.,
stimulus repetitions and studies of early semantic 1999) as well as during the subliminal presentation
processing indeed suggest an effect of stimulus rep- of faces with fearful expressions (Liddell et al.,
etition on the timing on meaning-related differences 2005). On a cortical level, its activity may be re-
in cortical activity: Pulvermüller and colleagues flected in transient early responses. Brief, sublim-
report neurophysiological evidence of differences in inal stimulation with fearful faces has recently
semantic processing from 100 ms post word onset been shown to result in a transient enhancement of
(Pulvermüller et al., 2001a). They used a task in the N2 and early P3a components, which, how-
which a single subject was repeatedly, over several ever, did not continue in later N4/P3b/LPC win-
days, presented with a set of 16 words that she had dows. For supraliminal stimulation, conversely,
to monitor and hold active memory as responses to N4/P3/LPC but not N2 components responded to
occasionally presented new words were required. emotional content (Liddell et al., 2004), suggesting
Thus, in above threshold presentation, preactiva- the operation of a slower, conscious processing
tion of the cortical networks coding for meaning by and evaluation route. Conceivably, subliminal
using tasks that require continuous attention to and stimuli receive a temporally limited amount of
working memory engagement with the stimuli as processing that wanes if it is not confirmed by
well as use of many repetitions may foster earliest further supraliminal input, much like in the case of
semantic processing differences, nonemotional and subliminal priming (Greenwald et al., 1996; Kiefer
emotional alike. and Spitzer, 2000). Recording ERPs during sub-
Further, a recent study on repetition effects in liminal and supraliminal semantic priming, Kiefer
symbol processing found an increase in N1 ampli- and Spitzer observe decay of subliminal semantic
tude (around 150 ms) across three repetitions of activation within 200 ms, a delay at which suprali-
initially unfamiliar symbol strings (Brem et al., minal priming effects can still be robustly demon-
2005), supporting the view that stimulus repetition strated. A plastic, maladaptive downregulation of
can amplify early cortical responses to word-like subcortical excitability may account for early re-
stimuli. Thus, repetition effects affecting emo- sponsiveness to unpleasant and disorder-related
tional stimuli more than neutral ones as a conse- words in clinical populations (see e.g. Pauli et al.,
quence of differential initial capture of attention 2005 for supportive evidence).
and rapid perceptual learning may account for Clearly, at present the operation of a fast sub-
some of the very early ERP effects in emotional cortical route from the thalamus and the amygdala
word processing. in emotional word processing that could account
Early effects of emotional content on brain re- for near or subthreshold emotion effects in visual
sponses to subliminally or near-subliminally pre- word processing remains a speculative conjecture.
sented stimuli have occasionally been accounted A most critical point is that such a mechanism
for by fast, subcortical short-cut routes (see Wiens, would require at least basic ‘reading abilities’ in
this volume, for a discussion of issues of sublim- the thalamus. While the case for stimuli such as
inal stimulus presentation). Evidence from animal faces or threatening scenes that by some are
160

assumed to be part of our ‘evolved fear module’ areas between 100 and 150 ms after word onset,
(Öhman and Mineka, 2001) can be made much whereas activation differentiating between subordi-
more easily, many would have a hard time believ- nate categories was evident only from 300 ms on.
ing in rather sophisticated subcortical visual ca- Dehaene (1995) observed the earliest ERP differ-
pacities allowing for the discrimination of written ences between words of different categories (verbs,
words. On the other hand, subcortical structures proper names, animals), 250–280 ms after word on-
are also subject to modifications by learning, and set. Semantic category differences were reflected in
by the time people take part in experiments they the scalp distribution of a left occipitotemporal
will usually have had about two decades of reading negativity. Using RSVP designs a similar occipito-
expertise. temporal negativity has been identified. This nega-
So far, most of the evidence for the subliminal tivity has been termed the ‘recognition potential’
processing of emotional stimuli is based on studies (RP). It is sensitive to semantic aspects of visual
with aversive material. Accordingly, the above- word processing and has its maximum around
reviewed studies evidence extremely early effects 250 ms after word (Rudell, 1992; Martin-Loeches et
primarily for unpleasant words (Flor et al., 1997; al., 2001; Hinojosa et al., 2004). The ‘RP’ responds
Knost et al., 1997; Bernat et al., 2001). to manipulations of depth of semantic analysis, its
An alternative explanation of some of these amplitude increasing with the meaningfulness and
early effects of enhancement by emotional content task-relevance of the presented word. Source anal-
in visual word processing that would not rely on ysis has placed the origin of the RP in the fusiform
subcortical by-pass routes and therefore on sub- gyrus (Hinojosa et al., 2001). Results from our lab-
cortical vision is reentrant connections between the oratory are consistent with the view that a word’s
so-called visual word form area (VWFA) and the emotional connotation enhances the associated rec-
emotion processing system. ognition potential (see Fig. 4).
During visual word recognition the earliest ac- Thus, a word’s emotional connotation could be
tivation of an invariant version of the visual word directly connected to the abstract representation of
form (i.e. the font-, size-, position-invariant repre- its visual form. Moreover, the combined evidence
sentation of the string of letters) occurs from about suggests that emotional content amplifies early
100 ms after stimulus onset (Sereno et al., 1998; stages of semantic analysis in much the same way
Assadollahi and Pulvermuller, 2003). Form invar- an instructed attention enhancing processing task
iant, abstract representations of highly over- would.
learned visual objects such as words and faces If enhanced semantic processing is an important
have been found to originate in the fusiform gyrus mechanism by which emotional content affects
(Haxby et al., 1994; Chao et al., 1999; Cohen et al., visual word processing, again, the question arises
2000; Dehaene et al., 2002). Electrophysiological as to the causative mechanism: back-projections
evidence with regard to the onset of word-specific from the anterior cingulate and the amygdala may
effects of fusiform activity varies, with some au- give rise to such processing enhancements. In sup-
thors reporting onsets around 120 ms (Tarkiainen port, amygdala lesions impair the enhanced detec-
et al., 1999; Assadollahi and Pulvermüller, 2001) tion of unpleasant words in an RSVP attentional
and others somewhat later around 170 ms (Bentin blink paradigm but not of identification enhance-
et al., 1999; Cohen et al., 2000). Timing differences ments caused by manipulation of target color
may be partly attributable to differences in word (Anderson and Phelps, 2001). Thus, assuming that
familiarity across experiments (King and Kutas, the amygdala plays a pivotal role in the preferen-
1998). tial processing of emotional words as recently sug-
Immediately after access of the visual word form, gested by several neuroimaging and lesion studies
meaning can be activated: Assadollahi and Rocks- (Isenberg et al., 1999; Anderson and Phelps, 2001;
troh (2005) showed that activation differences due Garavan et al., 2001; Hamann and Mao, 2002;
to super-ordinate categorical differences (animals Naccache et al., 2005), an alternative model to the
vs. plants) can be found in left occipitotemporal above described thalamo-amygdalo-cortical route
161

could account for most of the data. Emotional Chapman et al. (1978, 1979, 1980; see above) a
amplification of semantic processing would occur positivity occurring around 300 ms is discernible
after initial stimulus identification, caused by bi- and appears to be primarily related to the potency
directional reentrant communication between cor- dimension extracted from their data.
tical regions and the amygdale (Amaral et al., Using materials from the Begleiter et al. (1969,
2003). Crucially, cortical analysis would precede 1979) and Chapman et al. (1978, 1979) studies,
and spark subcortical amplification of cortical Vanderploeg et al. (1987) assessed ERP responses
processing. Clearly, a theoretically crucial priority to visually presented emotional (20 pleasant, 20
for future research is to determine the timing of unpleasant) and 20 neutral words and face draw-
subcortical mechanisms in relation to cortical en- ings (two per emotion category), which were eval-
hancement of ERP responses to emotional words. uated during viewing. The EEG was recorded from
Unlike for other semantic categories such as six electrodes referenced to linked ears in 10 male
movement-related verbs (Pulvermüller et al., 2000, subjects. During viewing, the visual stimuli were
2001b), ERP data for emotional words so far sug- presented for either 80 ms (words) or 100 ms (faces).
gest little consistent emotion-specific change in to- In the conditioning phase, the face drawings were
pography (but see Skrandies, 1998; Skrandies and shown for 1500 ms. For both faces and words
Chui, 2003; Ortigue, 2004). A distinct, emotion-as- clearly discernible emotion-category dependent
sociated topography might point to the existence of differences in ERP tracings appear from around
a homogeneous emotion lexicon localizable in dis- 300 ms after stimulus onset as parietal positivities.
tinct neuronal populations of the brain as has been A small but significant effect of emotional conno-
suggested for other semantic categories (Martin et tation of words but not faces on the spatial distri-
al., 1996). Rather, emotional content seems to am- bution of the ERP was also evident in the P2
plify cortical word processing, much in the same window. Interestingly, although sizeable in appear-
way as it enhances picture (Junghöfer et al., 2001) ance, the P3 effect of emotional connotation did not
or face processing (Schupp et al., 2004). However, reach statistical significance for words. For a later
functional neuroimaging techniques with better positivity (positive slow wave/late positive complex)
spatial resolution of especially deep cortical and a similar result was obtained; although visible in the
subcortical structures (see Cato and Crosson, this presented grand-averages, the difference in parietal
volume) and the more consistent use of dense array positivity between emotional, pleasant and unpleas-
EEG studies (Ortigue et al., 2004) may provide ad- ant, and neutral words does not reach significance
ditional information. in an analysis of the corresponding PCA factors
while it does for faces. The authors, in line with
Lifshitz’ (1966) early finding, suggest that words
Late components (after 300 ms) may be less powerful (or more heterogene-
ously evaluated) emotional stimuli than pictures
Healthy volunteers (Vanderploeg et al., 1987). Thus, ERPs from10
subjects may not yield enough statistical power
In relation to traditional stages of visual word to assess connotation-dependent ERP differences,
processing, effects occurring later than 300 ms particularly in studies using comparatively sparse
after word onset are less puzzling than the previ- electrode arrays. Also, the perceptual variance be-
ously discussed early ones. Enhanced late positivi- tween 20 words of a category may be higher than
ties in response to emotional word content have among two faces. Differential effects may, there-
most often, but not invariably, been found. In fore, also result from the greater consistency or
several of the already discussed studies reporting higher frequency of occurrence of the faces.
early ERP modulations as a function of emotional Indeed, subsequent studies have found robust
content, later enhanced positivities associated with effects of emotional connotation on later ERP
the emotional content of the word stimuli are also components. For instance, Naumann et al. (1992)
apparent. For instance, in the data shown by investigated late positive potentials to adjectives
162

varying in emotional content. Their key idea was was more pronounced for the emotional than for
that using ERPs it should be possible to dissociate the neutral adjectives. Thus, overall, an emphasis
emotional and cognitive processing and that, fol- on emotional processing (affective task) caused an
lowing LeDoux (1989), cognitive and emotional anterior shift of the scalp distribution. Further-
processing systems should be functionally and more, regardless of task, emotional stimuli led to
neuronally separable as reflected in distinct ERP more pronounced parietal peaks than neutral
scalp topographies. In an initial experiment, 30 ones. Again, the authors interpreted their results
prerated pleasant, unpleasant, and neutral adjec- as evidence for a functional and structural distinc-
tives were presented to 14 subjects who had to tiveness of affective and cognitive functions in the
either evaluate the words as pleasant, unpleasant, human brain as suggested by LeDoux (1989).
or neutral (affective task) or determine whether a However, a third demonstration of dissociable
word’s length was longer, shorter, or equaled six affective and cognitive processes in visual word
letters (structural task). The EEG was recorded processing failed. Naumann et al. (1997) examined
from three midline electrodes (Fz, Cz, and Pz), a sample of 54 students in three different tasks,
which were referenced to the left mastoid. ERPs namely letter search (structural task), concrete–ab-
were assessed between 300 and 700 ms after word stract decision (semantic task), and an unpleas-
presentation for the P3 component and between ant–neutral decision (affective task) on a set of
700 and 1200 ms for the later positive slow wave. nouns varying in emotional content. Fifty-six
For both components and all word categories, nouns were used that could be divided into sub-
ERP amplitudes were more positive going for the sets of seven words unambiguously belonging to
affective than for the structural task, particularly one of eight possible combinations of these at-
at electrodes Fz and Cz. Moreover, P3 amplitudes tributes. ERPs were now recorded from nine scalp
were also generally more positive in response to locations (Fz, Cz, Pz, and adjacent left and right
emotional than neutral adjectives. The spatial dis- parallels). There was indeed a considerably larger
tribution of the subsequent slow wave component P3 for unpleasant compared to neutral words, al-
varied with emotional category, displaying larger beit the effect was not general but restricted to the
amplitudes at Pz than at Cz and Fz for pleasant affective task. This casts doubt on the assumption
and unpleasant words but having equal amplitudes that cognitive and emotional word processing op-
at all three electrodes for the neutral words. This erate along completely separable routes and raises
pattern was taken as evidence for the hypothesized the question to what extent larger late positive
separateness of affective and cognitive processing potentials to emotional word stimuli occur outside
systems. the attentional focus. Given the early effects
Naumann et al. (1992) replicated this result in a (o300 ms) reported above, it would have been in-
second experiment having calculated an ideal sam- teresting to analyze the data with a focus on early
ple size of 106 subjects, minimizing the likelihood and possibly automatic impacts of emotion on
of false-negative results. In the replication study a word processing.
between groups design was used, assigning 53 sub- Naumann et al. (1997) raise a number of con-
jects each to either the structural or the affective ceivable reasons for the reduction of the effect,
task. Again, for both components and all word favoring a familiarity-based explanation. In the
types, more positive frontal ERPs were obtained experiments that had yielded ‘uninstructed’ and
for the affective than for the structural task. The topographically distinct late responses, subjects
P3 component was larger for both pleasant and had been familiar with the stimuli beforehand.
unpleasant than for neutral adjectives. Further- Thus, the affective differences between the stimuli
more, positivities in response to emotional adjec- may have already attracted the participants’
tives were particularly pronounced at Cz and Pz, attention. Moreover, the new design reduced the
and this gradient was most evident for the pleasant probability of occurrence for an emotional word,
words. For the slow wave, the scalp distribution possibly making this dimension less salient;
likewise exhibited a parietal maximum and this although the converse hypothesis, based on an
163

oddball effect, would be equally plausible. More- words following a sentence context. Students with
over, in their initial studies, Naumann et al. (1992) high-trait anxiety levels had a reduced N400 (or, in
had used adjectives that may produce somewhat Weinstein’s terminology enhanced P400) to words
different effects, given that ERP differences be- following a threatening sentence context, indicat-
tween word classes have been reported (e.g. Fed- ing facilitated integration of information within
ermeier et al., 2000; Kellenbach et al., 2002). threatening contexts. An alternative interpretation
Fischler and Bradley (this volume) report on a might suggest enhanced sustained attention to
series of well-controlled studies where effects of threatening information in highly anxious subjects,
word emotionality on late positivities are consist- if the potential described were taken to resemble a
ently found for both pleasantly and unpleasantly P3/LPC component, which is not entirely clear on
arousing words when the task requires semantic the basis of the presented data.
processing of the presented words but not otherwise. Personality-dependent changes in late cortical
Some studies also show larger late positive po- responses to emotional words have been subse-
tential effects restricted to pleasant words (Schap- quently replicated: Kiehl et al. (1999) tried to ex-
kin et al, 2000; Herbert et al., 2006), which were tend Williamson et al.’s (1991, see above) results of
missing in Naumann’s 1997 study. deficient early (P2) and late (LPC) ERP responses
For instance Schapkin et al. (2000, see above) re- to emotionally charged words in psychopathic
port larger P3 and late positive slow wave responses subjects. They assessed similarities and differences
to pleasant compared to both neutral and unpleas- in the processing of abstract–concrete vs. pleas-
ant words during evaluative decision. As mentioned ant–unpleasant words in psychopaths and com-
above, Herbert et al. (2006) report a study where parison subjects. To address the processing of
early ERP responses (P2, P3a) reflected the arousal emotional words a pleasant–unpleasant decision
dimension of the words, differentiating both pleas- task was used; although the initial study had not
ant and unpleasant from neutral words. The later revealed any valence differences. The altered task
LPC, however, differentiated pleasant from un- was apparently motivated by clinical observations
pleasant stimuli and was larger for pleasant words suggesting that psychopaths have difficulty in un-
(see Fig. 5). Bernat et al. (2001), on the other hand, derstanding abstract information and in distin-
report enhanced responses to unpleasant as com- guishing pleasant from unpleasant valence. Stimuli
pared to pleasant words across the entire analysis were controlled for word length and frequency,
window, until 1000 ms after word onset encompass- syllable number, and concreteness. Word presen-
ing P3 and LPC. tation was extended to 300 ms and words were
Schapkin et al. (2000) additionally assessed late presented only once, centrally and in a horizontal
negativities that were labeled N3 (around 550 ms) format. EEG was recorded from nine scalp posi-
and N4 (around 750 ms), finding no effect of emo- tions, again with a linked mastoids reference.
tional content. Data from our own studies do Analyses now focused on a 300–400 ms post-stim-
show a small effect of emotional content on N4 ulus window and a LPC window (400–800 ms).
amplitudes with larger N4 to neutral than to emo- Behaviorally, in both groups responses to pleasant
tional words possibly reflecting a contextual ex- words were faster and more accurate than those to
pectancy for an emotional content caused by unpleasant ones. Cortically, an N350 component
unequal stimulus probabilities. In both studies differentiated between pleasant and unpleasant
(Schapkin et al, 2000; Kissler et al., submitted), words but not between psychopaths and nonpsy-
two-thirds stimuli had emotional content (pleasant chopaths, being across groups larger for the pleas-
or unpleasant), only one-third was neutral. ant words. Moreover, the valence differentiation
was more pronounced over the left hemisphere. In
Late components — clinical studies the later time window (400–800 ms), unpleasant
words elicited more positive going brain waves
Weinstein (1995) is one of the few reports of a than pleasant ones. This left-hemispheric domi-
modulation of integration of emotionally charged nant differentiation was absent in psychopaths. In
164

effect, ERPs to unpleasant words were more pos- pictures (Keil et al., 2002), most researchers fo-
itive than ERPs to pleasant words across both time cused on an analysis of late positivities. The com-
windows, and the differentiation was reduced in parative paucity of reports on N400 modulation
psychopaths. It is unclear how the ERP patterns by emotional word content may partly reflect a
relate to the behavioral data (both groups were bias on the part of the investigators and appear
faster and more accurate for pleasant). But more surprising in view of the fact that N400 is often
positive-going late potentials for unpleasant stim- regarded as ‘the electrophysiological indicator’ of
uli in a binary pleasant–unpleasant decision are in semantic processes in the brain. On the other
line with data from Bernat et al. (2001). During a hand, it is becoming increasingly clear that the
lexical decision task, Williamson et al. (1991) re- N400 response does not index lexical access or se-
ported a larger LPC to emotional than to neutral mantic processing per se but reflects semantic in-
words in nonpsychopathic subjects and to a lesser tegration within a larger context, created by either
degree in psychopaths, but no differentiation bet- expectations on sentence content or other contex-
ween the pleasant and unpleasant words. Schapkin tual constraints within experiments (Kutas and
et al. (2000) and Herbert et al. (2006), by contrast, Federmeier, 2000). Thus, it is reasonable to as-
report larger late positivities for pleasant in com- sume that single-word studies will only result in
parison to both neutral and unpleasant words. N400 modulations if strong expectations on emo-
Note, that neither the Kiehl et al. (1999) nor the tional word content are established. Priming stud-
Bernat et al. (2001) studies report data on neutral ies or experiments establishing an emotional
stimuli. expectation within a sentence context may pro-
vide a better testing ground for the issue of N400
modulations by emotional word content. Indeed,
Comparing late emotional and late semantic word Weinstein (1995) followed this rationale establish-
processing ing emotional expectations on a sentence level.
Recent work from our laboratory also shows N400
A considerable number of studies have found am- modulation by emotional content in a lexical de-
plifying effects of emotional word content on cision task where an emotional expectation (pleas-
electrophysiological cortical activity later than ant or unpleasant) was established by a preceding
300 ms after word onset. In contrast to the very emotional picture. Of note, the pictures were of
early effects, they occur in a time range where similar emotional connotation as the subsequent
modulation of cortical responses by word meaning adjectives but the words were not descriptive of the
is not unusual in itself. By 300–400 ms after stim- picture content (Kissler and Kössler, in prepara-
ulus onset ERP tracings reflect conscious process- tion). A transient mood induction may have me-
ing stages (Halgren et al., 1994a, b) and clearly diated the effect; recently, effects of subjects’
vary with semantic expectancy (Kutas and Hill- emotional states on semantic processing have been
yard, 1980, 1984), task relevance (Sutton et al., reported (Federmeier et al., 2001). When subjects
1967), or depth of mental engagement (Dien et al., were in a mildly positive mood, their semantic
2004). Thus, it is not surprising that ERPs in this processing was facilitated as reflected by a smaller
time range can reflect processing differences bet- N400 potential to more distant members of given
ween words of different emotional content. From categories than when in a neutral mood. Thus, a
a semantics perspective, N400 might represent an number of studies suggest that both a word’s
appropriate ‘classical’ ERP component to assess emotional content and a subject’s emotional
for emotion effects. Indeed, some studies have state may affect the N400 ERP response (but see
found modulations of N400- or N400-like ERP Fischler and Bradley, this volume).
responses to words of emotional categories (Will- Still, so far the most consistently reported later
iamson et al., 1991; Weinstein, 1995; Kiehl et al., effects of emotional word categories on the ERP
1999). However, in line with ERP studies of affec- are seen in broadly distributed late positivities with
tive processing of faces (Schupp et al., 2004) and a parietal maximum (see also Fischler and Bradley,
165

this volume). Such late positivities have generally In emotion research, larger late positivities have
not been associated with specific aspects of seman- consistently been shown during free viewing of
tic processing but rather with task demands such emotional vs. neutral pictures (Keil et al., 2002)
as attentional capture, evaluation, or memory en- and faces (Schupp et al., 2004). If primary tasks
coding. In neurolinguistics, late positivities have distract participants from the emotional content of
repeatedly been suggested to index syntactic reanal- the visual stimuli, late positivities to emotional
ysis following morphosyntactic violations (Osterh- stimuli are often diminished reflecting competition
out et al., 1994; Friederici et al., 1996; Hagoort and for attentional resources. The degree to which and
Brown, 2000). Yet, some studies also report mod- circumstances under which emotion and attention
ulations of late positivities by semantic attributes of compete for resources, have additive effects, or
language. For instance, in antonym processing operate in parallel is a matter of ongoing debate
differential P3 and LPC responses were found, de- (see Schupp, et al., this volume, for a discussion).
pending on whether a word contained a given at- For visually presented word stimuli the picture
tribute or lacked it (Molfese, 1985). Contextual is similar; when the primary task requires an eval-
semantic constraints and stimulus abstractness have uative emotional decision (pleasant–unpleas-
also been reported to affect late positivites (Hol- ant–neutral, emotional–neutral) emotional words,
comb et al., 1999). Both contextually expected and like pictures or faces, are consistently associated
unexpected sentence-final words were associated with larger late positivities than neutral ones.
with larger positivities than contextually uncon- When the primary task requires structural stimulus
strained words, the effect being even more pro- processing the evidence is mixed, with some studies
nounced when the abstract words were contextually still finding larger positivities in response to emo-
unexpected. Münte et al. (1998) also find late pos- tional stimuli (Naumann et al., 1992) while others
itive responses to language semantics, thereby chal- do not (Naumann, 1997). During free viewing, a
lenging the account of specific morphosyntactic late recent study (Kissler et al., submitted manuscript)
positive shifts and corroborating the view that late finds a larger LPC to emotional words, suggesting
positivities reflect mental engagement and effortful that when subjects are free to allocate their
processing across a wide range of higher cognitive processing resources as they wish, they process
functions (Dien et al., 2004). emotional words more deeply than nonemotional
Late positivities are likely to share a proportion ones. Our results also indicate that late responses,
of neural generators and differ on others, reflecting around 500 ms, may be more affected by explicit
the extent to which the tasks that elicit them share attentional tasks than the simultaneously observed
or draw on different neural systems. Thus, topo- early effects around 250 ms. During lexical deci-
graphically distinct late positive shifts may relate sion (Wiliamson et al., 1991) and naming tasks
to different aspects of cognitive and emotional (Knost et al., 1997; Pauli et al., 2005), emotional
functioning, as suggested for instance by Nau- words have also been found to be associated with
mann et al. (1992). However, in order to unam- larger late positivities (but see Fischler and Brad-
biguously elucidate topographic changes that ley, this volume). Thus, when the task allows for or
reflect shifts in neural generator structure, simul- even requires semantic processing, emotional
taneous recordings from dozens of electrodes and words are processed more deeply than neutral
advanced data analysis techniques are necessary ones. When the task requires structural processing
(see also Junghöfer and colleagues, this volume). this processing advantage is considerably dimin-
From the extant studies on the emotional modu- ished (Naumann et al., 1997).
lation of late components in word processing, it is Clearly, the extent to which larger late positivities
hard to gauge the extent to which emotion induces to emotionally relevant words are driven by arousal,
genuine topographic changes indicative of the re- valence, or additional subject-, task-, or situation-
cruitment of additional distinct cortical structures specific factors is not quite settled. The matter is
or purely amplifies the activity of a unitary complicated by the fact that studies differed on the
processing system. instruments used to assess emotional word content
166

and the extent to which the pleasant and unpleasant that the approach system responds more strongly
dimension were differentiated or collapsed into one to relatively little input. The withdrawal system in
‘emotional’ category. A fair number of studies em- response to unpleasant input, in turn, is activated
ployed the empirically well-founded semantic differ- comparatively more at high levels of arousal,
ential technique to assess emotional content, or the this latter process being termed ‘negativity bias’
two-dimensional arousal  valence space, yet others (Caccioppo, 2000; Ito and Caccioppo, 2000). Vis-
do not even report the criteria by which the emo- ually presented words are likely to constitute less-
tional content of the material has been determined. arousing stimuli than complex colored pictures,
Although multidimensional models of affect are i.e., the word ‘cruel’ will be less arousing than a
empirically well founded and the use of numerical photograph of a corresponding scene, even if both
rating scales allows for the rapid assessment of stimuli receive comparable ratings. Therefore, in
large numbers of stimuli on many dimensions, an the absence of strong unpleasant personal
inherent problem with Likert-type scaling remains. associations for a given word, which may well be
Such scaling techniques assume that subjects will present in various clinical populations (see above),
meaningfully assign numbers to psychological a ‘positivity offset’ for written verbal material
stimuli, such that the quantitative relationships might be expected. Corresponding data are re-
between the numbers will correctly reflect the psy- ported, for instance, by Schapkin et al. (2000) or
chologically perceived relationships among the Herbert et al. (2006).
stimuli, including conservation of distance or con- Like for the early effects, the question arises how
servation of ratio, yielding interval or even ratio late effects of emotional word content on ERPs
scales. But these assumptions do not always hold, come about; subcortical activity has again been im-
such that it is unclear whether the psychological plicated. Nacchache et al. (2005) have recently for
distance between stimuli rated 2 and 4 on a given the first time recorded directly from the amygdala
scale is really the same as between stimuli rated 6 field potentials in response to emotional words.
and 8 (Luce and Suppes, 1965; Kissler and Bäuml, Three epilepsy patients with depth electrodes im-
2000; Wickelmaier and Schmid, 2004). Moreover, planted for presurgical evaluation performed an
the relationship between behavioral ratings and evaluative decision task (threatening–non-threaten-
physiological impact of emotional stimuli is likely ing) on a series of threat or nonthreat words pre-
to be nonlinear. sented subliminally or supraliminally. In all three
Nevertheless, the bulk of the data corroborates patients, larger amygdala potentials to threat than
the view that during earlier stages of processing, to nonthreat words could be identified around
emotion acts as a nonvalence-specific, arousal- 800 ms after word presentation in the subliminal
driven alerting system (see above). During later and around 500–600 ms in the supraliminal condi-
stages of processing (4300 ms), the patterns found tion. The study is pivotal in that it both directly
are more varied and may reflect flexible adapta- measures amygdala activity during emotional word
tions to contextual factors. In support, Herbert processing and provides clues as to the timing of this
et al. (2006) recently found arousal-driven ampli- activity. As detailed before, subcortical, primarily
fication of cortical responses to both pleasant amygdala activity may be a source of cortical am-
and unpleasant words within the first 300 ms and a plifying mechanisms in response to emotional stim-
divergent pattern that favors the processing uli visible in ERPs. Amygdala activity measured by
of pleasant material in a later time window (see depth electrodes around 600 ms after stimulus onset
Fig. 5). Keil (this volume) discusses a number of may provide a basis for LPC amplifications evident
task factors that contribute to processing advan- in the surface ERP. However, the timing of the re-
tages for pleasant or unpleasant material in turn. sponses poses new puzzles. If amygdala activity in
For language material with emotional content, a emotional word processing onsets around 600 ms,
general pleasant–unpleasant asymmetry in emo- how are early effects of emotional word content
tional processing may be important: At low levels generated (see discussion above)? Amplified cortical
of arousal, a ‘positivity offset’ is often found in ERP responses reflect the activation of larger
167

patches of cortex, indicating spread of activation in growing body of evidence demonstrates that some
a more densely packed neural network. These result aspects of word meaning must be active before the
from life-long associative learning mechanisms. The N400 is elicited.
effects of emotional learning can be seen in ampli- We propose that there is no fixed point in time
fied cortical ERP tracings whenever the correspond- where all semantic information is equally availa-
ing semantic network is accessed. Subcortical ble. Instead, subrepresentations can become acti-
mechanisms might be active primarily in the acqui- vated in a dynamic, possibly cascaded manner.
sition of emotional semantics, reflecting the role of Variations in timing of semantic signatures could
the amygdala in emotional learning even of abstract be interpreted in the light of an internal structure
representations (Phelps et al., 2001). Their impact of the entry of a mental lexicon. Different aspects
may be attenuated once a representation has been of semantics could be flexibly prioritized, depend-
acquired. Clearly, elucidating the mechanisms by ing on context, task, and motivation of the subject.
which amplified responses to emotional words are If the tenet of a monolithic lexical entry is given
generated is a vital issue for future research. up, the internal structure of a word’s lexical rep-
In sum, a considerable number of studies show resentation can be assessed by investigating the
enhanced late positive responses when people timing by which each subrepresentation is availa-
process emotionally laden words, pleasant and ble (Assadollahi and Rockstroh, 2005) or the con-
unpleasant alike. The responses are not as large as textual constraints that lead to the activation of a
for pictorial or face stimuli, but they have been particular subrepresentation at a given point in
reliably demonstrated across numerous studies. time (Sereno et al., 2003).
Major challenges for future research remain in de- Emotional semantics may be special; their con-
termining the relative role of arousal and valence nection to biologically significant system states
and their interactions with task demands. Finally, and behavioral output tendencies may ensure most
the question to what extent and at which points in rapid activation of the underlying neural network
time, processing of emotional words recruits spe- representations. A number of studies endorse a si-
cific cortical and subcortical neural circuitries mer- multaneous impact of emotional word content on
its further scientific attention. both cortical and peripheral responses (Flor et al.,
1997; Knost et al., 1997; Herbert et al., 2006),
corroborating the view that the neural networks
Processing emotional words — electrophysiological representing emotional concepts dynamically link
conjectures semantic and response information (Lang et al.,
1993, see Fig. 3). Many studies also show surpris-
The above review demonstrates that emotional ingly early cortical responses to emotional word
word content can amplify word processing at content. Thus, the subrepresentation of a word’s
all stages from access to word meaning (around emotional content may sometimes, though clearly
200 ms), to contextual integration (around not always, be activated before other lexical in-
400 ms), evaluation, and memory encoding formation is available, for example, whether a
(around 600 ms). Occasionally, emotionality-de- word denotes an animate or an inanimate entity.
pendent enhancements have been reported even Depending on personality, context, and task, emo-
before 200 ms. tional semantic networks may become activated
In neurolinguistics, the timing of lexical access is rapidly or gradually, operate over several hundred
heatedly debated. The reports about different milliseconds or their activation may decay quickly
points in time where some aspects of the lexical when not currently relevant.
information on a word are accessed vary between The specification of experimental factors con-
100 and 600 ms. Importantly, the interpretation of tributing to the generation of arousal-, valence-, or
the N400 has shifted from an index of semantic even potency-specific effects or to the temporal
access to a signature of the interaction between gradient of the processing enhancement caused by
single word semantics and context. Accordingly, a emotional connotation is lacking detail. As a
168

working hypothesis, emotional content amplifies situation is likely to determine its further process-
word processing at the earliest stage at which the ing, with tasks that require processing of semantic
visual word form is assembled, i.e., no sooner than content supporting sustained enhancement of
about 100 ms after word presentation. Learning, emotional stimuli and contextual effects determin-
top-down and priming processes may dynamically ing possible processing advantages for either va-
modify the time point at which the activation of lence. Under circumstances where structural
the representation takes place (see above). Arousal outweighs semantic processing or where the stim-
is likely to govern this amplification process within uli occur very briefly and are not followed by con-
the first 300 ms, reflecting a general orienting func- firmatory input, only transient enhancement of
tion of emotion. Under specific experimental ma- early brain responses and less, if any, amplification
nipulations, such as very brief stimulus duration, in later time windows will result.
or in populations with pronounced negative Subcortical structures, most prominently the
processing biases, earliest advantages for unpleas- amygdala, have been implied at all stages of this
ant material may be obtained, possibly reflecting emotional content-driven amplification process,
modality-independent operations of a rudimentary but the dynamics of the interplay between subcor-
rapid threat detection system (Öhman and Mine- tical and cortical structures in processing emo-
ka, 2001). In general, the neural representations of tional words await specification. Combining
abstract linguistic features are less likely to be ac- electrophysiological, functional magnetic reso-
tivated by such a superfast detection system. nance, and lesion approaches within a network
Later than 300 ms after stimulus onset the be- theory of emotion and language will be useful to
havioral relevance of the stimulus in a given clarify these issues.
Appendix

Table A1. A summary of experimental design, recording parameters and results of the reviewed studies on effects of emotional content on ERP measures of word
processing in healthy volunteers.

Study Subjects N, sex Task(s) Stimuli Presentation Recording and analysis ERP effects

Word class Stimulus duration Number of electrodes

Emotional content Position of presentation Reference

Control of additional Stimulus repetitions Epoch duration (post-


stimulus parameters stimulus)

Statistical analysis

Lifshitz (1966) 1, male Viewing, memorizing Word class not indicated 1000–2000 ms 4 Visual comparison dirty words versus neutral
words: no marked differences visible.
Emotional content Centrally Left-hemispheric leads,
determined by author: 40 reference unclear
Number of repetitions not
‘dirty words’ 40 neutral
words indicated 500 ms

Visual inspection
Controlled for: word
length

Begleiter and Platz 18, all males Viewing, word naming Nouns 10 ms 1 (O2) Amplitude: ‘taboo’4neutral & flash
(1969)
Emotional content Centrally Linked ears reference Naming4viewing (100–200 ms and
determined by authors: 2 200–300 ms)
‘taboo’ words (shit, fuck) 2 20 repetitions 1024 ms
Latency: ‘taboo’oneutral & flash
‘neutral’ words (tile, page) ANOVA t-test
blank flash (300–400 ms)

Controlled for: word


length

Kostandov and 37, sex not indicated Covert identification and ‘Conflict’ and ‘neutral’ Experiment 1: 200 ms 2 (Cz, O1) Experiment 1: Supraliminal: N2 (220 ms)
Arzumanov (1977) counting of stimulus repetitions words 50 repetitions? Emotional content on peak amplitude &
(‘all in state of jealousy’) Left mastoid reference latency n.s.
No further details on Experiment 2: 15 ms
Experiment 1 supraliminal 1000 ms
experimental stimuli given. 50 repetitions? P3a (270–320 ms)
23. O1: Peak amplitude: emotional4neutral
t-tests
Additional parameter
Experiment 2 subliminal, Peak latency: emotionaloneutral Cz: n.s.
control unclear
14. Experiment 2: Subliminal: N2 (220 ms)
Peak amplitude O1 and Cz:
emotional 4 neutral

P3a (300 ms)


O1 and Cz: Peak amplitude:

169
emotional 4 neutral
Peak latency, n.s.
170
Table A1. (Continued )

Study Subjects N, sex Task(s) Stimuli Presentation Recording and analysis ERP effects

Begleiter et al. 10, all males Letter identification Word class not indicated 20 ms 6 (F3, F4, C3, C4, P3, P4) N1–P2 peak-to-peak amplitude (140–200 ms)
(1979)
Affective rating Prerating on semantic Centrally Only P3 and P4 analyzed, Affective rating condition
differential: 62 pleasant, 62 linked ears reference Electrodes P3 and P4: pleasant 4 neutral
Single presentation
unpleasant, 62 neutral unpleasant 4 neutral pleasant 4 unpleasant
450 ms
Additional parameter Effects: left 4 right
ANOVA t-test
control unclear
Left electrode (P3) Affective rating 4 letter
identification

No emotion effect in letter identification

Chapman et al. 10, 4 males Word naming Word class not indicated 17 ms 1 (CPz) PCA/stepwise discriminant analysis
(1978) differentiate among all 6 stimulus types
Pre-rating on semantic Centrally Linked ears reference
within the first 500 ms
differentiala: 20 E+, 20
15 repetitions 510 ms
E–, 20 P+, 20 P–, 20 A+, Visual inspection of provided figures: N1, P2,
20 A– PCA of ERP, ANOVA on and P3a differentiate the+and – poles of the
PCA components E, P, A dimensions
Luminance, distribution of
discriminant analysis
letters

Chapman (1979) 10 Verbal affective rating Word class not indicated 17 ms 1 (CPz) PCA/stepwise discriminant analysis
differentiate among all 6 stimulus types
Pre-rating on semantic Centrally Linked ears reference within the first 500 ms
differentiala: 20 E+, 20
15 repetitions 510 ms
E–, 20 P+, 20 P–, 20 A+, Differentiation between semantic scales (E, P,
20 A– PCA of ERP, ANOVA on A) within 190 ms
PCA components
Additional parameter Differentiation among polarities of word
control unclear discriminant analysis content types (E+/, P +/, A+/) within
330 ms

Chapman et al. 10, 4 males Word naming Word class not indicated 17 ms 1 (CPz) PCA/stepwise discriminant analysis
(1980) differentiate among all 6 stimulus types
Verbal affective rating Pre-rating on semantic Centrally Linked ears reference
within the first 500 ms
differential: 20 E+, 20 E–,
20 P+, 20 P–, 20 A+, 20 15 repetitions 510 ms Visual inspection: N1, P2, P3a differentiate
A– PCA of ERP, ANOVA on the+and – poles of the E, P, A dimensions

Additional parameter PCA components


discriminant analysis
control unclear

Vanderploeg et al. 10, all males Affective rating Word class not indicated 80 ms 6 (Fz, Pz, F7, F8, T5, T6) n.s. trend: amplitude: emotional (pleasant &
(1987) unplesant) 4 neutral
Words from Begleiter et al. Centrally Linked ears reference
(1969, 1979) and Chapman On P3a (230–420 ms) and LPC (500–624) ms
Repeated until 32 artifact- 800 ms
et al., (1977, 1978)
free averages per category
PCA, ANOVA on PCA
Additional parameter obtained
components
control unclear
Naumann et al. (1992) Pilot study: 17, 8 males Word length decision Adjectives 125 ms 3 (Fz, Cz, Pz) Pilot study:
Left mastoid reference P3 amplitude (300–700 ms)
Replication study: 106, 50 Affective rating 90 words affective norms Centrally
Affective rating4word length decision.
males (Hager et al., 1985) 1500 ms post word Emotional content (pleasant &
Single presentation
30 pleasant, 30 unpleasant, ANOVAs on P3 and unpleasant) 4 neutral
30 neutral positive slow wave Positive slow wave amplitude (700–1200):
amplitude and peak
Controlled for: length, Affective rating4word length decision
latency
concreteness Electrode  emotional content
Larger positivities at central and parietal
electrodes for pleasant and unpleasant
Replication study:

P3 amplitude
Emotional content
Emotional (pleasant & unpleasant)4neutral
Electrode  task:
Fz: affective rating4word length decision

Electrode  emotional content


Larger effect for pleasant, particularly at Fz
and Cz
P3 peak latency:
Neutraloemotional (pleasant & unpleasant)

Positive slow wave amplitude


Electrode  emotional content
No parietal maximum for neutral

Naumann et al. 54 (26 males, between design, Letter detection Nouns 200 ms 9 (F4, C4, P4, Fz, Cz, Pz, Emotional content  task
(1997) 18 per condition) F3, C3, P3) Unpleasant4neutral
Concreteness decision Published affective norms Centrally
-Linked ears reference Effect from valence decision task only
(Schwibbe et al., 1981)
Valence decision Single presentation
1400 ms post word Valence decision task
28 unpleasant, 28 neutral
P3 amplitude:
ANOVA and Tukey HSD
Controlled for: length, Unpleasant4neutral
test on P3 and positive
concrete-ness
slow wave amplitudes and P3 peak latency
peak latencies Emotional content  task
Letter detection & concreteness decision:
unpleasantoneutral

Valence decision:
neutralounpleasant

Positive slow wave amplitude


No effects

171
172
Table A1. (Continued )

Study Subjects N, sex Task(s) Stimuli Presentation Recording and analysis ERP effects

Skrandies (1998) 22, 8 males Visualize and memorize words Nouns 1000 ms 30 (10–20 system) Complex sequence of effects, discriminating
among E, P, A dimensions and their polarity:
Pre-rating on semantic Centrally Average reference 1000 ms
differentiala (different P1 (80–130 ms):
population): 10 E+, 10 RSVPb ANOVAs on centroids of Peak latency A4E and P
scalp distributions and
E–, 10 P+, 10 P–, 10 A+, 40 repetitions
peak latencies Duncan Right negative centroid shift for E- words
10 A–
post hoc Left positive centroid shift for E- words
Controlled for: Word
N1 (130–195 ms):
length, word frequency
Increased peak latency for E and A words
Reduced global field power (GFP) for A
words

195–265 ms:
Peak latency PoE and A
Anterior centroid shift for E-, P-, A-
compared to E+, P+, A+

565–635 ms:
Peak latency:
E+, P+, A+4E-, P-, A-

635–765 ms:
Centroids: P+ and P- shifted anteriorally
and posteriorally

860–975 ms:
Peak latency:
E+ and P+4E- and P-A+oA-

Skrandies and 23, 10 males Visualize and memorize words Nouns 1000 ms 32 (10–20 system) Complex sequence of effects, discriminating
Chiu (2003) between E, P, A dimensions and their
Pre-rating on semantic Centrally Average reference polarity:
differentiala (different
RSVPb 1000 ms post word
population): 10 E+, 10 P1 (80–130 ms):
E–, 10 P+, 10 P–, 10 A+, 24(?) repetitions ANOVAs on peak latency Latency E+, P+, A+4E-, P-, A-
10 A– centroid position Global
N1 (130–195 ms):
field power (GFP)
Controlled for word GFP P-and A-4P+and A+
length, word frequency Posterior-anterior location of centroids:
A+anterior P+anterior E+and E-anterior
P-and A-

Left-right location of centroids:


Right-shift of A centroid

No effects after 300 ms


Schapkin et al. 15, 7 males Emotional vs. neutral Nouns 150 ms 14 (only C3, C4, P3, P4, P2 amplitude (200–300 ms)
(2000) classification O1, O2 analyzed) pleasant4unpleasant
Pre-rating on valence Left or right visual field
(different population) Linked ears reference P3 amplitude (300–450 ms)
8 repetitions (4 per visual
pleasant4unpleasant
6 pleasant, 6 unpleasant, 6 field) 800 ms post word
neutral Slow positive wave (1000–1800):
ANOVAs on base-to-peak Frontal
Controlled for: syllables, measures of N1, P2, P3,
pleasant4unpleasant
word length, word SPW post hoc t-tests Parietal:
frequency, concreteness,
pleasant4neutral
imagery, initial letter

Bernat et al. 17, 8 males Focus on fixation point Adjectives Subliminal (1 ms) 6 (F3, F4, P3, P4, CPz, Oz) Subliminal:
(2001) Supraliminal (40 ms) Amplitudes:
Pre-rating on semantic Linked ears reference
Emotion:
differential in same Centrally
1000 ms post word P2 (100–210 ms), P3 (200–500 ms), LPC
population.
12 repetitions (6 (500–900 ms): unpleasant4pleasant
ANOVAs on P1, N1, P2,
10 pleasant, 10 unpleasant, subliminal, 6 supraliminal) Emotion  hemisphere:
P3, LPC mean amplitudes
12 neutral
pleasant vs. unpleasant P1 (40–120 ms), N1 (80–170 ms), P3
‘word length, word (analysis of neutral words (200–500 ms):
frequency, luminance is not presented) Left: unpleasant4pleasant
Right: unpleasant4pleasant

Supraliminal:
Amplitudes:
Emotion:
P3 (200–500 ms) and LPC (500–900 ms):
unpleasant4pleasant

Emotion  hemisphere,
P1 (20–80 ms), N1 (52–150 ms),
Left: unpleasant4pleasant
Right: unpleasant4pleasant

P3 (200–500 ms)
Emotional Content  hemisphere:
unpleasant: left4right

Ortigue et al. 13, all males Divided field lexical decision Nouns 13 ms 123 (extended 10–20 Amplitude:
(2004) (in which visual field did a Prerating for emotionality system) 100–140 ms: emotional4neutral
Left and right visual field
word appear?) on seven-point scale
simultaneously Average reference Distinct right occipital spatial map for
(emotional–nonemotional)
emotional words presented in the right visual
30 repetitions 250 ms post word
16 words, (8 neutral 8 field
emotional) 96 ANOVAs on spatial
pseudowords configuration maps
LAURA source estimation
Controlled for: word
length, word frequency

173
174
Table A1. (Continued )

Study Subjects N, sex Task(s) Stimuli Presentation Recording and analysis ERP effects

Herbert et al. 26, 16 males Covert evaluation Adjectives 4000 ms 64 (extended 10–20 system) Amplitude:
(2006)
Pre-rated on valence and Centrally Average reference P2 (180–250 ms) and P3a (250–400 ms):
arousal by different emotional (pleasant & unpleasant)4neutral
population Single presentation 1000 ms post word
LPC (600–750 ms): pleasant4neutral
ANOVAs on channel
60 unpleasant, 60 pleasant,
60 neutral groups

Controlled for: word


length, word frequency

Kissler et al. 16, 8 males Silent reading Nouns 333 ms and 1000 ms 64 (extended 10–20 system) Amplitude: 200–300 ms occipito-temporal
(submitted) electrodes
Pre-rating of valence and RSVPb Average reference
arousal by different Emotional (pleasant ¼ unpleasant)4neutral
10 (5 per presentation 333 ms/1000 ms post word
population Left: emotional4neutral
speed) Right: emotional4neutral
ANOVA on groups of
60 unpleasant, 60 pleasant,
electrodes
60 neutral

Word length, word


frequency, concreteness

a
Dimensions of the semantic differential: E, evaluation; P, potency; A, activity. +/- indicate the positive and negative poles of these dimensions.
b
RSVP, rapid serial visual presentation of stimuli in a continuous consecutive stream without interstimulus interval.
Table A2. A summary of experimental design, recording parameters and results of the reviewed studies on effects of emotional content on ERP measures of word
processing in clinical populations.

Study Subjects N, sex Task(s) Stimuli Presentation Recording and ERP effects
analysis

clinical status Electrodes and ERP effects


reference analysis
window (post-
stimulus)
Word class Stimulus Statistics
duration
Emotional
content Position of
presentation
Control of
additional Stimulus
stimulus repetitions
parameters

Williamson et al. 16 — 8 psychopaths, Lexical decision Word class not 176 ms 5 (Fz, Cz, Pz, PT3, P240: 225–300 ms
(1991) 8 non-psychopaths, indicated PT4) Amplitude:
+/- 3 para-
all males Nonpsychopaths:
Affective norms foveally, Linked ears
emotional (pleasant &
(Toglia and vertical reference
unpleasant)4neutral.
Battig, 1978) format
2000 ms
Psychopaths: emotional
13 pleasant, 13 6 repetitions
ANOVA t-test (pleasant &
unpleasant, 13
unpleasant) ¼ neutral
neutral, 39
pseudowords LPC: 650–800 ms
Amplitude:
Word length,
Group  emotional
frequency,
Content  electrode
number of
site
syllables,
concreteness Midline:
Non-psychopaths:
emotional (pleasant &
unpleasant)4neutral
Psychopaths: emotional
(pleasant &
unpleasant)4neutral

175
176
Table A2. (Continued )

Study Subjects N, sex Task(s) Stimuli Presentation Recording and ERP effects
analysis

Weinstein (1995) 20 students 10 Decision on 20 threatening, 1100 ms per 3 (Fz, Cz, Pz) N100: 90–120 ms: Peak
highly anxious, 5 contextual fit 20 pleasant word amplitude: High
Linked ears
males 10 low sentences (5–7 anxiety group threat
Centrally reference
anxious 4 males words) priming4low anxiety
2 repetitions 650 ms group threat priming.
‘Probe words’:20
(Fz, Cz)
threat, 20 neutral ANOVA t-test
20 pleasant N400/P400: 400–500
(nouns & ms Peak amplitude:
adjectives) 2 Threat condition: High
pairings: fitting- anxiety group more
not fitting. positive than low
anxiety group (Cz,
Threat words
Pz).Peak latency: High
from MacLeod
anxiety groupolow
(1985).
anxiety group.
Emotional
content of (Plots of means appear
sentences and to diverge from the
words prerated reported statistics)
in different
population.
Control for
linguistic
parameters not
indicated
Kiehl et al. (1999), 29 prisoners, all Pleasant/unpleasant Word class not 300 ms 5 (Fz, Cz, Pz, PT3, N350: 300–400 ms
Task 3 males: 8 psycho- decision indicated PT4)Linked ears Amplitude:
Centrally
paths, 9 reference All groups:
Valence norms
nonpsychopaths, 12 Single Emotional content:
(Toglia and 1200 ms
mixed presentation pleasant4unpleasant
Battig, 1978)
ANOVA t-test on
Emotional
60 pleasant, 60 reaction times and
content  hemisphere:
unpleasant N350, LPC
Left:
amplitudes
Controlled for: pleasant4unpleasant
word length, Right:
frequency, pleasant4unpleasant
number of
LPC: 400–800 ms
syllabels,
Amplitude
imagery,
Emotional content:
concreteness
Non-psychopaths:
Unpleasant4pleasant
Psychopaths:
Unpleasant ¼ pleasant
Emotional
content  hemisphere:
Left:
unpleasant4pleasant
Right:
unpleasant4pleasant
Reaction times:
Pleasantounpleasant

Knost et al. (1997) 38:19 Detection and Adjectives Individual 11 (Fz, F3, F4, Cz, N100 (80–180 ms) at
prechronicpain naming perceptual C3, C4, Pz, P3, P4, electrode F3
Pre-rating on
patients, 11 threshold T3, T4) Prechronic pain
familiarity, body
males patients:
and pain Centrally Linked ears
Amplitude: pain
19 controls, 11 relatedness reference
Single words4body related &
males
40 pain-related, presentation 800 ms neutral
40 body-related,
ANOVA t-test LPC2: 600–800 ms
40 neutral
Amplitude both groups:
Controlled for: pain/body-
word frequency, related4neutral
word length

177
178
Table A2. (Continued )

Study Subjects N, sex Task(s) Stimuli Presentation Recording and ERP effects
analysis

Flor et al. (1997) 24:12 chronic pain Detection and Adjectives Perceptual 11 (Fz, F3, F4, Cz, N100 (80–140 ms) at
patients, 5 males 12 naming threshold C3, C4, Pz, P3, P4, electrodes F3, Fz, C3,
Pre-rating on
controls, 5 males T3, T4) P3, Pz
familiarity, body Centrally
and pain Linked ears Chronic pain patients
Single
relatedness reference only: pain words4body
presentation
related & neutral
40 pain-related, 800 ms
40 body-related, Both groups:
ANOVA t-test
40 neutral N100 (80–140 ms)
Pain words: left4right
Controlled for:
Neutral: right4left
word frequency,
word length N200 (140–200 ms) at
all electrodes:
Chronic pain
patients:pain
words4body related &
neutral
P200 (180–280 ms):
Chronic pain patients:
Larger responses to
pain words over right
than left
LPC1 (400–600 ms):
n.s. All: pain
words ¼ body
related ¼ neutral
LPC2 (600–800 ms)
n.s. All: pain
words ¼ body
related ¼ neutral.
Pauli et al. (2005) 50:25 panic patients, Detection and Adjectives and Perceptual 21 (only 9 analyzed: Threshold
9 males, 25 controls, naming verbs threshold and Fz, Cz, Pz, F3, F4, presentation:
9 males 1000 ms, C3, C4, P3, P4) Panic patients:
Pre-rating for
N200/P200
descriptiveness 2 presentations Linked ears pnce
(100–200 ms)
for panic- (one per
1000 ms Frontal effect panic
disorder duration)
words4neutral
symptoms by 2 ANOVA t-test
psychiatrists and P3a: (200–400 ms):
2 clinical panic words4neutral
psychologists
LPC1: 400–600 ms
48 panic related, Amplitude both
48 neutral groups: panic4neutral
Controlled for: 1000 ms presentation:
word frequency, LPC1: 400–600 ms
length, syllables, Amplitude both
distribution of groups: panic4neutral
word classes
LPC2: 600–800
Amplitude both
groups: panic4neutral
Emotional
content  hemisphere:
Left: panic4neutral
Right: panic4neutral

179
180

Acknowledgments Bradley, M.M. and Lang, P.J. (2000) Affective reactions to


acoustic stimuli. Psychophysiology, 37: 204–215.
Brem, S., Lang-Dullenkopf, A., Maurer, U., Halder, P., Buc-
This work was supported by a grant from the
her, K. and Brandeis, D. (2005) Neurophysiological signs of
Heidelberg Academy of Sciences (Mind and Brain rapidly emerging visual expertise for symbol strings. Neuro-
Program). We thank Anne Hauswald for help in report, 16: 45–48.
preparation of this manuscript and Christiane Caccioppo, J.T. (2000) Asymmetries in affect laden information
Beck, Susanne Kössler, Bistra Ivanona and Irene processing. In: Banaji, R. and Prentice, D.A. (Eds.), Per-
Winkler for assistance in the experimental work spectivism in Social Psychology: The Yin and Yang of Sci-
entific Progress. American Psychological Association Press,
described. Washington, DC, pp. 85–95.
Chao, L.L., Haxby, J.V. and Martin, A. (1999) Attribute-based
neural substrates in temporal cortex for perceiving and
References knowing about objects. Nat. Neurosci., 2: 913–919.
Chapman, R.M. (1979) Connotative meaning and averaged
Amaral, D.G., Behniea, H. and Kelly, J.L. (2003) Topographic evoked potentials. In: Begleiter, H. (Ed.), Evoked Brain Po-
organization of projections from the amygdala to the visual tentials and Behavior. Plenum Press, New York, pp. 171–197.
cortex in the macaque monkey. Neuroscience, 118: Chapman, R.M., McCrary, J.W., Chapman, J.A. and Bragdon,
1099–1120. H.R. (1978) Brain responses related to semantic meaning.
Anderson, A.K. and Phelps, E.A. (2001) Lesions of the human Brain Lang., 5: 195–205.
amygdala impair enhanced perception of emotionally salient Chapman, R.M., McCrary, J.W., Chapman, J.A. and Martin,
events. Nature, 411: 305–309. J.K. (1980) Behavioral and neural analyses of connotative
Assadollahi, R. and Pulvermüller, F. (2001) Neuromagnetic meaning: word classes and rating scales. Brain Lang., 11:
evidence for early access to cognitive representations. Ne- 319–339.
uroreport, 12: 207–213. Cohen, L., Dehaene, S., Naccache, L., Lehericy, S., Dehaene-
Assadollahi, R. and Pulvermuller, F. (2003) Early influences of Lambertz, G., Henaff, M.A. and Michel, F. (2000) The visual
word length and frequency: a group study using MEG. Ne- word form area: spatial and temporal characterization of an
uroreport, 14: 1183–1187. initial stage of reading in normal subjects and posterior split-
Assadollahi, R. and Rockstroh, B. (2005) Neuromagnetic brain brain patients. Brain, 123(Pt 2): 291–307.
responses to words from semantic sub- and supercategories. Davis, M. (1992) The role of the amygdala in fear and anxiety.
BMC Neurosci., 6: 57. Annu. Rev. Neurosci., 15: 353–375.
Barsalou, L.W. (1999) Perceptual symbol systems. Behav. Brain Dehaene, S. (1995) Electrophysiological evidence for category-
Sci., 22: 577–609 discussion 610–660. specific word processing in the normal human brain. Neuro-
Begleiter, H. and Platz, A. (1969) Cortical evoked potentials to report, 6: 2153–2157.
semantic stimuli. Psychophysiology, 6: 91–100. Dehaene, S., Cohen, L., Sigman, M. and Vinckier, F. (2005)
Begleiter, H., Projesz, B. and Garozzo, R. (1979) Visual evoked The neural code for written words: a proposal. Trends Cogn.
potentials and affective ratings of semantic stimuli. In: Beg- Sci., 9: 335–341.
leiter, H. (Ed.), Evoked Brain Potentials and Behavior. Ple- Dehaene, S., Le Clec, H.G., Poline, J.B., Le Bihan, D. and
num Press, New York, pp. 127–143. Cohen, L. (2002) The visual word form area: a prelexical
Bentin, S., Mouchetant-Rostaing, Y., Giard, M.H., Echallier, representation of visual words in the fusiform gyrus. Neuro-
J.F. and Pernier, J. (1999) ERP manifestations of processing report, 13: 321–325.
printed words at different psycholinguistic levels: time course Dien, J., Spencer, K.M. and Donchin, E. (2004) Parsing the late
and scalp distribution. J. Cogn. Neurosci., 11: 235–260. positive complex: mental chronometry and the ERP compo-
Bernat, E., Bunce, S. and Shevrin, H. (2001) Event-related nents that inhabit the neighborhood of the P300. Psycho-
brain potentials differentiate positive and negative mood ad- physiology, 41: 665–678.
jectives during both supraliminal and subliminal visual Dijksterhuis, A. and Aarts, H. (2003) On wildebeests and hu-
processing. Int. J. Psychophysiol., 42: 11–34. mans: the preferential detection of negative stimuli. Psychol.
Bower, G. (1981) Mood and memory. Am. Psychol., 36: Sci., 14: 14–18.
129–148. Everatt, J., McCorquidale, B., Smith, J., Culverwell, F., Wilks,
Bradley, M.M. and Lang, P.J. (1994) Measuring emotion: the A., Evans, D., Kay, M. and Baker, D. (1999) Association
Self-Assessment Manikin and the Semantic Differential. J. between reading ability and visual processes. In: Everatt, J.
Behav. Ther. Exp. Psychiatry, 25(1): 49–59. (Ed.), Reading and Dyslexia. Routledge, London, pp. 1–39.
Bradley, M., and Lang, P.J., 1998. Affective norms for English Federmeier, K.D., Kirson, D.A., Moreno, E.M. and Kutas, M.
words (ANEW): Instruction manual and affective ratings. (2001) Effects of transient, mild mood states on semantic
Technical report A-8, The Center for Research in Psycho- memory organization and use: an event-related potential in-
physiology, University of Florida. vestigation in humans. Neurosci. Lett., 305: 149–152.
181

Federmeier, K.D., Segal, J.B., Lombrozo, T. and Kutas, M. Isenberg, N., Silbersweig, D., Engelien, A., Emmerich, S., Ma-
(2000) Brain responses to nouns, verbs and class-ambiguous lavade, K., Beattie, B., Leon, A.C. and Stern, E. (1999) Lin-
words in context. Brain, 123(Pt 12): 2552–2566. guistic threat activates the human amygdala. Proc. Natl.
Flor, H., Knost, B. and Birbaumer, N. (1997) Processing of Acad. Sci. USA, 96: 10456–10459.
pain- and body-related verbal material in chronic pain pa- Ito, T.A. and Caccioppo, J.T. (2000) Electrophysiological ev-
tients: central and peripheral correlates. Pain, 73: 413–421. idence of implicit and explicit categorization processes. J.
Friederici, A.D., Hahne, A. and Mecklinger, A. (1996) Tem- Exp. Soc. Psychol., 36: 660–676.
poral structure of syntactic parsing: early and late event-re- Junghöfer, M., Bradley, M.M., Elbert, T.R. and Lang, P.J.
lated brain potential effects. J. Exp. Psychol. Learn. Mem. (2001) Fleeting images: a new look at early emotion discrim-
Cogn., 22: 1219–1248. ination. Psychophysiology, 38: 175–178.
Garavan, H., Pendergrass, J.C., Ross, T.J., Stein, E.A. and Keil, A., Bradley, M.M., Hauk, O., Rockstroh, B., Elbert, T.
Risinger, R.C. (2001) Amygdala response to both pos- and Lang, P.J. (2002) Large-scale neural correlates of affec-
itively and negatively valenced stimuli. Neuroreport, 12: tive picture processing. Psychophysiology, 39: 641–649.
2779–2783. Kellenbach, M.L., Wijers, A.A., Hovius, M., Mulder, J. and
Greenwald, A.G., Draine, S.C. and Abrams, R.L. (1996) Three Mulder, G. (2002) Neural differentiation of lexico-syntactic
cognitive markers of unconscious semantic activation. Sci- categories or semantic features? Event-related potential ev-
ence, 273: 1699–1702. idence for both. J. Cogn. Neurosci., 14: 561–577.
Hagoort, P. and Brown, C.M. (2000) ERP effects of listening to Kiefer, M. and Spitzer, M. (2000) Time course of conscious and
speech compared to reading: the P600/SPS to syntactic vi- unconscious semantic brain activation. Neuroreport, 11:
olations in spoken sentences and rapid serial visual presen- 2401–2407.
tation. Neuropsychologia, 38: 1531–1549. Kiehl, K.A., Hare, R.D., McDonald, J.J. and Brink, J. (1999)
Halgren, E., Baudena, P., Heit, G., Clarke, J.M., Marinkovic, Semantic and affective processing in psychopaths: an event-
K., Chauvel, P. and Clarke, M. (1994a) Spatio-temporal related potential (ERP) study. Psychophysiology, 36:
stages in face and word processing. 2. Depth-recorded po- 765–774.
tentials in the human frontal and Rolandic cortices. J. King, J.W. and Kutas, M. (1998) Neural plasticity in the dy-
Physiol. (Paris), 88: 51–80. namics of human visual word recognition. Neurosci. Lett.,
Halgren, E., Baudena, P., Heit, G., Clarke, J.M., Marinkovic, 244: 61–64.
K. and Clarke, M. (1994b) Spatio-temporal stages in face and Kissler, J. and Bäuml, K.H. (2000) Effects of the beholder’s age
word processing. I. Depth-recorded potentials in the human on the perception of facial attractiveness. Acta Psychol.
occipital, temporal and parietal lobes [corrected]. J. Physiol. (Amst.), 104: 145–166.
(Paris), 88: 1–50. Kissler, J., Herbert, C., Peyk, P. and Junghöfer, M. (submitted
Hamann, S. and Mao, H. (2002) Positive and negative emo- manuscript) Sex, crime and videotape—enhanced early cor-
tional verbal stimuli elicit activity in the left amygdala. Ne- tical responses to rapidly presented emotional words.
uroreport, 13: 15–19. Kissler, J., and Kössler, S., (in preparation) Pleasant pictures
Hauk, O., Johnsrude, I. and Pulvermuller, F. (2004) Somato- facilitate lexical decision.
topic representation of action words in human motor and Knost, B., Flor, H., Braun, C. and Birbaumer, N. (1997) Cer-
premotor cortex. Neuron, 41: 301–307. ebral processing of words and the development of chronic
Haxby, J.V., Horwitz, B., Ungerleider, L.G., Maisog, J.M., pain. Psychophysiology, 34: 474–481.
Pietrini, P. and Grady, C.L. (1994) The functional organi- Kostandov, E. and Arzumanov, Y. (1977) Averaged cortical
zation of human extrastriate cortex: a PET-rCBF study of evoked potentials to recognized and non-recognized verbal
selective attention to faces and locations. J. Neurosci., 14: stimuli. Acta Neurobiol. Exp. (Wars), 37: 311–324.
6336–6353. Kutas, M. and Federmeier, K.D. (2000) Electrophysiology re-
Herbert, C., Kissler, J., Junghöfer, M., Peyk, P. and Rockstroh, veals semantic memory use in language comprehension.
B. (2006) Processing emotional adjectives: evidence from Trends Cogn. Sci., 4: 463–470.
startle EMG and ERPs. Psychophysiology, 43(2): 197–206. Kutas, M. and Hillyard, S.A. (1980) Reading senseless sen-
Hinojosa, J.A., Martin-Loeches, M., Munoz, F., Casado, P. tences: brain potentials reflect semantic incongruity. Science,
and Pozo, M.A. (2004) Electrophysiological evidence of au- 207: 203–205.
tomatic early semantic processing. Brain Lang., 88: 39–46. Kutas, M. and Hillyard, S.A. (1984) Brain potentials during
Hinojosa, J.A., Martin-Loeches, M. and Rubia, F.J. (2001) reading reflect word expectancy and semantic association.
Event-related potentials and semantics: an overview and an Nature, 307: 161–163.
integrative proposal. Brain Lang., 78: 128–139. Lang, P.J. (1979) Presidential address, 1978A bio-infor-
Holcomb, P.J., Kounios, J., Anderson, J.E. and West, W.C. mational theory of emotional imagery. Psychophysiology,
(1999) Dual-coding, context-availability, and concreteness 16: 495–512.
effects in sentence comprehension: an electrophysiological Lang, P.J. (1994) The motivational organization of emotion:
investigation. J. Exp. Psychol. Learn. Mem. Cogn., 25: Affect-reflex connections. In: Van Goozen, S.H.M., Van de
721–742. Poll, N.E. and Sergeant, J.E. (Eds.), Emotions: Essays on
182

Emotion Theory. Lawrence Erlbaum Associates, Hillsdale, Osterhout, L., Holcomb, P.J. and Swinney, D.A. (1994) Brain
NJ, pp. 61–93. potentials elicited by garden-path sentences: evidence of the
Lang, P.J., Greenwald, M.K., Bradley, M.M. and Hamm, A.O. application of verb information during parsing. J. Exp. Psy-
(1993) Looking at pictures: affective, facial, visceral, and be- chol. Learn. Mem. Cogn., 20: 786–803.
havioral reactions. Psychophysiology, 30: 261–273. Pauli, P., Amrhein, C., Muhlberger, A., Dengler, W. and
LeDoux, J.E. (1989) Cognitive-emotional interactions in the Wiedemann, G. (2005) Electrocortical evidence for an early
brain. Cogn. Emotion, 3: 267–289. abnormal processing of panic-related words in panic disorder
LeDoux, J.E. (1995) Emotion: clues from the brain. Annu. Rev. patients. Int. J. Psychophysiol., 57: 33–41.
Psychol., 46: 209–235. Perfetti, C.A. (1998) Comprehending written language: a blue-
Liddell, B.J., Brown, K.J., Kemp, A.H., Barton, M.J., Das, P., print of the reader. In: Brown, C.M. and Hagoort, P. (Eds.),
Peduto, A., Gordon, E. and Williams, L.M. (2005) A direct The Neurocognition of Language. University Press, Oxford.
brainstem-amygdala-cortical ‘alarm’ system for subliminal Perfetti, C.A. and Sandak, R. (2000) Reading optimally builds
signals of fear. Neuroimage, 24: 235–243. on spoken language: implications for deaf readers. J. Deaf.
Liddell, B.J., Williams, L.M., Rathjen, J., Shevrin, H. and Stud. Deaf. Educ., 5: 32–50.
Gordon, E. (2004) A temporal dissociation of subliminal Phelps, E.A., O’Connor, K.J., Gatenby, J.C., Gore, J.C., Grill-
versus supraliminal fear perception: an event-related poten- on, C. and Davis, M. (2001) Activation of the left amygdala
tial study. J. Cogn. Neurosci., 16: 479–486. to a cognitive representation of fear. Nat. Neurosci., 4:
Lifshitz, K. (1966) The averaged evoked cortical response to 437–441.
complex visual stimuli. Psychophysiology, 3: 55–68. Posner, M.I., Abdullaev, Y.G., McCandliss, B.D. and Sereno,
Luce, R.D. and Suppes, P. (1965) Preference, utility and sub- S.C. (1999) Neuroanatomy, circuitry and plasticity of word
jective probability. In: Luce, R.D., Bush, R.R. and Galanter, reading. Neuroreport, 10: R12–R23.
E.H. (Eds.) Handbook of Mathematical Psychology, Vol. 3. Pulvermüller, F. (1999) Words in the brain’s language. Behav.
New York, Wiley, pp. 249–410. Brain Sci., 22: 253–279 discussion 280–336.
Martin, A., Wiggs, C.L., Ungerleider, L.G. and Haxby, J.V. Pulvermüller, F., Assadollahi, R. and Elbert, T. (2001a) Ne-
(1996) Neural correlates of category-specific knowledge. Na- uromagnetic evidence for early semantic access in word rec-
ture, 379: 649–652. ognition. Eur. J. Neurosci., 13: 201–205.
Martin-Loeches, M., Hinojosa, J.A., Gomez-Jarabo, G. and Pulvermüller, F., Harle, M. and Hummel, F. (2000) Neuro-
Rubia, F.J. (2001) An early electrophysiological sign of se- physiological distinction of verb categories. Neuroreport, 11:
mantic processing in basal extrastriate areas. Psychophysiol- 2789–2793.
ogy, 38: 114–124. Pulvermüller, F., Härle, M. and Hummel, F. (2001b) Walking
Molfese, D. (1985) Electrophysiological correlates of semantic or talking? Behavioral and neurophysiological correlates of
features. J. Psycholinguistic Res., 14: 289–299. action verb processing. Brain Lang., 78: 143–168.
Morris, J.S., Öhman, A. and Dolan, R.J. (1999) A subcortical Rudell, A.P. (1992) Rapid stream stimulation and the recog-
pathway to the right amygdala mediating ‘unseen’ fear. Proc. nition potential. Electroencephalogr. Clin. Neurophysiol., 83:
Natl. Acad. Sci. USA,, 96: 1680–1685. 77–82.
Münte, T.F., Heinze, H.J., Matzke, M., Wieringa, B.M. and Russel, J. (1980) A circumplex model of affects. J. Pers. Soc.
Johannes, S. (1998) Brain potentials and syntactic violations Psychol., 39: 1161–1178.
revisited: no evidence for specificity of the syntactic positive Schapkin, S.A., Gusev, A.N. and Kuhl, J. (2000) Categoriza-
shift. Neuropsychologia, 36: 217–226. tion of unilaterally presented emotional words: an ERP anal-
Naccache, L., Gaillard, R., Adam, C., Hasboun, D., Clemenc- ysis. Acta. Neurobiol. Exp. (Wars), 60: 17–28.
eau, S., Baulac, M., Dehaene, S. and Cohen, L. (2005) A Schendan, H.E., Ganis, G. and Kutas, M. (1998) Neurophys-
direct intracranial record of emotions evoked by subliminal iological evidence for visual perceptual categorization of
words. Proc. Natl. Acad. Sci. USA, 102: 7713–7717. words and faces within 150 ms. Psychophysiology, 35:
Öhman, A. and Mineka, S. (2001) Fears, phobias, and prepar- 240–251.
edness: toward an evolved module of fear and fear learning. Schupp, H.T., Öhman, A., Junghofer, M., Weike, A.I., Stock-
Psychol. Rev., 108: 483–522. burger, J. and Hamm, A.O. (2004) The facilitated processing
Ortigue, S., Michel, C.M., Murray, M.M., Mohr, C., Carbon- of threatening faces: an ERP analysis. Emotion, 4: 189–200.
nel, S. and Landis, T. (2004) Electrical neuroimaging reveals Sereno, S.C., Brewer, C.C. and O’Donnell, P.J. (2003) Context
early generator modulation to emotional words. Neuroim- effects in word recognition: evidence for early interactive
age, 21: 1242–1251. processing. Psychol. Sci., 14: 328–333.
Osgood, C.E., Miron, M.S. and May, W.H. (1975) Cross-Cul- Sereno, S.C., Rayner, K. and Posner, M.I. (1998) Estab-
tural Universals of Affective Meaning. University of Illinois lishing a time-line of word recognition: evidence from eye
Press, Urbana, Chicago, London. movements and event-related potentials. Neuroreport, 9:
Osgood, C.E., Suci, G.J. and Tannenbaum, P.H. (1957) The 2195–2200.
measurement of meaning. University of Illinois Press, Ur- Silvert, L., Delplanque, S., Bouwalerh, H., Verpoort, C. and
bana, Chicago, and London. Sequeira, H. (2004) Autonomic responding to aversive words
183

without conscious valence discrimination. Int. J. Psycho- Vanderploeg, R.D., Brown, W.S. and Marsh, J.T. (1987) Judg-
physiol., 53: 135–145. ments of emotion in words and faces: ERP correlates. Int. J.
Skrandies, W. (1998) Evoked potential correlates of semantic Psychophysiol., 5: 193–205.
meaning — a brain mapping study. Brain Res. Cogn. Brain Warrington, E.K. and Shallice, T. (1979) Semantic access dys-
Res., 6: 173–183. lexia. Brain, 102: 43–63.
Skrandies, W. and Chiu, M.J. (2003) Dimensions of affective Warrington, E.K. and Shallice, T. (1980) Word-form dyslexia.
semantic meaning — behavioral and evoked potential Brain, 103: 99–112.
correlates in Chinese subjects. Neurosci. Lett., 341: Weinstein, A. (1995) Visual ERPs evidence for enhanced
45–48. processing of threatening information in anxious university
Sutton, S., Tueting, P., Zubin, J. and John, E.R. (1967) Infor- students. Biol. Psychiatry, 37: 847–858.
mation delivery and the sensory evoked potential. Science, Wickelmaier, F. and Schmid, C. (2004) A Matlab function to
155: 1436–1439. estimate choice model parameters from paired-comparison
Tarkiainen, A., Helenius, P., Hansen, P.C., Cornelissen, P.L. data. Behav. Res. Methods Instrum. Comput., 36: 29–40.
and Salmelin, R. (1999) Dynamics of letter string perception Williamson, S., Harpur, T.J. and Hare, R.D. (1991) Abnormal
in the human occipitotemporal cortex. Brain, 122(Pt 11): processing of affective words by psychopaths. Psychophys-
2119–2132. iology, 28: 260–273.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 9

Event-related potential studies of language and


emotion: words, phrases, and task effects

Ira Fischler and Margaret Bradley

Psychology Department, PO Box 112250, University of Florida, Gainesville, FL 32611, USA

Abstract: This chapter reviews research that focuses on the effects of emotionality of single words, and
of simple phrases, on event-related brain potentials when these are presented visually in various tasks.
In these studies, presentation of emotionally evocative language material has consistently elicited a late
(c. 300–600 ms post-onset) positive-going, largely frontal–central shift in the event-related potentials (ERPs),
relative to neutral materials. Overall, affectively pleasant and unpleasant words or phrases are quite similar
in their neuroelectric profiles and rarely differ substantively. This emotionality effect is enhanced in both
amplitude and latency when emotional content is task relevant, but is also reliably observed when the task
involves other semantically engaging tasks. On the other hand, it can be attenuated or eliminated when the
task does not involve semantic evaluation (e.g., lexical decisions to words or orthographic judgments to the
spelling patterns) or when comprehension of phrases requires integration of the connotative meaning of
several words (e.g., compare dead puppy and dead tyrant). Taken together, these studies suggest that the
emotionality of written language has a rapid and robust impact on ERPs, which can be modulated by
specific task demands as well as the linguistic context in which the affective stimulus occurs.

Keywords: language; emotion; event-related potentials; motivation; phrase

There are good reasons for those who wish to un- dominance and control, with verbal threats and
derstand human emotion to be interested in emo- promises providing a virtual forum for the basic
tion and language. Our waking lives are filled with appetitive and defensive systems that motivate all
talk, to others and to ourselves — it has been es- behavior (e.g., Lang et al., 1997), and that the ex-
timated that the ‘‘stream of conscious experience’’ quisite informational capacity of language devel-
is largely a stream of self-talk (e.g., Klinger and oped from a regulatory, motivational base. The
Cox, 1987). We can be deeply moved, frightened, primacy of language as a skill unique to humans,
or aroused by what we hear or read. We actively and its ubiquity across societies and individuals,
seek out these kinds of materials, in the news ar- suggests that in order to understand emotional
ticles, stories, and novels we choose to read. We behavior and cognition the story told by language
might argue, in fact, that a primary evolutionary will be an important one.
purpose of language is motivational — to influence In the studies reported here, effects of emotion-
the thoughts, feelings, and therefore the actions of ality of single word and word phrases are explored
others in our group. Indeed, some have argued as they affect brain processes as reflected in event-
that language developed in part as a tool of social related potentials (ERPs). Individual words are an
appropriate starting point, as words and other
morphemes (e.g., the -ed in walked) are the build-
Corresponding author. E-mail: ifisch@ufl.edu ing blocks of meaning in language. But meaning

DOI: 10.1016/S0079-6123(06)56009-1 185


186

emerges from the sequence of words and phrases, showed an association between alpha waves
and often is more than the ‘‘sum of its lexical (10–13 Hz) and emotional calm (with even slower
parts’’ — dead is bad, and tyrant is bad, for ex- waves in sleep) and a progressive increase in fre-
ample, but a dead tyrant is not so bad. In this quency with increasing intensity of emotional
paper, we report data from a number of different arousal (from annoyance and anxiety to rage and
studies that varied the emotionality of individual panic). Other data from research with animal sub-
words, and of simple, two-word phrases. jects showed that, on the one hand, the brain stem
reticular formation (Moruzzi and Magoun, 1949)
modulated both EEG frequency and visceral
The motivational organization of emotion arousal, and on the other, that midbrain lesions
blocked emotion, prompting lethargic, apathetic
When studying emotion in an experimental con- behavior.
text, one immediate decision concerns selecting the The importance of emotional arousal in evalu-
affective parameters on which stimuli will vary, ative judgments and affective reactions could be
regardless of whether one uses words, pictures, due, in part, to a biphasic organization of emotion
sounds, movies, or other stimulus materials to in- arising from its evolutionary and functional her-
duce affect in the laboratory. In the work reviewed itage. For instance, Lang et al. (1997) proposed
here, we define emotion in terms of differences in that emotion stems from activation of one of two
rated pleasure and arousal of words and simple fundamental motive systems in the brain —
phrases, relying on a number of theories that pro- appetitive and defensive — that have evolved to
pose emotion is organized by dimensions of pleas- support behaviors that sustain the life of the in-
ure and arousal (e.g., Osgood et al., 1957; Lang dividual and the species, accounting for the pri-
et al., 1997). Supporting these views, empirical macy of the valence dimension in affective
factor analyses of evaluative language consistently expression. These two motivational systems are
have found that the primary dimension underlying associated with widespread cortical, autonomic,
people’s judgments of stimuli (that vary from sin- and behavioral activity that can vary in the inten-
gle words to abstract paintings) is hedonic valence, sity of activation. Differences in motivational in-
which accounts for the most variance in judgments tensity map onto the arousal parameter in
of meaning (e.g., Osgood et al., 1957; see also emotion’s dimensional organization. Importantly,
Mehrabian and Russell, 1974; Russell, 1980). He- arousal is not viewed as a separable system that is
donic valence is described by scales that, in aggre- independently modulated, but rather as represent-
gate, define a continuous dimension that ranges ing the activation (metabolic and neural) initiated
from unpleasant (unhappy, annoyed, despairing, by centrally organized appetitive or defensive sys-
etc.) to pleasant (happy, pleased, hopeful, etc.). tems that have evolved to protect and maintain
Arousal or intensity is the second major dimension life.
that emerges in these factor analyses, and scales In studying how language cues activate the
defining this dimension extend from an unaroused appetitive and defensive motivational systems un-
state (calm, relaxed, sleepy, etc.) to a state of high derlying emotion, it is important to categorize and
energy or arousal (excited, stimulated, wide- control verbal stimuli as they vary in hedonic va-
awake, etc.). lence and arousal. In a series of individual rating
The relative importance of hedonic valence and studies, we have collected affective reports of
arousal has been widely debated. Some have ar- pleasure and arousal for over 1000 English words
gued that arousal is the primary dimension of from many participants. These materials are dis-
emotion. Thus, Lindsley (1951) proposed an in- tributed to researchers in a collection called the
fluential ‘‘activation theory of emotion’’ (see also Affective Norms for English Words (ANEW;
earlier formulations by Duffy, 1941 and Arnold, Bradley and Lang, 1998). In each rating
1945), which was based, in part, on human study, participants are asked to judge the pleas-
electroencephalographic (EEG) studies that antness and arousal of each word using the
187

Self-Assessment Manikin (SAM) affective rating In Fig. 1, each word is plotted in a two-dimen-
system devised by Lang et al. (1980). SAM is a sional Cartesian space defined by its mean pleasure
graphic (nonverbal) figure that depicts each eval- and arousal rating. There are several characteristic
uative dimension using a nine-point scale, and features of the resulting affective space. First, these
ranges from a smiling, happy figure to a frowning, materials elicit ratings that completely span the
unhappy figure when representing the hedonic va- hedonic valence dimension ranging from extremely
lence dimension (see Fig. 1, ordinate); for arousal, pleasant to extremely unpleasant. Similarly, a wide
SAM ranges from an excited, wide-eyed figure to a range of arousal levels are elicited by these mate-
relaxed, sleepy figure (see Fig. 1, abscissa). The rials. Secondly, it is clear that pleasant words
SAM measures of pleasure and arousal correlate range continuously along the arousal dimension:
well with ratings on these evaluative dimensions the upper half of emotional space has exemplars at
obtained using a much longer, verbal semantic many positions along this dimension. Words labe-
differential scale (Bradley and Lang, 1994). ling unpleasant objects and events, however, tend

Affective Norms for English Words (ANEW)


Pleasure

Men
Women

Arousal
Fig. 1. Distribution of words in the two-dimensional affective space defined by each word’s mean pleasure and arousal rating (ANEW;
Bradley and Lang, 1998), using the Self-Assessment Manikin (SAM; Lang et al., 80).
188

to cluster in the high arousal quadrant of emo- In Chapman et al. (1980), both emotionally
tional space, suggesting that events and objects pleasant and unpleasant words elicited an in-
that threaten life are rarely perceived as calm or creased positivity in cortical ERPs, compared to
unarousing. The distribution of words in affective emotionally neutral words, as found in a number
space is remarkably similar to that of pictures, and of studies measuring ERPs when processing words
of sounds (Bradley and Lang, 2000), suggesting in isolation (e.g., Chapman et al., 1978; Begleiter et
that this distribution represents a relatively funda- al., 1979; Williamson et al., 1991; Skrandies, 1998;
mental organization of the affective world. Bernat et al., 2001). The emotionality effect in the
ERPs to words broadly resembles that obtained
when participants look at emotionally evocative
The potency of single words pictures (e.g., Cuthbert et al., 2000). In both cases,
ERPs to emotionally arousing stimuli diverge from
While a single word may not seem like a very po- those for neutral stimuli as early as 300 ms after
tent emotional stimulus, we know that even brief stimulus onset, have a wide topographic distribu-
presentations of individual emotionally evocative tion, and show little lateral asymmetry (though see
words can bias attention. The so-called emotional Cacioppo et al., 1996). The emotionality effect
Stroop effect, in which emotionally evocative may disappear within a few hundred milliseconds,
words, especially threatening words presented to and is rarely seen at latencies above 800 ms after
highly anxious, ‘‘hyperattentive’’ individuals, pro- word onset.
duces greater interference with naming of the color
of the word relative to neutral words, suggests an
early, possibly preattentive emotional response to Early vs. late components to emotional words
these words (see, e.g., Dalgleish, 1995; Williams
et al., 1996). Similarly, emotional words presented In view of the effects of word emotionality on
during a rapid series of briefly presented words tasks like the RSVP detection task noted above,
(the ‘‘RSVP’’ procedure) can ‘‘capture’’ attention, one might expect that ERPs to emotional words
improving detection of emotional words and tar- might be distinguishable from neutral words quite
gets, and impairing detection of other target words early in processing, and that these early compo-
that follow them, the ‘‘emotional blink’’ effect (see, nents might be less task dependent than later
e.g., Keil and Ihssen, 2004; Anderson, 2005; Kasc- differences, which in turn would be more respon-
hub and Fischler, 2006). sive to specific task demands. Bernat et al. (2001)
Chapman and colleagues (e.g., Chapman et al., reported an early (P1-N1) divergence of ERPs to
1980) explored the pattern of ERPs evoked by mood-descriptor adjectives (e.g., lonely, happy),
single emotional words presented in isolation. with unpleasant adjectives more positive than
Their resulting principle component analyses re- pleasant. The early difference was confined to the
vealed consistent ERP factors associated with each left hemisphere sites, and was observed both with
of Osgood’s three dimensions of ‘‘connotative supraliminal (40 ms) and subliminal (1 ms, un-
meaning’’ — evaluation, activity, and potency — masked) presentation of the words. On the other
corresponding roughly to the dimensions of he- hand, each of the 32 descriptors was shown mul-
donic valence, arousal, and dominance widely used tiple times during the study, and participants had
to standardize emotional materials (Bradley and also generated emotionality ratings for those terms
Lang, 1994). A similar set of components was on a number of occasions before the physiological
found when the words were being rated on dimen- phase of the study. In addition, neutral mood
sions of the semantic differential, as well as when terms were not included, and are important in or-
the words were simply read out loud, suggesting der to evaluate whether the effects were driven by
that the emotional response to these words was to emotional arousal, or by hedonic valence. More
some degree independent of the particular way the recently, Herbert et al. (in press) presented adjec-
participants needed to process them. tives that were pleasant, unpleasant, and neutral
189

and asked participants to make a covert judgment cases, there has been no primary task other than to
regarding their emotional response. The earliest read the words silently, a condition which, given
significant differences were for ERPs to emotional the nature of the materials, may default to a con-
words (pleasant and unpleasant) becoming more sideration of emotionality. In a series of separate
positive than neutral words in the P2 component experiments, we contrasted the nature of the en-
(180–250 ms post-onset), which continued into P3 coding task when processing the same set of
(250–400 ms). The difference was predominantly in words. In some of these studies, one or the other
the left hemisphere, and largest at the central and dimension of word emotionality was relevant. In
parietal sites. others, the emotional attributes of the words were
irrelevant to the decision. Normative ratings of
emotionality were carefully controlled, and we ex-
Pleasantness vs. emotionality
amined the consequence of task manipulations on
the latency and morphology of emotionality
Subjectively, the hedonic valence of a stimulus or
effects in the ERPs, which were recorded at five
event appears to be more salient than differences
scalp locations (see below).
in emotional arousal, especially for printed words.
In these studies, a set of 150 words was drawn
Indeed, as noted above, ratings of hedonic valence
from the ANEW (Bradley and Lang, 1998). The
for words range much more widely than ratings of
stimuli included five subsets of words from
arousal (Bradley and Lang, 1998). Although
different regions of affective space: pleas-
differences in ERPs between pleasant and unpleas-
ant, high-arousal words (e.g., passion), unpleas-
ant materials have sometimes been reported, there
ant, high-arousal words (e.g., vomit), pleasant,
is no consistent pattern that distinguishes between
low-arousal words (e.g., bunny), unpleasant, low-
them. More importantly, when neutral words
arousal words (e.g., ugly) and neutral, low-arousal
are included in the stimulus set, the ERP pat-
words (e.g., rattle). For clarity, this chapter fo-
tern typically indicates greater positivity over
cuses on data that compare ERPs for the most
central–parietal sensors for both pleasant and
arousing emotional stimuli (i.e., high-arousal
unpleasant words, relative to neutral words (e.g.,
pleasant and unpleasant words) as they differ
Williamson et al., 1991), with little difference be-
from neutral, low-arousal words. In each stimulus
tween the two classes of emotional words. When
set, both nouns and adjectives were included, but
differences have been found between pleasant and
there were no words directly describing emotional
unpleasant words, it is more commonly the case
states of either type (e.g., happy or happiness).
that unpleasant words elicit a larger and/or earlier
Word length, frequency, and imagery ratings were
divergence from neutral stimuli, than do the pleas-
also balanced across the sets.
ant words (e.g., Ito et al., 1998). Taken together,
In each of five studies, each of the 150 critical
these data suggest that it may be the intensity of
words was presented once, in random order, dur-
motivational activation — whether initiated by
ing the course of the experiment. Each trial began
activity in appetitive or defensive neural systems
with a fixation box signaling the start of the trial.
— that is critical in modulating the ERP during
Around 1/2 s later, one of the words was pre-
language processing.
sented, centered on the point of fixation. Word
duration was 175 ms, but the fixation box re-
Effects of encoding task on ERPs to single mained visible until 2 s after word onset. EEG was
emotional words sampled at a 125 Hz rate, with a bandpass of
0.1–40 Hz during the period from 100 ms prior to
One factor that could clearly impact the effects of word onset to 1000 ms after word onset. EEG was
emotionality on the ERP across studies is that of then rereferenced offline to the average of the
the encoding task. Most typically, participants are mastoids, and trials with blinks and other artifacts
overtly or covertly evaluating the emotionality or were rejected prior to averaging. Participants with
pleasantness of the presented words. In some greater than 15% rejection rate were discarded,
190

and sufficient sessions were conducted to obtain unpleasant words become negative through the
data for 30 participants in each experiment. slow wave (SW) region more rapidly than do the
ERPs for neutral stimuli. This overall pattern can
be seen at each of the sensors. The contrast bet-
Decisions about pleasantness ween emotional words (pleasant and unpleasant),
on the one hand, and neutral words, on the other,
We began the series by asking a group of partic- was significant both in the LPP and the later SW
ipants to judge word pleasantness at encoding. In a region, and the size of the difference did not vary
pilot study, this task was presented as a two-choice across sites. Importantly, there was no difference
bipolar decision (pleasant, unpleasant) that was to in ERPs between pleasant and unpleasant words at
be made as soon as the word was shown. Perhaps any time point or sensor site, despite the task rel-
not surprisingly, response time for neutral words evance of this affective dimension.
was significantly slower than for affective words, The lack of ERP differences before nearly 1/2 s
as the neutral items were neither pleasant nor un- after word onset suggests that the earlier onset of
pleasant, making the decision difficult. Thus, in the emotionality effects in prior studies of single words
actual Pleasantness Decision study, we allowed may be due either to repetition of the target words,
three choices at encoding — pleasant, unpleasant or (as we believe) to the particular task given the
or neutral — and delayed the overt response until participants. The lack of any differences in the
the end of each trial. ERPs for the pleasant and unpleasant words, as
The ERPs for the pleasant, neutral, and un- salient and as task relevant as the valence distinc-
pleasant words from the Pleasantness Decision tion was in this study, is striking, and is consistent
study are shown in Fig. 2. As illustrated in this with the primarily arousal-driven nature of the re-
figure, ERPs to the pleasant and unpleasant words sponse to these stimuli as seen in the ERPs. The
diverge from that for neutral words around 450 ms later SW negativity seen for the emotional words
after onset, taking the form of an enhanced compared to the neutral words may be nothing
late positive potential (LPP) peaking around more than an enhanced contingent negative var-
500 ms. Shortly thereafter, ERPs for pleasant and iation (CNV), reflecting the fact that the decision

Fig. 2. Event-related brain potentials at five scalp locations following presentation of affectively unpleasant, neutral, and pleasant
words. In this experiment, participants decided whether each word was pleasant, neutral, or unpleasant.
191

and response selection for the pleasant and un- This shift in the ERP emotionality effect indi-
pleasant words continues to be easier than for the cates that the intensity of motivation activation
neutral words. — whether appetitively engaging or defensively
activating — is the aspect of emotionality that is
first apparent in the ERP. Both appetitively and
Decisions about emotionality defensively engaging stimuli often signal the need
for immediate attention (and often action) and the
In the second experiment, participants were re- ERP appears to reflect the behavioral and attent-
quired to classify the words as emotional (whether ional adjustments to stimuli with motivational rel-
pleasant or unpleasant) or as unemotional, neutral evance. That is, the ERP appears to reflect cortical
words. Note that if this decision were made by first processes of heightened attention and preparation
detecting whether the word was pleasant or un- for action that is common for all motivationally
pleasant, and inferring that it should be therefore salient stimuli, rather than reflecting hedonic dis-
classified as emotional, the task becomes logically tinctions. Note, however, that even with this slight
identical to the Pleasantness Decision, with an shift in the ERP, there are still no apparent differ-
added step that might slow the overall decision, but ences in the early components that are commonly
should have no impact on the ERPs. Interestingly, associated with the processes that either precede or
however, the ERPs look quite different from those direct conscious attention and awareness.
of the first experiment. As illustrated in
Fig. 3, the ERPs for emotional words now diverge
earlier and are visible as early as 300 ms post-onset, Silent reading
are significant in the N400 region and continuing
into the LPP. Also in contrast to the results of the If, as we speculated earlier, simply reading a series
data from the Pleasantness Decision study, there is of words that vary in affect might result in an
now no sustained difference between the emotional emotionality effect in the ERPs, and that emo-
and the neutral words after about 600 ms. tional arousal, rather than the pleasantness, of

Fig. 3. Event-related brain potentials at five scalp locations following presentation of affectively unpleasant, neutral, and pleasant
words. In this experiment, participants decided if each word was emotionally arousing (whether pleasant or unpleasant), or
unemotional.
192

words is the dimension that dominates the ERPs, have less ability to automatically engage compre-
we might expect that silent reading of these words hension of the meaning of the words. Nonetheless,
with no decision at all might closely resemble the given that participants could choose to be rather
data from the Emotionality Decision study. passive in their response to the words, the contin-
In the third experiment, then, the same set of ued effect of emotionality in the LPP region is an
words was presented to participants who were now impressive demonstration that even for these
asked to read the words and ‘‘think about their wholly symbolic stimuli, the derivation of mean-
meaning.’’ No overt decision or response was re- ing is, to some degree, hard to avoid, as these sin-
quired, however. Results of this study are illus- gle words continued to elicit a measurable
trated in Fig. 4. In general, the magnitude of the emotional response.
ERP emotionality effect is somewhat diminished,
compared to the previous experiments, but the
pattern is clearly more similar to that found for the Semantic categorization
emotionality decision (Fig. 3) than for the pleas-
antness decision (Fig. 2). The contrast between the In each of the preceding experiments, word emo-
emotional words and neutral words was significant tionality was either an explicit or implicit focus of
in the LPP region (450–650 ms post-onset), and as the task. In each case, a robust, arousal-driven
for the data from the Emotionality Decision study, LPP effect was seen beginning as early as 300 ms
there is no difference among the word classes in after word onset, and showing little difference be-
the SW region of the ERPs. tween pleasant and unpleasant words. We won-
The somewhat smaller effects found in this ex- dered whether these effects would continue to be
periment compared to the previous one are likely seen if the decision involved an attribute of the
due to the lack of task demands that would engage words that was orthogonal to their emotionality.
participants in actually reading the words for In the fourth study in this series, we included a
meaning. In contrast to emotional pictures, for set of 60 target words, 30 from each of two tax-
example, it is not surprising that single words may onomic categories — tools and articles of clothing.

Fig. 4. Event-related brain potentials at five scalp locations following presentation of affectively unpleasant, neutral, and pleasant
words. In this experiment, participants read each word silently and made no decision or overt response.
193

Participants were asked to press a response button for the unpleasant, but not the pleasant, stimuli.
whenever they saw a word from one of these two This is quite interesting, particularly in view of the
target categories, and make no response to the re- equivalence of the ERPs for pleasant and unpleas-
maining (critical) words. The semantic decision is a ant words in the other experiments. It suggests that
relatively deep one, requiring that the meaning of the effect of hedonic valence in the present task is
each word be derived and a specific semantic at- not merely due to differences in emotional arousal
tribute evaluated. The use of two categories was — otherwise, it should have been present or en-
intended to discourage a strategy wherein partic- hanced when arousal was task relevant — but is in
ipants could maintain a list of common category fact due to the greater ability of the unpleasant
exemplars in working memory, and match these to words to attract attention under conditions when
presented words. emotionality is irrelevant to the task.
Results from this task are shown in Fig. 5. De-
spite the irrelevance of emotion to the task, there
were again differences due to word emotionality in Lexical decision
the ERPs in the LPP region that is maximal at the
Cz location, somewhat larger on the left posterior In the semantic decision task, a rather deep se-
(P3) site, and absent on the right anterior (F4) site. mantic analysis of the words was needed, in order
In contrast to when emotion was task relevant, to categorize them. The continued presence of an
however, there is no sign of any divergence until emotionality effect in the ERPs for the unpleasant
well into the rising portion of the LPP, and the words led us to wonder if a more superficial task
effect was not significant until 500 ms after word that still required lexical access would further di-
presentation. Moreover, the difference comes and minish or eliminate the emotionality effect. On the
goes within a span of 200 ms. Again, there is no sign one hand, deciding whether or not a string of let-
of any emotionality differences in the SW region. ters is a word does not require accessing the mean-
Also in contrast to the previous studies, the ERP ing of the word. On the other hand, a long
is only significantly different from neutral words tradition of research on semantic priming effects in

Fig. 5. Event-related brain potentials at five scalp locations following presentation of affectively unpleasant, neutral, and pleasant
words. In this experiment, participants pressed a key if a word was a member of one of two semantic categories (tools or articles of
clothing, represented by a set of filler words), and made no response to the critical words, none of which was from those categories.
194

the lexical decision task (e.g., Meyer and Schvane- evocative words elicit a fairly rapid brain response
veldt, 1971; Fischler, 1977; Holcomb, 1988) sug- as seen in the ERPs. For the most part, pleasant
gests that the core meaning of a word may be and unpleasant words affect the ERPs similarly,
automatically activated during this task. A final and there is little evidence for differences in he-
experiment in the series was therefore conducted in donic valence of word stimuli across these studies.
which participants were shown the same set of 150 The one exception, in the semantic decision task,
critical words, along with 60 nonwords that had suggests that unpleasant or threatening words are
been created by changing the target words from the more likely to elicit a response under conditions
two taxonomic categories (tools and clothing) by a when participants are not oriented toward dimen-
single letter. Participants were to press a response sions of emotionality. Clearly, too, there are some
button whenever a nonword was shown, and to limits to the automaticity of even this emotional
make no overt response to any of the words. attention capture by single words, as seen by the
The ERPs obtained during this lexical decision lack of emotionality differences in the ERP when
task are presented in Fig. 6. The robust emotion- people need only to make lexical decisions.
ality effect for the unpleasant words that was ob- The absence of ERP differences as a function of
tained in the Semantic Decision study has been word emotionality earlier than about 300 ms in
largely eliminated, even over the central sites. Al- any of these experiments, even when the encoding
though there is a slight trend toward an emotion- task directly focuses on the emotional meaning of
ality effect at the vertex, the ERP difference among words, is striking, and contrasts with at least some
the classes of words did not reach significance at demonstrations of such effects with single words
any site or in any temporal window. These results discussed earlier (e.g., Bernat et al., 2001). Given
suggest that when words are read but there is no that estimates of the time needed for ‘‘lexical ac-
need to consider their meaning, and emotionality cess’’ — the activation of the representation of a
is irrelevant to the decision, affective attributes of word as a linguistic object in memory — is of the
words have little effect on the reader. order of 200 ms, and that presumably the associ-
Overall, our studies of ERPs to single words ations that link that word to emotional experiences
show that under a variety of tasks, emotionally or ideas must also take time to be activated, it

Fig. 6. Event-related brain potentials at five scalp locations following presentation of affectively unpleasant, neutral, and pleasant
words. In this experiment, participants pressed a key if the item was a nonword, and made no response to the words.
195

perhaps would be more surprising than not to find word is processed, and may reflect the affective
emotionality effects in cortical ERPs for words meaning of the word pair rather than that of the
prior to 200 ms. second word alone. Hence, pairs such as starving
murderer and resting murderer should show no
emotionality effect at all for starving and resting,
When words combine: comprehension of simple but a larger emotionality effect, as measured on
emotion phrases the second word, for the pair starving murderer
than for the pair resting murderer.
Words are the fundamental semantic units of lan- The second question regarding word emotion-
guage. But the combination of words into phrases ality addressed in this research concerns the rela-
and sentences provides language with its expres- tionship between the hedonic valence of the first
sive power, as these combinations often convey and second words. Superimposed on any arousal
meanings that could not be predicted from the effects of the modifier and/or noun, the affective
meaning of the individual words (e.g., Murphy, congruence of the words of a pair may affect ERPs.
1990). The noun phrase, in a sense, is thus the Pairs with mismatched affect between modifier and
fundamental unit of discourse and communication noun (e.g., laughing murderer) may elicit a distinc-
through language. The second series of experi- tive ERP pattern compared to pairs that match in
ments to be reviewed here concerns how the hedonic valence (e.g., dead murderer), especially
affective response to emotionally potent simple when the task involves comprehension of the pair
two-word phrases compares to that obtained when as a coherent phrase. Note that this effect could be
words within a phrase are treated as isolated units. independent of any arousal-based effects of the
Given the stable pattern of ERPs to emotional first and second word separately, since emotional
words in isolation, what might we expect when arousal is the same for the matched and mis-
words are combined into meaningful phrases? It matched pairs.
may be that the effects of emotionality on the ERP On the one hand, there is some evidence for
to words within sentences are no more than the emotional congruence effects in word processing
sum of the ERP to each word presented in isola- tasks, including the so-called affective priming
tion. If this is the case, we should see an emotion- effect, in which decisions about emotional target
ality effect to the first word of the pair with the words are speeded by prior presentation of affec-
same time course as that for emotional words in tively congruent prime words (e.g., death– aggres-
isolation. Hence, a pair such as starving murderer sive) and slowed by incongruent prime words (e.g.,
should show an increased positivity beginning death– happy) (e.g., Fazio et al., 1986; Klauer,
around 250–300 ms to starving than to a neutral 1998; Wentura, 2000). This priming effect appears
word such as resting. Note that, as a consequence, to require a rather short (c. 200–300 ms) interval
the ERP to murderer may also differ in these cases between prime and target onset (stimulus onset
because of carry-over effects from the first to the asynchrony, or SOA; see Bargh et al., 1992).
second word. On the other hand, as we reviewed above, there
An alternative hypothesis, more consistent with is little evidence that affective valence modulates
the view that words in linguistically coherent se- the ERPs to individual emotional words upon
quences may be processed differently from those in which a valence-driven congruence effect might be
isolation (see Bellezza, 1984), is that the emotional based (although see Ito et al., 1998). Moreover, in
response of the first word may be deferred until it the present studies, a relatively long SOA between
is known what it is modifying. Compare, for ex- first and second words of a pair (750 ms) was used,
ample, a dead tyrant, or a dead puppy. A dead as we wanted to allow enough time for the first
tyrant may well be a good thing; a dead puppy is word to elicit the emotionality response, and to
usually not. On the other hand, a dead letter is detect this emotionality component if it is present.
fairly neutral. According to this view, effects of We also wanted to maximize the potential impact
emotionality may not develop until the second of the emotionality of the first word on the second
196

by presenting the second word at a point when the frequency, and rated imagery value. Three addi-
effects of arousal, based on our studies with word tional sets of twenty words were also selected from
in isolation, might be at a peak (500–700 ms post- the ANEW, again comparable in length, fre-
onset). For these reasons, we anticipated that the quency, and imagery, to be used as first words in
effects of matching or mismatching the hedonic the pairs (see Table 1). Each of these words was
valence of the modifier and noun of our pairs paired with a pleasant, neutral, and unpleasant
might have little impact on the ERPs to the second second word such that each combination made a
word in these experiments. relatively coherent (though sometimes semanti-
In each of the experiments presented here, par- cally odd) adjective–noun phrase (e.g., terrible su-
ticipants were shown a series of word pairs that icide, slow suicide, happy suicide). In addition, a set
formed modifier–noun phrases. In the first exper- of 20 word pairs was constructed such that the first
iment, the task was to judge if the pair formed a and second words did not form a ‘‘coherent’’
‘‘linguistically coherent phrase.’’ In the second phrase, for example, foot sentence.
study, in contrast, the same pairs were presented, Participants were told that the study concerned
but the task was to judge if either word came from brain activity associated with understanding simple
a target semantic category. Our hypothesis was phrases and making decisions regarding the coher-
that the phrase task would produce diminished ence of simple phrases. No mention was made of
emotionality effects to the first word, and en- the emotionality of the words. Examples of linguis-
hanced effects to the second, compared to the tically ‘‘coherent’’ and ‘‘incoherent’’ word pairs
word task. were given, and it was explained that an overt
response should be made only to pairs judged as
incoherent, which would occur fairly infrequently.
Comprehension of noun phrases On each trial, a fixation mark was followed by the
first word, centered at the point of fixation. The
Nine sets of 20-word pairs each were created, with word remained visible for 250 ms. The second word
emotionality of the first and of the second words was presented in the same location 750 ms after first
(pleasant, neutral, and unpleasant) controlled and word onset, and also remained visible for 250 ms.
varied factorially. Three sets of twenty nouns each Grand averaged ERPs were obtained across 20
were selected from the ANEW (Bradley and Lang, participants for each of the nine experimental
1998) to be used as second words in these noun word pair conditions. The ERPs to the first words
phrase pairs. The sets varied in rated pleasantness, as a function of emotional class (and averaged
with both pleasant (e.g., lover) and unpleasant across second word class), are shown for five cor-
(e.g., suicide) words rated similarly higher in tical sites in Fig. 7. Most notably, as predicted,
arousal than were neutral words (e.g., banner, there was no difference between the pleasant, neu-
see Table 1). The sets were comparable in length, tral, and unpleasant words at any point prior to
onset of the second word.
Table 1. Mean ratings of valence (pleasantness) and arousal The ERPs to the second words of the phrase,
(from the Bradley & Lang, 1998, norms) for the words used in averaged across first word emotionality, are shown
the Word Pair experiments (see text)
in Fig. 8. As illustrated in the figure, emotional
First Words Second Words second words — both pleasant and unpleasant —
elicit a more positive-going ERP than do neutral
Valence Arousal Valence Arousal
words, diverging around 350 ms post-onset at
Pleasant 7.9 6.0 8.2 6.3 centroparietal sites and continuing through the
Neutral 4.8 4.4 4.9 4.7 end of the recording epoch. The ERP emotionality
Unpleasant 2.1 5.5 1.8 6.3 effect for the word completing the phrase appears
Note: A 1–9 scale is used for the SAM ratings, with 9 the most
broadly similar to that observed when single words
pleasant/arousing. are presented and participants are judging their
emotionality (see Fig. 4). Effects of emotionality
197

Fig. 7. Event-related brain potentials at five scalp locations following presentation of an affectively unpleasant, neutral, and pleasant
first word of a two-word phrase. In this experiment, participants pressed a key when they judged the complete pair to be ‘‘linguistically
incoherent’’ (represented by a set of filler pairs; see text).

Fig. 8. Event-related brain potentials at five scalp locations following presentation of an affectively unpleasant, neutral, and pleasant
second word of a two-word phrase. In this experiment, participants pressed a key when they judged the complete pair to be ‘‘lin-
guistically incoherent’’ (represented by a set of filler pairs; see text).

were significant in a 350–450 ms time window, in words, although the comparison between pleasant
the LPP time window (450–600 ms), and in the SW and neutral nouns was only marginal (po0.06) in
window (600–750 ms). In each case, the ERPs to the earliest time window. The difference between
the unpleasant and pleasant nouns were signifi- pleasant and unpleasant words never approached
cantly more positive than they were to the neutral significance. Moreover, the effects were present at
198

each recording site, with no significant topo- presenting occasional target words from one of two
graphic differences in the ANOVAs on the scaled prespecified semantic categories, and requesting
amplitudes. participants to monitor each word pair for the oc-
An effect of emotional congruence on the ERPs currence of one of these target words. There is sub-
in these analyses would have appeared as a sig- stantial evidence that, for sentences at least,
nificant interaction of first and second word va- comprehension of meaning at the phrase level
lence in the ERPs to the second word, with may not be automatic. For example, Fischler et
matched-valence ERPs (pleasant/pleasant and un- al. (1985; see also Fischler and Raney, 1991) found
pleasant/unpleasant) diverging from mismatched that the truth value of sentences such as A robin is a
ERPs (pleasant/unpleasant and unpleasant/pleas- bird/A robin has a bird modulated the size of the
ant). This interaction never reached significance, N400 priming effect when the task was sentence
for any site or interval. Also, there were no main verification, but not when the task was to judge the
effects of first word valence on the ERPs to the semantic relatedness of the first and second nouns
second word. (see also Brega and Healy, 1999).
Taken together, the pattern of ERPs to the Our hypothesis, then, was that the focus on in-
emotional word pairs in this experiment was dividual words rather than phrases in the second
straightforward. First, emotionality of the first experiment in this series would elicit emotionality
(modifier) words had no impact on the ERPs to effects in the ERPs to the first words. In contrast,
these words, despite the use of words whose emo- if the adjectives used were simply not potent
tionality ratings were at least as strong, in hedonic enough to elicit an emotional response, then
valence and emotional arousal, as in earlier studies changing the task should have little effect on the
of words in isolation, and despite the use of a task ERP results, with first words again showing no
requiring semantic analysis of both words, and an emotionality effect.
interval between first and second words in which Regarding the pattern of congruence effects,
emotionality effects had been clearly seen in prior there were several possibilities. Klauer and Stern
work. Second, there were clear effects of emotion- (1992) have suggested that affective congruence
ality of second words — the modified nouns — on effects depend on the ease with which the prime
ERPs. Third, as with earlier studies, these ERP and target words can be implicitly thought of as a
effects were driven by the emotional intensity pair. This would imply that congruence effects
(arousal) of the words rather than by differences in should have been observed in the phrase task,
hedonic valence; in no case did the pleasant and however. If we are successful in eliciting an emo-
unpleasant second words systematically differ tionality response to the first word by requiring a
from each other in the ERP. decision to it, this might set the stage for such
match/mismatch effects between the first and sec-
ond words. Nonetheless, the SOA remains sub-
Comprehension of single word within noun phrases stantially longer than those typically found to
produce affective priming effects (see Bargh et al.,
The lack of emotionality effects for the first words 1992; Wentura, 2000), and thus congruence effects
in this experiment is presumably a consequence of were not strongly expected in this experiment.
the coherence encoding task, which required com- The design and materials of this experiment were
prehension of the word pair as a phrase. The affec- identical to that of the phrase coherence task, with
tive meaning of the first word thus must await the the exception that the twenty incoherent pairs were
completion of the entire phrase. To test this hy- modified so that one of the two words came from
pothesis, the same word pairs were presented with one of two semantic categories: tools and articles of
the same timing, using an encoding task that di- clothing. Participants were told to read each word
rected attention to the meaning of each individual of each pair, and press a response button when a
word of the pair, rather than the phrase as a unit. In word from one of the two target categories was
the last experiment to be reviewed, this was done by shown. Several examples of category instances were
199

Fig. 9. Event-related brain potentials at five scalp locations following presentation of an affectively unpleasant, neutral, and pleasant
first word of a two-word phrase. In this experiment, participants pressed a key when they judged either of the two words of the pair to
be a member of one of two semantic categories (tools or articles of clothing, represented by a set of filler pairs; see text).

given. As before, the SOA was 750 ms, and each valence) are shown for each recording site in
participant saw all 200 pairs, in random order. Fig. 10. The only time window resulting in a sig-
The ERPs to the first words, as a function of nificant effect of word emotionality is 450–600 ms
hedonic valence (and averaged across second word after onset of the second word. As before, con-
valence), are shown for the five cortical sites in trasts showed that the difference is due to greater
Fig. 9. The major morphological features of the positivity for ERPs to pleasant and unpleasant
ERPs at the frontal and central sites, as well as the words, compared to neutral words, with no differ-
more complex waveform parietally, correspond ence in the ERPs between the emotional words.
closely to those observed in the noun phrase task. The effect did not differ across electrode site.
The most notable overall difference from the pair There was little evidence of an interaction between
coherence experiment is that now there is a sub- first and second word valence at any site except for a
stantial emotionality effect in the ERPs. Emo- slight trend in the 600–750 ms time window at P3.
tional words (either pleasant or unpleasant) again The pattern of amplitudes across conditions at this
prompt more positive potentials, compared to interval was complex, however, and did not easily
neutral words, which is apparent around 400 ms lend itself to an affective priming interpretation.
and continues into the LPP region at centrofrontal In sum, the data clearly show that when the
sites. The effects, which were significant in the words of a pair that potentially form a coherent
350–450 ms interval and were marginal just before phrase must be comprehended as individual
and after this time window, were entirely due to a words, the emotionality of the first as well as the
difference between emotional and neutral condi- second word results in an enhanced positive-going
tions, with ERPs to pleasant and unpleasant words ERP. This contrasts with the ERPs to the identical
essentially identical to each other and both differ- words when the two items are treated as a noun
ent from that elicited for neutral words. phrase. On the other hand, there was no evidence
The ERPs to the second words as a function of that emotional congruence of the items affected
hedonic valence (and averaged across first word ERPs to the second word in the pairs.
200

Fig. 10. Event-related brain potentials at five scalp locations following presentation of an affectively unpleasant, neutral, and pleasant
second word of a two-word phrase. In this experiment, participants pressed a key when they judged either of the two words of the pair
to be a member of one of two semantic categories (tools or articles of clothing, represented by a set of filler pairs; see text).

Emotionality effects on the ERPs to the first and rather than determined by a sequential cascade of
second words of a word pair, then, was critically emotionality effects to each word in turn. The emo-
affected by the task. When the word pairs were tionality effect for second words in the noun-phrase
considered as phrases, effects of emotionality of task was greatest, in fact, not to the most emotion-
the first word (i.e., the modifier) were wholly ab- ally arousing individual words, but to the most
sent, but when the pairs were treated as two sep- emotionally arousing phrases, as shown by a post
arate words, a significant emotionality effect was hoc analysis of emotionality ratings of the phrases.
observed in the ERP elicited by the first word. This The contrasting effects of emotionality in the
demonstrates that the absence of these effects in two experiments suggest limits to the automaticity
the noun phrase task is not due to the particular of comprehension of emotional aspects of word
words used, or to the word class of adjectives vs. meaning, in at least two ways. First, consistent
nouns. Rather, we conclude that when word se- with the findings of our studies of single words,
quences can form phrases, and are treated as such, even when the meaning of words is being proc-
the emotional response to the modifying adjectives essed, as in both experiments here, the response to
is deferred until, so to speak, one knows what is their emotionality depends on the current task,
being modified: is it a tyrant, a puppy or a letter with some tasks failing to elicit any emotionality
that’s dead? That is, the hedonic valence is only effects in the ERPs. Second, consistent with the
apparent at the level of the word phrase, rather findings of Fischler et al. (1985) and others, the
than at the level of the individual words. meaning of phrases and clauses requires attention
The shift from phrase- to word-level tasks had at that level of analysis. When words in a sequence
the opposite effect on the emotional response to the are treated as isolated lexical events, on the
second words, reducing the ERP differences in size, other hand, even if they have the potential to be
as well as in temporal and topographic extent. This integrated into phrases, the impact of phrase-level
reduction is consistent with the view that for meaning is attenuated.
phrases, the emotionality effect in the ERPs is In contrast to the effects of word emotionality
wholly driven by the emotionality of the phrase, and task on the ERPs, there was little effect of the
201

congruence or incongruence of emotional valence parietal sites, with only a slight trend toward a left-
on the ERPs in either experiment. Since we did maximal lateral asymmetry. With few exceptions,
observe what might be considered a congruence there is little difference between pleasant and un-
effect in the behavioral data — the coherence rat- pleasant words in these ERPs.
ings were higher, and the probability of ‘‘incoher- This pattern of ERP modulation is strikingly
ent’’ responses was lower for emotionally similar to that found when people view emotional
congruent than incongruent pairs — it may be pictures: in both cases, a late positive potential,
that the ERP measure is simply insensitive to the emerging around 300–400 ms after stimulus onset
hedonic valence dimension of emotionality of and maximal over central–parietal sensors is ap-
words. However, this seems unlikely given the sev- parent, which is heightened for emotionally en-
eral positive findings of congruence effects in very gaging, compared to neutral, stimuli (e.g.,
different paradigms (e.g., Cacioppo et al., 1993; Cuthbert et al., 2000). For both pictures and
Chung et al., 1996). Alternatively, such effects with words, ERPs primarily reflect the intensity or
the present materials and tasks may be observable arousal parameter of emotion, rather than differ-
only when the SOA between first and second ences in hedonic valence. One interpretation is that
words is much shorter, as is the case for the affec- the late positive component of the ERP reflects
tive priming studies (see above) — under 300 ms. heightened attention allocation to motivationally
To summarize the main results, the pattern of salient stimuli, which is implemented in cortical
emotionality effects for ERPs suggests that when structures and initiated by activity in subcortical
words are combined into phrases, and treated as structures differentially engaged by appetitive and
phrases, the emotional response to these pairs is defensive stimuli. Because the ERP only weakly
driven by the emergent meaning of the pair as a reflects differential subcortical activity, however, it
linguistic unit, rather than a sequential unfolding of instead reflects the heightened attention and en-
emotionality effects to individual words that cas- gagement that is found when emotional (whether
cades into an overall emotionality effect. This is pleasant or unpleasant) cues are encoded.
seen in the absence of emotionality effects to the Collectively, the studies reviewed here also show
first words and the enhanced emotionality effects to how the specific encoding task, and also whether
the second word (compared to our prior work, at the words are presented in isolation or embedded
least). In this view, the absence of systematic con- in other words as part of phrases impact the pres-
gruence effects between first and second words also ence, timing and extent of ERP emotionality
makes sense: if there is no emotionality response to effects. It is a reminder that, as claimed by Ge-
the first word as such, then there is nothing to be stalt psychologists and ‘‘interactionist’’ psycholin-
congruent or incongruent with the second word. guists, a sentence is more than the sum of its lexical
What matters is the meaning of the phrase. parts. To understand how emotion is compre-
hended through language, it is necessary to con-
sider both the local and global aspects of language
Comprehending words, phrases, and sentences processing, as well as the larger context in which
we hear or read of emotional events.
Across these studies, there is consistency in how
emotionality affects the brain’s response to words.
ERP differences generally took the form of a Acknowledgments
broad increase in the magnitude of a late positive
potential for pleasant and unpleasant words, rel- The research described here was supported by
ative to neutral words. This emotionality effect NIMH grants to each author as investigators in
occurs around a half second after word onset, al- the NIMH Center for the Study of Emotion and
though it can emerge as early as 300 ms post-onset, Attention (P50-MH52384) at the University of
and can endure for several hundred milliseconds Florida. The first series of experiments, on
or longer. The effect is maximal at central and effects of single emotional words, was done in
202

collaboration with Romko Sikkema, Vincent van Chapman, R.M., McCrary, J.W., Chapman, J.A. and Bragdon,
Veen, and Michelle Simmons. The second series of H.R. (1978) Brain responses related to semantic meaning.
Brain Lang, 5: 195–205.
experiments, on effects of emotional noun phrases,
Chapman, R.M., McCrary, J.W., Chapman, J.A. and Martin,
was done in collaboration with Michael McKay, J.K. (1980) Behavioral and neural analyses of connotative
David Goldman, and Mireille Besson. Portions of meaning: word classes and rating scales. Brain Lang, 11:
the first noun phrase experiment formed the basis 319–339.
of David Goldman’s senior thesis at the University Chung, G., Tucker, D.M., West, P., Potts, G.F., Liotti, M. and
of Florida. We acknowledge the assistance of nu- Luu, P. (1996) Emotional expectancy: brain electrical activity
associated with an emotional bias in interpreting life events.
merous undergraduate students in data collection Psychophysiology, 33: 218–233.
and analysis, notably Michael Mattingly, Moshe Cuthbert, B.N., Schupp, H.T., Bradley, M.M., Birbaumer, N.
Feldman, Kim Anderson, Carla Hill, Candice and Lang, P.J. (2000) Brain potentials in affective picture
Mills, and William Lowenthal. processing: covariation with autonomic arousal and affective
report. Biol. Psychol., 52: 95–111.
Dalgleish, T. (1995) Performance on the emotional Stroop task
References in groups of anxious, expert, and control subjects: a com-
parison of computer and card presentation formats. Cogn.
Anderson, A.K. (2005) Affective influences on the attentional Emotion, 9: 341–362.
dynamics supporting awareness. J. Exp. Psychol. Gen., 134: Duffy, E. (1941) An explanation of ‘‘emotional’’ phenomena
258–281. without the use of the concept ‘‘emotion’’. J. Gen. Psychol.,
25: 283–293.
Arnold, M.B. (1945) Physiological differentiation of emotional
states. Psychol. Rev., 52: 35–48. Fazio, R.H., Sanbonmatsu, D.M., Powell, M.C. and Kardes,
Bargh, J.A., Chaiken, S., Govender, R. and Pratto, F. (1992) F.R. (1986) On the automatic activation of attitudes. J. Pers.
The generality of the automatic attitude activation effect. Soc. Psychol., 50: 229–238.
Fischler, I. (1977) Associative facilitation without expectancy in
J. Pers. Soc. Psychol., 62: 893–912.
Begleiter, H., Porjesz, B. and Garosso, R. (1979) Visual evoked a lexical decision task. J. Exp. Psychol. Hum. Percept. Per-
potentials and affective ratings of semantic stimuli. In: Beg- form., 3: 18–26.
Fischler, I., Boaz, T., Childers, D. and Perry, N.W.J.
leiter, H. (Ed.), Evoked Brain Potentials and Behavior. Ple-
num Press, New York. (1985) Lexical and propositional components of priming
Bellezza, F.S. (1984) Reliability of retrieval from semantic during sentence verification [Abstract]. Psychophysiology,
memory: noun meanings. Bull. Psychon. Soc., 22: 377–380. 22: 576.
Fischler, I. and Raney, G.E. (1991) Language by eye: behavi-
Bernat, E., Bunce, S. and Shevrin, H. (2001) Event-related
brain potentials differentiate positive and negative mood ad- oral and psychophysiological approaches to reading. In:
jectives during both supraliminal and subliminal visual Jennings, J.R. and Coles, M.G.H. (Eds.), Handbook of Cog-
processing. Int. J. Psychophysiol., 42: 11–34. nitive Psychophysiology: Central and Autonomic Nervous
System Approaches. Wiley, Oxford, England, pp. 511–574.
Bradley, M.M. and Lang, P.J. (1994) Measuring emotion: the
Self-Assessment Manikin and the semantic differential. Herbert, C., Kissler, J., Junghoefer, M., Peyk, P. and Rocks-
J. Behav. Ther. Exp. Psychiatry, 25: 49–59. troh, B. (2006) Processing of emotional adjectives — evidence
from startle EMG and ERPs. Psychophysiology, 43:
Bradley, M.M. and Lang, P.J., (1998). Affective norms for
English words (ANEW): instruction manual and affective 197–206.
ratings. Technical Report A-8, The Center for Research in Holcomb, P.J. (1988) Automatic and attentional processing: an
Psychophysiology, University of Florida. event-related brain potential analysis of semantic priming.
Brain Lang, 35: 66–85.
Bradley, M.M. and Lang, P.J. (2000) Measuring emotion: be-
havior, feeling and physiology. In: Lane, R. and Nadel, L. Ito, T.A., Larsen, J.T., Smith, N.K. and Cacioppo, J.T. (1998)
(Eds.), Cognitive Neuroscience of Emotion. Oxford Univer- Negative information weighs more heavily on the brain: the
sity Press, New York. negativity bias in evaluative categorizations. J. Pers. Soc.
Psychol., 75: 887–900.
Brega, A.G. and Healy, A.F. (1999) Sentence interference in the
Stroop task. Mem. Cogn., 27: 768–778. Kaschub, C. and Fischler, I., 2006. The emotional blink to
Cacioppo, J.T., Crites, S.L., Berntson, G.G. and Coles, M.G. words without perceptual or semantic cuing. Manuscript
(1993) If attitudes affect how stimuli are processed, should submitted for publication.
Keil, A. and Ihssen, N. (2004) Identification facilitation for
they not affect the event-related brain potential? Psychol.
Sci., 4: 108–112. emotionally arousing verbs during the attentional blink.
Cacioppo, J.T., Crites Jr., S.L. and Gardner, W.L. (1996) At- Emotion, 4: 23–35.
Klauer, K.C. (1998) Affective priming. In: Stroebe, W. and
titudes to the right: evaluative processing is associated with
lateralized late positive event-related brain potentials. Pers. Hewstone, M. (Eds.) European Review of Social Psychology,
Soc. Psychol. Bull., 22: 1205–1219. Vol. 8. Wiley, New York, pp. 67–103.
203

Klauer, K.C. and Stern, E. (1992) How attitudes guide mem- Moruzzi, G. and Magoun, H.W. (1949) Brain stem reticular
ory-based judgments: a two-process model. J. Exp. Soc. Psy- formation and activation of the EEG. Electroenceph. Clin.
chol., 28: 186–206. Neurophysiol., 1: 455–473.
Klinger, E. and Cox, W.M. (1987) Dimensions of thought flow Murphy, G.L. (1990) Noun phrase interpretation and concep-
in everyday life. Imagin. Cogn. Pers., 7: 105–128. tual combination. J. Mem. Lang, 29: 259–288.
Lang, P.J., Bradley, M.M. and Cuthbert, M.M. (1997) Moti- Osgood, C.E., Suci, G.J. and Tannenbaum, P.H. (1957)
vated attention: affect, activation and action. In: Lang, P.J., The Measurement of Meaning. University of Illinois Press,
Simons, R.F. and Balaban, M.T. (Eds.), Attention and Ori- Oxford.
enting: Sensory and Motivational Processes. Lawrence Er- Russell, J.A. (1980) A circumplex model of affect. J. Pers. Soc.
lbaum Associates, Hillsdale, NJ. Psychol., 39: 1161–1178.
Lang, P.J., Kozak, M.J., Miller, G.A., Levin, D.N. and Skrandies, W. (1998) Evoked potential correlates of semantic
McLean Jr., A. (1980) Emotional imagery: conceptual struc- meaning — a brain mapping study. Cogn. Brain Res., 6:
ture and pattern of somato-visceral response. Psychophysi- 173–183.
ology, 17: 179–192. Wentura, D. (2000) Dissociative affective and associative prim-
Lindsley, D.B. (1951) Emotion. In: Stevens, S.S. (Ed.), Hand- ing effects in the lexical decision task: yes versus no responses
book of Experimental Psychology. Wiley, Oxford, pp. to word targets reveal evaluative judgment tendencies. J. Exp.
473–516. Psychol. Learn. Mem. Cogn., 26: 456–469.
Mehrabian, A. and Russell, J.A. (1974) A verbal measure of Williams, J.M.G., Mathews, A. and MacLeod, C. (1996) The
information rate for studies in environmental psychology. emotional Stroop task and psychopathology. Psychol. Bull.,
Environ. Behav., 6: 233–252. 120: 3–24.
Meyer, D.E. and Schvaneveldt, R.W. (1971) Facilitation in Williamson, S., Harpur, T.J. and Hare, R.D. (1991) Abnormal
recognizing pairs of words: evidence of a dependence between processing of affective words by psychopaths. Psychophys-
retrieval operations. J. Exp. Psychol., 90: 227–234. iology, 28: 60–273.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 10

Emotional connotation of words: role of emotion in


distributed semantic systems

M. Allison Cato Jackson1, and Bruce Crosson2,3

1
Nemours Children’s Clinic, Neurology Division, 807 Children’s Way, Jacksonville, FL 32207, USA
2
Brain Rehabilitation Research Center, Malcom Randall VA Medical Center, 1601 SW Archer RD, Gainesville,
FL 32608-1197, USA
3
Department of Clinical & Health Psychology, University of Florida, Gainesville, FL, USA

Abstract: One current doctrine regarding lexical–semantic functions asserts separate input and output
lexicons with access to a central semantic core. In other words, processes related to word form have
separate representations for input (comprehension) vs. output (expression), while processes related to
meaning are not split along the input–output dimension. Recent evidence from our laboratory suggests that
semantic processes related to emotional connotation may be an exception to this rule. The ability to
distinguish among different emotional connotations may be linked distinctly both to attention systems that
select specific sensory input for further processing and to intention systems that select specific actions for
output. In particular, the neuroanatomic substrates for emotional connotation on the input side of the
equation appear to differ from the substrates on the output side of the equation. Implications for semantic
processing of emotional connotation and its relationship to attention and motivation systems are discussed.

Keywords: language; emotion; functional magnetic resonance imaging; semantic processing; neuroimaging

In this paper, we discuss the semantics of the emo- three studies inform theories of semantic and emo-
tional connotation (EC) of words and how the EC tion processing.
of words (positive or negative) is processed in the Before we begin, however, it is necessary to define
brain relative to words with neutral EC. We will EC and how processing EC differs from the actual
review findings from three functional magnetic res- experience of emotion. EC refers to knowledge
onance imaging (fMRI) studies that support unique about the emotional property of an object, action,
brain activity associated with processing and pro- or event. It is not a direct reference to an emotion
ducing words with EC. This series of studies in- such as happiness, sadness, anger, or fear. Rather, it
forms theories of semantic organization in the brain is implied knowledge that an object, action, or
as well as theories of emotional processing. We will event is likely to evoke a specific emotion. For ex-
review relevant theoretical background in semantics ample, our knowledge of different animals indicates
and emotional processing, prior to the presentation that bears are dangerous animals and, therefore, are
of findings from the three neuroimaging studies. In to be feared. Or, our knowledge about birthday
the ‘Discussion’ section, we will discuss how these parties indicates that these are (usually) happy
events. While such knowledge (i.e., EC) undoubt-
edly is derived, at least in part, from experience with
Corresponding author. Tel.: +1-904-390-3665; birthday parties, bears, and other objects, actions,
Fax: +1-904-390-3470; E-mail: acato@nemours.org or events that evoke emotion, a distinction can be

DOI: 10.1016/S0079-6123(06)56010-8 205


206

drawn between EC and emotional experience. heard word seen object printed word
While it is unclear whether emotions must be in-
voked to some degree to understand EC, it is safe to
phonologic structural orthographic
say that it is not necessary to experience an emotion description
input input
to understand the emotional significance of an ob- lexicon system lexicon
ject, action, or event. In other words, one does not
have to be subjected to the intense fear that one
might experience when standing face-to-face with a
bear to understand that a bear should evoke this semantics
emotion. Indeed, the fact that EC can be under-
stood and evaluated in the absence of intense emo-
tion can be a useful tool for learning about the
emotional significance of potentially dangerous sit-
phonologic orthographic
uations so that one can take steps to avoid these output output
and the strong emotions they would likely provoke. lexicon lexicon

spoken word written word


Relevant theories of semantic processing
Fig. 1. Model of lexical–semantic processing. Adapted from
Theories of lexical–semantic processing attempt to Ellis and Young (1988).
explain the relationship between stored knowledge
and the means by which we gain, access, and express access to stored knowledge. Although many case
knowledge. One current doctrine regarding lexi- studies in the literature support aspects of this
cal–semantic functions (e.g., Ellis and Young, 1988) model, in particular, the occurrence of isolated im-
asserts the existence of separate input and output pairment or preservation of one of the lexicons, the
lexicons, each of which has access to a central se- model has significant drawbacks. First, it can be of
mantic core that processes the meaning of objects, limited utility in understanding broader and more
actions, and words (see Fig. 1). An implication of common impairments of language. Second, evi-
the model is that semantic knowledge would be dence suggests that the semantic system is neither
gained primarily through heard or written words or modular nor unitary. Third, the model fails to take
through seen objects. On the input side, information into account attentional systems that select one
is filtered through the appropriate lexicon (phono- source of information among multiple sources for
logical, orthographic, or structural) before meaning further processing or intentional systems that select
is processed at the level of semantics. On the output one action for execution among multiple possible
side, a concept from semantic processing can be ex- actions.
pressed verbally through output lexicons (most For a long time, lesion studies were the primary
commonly, phonologic, or orthographic). means of building and examining models such as
According to this model, separate representations that of Ellis and Young (1988). However, over the
(i.e., lexicons) exist in the brain for each input and past decade and a half, functional neuroimaging
output modality. Thus, it is possible to have an techniques have provided in vivo methods for
impairment in one input or output lexicon but leave measuring brain activity during cognitive challenge.
other lexicons intact. Further, semantic knowledge For example, the most commonly used fMRI tech-
about objects, actions, and words can be retained in nique takes advantage of a natural contrast agent,
the face of impairment in any or all of the lexicons. deoxyhemoglobin, which occurs in red blood cells
This model also suggests that access to the meaning and is paramagnetic. A dynamic change in the pro-
of an object becomes independent of the means by portion of this natural contrast agent to oxygenated
which we learn or access that information. That is, hemoglobin creates the blood oxygen level depend-
each input or output modality has equal and full ent (BOLD) signal. The BOLD signal typically
207

increases as blood flow increases. Thus, when blood during semantic processing, and the cortex activated
flow increases to perfuse brain regions that subserve may even vary depending upon the body part used
a given activity (e.g., primary motor cortex during for the action in question (Pulvermuller, 2005).
finger tapping), the BOLD signal increases as well, However, these neuroimaging studies have not dis-
revealing active brain regions in fMRI studies. sociated attribute, category, or modality. The prob-
Over the past 25 years, lesion and subsequently lem is that the categories of items studied have been
neuroimaging evidence have contributed to our closely associated with particular attributes and/or
knowledge about the architecture of the semantic modalities, making it difficult to separate and ex-
system. Semantic attributes (e.g., visual form or plore the different components of the semantic sys-
function), semantic category (e.g., living vs. nonliv- tem. For example, form (i.e., shape) is usually
ing), and modality of processing (e.g., visual or processed in the visual modality, and when visual
verbal) have been shown to be important compo- cortex is activated for objects relying upon shape for
nents of the semantic system. Warrington and identification (e.g., animals), it is difficult to deter-
Shallice (1984) had linked semantic attributes to mine if this cortex is activated because of the specific
semantic categories, explaining category-specific feature being processed (shape) or because of the
deficits by the fact that some objects are more modality in which the feature is typically repre-
closely associated with specific semantic attributes sented. It should also be noted that few studies have
than other categories. For example, function is explored the impact of engaging attentional vs. in-
more important for distinguishing between nonliv- tentional systems in semantic processing.
ing items, such as tools (i.e., implements), and vis- We became interested in EC as a semantic attri-
ual form is more important for distinguishing bute because it is relatively independent of modality
between living items such as animals. Thus, when of processing and category of the object being proc-
visual processing is impaired, semantic processing essed. For example, one can experience disgust
of animals should be more affected than processing when presented with numerous living and nonliving
of tools. Warrington and McCarthy (1987) later objects. Further, one can experience disgust by see-
added an action-oriented domain to explain differ- ing, hearing, touching, smelling, tasting, or even
ences in processing small objects that could be ma- talking about the disgusting object, action, or event.
nipulated as opposed to large objects that could not When an object possesses an EC (whether pleasing,
be manipulated. While there is much validity to disgusting, frightening), it is a defining characteristic
these observations, they suggest an interdependence of that object, i.e., one of the object’s semantic at-
among semantic attribute, semantic category, and tributes. Although one does not have to experience a
even to some degree modality. Yet, one particularly full-blown emotion to understand EC, it is useful to
seminal lesion study by Hart and Gordon (1992) understand some aspects of emotional processing.
demonstrated that modality of processing, semantic Such an understanding will enable us to draw im-
attributes, and semantic category could be frac- portant distinctions between emotional experience
tionated from one another, suggesting that the neu- and EC. Thus, before discussion of our observations
ral substrates for these dimensions might differ in regarding brain activity during processing words
some important ways. with positive or negative EC relative to words with
In functional neuroimaging studies, disparate pat- neutral EC, a brief discussion of relevant emotion
terns of brain activity have been demonstrated for theory is offered.
processing information about animals vs. processing
information about tools (e.g., Damasio et al., 1996;
Martin et al., 1996). One such focus has been in the Relevant theories of emotion processing
ventral visual stream, where different regions appear
to be involved in processing living vs. nonliving Emotion processing theories attempt to predict the
things (e.g., Chao et al., 1999; Ishai et al., 1999). nature and location of processing emotionally sali-
Further, processing movement or action-related ent information in the brain. While it is well estab-
words tends to activate motor or premotor cortices lished that limbic areas, including the orbitofrontal
208

cortex, the amygdala, and the hypothalamus are in- did not address whether the brain regions that
volved in the experience of emotions such as fear, process words with EC differ when the task at
rage, and joy, it is not as clear what brain regions hand weighs more heavily on intentional or at-
mediate processing of EC. With respect to viewing tentional demands.
emotionally salient pictures, evidence has been Intention refers to the selection of one among
found to support two competing hypotheses. The multiple potential actions for execution and initi-
right hemisphere hypothesis states that emotionally ation of that action (Fuster, 2003; Heilman et al.,
salient information is biased for processing in the 2003; Crosson et al., in press). The relevance of
right hemisphere, relative to emotionally neutral intention to emotion is clear. Emotions have been
material. On the other hand, the bivalent hypothesis defined as action dispositions (Lang, 1995). Be-
suggests that information that is negatively valenced cause emotions act as motivators for action, they
is biased for processing in the right hemisphere, will be a factor in selecting one course of action vs.
while the left hemisphere mediates processing of others. Specifically, objects, actions, and events
positive EC. Evidence for these hypotheses come with negative ECs generally will be avoided, and
from lesion studies (e.g., Blonder et al., 1991; Peper objects, actions, and events with positive ECs nor-
and Irle, 1997) and EEG and neuroimaging studies mally will be sought out. Attention is the selection
(e.g., Davidson, 1995; Lane et al., 1997a, b; Canli et of one source of information among multiple
al., 1998; Lang et al., 1998) in healthy adults. In sources of information for further processing and
most cases, pictures with emotionally salient vs. initiation of that processing. Emotionally salient
neutral material have been used to examine brain stimuli will be attended to because of their impli-
processing of EC. However, processing words with cations for guiding behavior. We will develop these
EC is less likely to evoke emotional reactions in in- themes in greater detail as we discuss the implica-
dividuals, relative to viewing pictures. Thus, it is tions of our findings.
likely that processing words with EC will lead to
different patterns of activity than viewing pictures of
actual items with EC. Fewer studies have examined Review of three studies from our laboratory
brain correlates of processing words with EC.
Maddock and Buonocore (1997) used fMRI to By conceptualizing EC as a semantic attribute and
examine brain activation during processing of with support of previous neuroimaging studies of
threat-related vs. neutral words. They found left processing words with EC, we predicted that
retrosplenial activity that was unique to processing processing words with EC would lead to a pattern
threat-related relative to neutral words. They did of brain activity in the left hemisphere different from
not find evidence to support either the right hemi- that of processing words with no strong EC. In
sphere or bivalent hypothesis. In a follow-up particular, processing the EC of words should
study, Maddock et al. (2003) found rostral fron- engage areas of the brain with access to emotions
tal and posterior cingulate activity when evaluat- and/or emotional knowledge. In addition, we sus-
ing either pleasant or unpleasant compared to pected that the brain regions associated with
neutral words; however, these authors could not processing words with EC would be the same
distinguish between the roles of these two regions whether the task was more heavily weighted toward
on the basis of their data. Similarly, Beauregard et processing incoming information (i.e., engaged at-
al. (1997) used PET to investigate brain activity tentional systems) vs. more heavily weighted toward
during viewing of words with EC. Findings of producing a response (i.e., engaged intentional sys-
unique brain activation associated with processing tems). In other words, we predicted that the brain
EC included unique activity in left prefrontal and regions processing EC would be invariant across
left orbital frontal cortex. Thus, previous neuroim- task demands that emphasize input vs. output.
aging studies of EC of words suggested that left To examine the role of EC in lexical–semantic
frontal and posterior cingulate regions mediate processing, our group developed three paradigms.
verbal processing of EC. These previous studies Two of these tasks involved generation of multiple
209

words from given semantic categories and a third motivation into action. Attention properly selects
semantic monitoring task involved making seman- one among several competing sources of sensory
tic judgments about given words. All tasks en- information for further processing and regulates ac-
gaged the phonological input lexicon to some tivity in the posterior realm of sensory processing.
degree (spoken words were the input stimulus in Heilman et al. (2003) have suggested that posterior
all cases). However, the tasks differed by the rel- cingulate cortex (including retrosplenial cortex) is
ative emphasis on attentional vs. intentional task involved in selecting the stimuli that are motiva-
demands. The two word generation tasks relied tionally/emotionally significant.
more on the intentional system — the system that This analysis suggests an alternative to our orig-
mediates response selection and production. The inal hypothesis of no difference in processing EC
semantic monitoring task placed heavier demands based on input or output. If processing EC is in-
on the ability to attend to spoken words and make timately associated with attention and intention, the
semantic decisions. This task involved some re- degree to which a task emphasizes sensory process-
sponse selection as well, but the primary process- ing vs. action may determine what mechanisms are
ing demands were on the input side. The results of engaged to perform the task. Across the three stud-
these three studies are previously published and ies that follow, we present findings that suggest
described in great detail elsewhere (Crosson et al., (contrary to our original predictions) that the re-
1999, 2002; Cato et al., 2004). trosplenial region, in more posterior cortex, may
In this paper, we will evaluate the results of the have a more sensory, or attentional, role in seman-
three fMRI tasks to determine what conclusions we tic processing in contrast with the frontal region.
can draw about the role of EC in lexical–semantic The findings also suggest that the rostral frontal
processing. In particular, we will examine the re- region may have more of a role in the intentional
sults to address whether (1) unique brain activity aspects of word production.
was consistently associated with processing EC and
whether (2) the pattern of activity differed across
tasks that were more heavily weighted towards Question 1: Does processing emotional connotation
attentional vs. intentional demands. of words lead to unique patterns of brain activity
As noted above, EC is intimately associated with relative to processing words with neutral emotional
intention, providing the motive to select which ac- connotation?
tion is performed in many circumstances. In other
words, emotions provide an important substrate for In the first and the second studies, only the left
action. Likewise, EC is intimately associated with hemisphere was scanned due to limitations in the
attention, providing the motive to attend to some number of slices we could acquire. In the first study
stimuli and not others. For these reasons, the an- (Crosson et al., 1999), 17 right-handed healthy
atomical units involved in intention and attention adults alternated between generating words and re-
will be briefly addressed. Following the suggestion peating words with neutral EC in each of two func-
of the Russian neuroanatomist Betz from the late tional runs (Fig. 2). Each functional run consisted
19th century, Fuster (2003) and Luria (1973) have of 6.4 cycles, and word generation and word rep-
both described the cerebral cortex as being divided etition half cycles lasted 18.4 s each. During each
into an anterior realm involved primarily in action half cycle participants heard a cue, either ‘generate’
and a posterior realm involved primarily in sensory or ‘repeat’ followed by either a category, or a list of
processing. Further, Heilman et al. (2003) and emotionally neutral words to repeat, respectively.
Fuster (2003) have divided attention into two When just provided a category, participants silently
realms. As noted above, intention selects one among generated as many words as possible until they were
several competing actions for execution and regu- told to stop generating words 18.4 s after presenta-
lates activity in the anterior realm of action. Heil- tion of the category. One run consisted of neutral
man et al. (2003) have suggested that anterior categories and the other run consisted of categories
cingulate cortex is involved in translating emotion/ with positive and negative ECs. Rate of generation
210

18.4 s 18.4 s 18.4 s 6.4 more than two legs and primarily land-dwelling
cycles (e.g., grasshopper). The baseline task consisted of
gen rep gen
tone monitoring. During this task, participants
pressed a button each time two or more high-
pitched tones were played in a tone sequence. Tar-
get density matched across tone- and word-moni-
Fig. 2. In each of two functional runs during the first experi-
ment, participants alternated between word generation and word toring tasks. The task demands were identical
repetition. Each functional run consisted of 6.4 cycles, and word among the three word-monitoring tasks other than
generation and word repetition half cycles lasted 18.4 –s each. the rated emotional valence and arousal of the
During the word generation task, when provided a category, words. We found that monitoring words with EC
participants silently generated as many words as possible until
for the two characteristics led to unique brain re-
told to stop generating. One run used emotionally neutral cat-
egories and the other run used categories with positive and neg- gions relative to animal or implement monitoring.
ative emotional connotations. (From Crosson et al., 1999.) When monitoring for words with EC was directly
compared to monitoring implements or animals, a
left rostral frontal region of brain activity was
of words to emotional and neutral categories were uniquely activated by semantic processing of EC
closely matched by selecting categories with ECs (see Fig. 5). This region encompassed the same left
and neutral categories that yielded similar response rostral frontal area as in the first study, in which
rates during a pilot study. As measured by oral subjects generated words with EC.
generation after the scanning session, the rate of In the third study (Cato et al., 2004), we repeated
generation for categories with ECs and emotionally the word generation paradigm but included whole
neutral categories were closely matched to the rate brain coverage and enough trials to compare va-
of word repetition, with no significant differences in lence (positive or negative) of the categories. In
rate (see Crosson et al., 1999 for details). In data each of four functional runs (Fig. 6), 26 healthy,
analyses, no distinction was made between positive right-handed adults alternated between word gen-
vs. negative EC. When generation of words for eration and word repetition. At the beginning of
categories with EC was compared to generation of each trial within a run, participants heard a cue,
words for categories with neutral EC, an area in the either ‘generate’ or ‘repeat’ followed by either a
left rostral frontal lobe was active (see Fig. 3). category, or a list of emotionally neutral words to
In the second study (Crosson et al., 2002), we repeat, respectively. When provided with a cate-
compared semantic monitoring of words with EC gory, participants silently generated as many words
to semantic monitoring of words with neutral EC as possible in 16.5 s. Length of word repetition tri-
(Fig. 4). In each of three functional runs, 16 als varied between 9.9 and 23.1 s (Fig. 6). Positive,
healthy, right-handed adults alternated between negative, and neutral word generation were inter-
monitoring words and monitoring tones. Each spersed in each run. Rate of generation of words to
functional run consisted of eight cycles, and tone- both emotional categories and to neutral categories
and word-monitoring half cycles lasted 28 s each. were closely matched by selecting categories with
During each 28 s half cycle of word monitoring, ECs and neutral categories that yielded similar re-
participants heard nine words. During one run, sponse rates during a pilot study. As measured by
participants listened to words with positive and oral generation after the scanning session, the rate
negative ECs. For this run, they pressed a button of generation for categories with ECs and emo-
each time the word was both nonliving and had a tionally neutral categories showed no significant
negative connotation (e.g., vomit). The other two differences (see Cato et al., 2004 for details). Rated
runs consisted of emotionally neutral words. In one arousal was matched between categories with pos-
neutral run, the participants monitored implements itive or negative EC based on the Affective Norms
for those that required only one hand to use and are for English Words (ANEW; Bradley et al., 1988).
used primarily outdoors (e.g., baseball). The other Subjects’ ratings of categories demonstrated
neutral run consisted of monitoring animals having significant differences between positive, negative,
211

Frontal Pole
2.5

2 Emogen
Neutgen
1.5

0.5

0
1 2 3 4 5 6 7 8 9 10

Fig. 3. Activity near the left frontal pole was associated with generating words with emotional connotations vs. generating emotionally
neutral words (Crosson et al., 1999).

28 s 28 s 28 s In summary, across the three studies, the left


8
rostral frontal region was associated with semantic
cycles
word tone word processing of EC. In all three studies, we made
careful choice of baseline tasks in order to isolate
the semantic system. For the semantic monitoring
task, to eliminate the brain activity associated with
Fig. 4. In each of three functional runs during the second ex- auditory processing of sounds, the tone-monitoring
periment, participants alternated between monitoring words baseline task was used. As both the semantic mon-
and monitoring tones. Each functional run consisted of eight
itoring and the tone-monitoring tasks required a
cycles, and tone- and word-monitoring half cycles lasted 28 s
each. During each 28 s half cycle of word monitoring, partic- button press, brain activity associated with execut-
ipants heard nine words. During one run, participants made a ing the response was also eliminated. In the case of
semantic judgment about words with positive and negative the two word generation tasks, use of a neutral EC
emotional connotations. The other two runs consisted of emo- word repetition baseline task subtracted elements
tionally neutral words. In one, the participants made semantic
of the phonological input as well as elements of
judgments about implements, and in the other, they made se-
mantic judgments about animals. The comparison task con- making a (silent) verbal response. Thus, in answer
sisted of monitoring tone sequences for two or more high- to the first question, processing of EC does appear
pitched tones. Target density matched across tone- and word- to require unique brain activity in the rostral fron-
monitoring tasks. (From Crosson et al., 1999.) tal lobe. We concluded that this rostral frontal area
mediates semantic processing of EC.
and neutral categories, and arousal ratings were
significantly greater for the emotional than the
neutral categories (see Cato et al., 2004 for details). Question 2: Does the nature of the task influence the
For a third time, robust rostral frontal activity was pattern of brain activity associated with processing
found, this time bilaterally, but in the left more emotional connotation?
than the right hemisphere (Fig. 7). This activity was
associated with positive and negative, but not A cursory look at these three studies suggests that
neutral EC. No significant differences in activity the nature and location of processing EC does not
were found according to valence of the categories change when task demands are varied. In fact, a
with EC. comparison between studies 1 and 2 reveals that the
212

Fig. 5. Sagittal and coronal images reveal activity in the left rostral frontal pole that is unique to monitoring words with emotional
connotation relative to monitoring tool names or animal names (Crosson et al., 2002).

16.5” 9.9” 16.5” 16.5” 16.5” 23.1”


L2
A58
“Weapons” “Vacations” “Buildings”

gen rep-4 gen rep-7 gen rep-10

Emotional versus Neutral Words

9.5 cycles
L3 A60
Fig. 6. In each of four functional runs during the third study,
participants alternated between word generation and word rep-
etition trials. At the beginning of each trial within a run, par-
ticipants heard a cue, either ‘generate’ or ‘repeat’ followed by
either a category, or a list of emotionally neutral words to re-
peat, respectively. When provided with a category, participants Positive versus Neutral Words
silently generated as many words as possible in 16.5 s. Length of
word repetition trials varied between 9.9 and 23.1 s. Positive,
negative, and neutral word generation trials were interspersed L4
in each run. Rate of generation of words to emotional and A60
neutral categories were closely matched. Rated arousal was
matched between categories with positive or negative emotional
connotation. (From Cato et al., 2004.)

Negative versus Neutral Words


same frontal region is active for semantic processing
Fig. 7. In the third study comparing word generation to pos-
of EC during an attentional task and during an in- itive, negative, vs. neutral emotional connotations, the left
tentional task. However, in the third study (Cato frontal pole was significantly active during processing of pos-
et al., 2004), involving a larger sample size and a itive and negative, but not neutral emotional connotations.
larger number of blocks, we found another major
cluster of functional activity associated with seman- significantly differed between the rostral frontal and
tic processing in the left retrosplenial region. This retrosplenial areas (see Fig. 7). The rostral frontal
region has been implicated frequently in the litera- region remained active throughout the word gener-
ture as a candidate region for semantic processing ation blocks and activity in this region declined to
of EC (Maddock and Buonocore, 1997; for baseline levels only after the end of the trial. The
review, Maddock, 1999). Surprisingly, time courses retrosplenial area was active upon hearing the
213

X = -2
1.4
1.2

1
0.8

0.6
0.4

red = p < 0.005, yellow = p < 0.001 0.2

Signal
1.4 0
1 2 3 4 5 6 7 8 9 10 11
1.2 -0.2

1 -0.4

0.8 -0.6

0.6 -0.8

0.4 -1

0.2 -1.2
Image Number
Signal

0
1 2 3 4 5 6 7 8 9 10 11
-0.2
Positive Categories
-0.4
Negative Categories
-0.6 Neutral Categories

-0.8
-1
-1.2
Image Number

Fig. 8. Time course analysis revealed that rostral frontal cortex and retrosplenial cortex may play different roles in processing
emotional, but not neutral, emotional connotations (Cato et al., 2004).

category cue at the beginning of word generation Discussion


trials, but the activity in this area returned to base-
line during the responses made by the participant. Returning to the Ellis and Young (1988) model
The fact that both areas demonstrated activity dur- described at the beginning of the paper, it appears
ing generation of words with EC but not during that new evidence from brain imaging studies in-
generation of emotionally neutral words suggested forms a number of the assumptions of this mod-
that both areas are involved in semantic processing ular model. First, the semantic system does not
of EC. However, the differences in time courses in- appear to be unitary in the sense that it consists of
dicated that the rostral frontal region was more in- a single module. Brain activity differs during se-
volved in semantic processing of EC during word mantic processing as a function of the modality,
production, while the retrosplenial region was more category, and attributes of the information being
involved when attention was directed towards in- processed. In particular, processing of semantic
coming (in this case, heard) categories with EC (Fig. attributes is widely distributed with the topogra-
8). The implications for intention and attention in phy dependent upon the nature of the particular
processing EC will be discussed below. attribute. For example, our studies indicate that
214

emotional attributes are processed in rostral fron- engage in procreation. As humans have developed
tal and retrosplenial cortex. The findings of Hauk language, it has become a tool to assist in these
et al. (2004) suggest not only that human action activities. Thus, the attentional and intentional
attributes of words are processed in motor and systems may be relatively more invoked during
premotor cortex, but also that the part of the body processing of words with EC, leading to multiple
involved in the action determines the dorsal and clusters of brain activity in posterior (sensory) and
ventral location of the activity in keeping with the anterior (action) regions.
motor homunculus. The findings of Chao et al. With respect to the topography for processing
(1999) indicate that processing of visual attributes EC, activity in the retrosplenial region is consistent
is associated with activity in the ventral visual with the model invoking posterior cingulate and re-
stream, even when the input is words. This con- trosplenial cortex in the interface between motiva-
clusion also is consistent with the lesion study of tion/emotion and attention (Heilman et al., 2003).
Hart and Gordon (1992). Second, each input and However, the frontal regions involved in generating
output sensory modality does not always have words with ECs was anterior to the region that
equal access to all parts of the semantic system. might be predicted from this model. Nonetheless,
Rather, in some cases, the nature of the interaction Tranel (2002) suggested that this region of the fron-
of meaning with attention (sensory) and intention tal lobes is involved using emotions to facilitate de-
(response) systems may determine how some se- cision making, in other words, in selecting which
mantic information is processed. actions to perform (i.e., intention). We suggest that
From our recent studies it specifically appears because this area is involved in evaluating emotions
that processing EC requires unique activity in both during intentional processes, it well suited to
attention systems that select specific sensory input processing EC of words. This synergy between its
for further processing and intention systems that intentional role and its role in processing EC is sim-
select specific actions for output. In particular, the ilar to the synergy in the role of the motor and
neuroanatomic substrates for EC on the input side premotor cortex in executing actions and in process-
of the equation appear to differ from the subst- ing action attributes of words (Hauk et al., 2004).
rates on the output side of the equation. This con- It should be noted that the topography for
clusion suggests that the semantic system, at least processing EC was similar to the topography of
in the case of EC, is not neatly separated from Maddock et al.’s (2003) recent study; only the ros-
input and output systems, as Ellis and Young’s tral frontal activity in our studies was somewhat
model suggests. Rather, attentional and inten- superior to that of Maddock et al. Further, this lat-
tional systems play a role in semantic processing of ter study was not able to distinguish between the
EC. It is not clear if there are other semantic at- role of the rostral frontal and retrosplenial cortex on
tributes that share this characteristic with EC. In the basis of their data, whereas the time courses in
addition, our findings support a left hemisphere our most recent study (Cato et al., 2004) suggested
dominance for processing words with EC. This an explanation. Also, not all studies of processing
suggests that processing words with EC is different the EC of words have produced this topography.
than processing pictures with EC where support For example, a recent study by Kensinger and
for the right hemisphere and bivalent hypotheses Corkin (2004) attempted to dissociate valence and
have been found (e.g., Canli et al., 1998). arousal by comparing encoding of unpleasant–low
EC is a fundamental attribute associated with a arousal words, unpleasant high-arousal words, and
vast array of semantic information. It is linked to neutral low-arousal words. Findings for neither va-
survival. Without knowledge of emotional salience lence nor arousal mapped well onto our most recent
of incoming information we are stripped of a ma- findings (Cato et al., 2004), most likely because of
jor aid to safety, sustenance, and procreation. In significant differences in task.
the evolution of the species, it has been adaptive to Finally, it is unclear whether other types of infor-
respond quickly and to remain alert in threatening mation also require processing in both sensory and
circumstances, to seek means of sustenance, and to action planning areas during semantic processing, or
215

whether this feature is unique to EC. In addition, the Cato, M.A., Crosson, B., Gokcay, D., Soltysik, D., Wierenga,
series of studies reported here all used the same input C., Gopinath, K., Himes, N., Belanger, H., Baver, R.M.,
Fischler, I.S., Gonzalez-Rothi, L. and Briggs, R.W. (2004)
modality and did not examine the effects of using
Processing words with emotional connotation: an FMRI
visual, rather than auditory input. This preliminary study of time course and laterality in rostral frontal and re-
evidence does suggest that more distributed models trosplenial cortices. J. Cogn. Neurosci., 16: 167–177.
of lexical–semantic processing are needed. Such Chao, L.L., Haxby, J.V. and Martin, A. (1999) Attribute-based
models should preserve the anatomical precision neural substrates in temporal cortex for perceiving and
that is afforded by fMRI investigations (maintaining knowing about objects. Nat. Neurosci., 2: 913–919.
Crosson, B., Benjamin, M. and Levy, I.F. (in press) Role of the
some aspect of modularity), but also should address basal ganglia in language and semantics: supporting cast. In:
the interface between attentional, semantic, and in- J. Hart, Jr. and M. Kraut (eds.), Neural Bases of Semantic
tentional systems to reflect the distributed nature of Memory. Cambridge University Press, Cambridge, UK.
information flow during semantic processing. It ap- Crosson, B., Cato, M.A., Sadek, J.R., Gokcay, D., Bauer,
R.M., Fischler, I.S., et al. (2002) Semantic monitoring of
pears that semantic information may be used in
words with emotional connotation during fMRI: contribu-
different ways by attentional and intentional systems tion of anterior left frontal cortex. J. Int. Neuropsychol. Soc.,
to assure quick and appropriate response to incom- 8: 607–622.
ing, salient information. More work is needed on Crosson, B., Radonovich, K., Sadek, J.R., Gokcay, D., Bauer,
modeling information flow and how the human R.M., Fischler, I.S., et al. (1999) Left-hemisphere processing
brain filters information on line for appropriate re- of emotional connotation during word generation. Neurore-
port, 10: 2449–2455.
sponse selection and action. Damasio, H., Grabowski, T.J., Tranel, D., Hichwa, R.D. and
Damasio, A.R. (1996) A neural basis for lexical retrieval.
Nature, 380: 499–505.
Acknowledgments Davidson, R.J. (1995) Cerebral asymmetry, emotion, and affec-
tive style. In: Davidson, R.J. and Hugdahl, K. (Eds.), Brain
Asymmetry. MIT Press, Cambridge, MA.
This research was supported in part by grant no. Ellis, A.W. and Young, A.W. (1988) Human Cognitive Neuro-
DC 03455 from the National Institute on Deafness psychology. Erlbaum, London.
and Other Communication Disorders (PI: Crosson), Fuster, J.M. (2003) Cortex and Mind: Unifying Cognition.
a grant from the McKnight Institute at the Univer- Oxford University Press, New York.
sity of Florida (PI: Crosson), and Research Career Hart, J. and Gordon, B. (1992) Neural subsystems for object
knowledge. Nature, 359: 60–64.
Scientist Award no. B3470S to Dr. Crosson from
Hauk, O., Johnsrude, I. and Pulvermuller, F. (2004) Somato-
the Department of Veterans Affairs Rehabilitation topic representation of action words in human motor and
Research and Development Service. Dr. Cato premotor cortex. Neuron, 41: 301–307.
wishes to thank Nemours Children’s Clinic for its Heilman, K.M., Watson, R.T. and Valenstein, E. (2003) Ne-
support during the preparation of this manuscript. glect and related disorders. In: Heilman, K.M. and Valen-
stein, E. (Eds.), Clinical Neuropsychology (4th ed.). Oxford
University Press, New York, pp. 296–346.
Ishai, A., Ungerleider, L.G., Martin, A., Schouten, J.L. and
References Haxby, J.V. (1999) Distributed representations of objects in
the human ventral visual pathway. Proc. Natl. Acad. Sci.
Beauregard, M., Chertkow, H., Bub, D., Murtha, S., Dixon, S. USA, 96: 9379–9384.
and Evans, A. (1997) The neural substrate for concrete, ab- Kensinger, E.A. and Corkin, S. (2004) Two routes to emotional
stract, and emotional word lexica: a positron emission tomo- memory: distinct neural processes for valence and arousal.
graphy study. J. Cogn. Neurosci., 9: 441–461. Proc. Natl. Acad. Sci. USA, 101: 3310–3315.
Blonder, L.X., Bowers, D. and Heilman, K.M. (1991) The role Lane, R.D., Reiman, E.M., Ahern, G.L., Schwartz, G.E. and
of the right hemisphere in emotional communications. Brain, Davidson, R.J. (1997a) Neuroanatomical correlates of hap-
114: 1115–1127. piness, sadness, and disgust. Am. J. Psychiatr., 154: 926–933.
Bradley, M.M., Cuthbert, B.N. and Lang, P.J. (1988) Affective Lane, R.D., Reiman, E.M., Bradley, M.M., Lang, P.J., Ahern,
Norms for English Words (ANEW): Technical Manual and G.L., Davidson, R.J. and Schwartz, G.E. (1997b) Neuroan-
Affective Ratings. University of Florida, Gainesville, FL. atomical correlates of pleasant and unpleasant emotion.
Canli, T., Desmond, J.E., Zhao, Z., Glvoer, G. and Gabrieli, Neuropsychologia, 35: 1437–1444.
J.D.E. (1998) Hemispheric asymmetry for emotional stimuli Lang, P.J. (1995) The emotion probe: studies of motivation and
detected with fMRI. Neuroreport, 9: 3323–3339. attention. Am. Psychol., 50: 372–385.
216

Lang, P.J., Bradley, M.M., Fitzsimmons, J.R., Cuthbert, B.N., Martin, A., Wiggs, C.L., Ungerleider, L.G. and Haxby, J.V.
Scott, J.D., Moulder, B. and Nangia, V. (1998) Emotional (1996) Neural correlates of category-specific knowledge.
arousal and activation of the visual cortex: an fMRI analysis. Nature, 379: 649–652.
Psychophysiology, 35: 199–210. Peper, M. and Irle, E. (1997) Categorical and dimensional
Luria, A.R. (1973) The Working Brain: An Introduction to decoding of emotional intonations in patients with focal
Neuropsychology. Basic Books, New York (translated by brain lesions. Brain Lang., 58: 233–264.
Basil Haigh). Pulvermuller, F. (2005) Brain mechanisms linking language and
Maddock, R.J. (1999) The retrosplenial cortex and emotion: action. Nat. Rev. Neurosci., 6: 576–582.
new insights from functional neuroimaging of the human Tranel, D. (2002) Emotion, decision making, and the ventro-
brain. Trends Neurosci., 22: 310–316. medial prefrontal cortex. In: Stuss, D.T. and Knight, R.T.
Maddock, R.J. and Buonocore, M.H. (1997) Activation of left (Eds.), Principles of Frontal Lobe Function. Oxford Univer-
posterior cingulate gyrus by the auditory presentation of threat- sity Press, New York, pp. 338–353.
related words: an fMRI study. Psychiatry Res., 75: 1–14. Warrington, E.K. and McCarthy, R.A. (1987) Categories of
Maddock, R.J., Garret, A.S. and Buonocore, M.H. (2003) knowledge: further fractionation and an attempted integra-
Posterior cingulate cortex activation by emotional words: tion. Brain, 110: 1273–1296.
fMRI evidence from a valence decision task. Hum. Brain Warrington, E.K. and Shallice, T. (1984) Category-specific se-
Mapp., 18: 30–41. mantic impairments. Brain, 107: 829–853.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 11

Macroscopic brain dynamics during verbal and


pictorial processing of affective stimuli

Andreas Keil

Department of Psychology, University of Konstanz, PO Box D23, D-78457 Konstanz, Germany

Abstract: Emotions can be viewed as action dispositions, preparing an individual to act efficiently and
successfully in situations of behavioral relevance. To initiate optimized behavior, it is essential to accurately
process the perceptual elements indicative of emotional relevance. The present chapter discusses effects of
affective content on neural and behavioral parameters of perception, across different information channels.
Electrocortical data are presented from studies examining affective perception with pictures and words in
different task contexts. As a main result, these data suggest that sensory facilitation has an important role
in affective processing. Affective pictures appear to facilitate perception as a function of emotional arousal
at multiple levels of visual analysis. If the discrimination between affectively arousing vs. nonarousing
content relies on fine-grained differences, amplification of the cortical representation may occur as early as
60–90 ms after stimulus onset. Affectively arousing information as conveyed via visual verbal channels was
not subject to such very early enhancement. However, electrocortical indices of lexical access and/or
activation of semantic networks showed that affectively arousing content may enhance the formation of
semantic representations during word encoding. It can be concluded that affective arousal is associated
with activation of widespread networks, which act to optimize sensory processing. On the basis of pri-
oritized sensory analysis for affectively relevant stimuli, subsequent steps such as working memory, motor
preparation, and action may be adjusted to meet the adaptive requirements of the situation perceived.

Keywords: emotion; attention; oscillatory brain activity; rapid categorization

Introduction stimulus modality (e.g., visual, auditory), infor-


mation channel (e.g., verbal, pictorial), and exper-
Studies capitalizing on the neural correlates of imental task context (e.g., passive viewing,
affective stimulus perception have been motivated concurrent task processing). With the body of
by a multitude of research questions and traditions evidence continuously increasing, these variables
of science. To date, this has led to a broad range of have proven crucial for predicting the effects that
experimental approaches used in affective percep- emotional content exerts on the behavioral and
tion research. Some of the most obvious choices neural levels. In particular, discrepancies between
researchers must take in designing experiments verbal and pictorial information channels in the
on affective stimulus processing are with respect visual modality have attracted attention, as results
to the nature of the stimulus material. Indeed, appear mixed and available theoretical accounts
empirical findings point to a prominent impact of fail to account for this lack of consistency. The
question arises, what is different when emotional
Corresponding author. Tel.: +49-7531-882696; Fax: +49- responses are evoked by means of verbal vs.
7531-882891; E-mail: Andreas.Keil@uni-konstanz.de pictorial content?

DOI: 10.1016/S0079-6123(06)56011-X 217


218

In the current chapter, this question is discussed observed as a function of hedonic valence and
in a framework of dynamic perception and action response mode (e.g., Neumann and Strack, 2000).
regulation in presence of behaviorally relevant Most interestingly, many effects depend upon in-
information, which can be conveyed via pictorial terindividual differences and are reliably seen in
and verbal objects. Emotions are viewed as imple- anxious participants, but not necessarily in healthy,
mented in neuronal networks encompassing low-anxious controls (Bradley et al., 1995). In ad-
aspects of stimulus representations, memories, dition, the channels of information used in this
and actions, among others. Behavioral and elect- literature (words, faces, scenes) differ in physio-
rophysiological evidence is presented, supporting logical and self-rated affective intensity (Sloan et
the view that macroscopic (neural mass) oscilla- al., 2002), which contributes to the inconsistency
tions are crucial for signal transmission and plastic found in the literature.
changes in these networks. This tenet is illustrated Surprisingly high variability of findings in motor
with experimental data from studies of emotional responses to affective stimuli has been observed
perception and emotion–attention interactions. even in tasks having low complexity. For instance,
Potential distinctions of verbal and pictorial chan- simple response tasks as well as choice responses
nels are viewed in the light of (1) differences in the to affective stimuli have failed to yield converging
spatiotemporal activation patterns in verbal vs. evidence (Megibow and De Jamaer, 1979; Wedding
picture processing and (2) differences in terms of and Stalans, 1985). With respect to choice reac-
emotional intensity. The chapter concludes with tions, it has recently been argued that only pleas-
remarks on potential elements for a model of ant target content leads to facilitated response time
plastic perception-action regulation as mediated and accuracy, whereas aversive content is associ-
by dynamic cortical networks. ated with delayed motor responding (Leppänen
and Hietanen, 2003, 2004). In addition, pleasure
facilitation of choice reactions has been observed
Affective words and pictures: behavioral data with affective pictures depicting more complex
scenes (Lehr et al., 1966). This is in contrast to
Experimental tasks involving the presence of affec- visual search tasks in which aversive stimuli are
tive stimuli have a long tradition in experimental typically detected faster than pleasant or neutral
psychology. A significant part of this literature targets (Hansen and Hansen, 1988; Öhman et al.,
tested predictions derived from motivational the- 2001a, b). The most widely used choice reaction
ories on human behavior, examining speed, and paradigm involving words is the lexical decision
accuracy as a function of stimulus intensity and task, which requires subjects to make a choice
hedonic valence (see e.g., Kamin, 1955). In partic- response, indicating if a stimulus visible on the
ular, trait characteristics of different populations screen is or is not part of a language’s lexicon, i.e.,
(e.g., patient groups with anxiety disorders) were whether it is a word or not (Calvo and Castillo,
at the core of this early literature on reaction time 2005). Using this design, most researchers have
and accuracy in response to emotional stimuli observed advantages in speed and accuracy for
(Castaneda, 1956). Despite this long tradition, affectively arousing words, compared to low-
data describing the effects of affectively arousing arousing control words (e.g., Altarriba and Bauer,
information on behavioral performance are not 2004). Thus, facilitation has been associated with
conclusive. There have been contradictory results, facilitation for pleasant and unpleasant, compared
depending on the type of behavioral task as well to neutral targets (Williamson et al., 1991).
as on the type of stimuli used (see Kitayama, As a possible explanation for this divergent
1990). Empirical findings are at variance with each pattern of findings, one may put forward the
other, suggesting either facilitation (Williamson differences in physiological arousal related to
et al., 1991) or inhibition (Watts et al., 1986) of viewing words, faces, and scenes (cf., Keil et al.,
behavioral responses in the presence of aversive 2005a). Pictures are capable of inducing a high
or appetitive stimuli. Often, differences were amount of physiological arousal in the viewer
219

(Lang, 1994) and thus can be expected to differen- the left hemifield and was accompanied by right-
tially affect behavioral performance when hedonic hemispheric electrocortical facilitation. The emo-
valence (i.e., pleasant vs. unpleasant content) is tional Stroop is one of the most widely used
manipulated. Compared to affective pictures, word paradigms to show such interference of affective
and face stimuli are related to a lesser degree in content with a foreground (e.g., color naming)
arousal, both physiologically and in terms of sub- task (Mogg et al., 1989). Generally, these studies
jective ratings (Lang et al., 1998). Besides differ- have produced evidence for strong interference as
ences in arousal, the higher frequencies of pleasant a function of affective content, with responses to a
faces as well as the role of context generated by the task-relevant dimension being delayed by task-
stimuli have been put forward as possible expla- irrelevant affective content (Watts et al., 1986).
nations for this divergent pattern of findings Interestingly, these effects have also been observed
(Leppänen and Hietanen, 2004). Several studies with nonwords that gained affective significance
observing differences between pleasant and un- by means of classical conditioning (Richards and
pleasant words (reviewed in Kissler et al, this Blanchette, 2004). This latter finding might be
volume) make the situation more complicated. taken as evidence for a more consistent pattern of
Results from priming paradigms suggest, however, interference tasks, showing similar effects across
that congruent affective content facilitates subse- stimulus kinds. Indeed, Stroop designs using
quent choice reactions, even when prime and tar- picture representations of objects or faces with a
get stimuli are faces vs. words and vice versa strong affective connotation have converged with
(Carroll and Young, 2005). Thus, while differences visual word Stroop (Anes and Kruer, 2004), sug-
in the stimulus channel may lead to different gesting that affective interference on concurrent
experimental outcomes, the activation of seman- task behavior can be more easily predicted.
tic information related to a specific emotional The behavioral literature as briefly reviewed
disposition may be compatible across information above suggests that both affective faces and pic-
channels. tures are reliably capable of creating interference
Summarizing these results, it turns out that in across paradigms. Affective words might be capa-
choice reaction tasks with affectively arousing pic- ble to automatically withdraw resources from a
tures, there is a tendency to find facilitation for concurrent task specifically in paradigms, in which
pleasant, compared to neutral and unpleasant, attention is readily available to process verbal
content; whereas lexical decision is facilitated as a information. In contrast, interference effects of
function of emotional arousal associated with tar- emotional words may be diminished in tasks that
get words. The situation changes considerably require processing a high number of stimuli in a
when more complex tasks are considered. In par- short amount of time (Harris and Pashler, 2004;
ticular, tasks have been widely used that require Harris et al., 2004), with affective stimuli being
a response to a stimulus, which is competing for outside the spotlight of attention (Calvo and
resources with a concurrently visible affective Castillo, 2005). Thus, activation of concepts by
stimulus. As a main finding, interference effects lexico-semantic routes may differ from activation
of to-be-ignored affective pictures were repeatedly on the basis of object recognition.
reported (Hartikainen et al., 2000). Hence, behavi- The ‘‘attentional blink’’ (AB) paradigm (Raymond
oral performance in a concurrent task declines, as et al., 1992) represents a potential avenue to
the amount of affective arousal present, e.g., in the examine these aspects. The term AB refers to an
background of the task array, is increased impairment in reporting the second target (T2) of
(Schimmack and Derryberry, 2005). Combining two targets presented in rapid succession, if the
interference of picture affective content with a temporal distance between T1 and T2 is in a range
spatial attention task, Keil et al. (2005a) found between 200 and 500 ms. Thus, if words are shown
response time interference effects of unpleasant in a rapid series (i.e., 6–10 per second), the second
relative to neutral picture content present in a of two target words identified by a certain feature
background stream. This difference was specific to (e.g., color) will be reported less likely when
220

appearing in the AB time window. Using visual survey of the neurobehavioral correlates of affec-
word stimuli, it has been shown that affectively tive picture viewing is given, focusing on scene
arousing T2 s are less affected by this effect, being perception under varying task instructions. Par-
associated with more accurate report than affec- ticular attention is paid to oscillatory measures of
tively neutral T2 s (Anderson and Phelps, 2001; large-scale brain processes.
Keil and Ihssen, 2004; Anderson, 2005). Interest-
ingly, enhancing the affective arousal of T1 stimuli
leads to greater interference with subsequent target Neural facilitation and amplification by affective
processing as indicated in poor T2 identification pictures: the role of emotional arousal
accuracy (Ihssen and Keil, submitted). Most
recently, interference of arousing first targets on As indicated earlier in the chapter, one way to
T2 detection has been shown with affective picture conceptualize emotions is to view them as action
T1 s (Most et al., 2005), which is consistent with dispositions, facilitating evolutionary adaptive
results from other interference tasks reported behaviors (Lang, 1979; Frijda, 1988). On the the-
above. The AB paradigm offers the manipulation oretical level, this concept has been described in
of time intervals and is in accordance with the terms of affective networks, linking different
needs of recording electrocortical oscillations. This aspects of affective perception and behavior (e.g.,
paradigm is, therefore, considered in more detail Lang, 1979). Importantly, network models allow
later in this chapter, when discussing neurobehavi- for different routes capable of initiating an affec-
oral correlates of affective word and picture tive episode, being based, for instance, on imagery,
processing. visual objects, or physiological processes (see
Taken together, these findings raise two ques- Fig. 1). As a basis of adaptive behavior, the per-
tions regarding the effects of affective content on ception of emotionally arousing stimuli has been
perception and performance: First, what is the role associated with acting efficiently and successfully
of affective arousal as communicated via different in situations indexed as motivationally significant
information channels (scenes, faces, words) for by properties of the visual scene. The prioritized
perceptual and behavioral facilitation vs. interfer- selection of features representing affective arousal
ence? Second, what is the differential role of affec- may occur in an automatic manner and has been
tive arousal when affective stimuli serve as targets referred to as ‘‘motivated attention’’ (Lang et al.,
vs. distractors. The remaining paragraphs of the 1997b). Consistent with the network perspective
present chapter address these questions, based on taken here, this behavioral advantage may arise as
a multivariate database that includes behavioral, a consequence of optimized processing throughout
peripheral physiological, and brain measures. the parts of the emotional network, beginning with
prioritized perception of affectively relevant sen-
sory features. Given the massively parallel and
Neurobehavioral correlates of affective picture bidirectional nature of the human visual system,
processing the sensory part of the network may in turn be
subject to top-down regulation exerted by higher
The availability of standardized sets of pictures parts of the emotion network, coding working
showing affective faces (Lundqvist et al., 1998) or memory and action representations (Keil, 2004). As
scenes (Lang et al., 1999) has stimulated a plethora a consequence, the sensory representation of an
of studies addressing the multifaceted aspects of emotional stimulus together with its neurophysio-
emotional processing. The International Affective logical underpinnings may represent a useful target
Picture System (IAPS, Lang et al., 1999), for for research into the spatiotemporal dynamics of
instance, provides normative ratings of emotional affective processing. To study these phenomena, it
valence, arousal, and dominance, for more than is necessary to monitor brain dynamics at a high
1000 colored pictures, and is referred to exten- temporal rate. Electrocortical measures provide
sively within this volume. In the present section, a one avenue to study fast changes in the neuronal
221

Fig. 1. Schematic representation of an emotion network, encompassing different aspects of the affective response. Activation of the
entire network can be accomplished via different routes such as the stimulus representation, the imagination of emotional episodes in
memory, and physiological processes. These elements are organized in a parallel fashion and may undergo changes over time, adapting
the emotional response to environmental or internal demands. Such changes may be conceptualized as changes of weights in the
functional architecture of the network.

mass activity of cortical cells and are reviewed in as skin conductance or heart rate changes as well
the subsequent paragraphs. as a strong relationship with self-report question-
naire data (Cuthbert et al., 2000). Interestingly,
this pattern of results has been stable across visual
Evidence from electrocortical measures: time-domain stimuli depicting affective scenes (Palomba et al.,
analyses 1997), faces (Schupp et al., 2004), conditioned
As a robust empirical finding, passive viewing of stimuli (Pizzagalli et al., 2003), and specific content
affectively arousing, compared to nonarousing such as dermatological disease (Kayser et al.,
pictures, has been associated with amplification 1997). Late ERP differences have been associated
(enhancement) of electrocortical parameters with activity in the extended visual cortex (Keil
(Schupp et al., 2000). Recently, neuroimaging et al., 2002) and in consequence have been theo-
and electrocortical data have converged to show retically related to the concept of motivated
that the extended visual cortex is a key structure attention, in which motivationally relevant stimuli
demonstrating facilitation effects as a function of automatically arouse attentional resources being
affective arousal (Keil et al., 2002; Bradley et al., directed to the relevant locations and features
2003). Studies investigating modulations of the present in the field of view (Lang et al., 1997b).
event-related potential (ERP) derived from stimulus- Given this view, it is interesting to consider
locked time-domain averaging of electroencepha- electrophysiological research on instructed as
logram (EEG) epochs recorded during picture opposed to ‘‘natural’’ attention to task-relevant
viewing have repeatedly observed such enhance- stimulus properties. In this field, authors have
ment related to visual cortex activity (Cacioppo demonstrated early in addition to late modulations
and Gardner, 1999; Keil et al., 2002). ERP differ- of ERP data (Hillyard and Anllo-Vento, 1998),
ences were particularly robust in late time segments suggesting an early locus of several types of
(i.e., on ERP components such as the P3 or the selective attention. For instance, attending to a
slow wave), following picture onset by approxi- particular location in space may amplify the elect-
mately 300 ms (Mini et al., 1996). This modulation rocortical response as early as at the stage of the
has been shown to closely correlate with other P1 component of the ERP, typically starting
physiological measures of affective processing such around 80 ms after onset of a stimulus (Martinez
222

et al., 1999). Instructing a subject to attend to a data as reviewed in the present volume (see e.g.,
particular set of features, such as color or shape, is the chapter by Sabatinelli et al., this volume) have
often associated with relative enhancement of the been supportive of this notion and have provided
so-called N1 ERP component at a latency of more precise information on the anatomical struc-
140–200 ms (Luck and Hillyard, 1995). tures, which may mediate such reentrant modula-
In line with these findings from selective atten- tion of visual cortex.
tion research, affectively arousing stimuli appear The topic of early attention to affectively arous-
to attract attentional resources at early stages of ing stimuli raises questions related to the involve-
processing, leading to relative amplification, which ment of primary visual (i.e., striate) cortex. For
occurs earlier than the late potential reported instance, Pourtois et al. (2004) employed a visual
above (Junghöfer et al., 2001; Schupp et al., hemifield paradigm with covert orienting to emo-
2003b). This facilitation may be manifest at the tional faces. For fearful compared to happy faces,
level of the P1/N1 component of the ERP, also these authors reported enhancement of the first
modulating oscillatory responses in visual cortex detectable visual ERP deflection at 90 ms after
around 80–100 ms post-stimulus (Keil, 2004). For stimulus onset, likely originating in the striate cor-
instance, using a hemifield design with colored tex. They concluded that the emotional relevance
affective pictures, Keil et al. (2001b) reported of the stimuli might be associated with increased
enhancement of the N1 amplitude (160 ms post- activation in the primary visual cortex, possibly
stimulus) for arousing, compared to neutral stim- due to an interaction with subcortical structures.
uli. Other authors have demonstrated similar Following on these findings, Stolarova et al. (2006)
effects of emotional content on very early ERP presented grating patterns to the hemifields, paired
components (Pizzagalli et al., 2003; Smith et al., with unpleasant or neutral pictures in a delayed
2003), raising the question as to the mechanisms classical conditioning paradigm. This procedure
mediating such fast differential responding. In allowed to control for the physical stimulus fea-
analogy to findings in the field of selective atten- tures of conditioned stimuli (i.e., grating patterns)
tion (Hillyard and Anllo-Vento, 1998), a ‘sensory and to explore the development of perceptual
gain’ mechanism has been hypothesized to amplify facilitation across blocks of trials, as a function of
sensory processing according to the importance of learning. As a main finding, this study resulted in
the stimulus for the organism. For the case of greater amplitude of the visual ERP for threat-
spatial selective attention, these functions are me- related stimuli vs. neutral stimuli occurring as
diated by the joint action of a widespread network early as 65–90 ms following stimulus onset (i.e., the
of cortical and deep structures, including areas in C1 component of the ERP; see Martinez et al.,
occipital, parietal and frontal cortex as well as the 1999). Interestingly, this difference appeared to
striatum and superior colliculi (Mesulam, 1998). increase with continuing acquisition of affective
Clearly, ERP work reported so far is consistent meaning, being more pronounced in a second
with the involvement of such a widespread cortical conditioning block, compared to the first block of
network in early discrimination of features related trials. This suggests a role of short-term cortical
to affective content. Confirming evidence comes plasticity for early affective perception, tuning
from analyses of the topographical distribution in sensory cortex to amplify specific visual features
dense array ERP data, pointing to generators in that are related to motivationally relevant infor-
higher order visual, as well as right-parietal mation in a specific context.
cortices (Junghöfer et al., 2001; Keil et al., 2002).
These effects have often been interpreted as man-
ifestations of reentrant tuning of visual cortex, Evidence from electrocortical measures: frequency-
which may be mediated by deep cortical structures domain analyses
such as the amygdala, or by temporal, parietal, While time-domain averaging as typically
and prefrontal cortical structures, among others employed in ERP research provides important in-
(Keil, 2004; Sabatinelli et al., 2005). Neuroimaging sights into the timing of neural mass action on a
223

millisecond scale, there is also complementary state of the nervous system can be identified and
information in electrocortical recordings that can characterized. Moreover, instances of synchroni-
be examined using frequency-domain approaches. zation across recording sites or model dipoles may
These methods take advantage of the fact that be evaluated using wavelet algorithms (Lachaux
EEG readings are oscillatory in nature, containing et al., 1999; Gruber et al., 2002). In particular, the
a wealth of characteristic rhythmic processes in gamma band range of the temporal spectrum has
any time window considered. Frequency-domain attracted attention (DePascalis et al., 1987). The
analysis techniques allow assessing two measures terms ‘gamma-band oscillations’ or ‘gamma-band
of oscillations at a given frequency: (i) the spectral activity’ (GBA) refer to those oscillations in elect-
power or amplitude, which is a measure of the rophysiological recordings that lie in the higher
magnitude of the oscillatory response and (ii) frequency range of the temporal spectrum, typi-
the phase, which can be considered reflecting the cally above 20 Hz.Theoretically, these oscillation
latency of an oscillation with respect to a reference have been associated with activation of an object
function. representation (Tallon-Baudry and Bertrand, 1999),
Using amplitude and phase measures, one can possibly integrating different features of an object
obtain estimates of the amount of energy present into one percept (Keil et al., 1999). Elevated GBA
at a known frequency during a given recording has been reported during viewing affective stimuli
epoch. Moreover, it is possible to examine tempo- (Aftanas et al., 2004). Specifically, gamma responses
ral dynamics across sensors or model dipoles, were markedly enhanced in the right hemisphere as
helping to elucidate the organization of functional a function of affective arousal (Müller et al., 1999).
networks that connect different areas of the Furthermore, viewing affective pictures in the vis-
cerebral cortex (Gruber et al., 2002). In particular, ual hemifields indicated that right-hemispheric
measures reflecting stability of phase differences cortical areas are more sensitive to affectively
between recording sites or brain areas across time arousing content than left-hemispheric sites in late
may indicate epochs of large-scale synchronization time ranges between 400 and 600 ms after stimulus
occurring during specific cognitive or behavioral onset (Keil et al., 2001b). Early oscillatory re-
processes (Gevins et al., 1999). In the field of sponses pointed to involvement of basic visual
affective perception, several experimental strate- analysis in affective modulation, showing greater
gies have been used to examine electrocortical amplitudes for aversive pictures. This difference
oscillations in the frequency domain. Many studies was increased in the second block of trials, com-
have compared the properties of the frequency pared to the first (Keil et al., 2001a,b), thus
spectrum of EEG data during viewing of stimuli paralleling the time course of early ERP modula-
different in affective content (for a review, see Keil tion reported earlier (Stolarova et al., 2006).
et al., 2001a). Early work has shown evidence of a In addition to time-frequency analyses of the
general reduction of the alpha and enhancement of intrinsic oscillatory activity considered so far,
the beta frequency range as a function of emo- macroscopic electrocortical oscillations could also
tional arousal (Tucker, 1984; Ray and Cole, 1985; be driven by extrinsic stimulation. Work employ-
Collet and Duclaux, 1987). ing steady-state visual evoked potentials (ssVEPs)
With the advent of dense-array electrode arrays capitalizes on the fact that the human brain
and novel time-frequency approaches to data responds to rapid visual stimulation at a fixed
reduction, analyses of the EEG power spectrum rate (i.e., flickering of a stimulus at five cycles per
could be refined to yield better resolution both in second or greater) with a near-sinusoidal response
the spatial and the time domains (Keil et al., at the same frequency. This brain response can be
2001a). For instance, wavelet analyses of EEG extracted at the known frequency and is referred to
epochs provide an estimate of the evolutionary as the ssVEP. Investigations using source analysis
spectrum, which reflects changes in different fre- techniques to enhance spatial specificity of scalp
quency bands across time (Bertrand et al., 1994). signals have suggested that the ssVEP is generated
Thus, epochs of rapid variability in the oscillatory in parieto-occipital cortices. Source configurations
224

might depend on the stimulation frequency, how- reduction for ssVEPs elicited by flickering full-field
ever (Müller et al., 1997). Importantly, the ssVEP stimulation, when concurrently presented picture
reflects multiple excitations of the visual system stimuli were emotionally engaging as opposed to
with the same stimulus in a short amount of time. having neutral content (Kemp et al., 2002, 2004).
As a consequence, it is capable of reflecting changes This opposite pattern is in line with the results
in neural mass activities, which are related to both reported above, as the full-field flicker represents a
initial sensory processing and reentrant, top-down concurrent stimulus that competes for resources
modulation on the basis of stimulus features and with the affective pictures. Thus, engaging stimuli
higher order processing (Silberstein et al., 1995). can be expected to draw attentional resources not
It has been demonstrated that the amplitude of available for processing of the flicker stimulus,
ssVEPs is sensitive to several types of visual selec- which decreases the flicker-evoked ssVEP.
tive attention, showing an enhancement with Given their excellent signal-to-noise ratio,
increased attentional load (Morgan et al., 1996; ssVEPs are a useful tool to examine changes of
Müller et al., 1998, 2003). In addition, amplitude neural responding during acquisition of affective
and phase alterations of ssVEPs were associated significance across small numbers of trials. This
with affective stimulus characteristics. It has property is relevant for studies of classical fear
repeatedly been observed that the ssVEP ampli- conditioning in which initially neutral cues acquire
tude was enhanced when viewing flickering picture signal function for unpleasant contingencies. In a
stimuli rated as being emotionally arousing. Pre- series of studies, Moratti and colleagues have
senting pleasant, neutral, and unpleasant pictures shown enhancement of ssVEF amplitude to stim-
from the IAPS at a rate of 10 Hz, Keil et al. (2003) uli predicting aversive events after a small number
found higher ssVEP amplitude and accelerated of learning trials. Relating these findings to con-
phase for arousing, compared to calm pictures (see currently measured heart rate changes, they sug-
Fig. 2). These differences were most pronounced at gested that enhancement of the visual response
central posterior as well as right parieto-temporal depends on the development of a fear response to
recording sites. In another study, the steady-state the aversive conditioned stimulus (Moratti and
visual evoked magnetic field (ssVEF), which is the Keil, 2005). Hence, with increasing heart rate
magnetocortical counterpart of the ssVEP, also acceleration in response to the aversive stimulus,
varied as a function of emotional arousal (Moratti the cortical response amplitude was also enhanced.
et al., 2004). Source estimation procedures pointed This pattern of results was replicated in a second
to involvement of parieto-frontal attention networks study in which all subjects were fully aware of the
in arousing and directing attentional resources to- contingencies, reducing the variability due to atten-
ward the relevant stimuli. In line with this inter- tion to the fear signal (Moratti et al., 2006). In line
pretation, Kemp and collaborators found amplitude with observations made in affective picture viewing,

Fig. 2. Steady-state VEP amplitude differences for affective pictures differing in content. Amplitude was enhanced for both pleasant
and unpleasant picture content, compared with neutral content. Topographies represent a grand mean across 19 participants.
225

processing of conditioned fear stimuli activated contralateral to the attended visual hemifield.
frontal, parietal, and occipital cortical areas. Given Differences were most pronounced during selec-
the oscillatory nature of ssVEPs, this suggests tion of the left visual hemifield, at right temporal
coactivation in a common network, with structures electrodes. Consistent with this finding, phase in-
interacting in a massively parallel manner. formation revealed accelerated processing of avers-
Ongoing studies have analyzed ssVEPs on the ive compared to affectively neutral pictures.
level of single trials, estimating the relevant param- Results suggest that affective stimulus properties
eters for epochs reflecting 6 s of single-picture modulate the spatio-temporal process of informa-
viewing. Results showed surprisingly high reliabil- tion propagation along the ventral stream, being
ities for ssVEP amplitude and stable topographical associated with amplitude amplification and tim-
distributions of phase values. Interestingly, phase ing changes in posterior and fronto-temporal
synchrony between brain areas, as estimated by cortex. Importantly, even nonattended affective
means of a distributed source model, showed pictures showed electrocortical facilitation through-
alignment of phase values between fronto-temporal out these sites. In line with network perspectives of
and occipito-parietal regions (Keil et al., 2005b). emotion, this pattern of results supports the notion
This synchronization of phase values on the ssVEP that emotional content affects both initial sensory
frequency was enhanced as a function of affective processing and stages of more complex processing,
arousal. Furthermore, we found a linear relation- depending on visual objects’ affective and motiva-
ship between picture arousal as measured by skin tional significance (Lang, 1979; Bower et al., 1994).
conductance and ssVEP amplitude, being most Analyses of intersite phase synchrony, which are
pronounced in frontal cortical regions (Smith under way, may shed further light on the spatio-
et al., 2005). temporal relationships between sensory and higher
To summarize, there is now a large body of order processing (Keil et al., 2005b).
evidence in psychophysiology, strongly suggesting
that emotional picture content guides attentional
resources in a rapid, and possibly automatic, fash- Neurobehavioral correlates of emotional language
ion. Several authors have addressed this topic by processing
manipulating affective and attentional load directed
to experimental stimuli (e.g., Schupp et al., 2003a). As outlined in the introduction of the chapter,
The chapter by Harald Schupp (this volume) pro- behavioral studies have pointed to a facilitatory
vides a comprehensive review of this work. Here, influence of affective content in word identifi-
the specific effects of attention–emotion interac- cation task such as the lexical decision paradigm
tions on the ssVEP are briefly considered, as these (Williamson et al., 1991). Thus, paralleling re-
approaches may help to elucidate the role of search with affective pictures, an important ques-
bottom-up and top-down processes, which act to tion in the literature, relates to the temporal
organize the integrated functioning of affective dynamics of potential facilitation for affectively
networks. In a study that crosses the factors of arousing verbal information. As opposed to pic-
selective spatial attention and affective content, ture work, semantic information conveyed by means
Keil et al. (2005a) investigated the ssVEP ampli- of verbal channels requires differential processing
tude and phase as a function of attention–emotion steps such as sublexical encoding of the language
interactions. Participants silently counted random- stimulus and subsequent lexical access (Ramus,
dot targets embedded in a 10-Hz flicker of colored 2001). Hence, there is no singleton set of features
pictures presented to both hemifields. An increase that can be associated with a specific affective
of ssVEP amplitude was observed as an additive content, but affective content may be resulting
function of spatial attention and emotional con- from — and potentially co-occurring with and
tent. Mapping the statistical parameters of this determining — lexical processing (Kitayama, 1990).
difference on the surface of the scalp, the authors In the present framework, this means that activat-
found occipito-temporal and parietal activation ing an emotion network via lexical information is
226

different from affective object recognition regarding (Luck et al., 1996). Specifically, detected T2 s
several aspects. (i) Depending on the familiarity of are associated with normal electrophysiological
the verbal material, the symbolic operations re- markers of conscious identification (Kranczioch
sulting in semantic access are expected to take et al., 2003). Thus, early ERP components in re-
longer than sensory encoding of known objects in sponse to missed T2 stimuli were not different
the visual scene. For instance, words having high from T2 s outside the blink period. In contrast, the
vs. low frequency in a given language are distin- P3 component was reduced for missed T2 s (Vogel
guished as early as 120–160 ms post-stimulus (Ass- et al., 1998). This pattern of results may be taken
adollahi and Pulvermüller, 2001). This suggests as evidence for a postperceptual, working-
that the representation corresponding to verbal memory-related interference process, which im-
material is accessed relatively early, but is accessed pairs performance for T2 s within the AB period.
later than semantics related to visual objects. (ii) Interestingly, T2 stimuli carrying ‘salient’ infor-
The locus of facilitation by affective picture mation, such as a participant’s own name, appear
content should be early in the cascade of visual to be reported more accurately, compared to less-
processes and might extend as early as initial visual salient stimuli (Shapiro et al., 1997).
analysis. This should affect later stages as well. In a series of studies with affective visual verbs,
Alternatively, one might predict that coding lexical Ihssen and Keil (2004) set out to investigate post-
information should be facilitated later, i.e., on the perceptual facilitation for arousing compared to
level of lexical analysis, which again may act back neutral visual verbs during the AB. Verbal stimuli
to visual analysis, but can be expected to have were selected on the basis of subjective ratings of
different spatio-temporal dynamics. The following their affective valence and arousal. These ratings
paragraphs examine these issues in the context of for German verbs showed a typical distribution in
the AB task with affective words, which is suitable a two-dimensional affective space spanned by the
to examine facilitation and interference at different dimensions of valence and arousal, as has been
intensities. On the electrophysiological level, the previously reported for pictorial (Lang et al.,
AB is compatible with recording of ssVEPs, which 1997a) and auditory stimuli (Bradley and Lang,
is another advantage for analyses focusing on 2000). As expected, Ihssen and Keil found strong
frequency domain analyses, including synchrony reduction of report accuracy with short T1–T2
between different cortical regions. intervals (232 ms), throughout experiments. While
intermediate intervals (i.e., 464 ms) showed less
impairment of verb identification than the 232 ms
Affective facilitation during the AB SOA condition, long T1–T2 distance of 696 ms
was associated with almost unimpaired accuracy
As introduced above, spread of activation and of report.
consolidation of a stimulus representation in In terms of affective modulation, results sug-
working memory can be investigated using the gested identification as a function of affective
so-called AB design, which requires the selection arousal especially in the short-interval condition,
of two targets in a rapid stream of stimuli. Usually, in which performance was most impaired. Thus,
a first target (T1) is followed by a second target motivationally or affectively relevant material was
(T2), with processing of T1 being critical for selected preferentially from a temporal stream of
occurrence of the AB (Chun and Potter, 1995; verbal information. This difference disappeared
Potter et al., 1998). It has been repeatedly shown when affective valence was manipulated while
that at high rates of visual presentation (e.g., at arousal was held constant at a low level (Keil
6 Hz and higher), T2 s being presented in an inter- and Ihssen, 2004). Systematically manipulating T1
val between 180 and 500 ms after a given T1 are and T2 affective content in AB tasks with and
reported less accurately. However, there is evi- without explicit target definition, differential
dence that T2 items are perceived, and semantic effects of T1 and T2 emotional intensity emerged
information may be processed to a certain degree (Ihssen and Keil, submitted). First, emotional
227

and synchrony measures (Pulvermüller, 1996). In a


study using steady-state auditory fields generated
by an amplitude-modulated word stream at 40 Hz,
Härle et al. (2004) showed that during word com-
prehension, there is a specific increase in large-
scale synchrony that encompasses frontal and
parietal, in addition to perisylvian cortices.
The EEG can be recorded during AB tasks,
and data can be evaluated using time-frequency
domain techniques. In the case of manipulating
affective content of T2, enhancement of the ssVEP
amplitude was observed for affectively arousing
T2 verbs, beginning at around 150 ms following
T2 onset (Ihssen et al., 2004). This difference was
most pronounced at parietal sites, where the
ssVEP amplitude was most pronounced. In addi-
tion, it was restricted to the early lag condition,
paralleling behavioral facilitation effects, which
Fig. 3. Response accuracy for the identification of second Tar- were also specific to short T1–T2 intervals (Keil
gets (T2) in an attentional blink task with affective words. Par- and Ihssen, 2004). As displayed in Fig. 4, the
ticipants (N ¼ 13) were asked to identify two green target words 8.7-Hz amplitude of the ssVEP was enhanced in a
in a rapid stream of words shown at a rate of 8.7 words per time range between 140 and 200 ms after T2 onset,
second. With a T1–T2 stimulus onset asynchrony of 232 ms,
performance was substantially impaired, compared to longer
whereas earlier differences were not detected.
T1–T2 intervals. This impairment was less pronounced for Thus, modulation of the ssVEP amplitude during
affectively arousing T2 words. the AB suggests enhanced processing of emotion-
ally arousing T2 verbs at latencies that are con-
arousal of an explicit target (both T1 and T2) sistent with lexical access (Assadollahi and
defined by a relevant dimension (i.e., color) relia- Pulvermüller, 2001). In line with that notion, the
bly facilitated conscious identification of that item topography of the ssVEP showed a left occipito-
(see Fig. 3). Second, arousing content of T1 words temporal preponderance, compared to the right
interfered with identification of the subsequent hemisphere, in which amplitudes were generally
target when the rate of T2 identification was well lower (see Fig. 4).
below 100%, i.e., when the T2 task was difficult Manipulating the content of T1 instead of T2,
enough. Affective interference effects did not de- behavioral effects reported above were replicated
pend on an explicit target definition. For instance, (Ihssen et al., 2005). Thus, interference of T1 on
interference effects were established for emotional T2 performance increased as a function of affec-
nontarget words preceding targets in a semantic tive arousal. In contrast to T2 facilitation, T1
categorization task (Ihssen and Keil, submitted). interference was present across T1–T2 intervals.
Preliminary electrophysiological results in nine
participants indicated that behavioral interference
Steady-state visual evoked potentials during the AB was paralleled by amplitude reduction in the time
Given their sensitivity to rapidly changing oscilla- range, in which facilitation for affectively arousing
tions directly related to experimental manipula- T2 s was observed. Again, topographical distribu-
tions, steady-state potentials have been used to tions suggest that this effect is related to differen-
address different aspects of language processing. tial lexical access for affectively arousing,
Lexical access is assumed to be effected by the in- compared to calm words. This interpretation is
tegrated dynamics of various areas in the brain corroborated by analyses of phase synchrony,
that may be examined by means of time-frequency which allow evaluating epochs of synchronization
228

Fig. 4. Steady-state VEP amplitude differences for second targets in an attentional blink task, differing in content. The spline-
interpolated topographies shown here cover a time range between 140 and 200 ms following onset of the T2 stimulus. Amplitude is
enhanced for both pleasant and unpleasant word content, compared with neutral content. Topographies represent a grand mean across
nine participants.

between occipito-parietal and fronto-temporal stimuli. ERP data revealed amplitude modulations
regions. These ongoing analyses have revealed of a late component, the N2, as a function of
enhanced coupling across the left hemisphere at stimulus properties (i.e., CS+ vs. CS). Over right-
the steady-state frequency, which might reflect hemispheric sensors, negativity pertinent to the
facilitated integration within the semantic network latency range of about 250 and 310 ms was spe-
representing the affective content. cifically enhanced for the CS+ during intermittent
aversive conditioning. During extinction trials, this
conditioning effect was paralleled by gamma-band
Learning to associate language stimuli with affective (18–40 Hz) oscillations occurring as early as
content 80–120 ms after stimulus onset. Thus, already
early stages of syllable processing may be affected
One conclusion from the data presented so far is by the motivational significance of a given lan-
that spatial and temporal dynamics of affective guage stimulus, if this stimulus gains affective
stimulus processing depend on the channel em- relevance by contingent pairing with external
ployed. Affective picture perception involves early events. This would imply a process different from
sensory cortical processes as well as higher order facilitation related to the semantic content or
stages. Affective information as conveyed via meaning of a language stimulus.
lexical channels in contrast seems to affect the
earliest steps in lexical access, which may be facil-
itated as a function of emotional arousal. The Conclusions
question arises, whether learned significance as
associated with simple language stimuli may lead Taken together, the selective review presented
to early sensory amplification as well. In fact, sup- above suggests that both picture and word stimuli
portive evidence has recently been suggested by a are processed in a facilitated manner, if their con-
classical conditioning experiment involving syn- tent is emotionally arousing. There are important
thesized stop consonant–vowel syllables (Heim differences in the spatial and temporal dynamics,
and Keil, in press). Here, aversive white noise however, as indicated by their large-scale oscilla-
(unconditioned stimulus, US) was paired with tory correlates. In the laboratory, very early
two /ba/ syllables (conditioned stimulus, CS+). sensory modulation appears for stimuli with sali-
Two /da/ syllables served as the CS- and indicated ent features that have been experienced multiple
absence of the noxious US. The EEG was recorded times. The more immediate relationship of pictures
while healthy volunteers passively listened to the to natural objects allows for a fast activation of
229

object representations on the basis of character- Affective Science. Oxford University Press, New York, NY,
istic features, even with short presentation times. USA, pp. 301–318.
Bradley, B.P., Mogg, K., Millar, N. and White, J. (1995)
Affective picture viewing typically shows more
Selective processing of negative information: effects of clin-
similarities with the real-world situation and is ical anxiety, concurrent depression, and awareness. J.
related to higher physiological arousal, compared Abnorm. Psychol., 104: 532–536.
to linguistic stimuli. Affective arousing content of Bradley, M.M. and Lang, P.J. (2000) Affective reactions to
words seems to facilitate lexical access and subse- acoustic stimuli. Psychophysiology, 37: 204–215.
quent processing such as consolidation in working Bradley, M.M., Sabatinelli, D., Lang, P.J., Fitzsimmons, J.R.,
King, W. and Desai, P. (2003) Activation of the visual cortex
memory. When paired with distinct external affec- in motivated attention. Behav. Neurosci., 117: 369–380.
tive consequences, features of words, however, Cacioppo, J.T. and Gardner, W.L. (1999) Emotion. Annu. Rev.
may acquire more direct access to affective net- Psychol., 50: 191–214.
works, acting more rapidly to enable fast visual Calvo, M.G. and Castillo, M.D. (2005) Foveal vs. parafoveal
attention-grabbing power of threat-related information. Exp.
analysis. Both pictures and words thus underlie
Psychol., 52: 150–162.
dynamic changes in affective processing, showing Carroll, N.C. and Young, A.W. (2005) Priming of emotion
dependencies of the context, internal states of the recognition. Q. J. Exp. Psychol. A, 58: 1173–1197.
individual, and previous experience. Dynamic Castaneda, A. (1956) Reaction time and response amplitude as
network perspectives on affective processing are a function of anxiety and stimulus intensity. J. Abnorm.
capable of accounting for these findings. They allow Psychol., 53: 225–228.
Chun, M.M. and Potter, M.C. (1995) A two-stage model for
generating hypotheses for future work investigat- multiple target detection in rapid serial visual presentation.
ing the dynamic and plastic properties of affective J. Exp. Psychol. Hum. Percept. Perform., 21: 109–127.
processing. Collet, L. and Duclaux, R. (1987) Hemispheric lateralization of
emotions: absence of electrophysiological arguments. Phy-
siol. Behav., 40: 215–220.
References Cuthbert, B.N., Schupp, H.T., Bradley, M.M., Birbaumer, N.
and Lang, P.J. (2000) Brain potentials in affective picture
Aftanas, L.I., Reva, N.V., Varlamov, A.A., Pavlov, S.V. and processing: covariation with autonomic arousal and affective
Makhnev, V.P. (2004) Analysis of evoked EEG synchroni- report. Biol. Psychol., 52: 95–111.
zation and desynchronization in conditions of emotional DePascalis, V., Marucci, F.S., Penna, P.M. and Pessa, E. (1987)
Hemispheric activity of 40 Hz EEG during recall of emo-
activation in humans: temporal and topographic character-
istics. Neurosci. Behav. Physiol., 34: 859–867. tional events: differences between low and high hypnotiza-
Altarriba, J. and Bauer, L.M. (2004) The distinctiveness of bles. Int. J. Psychophysiol., 5: 167–180.
emotion concepts: a comparison between emotion, abstract, Frijda, N.H. (1988) The laws of emotion. Am. Psychol.,
43: 349–358.
and concrete words. Am. J. Psychol., 117: 389–410.
Anderson, A.K. (2005) Affective influences on the attentional Gevins, A., Smith, M.E., McEvoy, L.K., Leong, H. and Le, J.
dynamics supporting awareness. J. Exp. Psychol. Gen., 134: (1999) Electroencephalographic imaging of higher brain
function. Philos. Trans. R. Soc. Lond. B. Biol. Sci., 354:
258–281.
Anderson, A.K. and Phelps, E.A. (2001) Lesions of the human 1125–1133.
amygdala impair enhanced perception of emotionally salient Gruber, T., Muller, M.M. and Keil, A. (2002) Modulation of
events. Nature, 411: 305–309. induced gamma band responses in a perceptual learning task
in the human EEG. J. Cogn. Neurosci., 14: 732–744.
Anes, M.D. and Kruer, J.L. (2004) Investigating hemispheric
specialization in a novel face-word Stroop task. Brain Lang., Hansen, C.H. and Hansen, R.D. (1988) Finding the face in the
89: 136–141. crowd: an anger superiority effect. J. Pers. Soc. Psychol., 54:
Assadollahi, R. and Pulvermüller, F. (2001) Neuromagnetic 917–924.
Harris, C.R. and Pashler, H. (2004) Attention and the process-
evidence for early access to cognitive representations.
Neuroreport, 12: 207–213. ing of emotional words and names: not so special after all.
Bertrand, O., Bohorquez, J. and Pernier, J. (1994) Time– Psychol. Sci., 15: 171–178.
frequency digital filtering based on an invertible wavelet Harris, C.R., Pashler, H.E. and Coburn, N. (2004) Moray re-
visited: high–priority affective stimuli and visual search. Q. J.
transform: an application to evoked potentials. IEEE Trans.
Biomed. Eng., 41: 77–88. Exp. Psychol. A, 57: 1–31.
Bower, G.H., Lazarus, R., LeDoux, J.E., Panksepp, J., Dav- Hartikainen, K.M., Ogawa, K.H. and Knight, R.T. (2000)
Transient interference of right hemispheric function due to
idson, R.J. and Ekman, P. (1994) What is the relation
between emotion and memory? In: Ekman, R.J.D.P. (Ed.), automatic emotional processing. Neuropsychologia, 38:
The Nature of Emotion: Fundamental Questions. Series in 1576–1580.
230

Heim, S. and Keil, A. (in press) Effects of classical conditioning Keil, A., Müller, M.M., Ray, W.J., Gruber, T. and Elbert, T.
on identification and cortical processing of speech syllables. (1999) Human gamma band activity and perception of a
Exp. Brain Res., DOI: 10.1007/s00221-006-0560-1. gestalt. J. Neurosci., 19: 7152–7161.
Hillyard, S.A. and Anllo-Vento, L. (1998) Event-related brain Keil, A., Smith, J.C., Wangelin, B.C., Sabatinelli, D., Bradley,
potentials in the study of visual selective attention. Proc. M.M. and Lang, P.J. (2005b) Source space analysis of single
Natl. Acad. Sci. USA, 95: 781–787. EEG epochs in the frequency domain: an application for
Härle, M., Rockstroh, B.S., Keil, A., Wienbruch, C. and steady-state potentials. Psychophysiology, 42: S73.
Elbert, T.R. (2004) Mapping the brain’s orchestration during Kemp, A.H., Gray, M.A., Eide, P., Silberstein, R.B. and
speech comprehension: task-specific facilitation of regional Nathan, P.J. (2002) Steady-state visually evoked potential
synchrony in neural networks. BMC Neurosci., 5: 40. topography during processing of emotional valence in
Ihssen, N., Heim, S. and Keil, A. (2005) Electrocortical corre- healthy subjects. Neuroimage, 17: 1684–1692.
lates of resource allocation during the attentional blink. Kemp, A.H., Gray, M.A., Silberstein, R.B., Armstrong, S.M.
Psychophysiology, 42: S67. and Nathan, P.J. (2004) Augmentation of serotonin enhances
Ihssen, N. and Keil, A. (submitted) Affective facilitation and pleasant and suppresses unpleasant cortical electrophysio-
inhibition of conscious identification: an investigation with logical responses to visual emotional stimuli in humans.
the attentional blink. Neuroimage, 22: 1084–1096.
Ihssen, N., Heim, S. and Keil, A. (2004) Emotional arousal of Kitayama, S. (1990) Interaction between affect and cognition in
verbal stimuli modulates the attentional blink. Psychophys- word perception. J. Pers. Soc. Psychol., 58: 209–217.
iology, 41: S36. Kranczioch, C., Debener, S. and Engel, A.K. (2003) Event-
Junghöfer, M., Bradley, M.M., Elbert, T.R. and Lang, P.J. related potential correlates of the attentional blink phenom-
(2001) Fleeting images: a new look at early emotion discrim- enon. Brain Res. Cogn. Brain Res., 17: 177–187.
ination. Psychophysiology, 38: 175–178. Lachaux, J.P., Rodriguez, E., Martinerie, J. and Varela, F.J.
Kamin, L.J. (1955) Relations between discrimination, appara- (1999) Measuring phase synchrony in brain signals. Hum.
tus stress, and the Taylor scale. J. Abnorm. Psychol., Brain Mapp., 8: 194–208.
51: 595–599. Lang, P.J. (1979) A bioinformational theory of emotional
Kayser, J., Tenke, C., Nordby, H., Hammerborg, D., Hugdahl, imagery. Psychophysiology, 16: 495–512.
K. and Erdmann, G. (1997) Event-related potential (ERP) Lang, P.J. (1994) The motivational organization of emotion:
asymmetries to emotional stimuli in a visual half-field par- affect-reflex connections. In: Van Goozen, S.H.M., Van de
adigm. Psychophysiology, 34: 414–426. Poll, N.E. and Sergeant, J.E. (Eds.), Emotions: Essays on
Keil, A. (2004) The role of human prefrontal cortex in Emotion Theory. Lawrence Erlbaum Associates, Hillsdale,
motivated perception and behavior: a macroscopic perspec- NJ, pp. 61–93.
tive. In: Otani, S. (Ed.), Prefrontal Cortex: From Synaptic Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1997a) Inter-
Plasticity to Cognition. Kluwer, New York. national Affective Picture System (IAPS): Technical Manual
Keil, A., Bradley, M.M., Hauk, O., Rockstroh, B., Elbert, T. and Affective Ratings. The Center for Research in Psycho-
and Lang, P.J. (2002) Large-scale neural correlates of affec- physiology, University of Florida, Gainesville.
tive picture processing. Psychophysiology, 39: 641–649. Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1997b)
Keil, A., Gruber, T. and Muller, M.M. (2001a) Functional Motivated attention: affect, activation, and action. In: Lang,
correlates of macroscopic high-frequency brain activity in P.J., Simons, R.F. and Balaban, M.T. (Eds.), Attention and
the human visual system. Neurosci. Biobehav. Rev., 25: Orienting: Sensory and Motivational Processes. Lawrence
527–534. Erlbaum Associates, Hillsdale, NJ, pp. 97–135.
Keil, A., Gruber, T., Muller, M.M., Moratti, S., Stolarova, M., Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1998) Emotion,
Bradley, M.M. and Lang, P.J. (2003) Early modulation of motivation, and anxiety: brain mechanisms and psychophys-
visual perception by emotional arousal: evidence from iology. Biol. Psychiatry, 44: 1248–1263.
steady-state visual evoked brain potentials. Cogn. Affect. Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1999) Interna-
Behav. Neurosci., 3: 195–206. tional Affective Picture System (IAPS): Instruction Manual
Keil, A. and Ihssen, N. (2004) Identification facilitation for and Affective Ratings, Technical Report. The Center for
emotionally arousing verbs during the attentional blink. Research in Psychophysiology, University of Florida,
Emotion, 4: 23–35. Gainesville, FL.
Keil, A., Moratti, S., Sabatinelli, D., Bradley, M.M. and Lang, Lehr, D.J., Bergum, B.O. and Standing, T.E. (1966) Response
P.J. (2005a) Additive effects of emotional content and spatial latency as a function of stimulus affect and presentation
selective attention on electrocortical facilitation. Cereb. order. Percept. Mot. Skills, 23: 1111–1116.
Cortex, 15: 1187–1197. Leppänen, J.M. and Hietanen, J.K. (2003) Affect and face
Keil, A., Müller, M.M., Gruber, T., Wienbruch, C., Stolarova, perception: odors modulate the recognition advantage of
M. and Elbert, T. (2001b) Effects of emotional arousal in happy faces. Emotion, 3: 315–326.
the cerebral hemispheres: a study of oscillatory brain Leppänen, J.M. and Hietanen, J.K. (2004) Positive facial
activity and event-related potentials. Clin. Neurophysiol., expressions are recognized faster than negative facial expres-
112: 2057–2068. sions, but why? Psychol. Res., 69: 22–29.
231

Luck, S.J. and Hillyard, S.A. (1995) The role of attention in Neumann, R. and Strack, F. (2000) Approach and avoidance:
feature detection and conjunction discrimination: an electro- the influence of proprioceptive and exteroceptive cues on en-
physiological analysis. Int. J. Neurosci., 80: 281–297. coding of affective information. J. Pers. Soc. Psychol., 79:
Luck, S.J., Vogel, E.K. and Shapiro, K.L. (1996) Word mean- 39–48.
ings can be accessed but not reported during the attentional Öhman, A., Flykt, A. and Esteves, F. (2001a) Emotion drives
blink. Nature, 383: 616–618. attention: detecting the snake in the grass. J. Exp. Psychol.
Lundqvist, D., Flykt, A. and Öhman, A. (1998) Karolinska Gen., 130: 466–478.
Directed Emotional Faces. Dept. of Neurosciences, Karo- Öhman, A., Lundqvist, D. and Esteves, F. (2001b) The face in
linska Hospital, Stockholm. the crowd revisited: a threat advantage with schematic stim-
Martinez, A., Anllo-Vento, L., Sereno, M.I., Frank, L.R., uli. J. Pers. Soc. Psychol., 80: 381–396.
Buxton, R.B., Dubowitz, D.J., Wong, E.C., Hinrichs, H., Palomba, D., Angrilli, A. and Mini, A. (1997) Visual evoked
Heinze, H.J. and Hillyard, S.A. (1999) Involvement of striate potentials, heart rate responses and memory to emotional
and extrastriate visual cortical areas in spatial attention. Nat. pictorial stimuli. Int. J. Psychophysiol., 27: 55–67.
Neurosci., 2: 364–369. Pizzagalli, D.A., Greischar, L.L. and Davidson, R.J. (2003)
Megibow, M.M. and De Jamaer, K.A. (1979) Hemispheric Spatio-temporal dynamics of brain mechanisms in aversive
equality in reaction times to emotional and non-emotional classical conditioning: high-density event-related potential
nouns. Percept. Mot. Skills, 49: 643–647. and brain electrical tomography analyses. Neuropsychologia,
Mesulam, M.M. (1998) From sensation to cognition. Brain, 41: 184–194.
121: 1013–1052. Potter, M.C., Chun, M.M., Banks, B.S. and Muckenhoupt, M.
Mini, A., Palomba, D., Angrilli, A. and Bravi, S. (1996) (1998) Two attentional deficits in serial target search: the
Emotional information processing and visual evoked brain visual attentional blink and an amodal task-switch deficit.
potentials. Percept. Mot. Skills, 83: 143–152. J. Exp. Psychol. Learn. Mem. Cogn., 24: 979–992.
Mogg, K., Mathews, A. and Weinman, J. (1989) Selective Pourtois, G., Grandjean, D., Sander, D. and Vuilleumier, P.
processing of threat cues in anxiety states: a replication. (2004) Electrophysiological correlates of rapid spatial orient-
Behav. Res. Ther., 27: 317–323. ing towards fearful faces. Cereb. Cortex, 14: 619–633.
Moratti, S. and Keil, A. (2005) Cortical activation during Pulvermüller, F. (1996) Hebb’s concept of cell assemblies and
Pavlovian fear conditioning depends on heart rate response the psychophysiology of word processing. Psychophysiology,
patterns: an MEG study. Brain Res. Cogn. Brain Res., 25: 33: 317–333.
459–471. Ramus, F. (2001) Outstanding questions about phonological
Moratti, S., Keil, A. and Miller, G.A. (2006) Fear but not processing in dyslexia. Dyslexia, 7: 197–216.
awareness predicts enhanced sensory processing in fear con- Ray, W.J. and Cole, H.W. (1985) EEG alpha activity reflects
ditioning. Psychophysiology, 43: 216–226. attentional demands, and beta activity reflects emotional and
Moratti, S., Keil, A. and Stolarova, M. (2004) Motivated cognitive processes. Science, 228: 750–752.
attention in emotional picture processing is reflected by Raymond, J.E., Shapiro, K.L. and Arnell, K.M. (1992) Tem-
activity modulation in cortical attention networks. Neuro- porary suppression of visual processing in an RSVP task: an
image, 21: 954–964. attentional blink? J. Exp. Psychol. Hum. Percept. Perform.,
Morgan, S.T., Hansen, J.C. and Hillyard, S.A. (1996) Selective 18: 849–860.
attention to stimulus location modulates the steady-state Richards, A. and Blanchette, I. (2004) Independent manipula-
visual evoked potential. Proc. Natl. Acad. Sci. USA, 93: tion of emotion in an emotional stroop task using classical
4770–4774. conditioning. Emotion, 4: 275–281.
Most, S.B., Chun, M.M., Widders, D.M. and Zald, D.H. (2005) Sabatinelli, D., Bradley, M.M., Fitzsimmons, J.R. and Lang,
Attentional rubbernecking: Cognitive control and personal- P.J. (2005) Parallel amygdala and inferotemporal activation
ity in emotion-induced blindness. Psychon. Bull. Rev., 12: reflect emotional intensity and fear relevance. Neuroimage,
654–661. 24: 1265–1270.
Müller, M.M., Keil, A., Gruber, T. and Elbert, T. (1999) Process- Schimmack, U. and Derryberry, D. (2005) Attentional inter-
ing of affective pictures modulates right-hemispheric gamma ference effects of emotional pictures: threat, negativity, or
band EEG activity. Clin. Neurophysiol., 110: 1913–1920. arousal? Emotion, 5: 55–66.
Müller, M.M., Malinowski, P., Gruber, T. and Hillyard, S.A. Schupp, H.T., Cuthbert, B.N., Bradley, M.M., Cacioppo, J.T.,
(2003) Sustained division of the attentional spotlight. Nature, Ito, T. and Lang, P.J. (2000) Affective picture processing: the
424: 309–312. late positive potential is modulated by motivational rele-
Müller, M.M., Picton, T.W., Valdes-Sosa, P., Riera, J., Teder- vance. Psychophysiology, 37: 257–261.
Salejarvi, W.A. and Hillyard, S.A. (1998) Effects of spatial Schupp, H.T., Junghofer, M., Weike, A.I. and Hamm, A.O.
selective attention on the steady-state visual evoked potential in (2003a) Attention and emotion: an ERP analysis of facilitated
the 20–28 Hz range. Brain Res. Cogn. Brain Res., 6: 249–261. emotional stimulus processing. Neuroreport, 14: 1107–1110.
Müller, M.M., Teder, W. and Hillyard, S.A. (1997) Magneto- Schupp, H.T., Junghofer, M., Weike, A.I. and Hamm, A.O.
encephalographic recording of steady-state visual evoked (2003b) Emotional facilitation of sensory processing in the
cortical activity. Brain Topogr., 9: 163–168. visual cortex. Psychol. Sci., 14: 7–13.
232

Schupp, H.T., Ohman, A., Junghofer, M., Weike, A.I., Stolarova, M., Keil, A. and Moratti, S. (2006) Modulation
Stockburger, J. and Hamm, A.O. (2004) The facilitated of the C1 visual event-related component by conditioned
processing of threatening faces: an ERP analysis. Emotion, stimuli: evidence for sensory plasticity in early affective
4: 189–200. perception. Cereb. Cortex, 16: 876–887.
Shapiro, K.L., Caldwell, J. and Sorensen, R.E. (1997) Tallon-Baudry, C. and Bertrand, O. (1999) Oscillatory gamma
Personal names and the attentional blink: a visual ‘‘cock- activity in humans and its role in object representation.
tail party’’ effect. J. Exp. Psychol. Hum. Percept. Perform., Trends Cogn. Sci., 3: 151–162.
23: 504–514. Tucker, D.M. (1984) Lateral brain function in normal and
Silberstein, R.B., Ciorciari, J. and Pipingas, A. (1995) disordered emotion: interpreting electroencephalographic
Steady-state visually evoked potential topography during evidence. Biol. Psychol., 19: 219–235.
the Wisconsin card sorting test. Electroencephalogr. Clin. Vogel, E.K., Luck, S.J. and Shapiro, K.L. (1998) Electrophys-
Neurophysiol., 96: 24–35. iological evidence for a postperceptual locus of suppression
Sloan, D.M., Bradley, M.M., Dimoulas, E. and Lang, P.J. during the attentional blink. J. Exp. Psychol. Hum. Percept.
(2002) Looking at facial expressions: dysphoria and facial Perform., 24: 1656–1674.
EMG. Biol. Psychol., 60: 79–90. Watts, F.N., McKenna, F.P., Sharrock, R. and Trezise, L.
Smith, J.C., Keil, A., Wangelin, B.C., Sabatinelli, D. and Lang, (1986) Colour naming of phobia-related words. Br. J.
P.J. (2005) Single-trial analyses of steady-state visual poten- Psychol., 77(Pt 1): 97–108.
tials: effects of emotional arousal. Psychophysiology, Wedding, D. and Stalans, L. (1985) Hemispheric differences in
42: S116. the perception of positive and negative faces. Int. J. Neuro-
Smith, N.K., Cacioppo, J.T., Larsen, J.T. and Chartrand, T.L. sci., 27: 277–281.
(2003) May I have your attention, please: electrocortical re- Williamson, S., Harpur, T.J. and Hare, R.D. (1991) Abnormal
sponses to positive and negative stimuli. Neuropsychologia, processing of affective words by psychopaths. Psychophys-
41: 171–183. iology, 28: 260–273.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 12

Intonation as an interface between language


and affect

Didier Grandjean, Tanja Bänziger and Klaus R. Scherer

Swiss Center for Affective Sciences, University of Geneva, 7 rue des Battoirs, 1205 Geneva, Switzerland

Abstract: The vocal expression of human emotions is embedded within language and the study of into-
nation has to take into account two interacting levels of information — emotional and semantic meaning.
In addition to the discussion of this dual coding system, an extension of Brunswik’s lens model is proposed.
This model includes the influences of conventions, norms, and display rules (pull effects) and psychobi-
ological mechanisms (push effects) on emotional vocalizations produced by the speaker (encoding) and the
reciprocal influences of these two aspects on attributions made by the listener (decoding), allowing the
dissociation and systematic study of the production and perception of intonation. Three empirical studies
are described as examples of possibilities of dissociating these different phenomena at the behavioral and
neurological levels in the study of intonation.

Keywords: prosody; intonation; emotion; attention; linguistic; affect; brain imagery

Emotions are defined as episodes of massive, syn- The coding of information in prosody
chronous recruitment of mental and somatic re-
sources to adapt to or cope with stimulus events One of the most interesting issues in vocal affect
that are subjectively appraised as being highly signalling is the evolutionary continuity bet-
pertinent for an individual and involve strong mo- ween animal vocalizations of motivational/emo-
bilization of the autonomous and somatic nervous tional states and human prosody. Morton (1982)
systems (Scherer, 2001; Sander et al., 2005). The proposed universal ‘‘motivational-structural rules’’
patterns of activations created in those systems in an attempt to understand the relationship bet-
will be reflected, generally, in expressive behavior ween fear and aggression in animal vocalizations.
and, specifically, will have a powerful impact on Morton tried to systematize the role of fundamental
the production of vocal expressions. Darwin frequency, energy, and quality (texture) of vocali-
(1998/1872) observed that in many species, voice zation for the signalling of aggressive anger and
is moreover exploited as an iconic affective signal- fear. The major differences between these charac-
ling device. Vocalizations, as well as other emo- teristics of anger and fear are the trajectory of the
tional expressions, are functionally used by contours of the respective sounds and the roughness
conspecifics and sometimes even members of other or the ‘‘thickness’’ of the sounds. For instance, the
species to make inferences about the emotional/ ‘‘fear endpoint’’ (no aggression) is characterized by
motivational state of the sender. continuous high pitch compared with low pitch and
roughness at the ‘‘aggressive endpoint’’ (see Fig. 1).
Between these two extremes, Morton described var-
Corresponding author. Tel.: +41223799213; iations corresponding to mixed emotions character-
Fax: +41223799844; E-mail: Didier.Grandjean@pse.unige.ch ized by different combinations of the basic features.

DOI: 10.1016/S0079-6123(06)56012-1 235


236

Fig. 1. Motivational-structural rules (Morton, 1982). Modelling the modifications of F0 and energy of vocalizations signalling
aggression (horizontal axis) or fear/appeasement (vertical axis).

This kind of iconic affective signalling device, has suggested the notion of ‘‘domestication of affect
common to most mammals, has changed in the sounds’’ constituted by this mixture of natural and
course of human evolution in the sense that it be- cultural aspects. It is particularly important to dis-
came the carrier signal for language. This commu- tinguish ‘‘push effects’’ and ‘‘pull effects’’ (Scherer,
nication system is discretely and arbitrarily coded 1986). The push effects represent the effect of un-
and uses the human voice and the acoustic varia- derlying psychobiological mechanisms, for instance,
bility it affords to build the units of the language the increase of arousal that increases muscle tension
code. This code has been superimposed on the af- and thereby raises fundamental frequency (F0). The
fect signalling system, which continues to be used equivalence of ‘‘push’’ on the attribution side is
in human communication, for example, in affect some kind of schematic recognition, which is prob-
bursts or interjections (Scherer, 1985). Many as- ably also largely innate. These psychobiological ef-
pects of the primitive affect signalling system have fects are complemented by ‘‘pull effects,’’ that is,
been integrated into the language code via prosody. conventions, norms, and display rules that pull the
It is not only that the two codes basically coexist, voice in certain directions. The result is a ‘‘dual
but also that speech prosody has integrated some code,’’ which means that, at the same time, several
of the same sound features as described in Mor- dimensions are combined to produce speech. This
ton’s examples. These vocal productions are not complex patterning makes it difficult to understand
pure, but they are shaped by both physiological which features are specific to particular emotions in
reactions and cultural conventions. Wundt (1900) the acoustic signal.
237

Scherer et al. (1984) have proposed distinguish- resulting in emotion inference. The model encour-
ing coding via covariation (continuous affect sig- ages voice researchers to measure the complete
nalling), determined mostly by biopsychological communicative process, including (a) the emo-
factors and configuration (discrete message types), tional state expressed, (b) the acoustically meas-
shaped by linguistic and sociocultural factors. This ured voice cues, (c) the perceptual judgments of
distinction is very important but unfortunately voice cues, and (d) the process that integrates all
often neglected in the literature. cues into a judgment of the encoded emotion.
Given the complex interaction of different de- Figure 2 further illustrates that both psychobio-
terminants in producing intonation patterns, it is logical mechanisms (push effects) and social norms
of particular importance to clearly distinguish be- (pull effects) will influence the expression (encod-
tween production or encoding and perception or ing) of vocal expressions and, conversely, will be
decoding of intonational messages. Brunswik’s transposed into rules used for impression forming
‘‘lens model’’ allows one to combine the study of (decoding).
encoding and decoding of affect signals (Brunswik, The patterns of intonation produced during
1956). This model (see Fig. 2) presumes that emo- a social interaction are strongly influenced by con-
tion is expressed by distal indicator cues (i.e., ventions shaped by different social and cultural
acoustic parameters), which can be extracted from rules (Ishii et al., 2003). Moreover, for tone lan-
the acoustic waveforms. These objective distal cues guages, different tones may convey specific seman-
are perceived by human listeners who, on the basis tic meaning (Salzmann, 1993). In these cases, tones
of their proximal percept, make a subjective attri- with semantic functions interact in a complex man-
bution of what underlying emotion is expressed by ner with emotional influences. It is very important
the speaker, often influenced by the context. This to keep apart these different aspects because they
model allows one to clearly distinguish between are governed by very different principles (Scherer
the expression (or encoding) of emotion on the et al., 2003).
sender side, the transmission of the sound, and Given that the underlying biological processes
the impression (or decoding) on the receiver side, are likely to be dependent on both the idiosyncratic

Fig. 2. Adaptation of Brunswik’s lens model, including the influences of conventions, norms, and display rules (pull effects) and
psychobiological mechanisms (push effects) on emotional vocalizations produced by the speaker (encoding) and the reciprocal in-
fluences of these two aspects on attributions made by the listener (decoding).
238

nature of the individual and the specific nature of muscle tension, influence not only the source
the situation, relatively strong interindividual dif- (namely, the vocal cords), but also the resonances
ferences in the expressive patterns will result from produced by the modifications of the vocal tract
push effects. Conversely, for pull effects, a very shapes. Spectral energy distribution is influenced
high degree of symbolization and conventionaliza- by the vocal tract of the speaker and contributes to
tion, and thus comparatively few and small indi- the possibilities of identifying an individual by the
vidual differences, are expected. With respect to sound of the voice. Recently, a study conducted by
cross-cultural comparison, one would expect the Ghazanfar (2005) has shown that monkeys are
opposite: Very few differences between cultures for able to attribute the size of the body from the vo-
push effects and large differences for pull effects. cal production of a conspecific. Spectral energy
distribution is also modified by emotional proc-
esses, with the relative amount of energy in high-
What is the role of intonation in vocal affect and low-frequency bands changing for different
communication? emotions. For example, the voice quality can be
modified in terms of roughness or sharpness dur-
First, we propose defining different terms related ing an emotional process, information that can be
to intonation to clarify the concepts. We suggest used by the listener to infer the emotional state of
that the term ‘‘prosody’’ refers to all supraseg- the interlocutor. It is very important to take these
mental changes in the course of a spoken utterance aspects into account when we want to understand
— intonation, amplitude, envelope, tempo, the organism’s ability to attribute emotional states
rhythm, and voice quality. We talk about intona- to conspecifics and even to other species.
tion as the contour of F0 in the utterance. The Different features of prosody might be coded
amplitude envelope is determined by the contour differentially in terms of the distinction mentioned
of acoustic energy variations over the utterance above between continuous or discrete types of
and will most of the time be correlated with F0. coding. Thus, Scherer et al. (1984) have shown
The tempo consists of the number of phonemic that F0 is coded continuously, whereas many in-
segments per time unit. The rhythm corresponds tonation contour shapes are coded configuration-
to the structure of F0 accents, amplitude peaks, ally. On the left panel of Fig. 3, the pitch range of
and pausing distribution in the utterance. Finally, an utterance is continuously varied using copy
voice quality is defined by the distribution of en- synthesis (following the covariation principle);
ergy in the spectrum as produced by different with wide pitch range, there is a continuous in-
phonation modes. The modifications of the vocal crease in the amount of emotional content
tract characteristics, for example, the overall perceived in the utterance. On the right panel of

Fig. 3. Effects of F0 range manipulation (covariation principle, left panel) and different contour shapes (configuration principle) for
‘‘Wh’’ and ‘‘Y/N’’ questions (right panel) on the perception of emotion.
239

Fig. 3, configuration effects are demonstrated; processes at the central nervous system (CNS) by
whether the utterance is perceived as challenging electroencephalography (EEG). Finally, the rela-
or not depends on whether final fall or final rise is tionship between emotion and attention in pro-
used. This depends also on the context: It makes a sodic decoding is addressed in a study using
major difference if a ‘‘Wh’’ question is pronounced functional magnetic resonance imagery (fMRI).
(‘‘Where are you going’’) with a final fall that is
not really challenging versus a final rise. The effect
depends on configuration features rather than Are the emotions specific intonation contours?
continuous covariation.
Several distinctions are crucial and should be The question of how specifically intonation
addressed or at least taken into account in future codes emotion has been addressed in a study in-
research related to vocal emotional communica- vestigating the contribution of intonation to the
tion. The distinction between coding based on an vocal communication of emotions (Bänziger and
ancient affect signalling system and coding based Scherer, 2005). Intonation is defined in this context
on language should be systematically taken into as pitch (or F0) fluctuations over time. Other pro-
account, particularly with respect to the distinction sodic aspects such as rhythm, tempo, or loudness
between covariation and configuration principles fluctuations were not included in this study.
(see also Scherer, 2003). Because of the possibility Along the lines of the distinction outlined by
that different coding principles underlie different Scherer et al. (1984), authors from different
features of prosody, we also need to distinguish research backgrounds independently postulated
very carefully between intonation as a phenome- that (a) specific configurations of pitch patterns
non as carried by F0, tempo, amplitude, and voice (pitch contours) reflect and communicate specific
quality; all of these aspects interact but it is pos- emotional states (e.g., Fonagy and Magdics, 1963)
sible to pull them apart. We also need to distin- and (b) continuous variation of pitch features
guish between push and pull effects; in other (such as pitch level or pitch range) reflect and
words, the question as to what extent there are communicate features of emotional reactions, such
differences in physiological push, producing as emotional arousal (e.g., Pakosz, 1983). Evi-
certain vocal effects, versus the effects produced dence supporting the first claim (existence of emo-
essentially by the speaker, modifying the command tion-specific pitch contours) consists mostly of
structure for the vocalizations on the basis of tem- selected examples rather than of empirical exam-
plates given by linguistic or cultural schema (pull). ination of emotional speech recordings. On the
Finally, we need to distinguish much more care- other hand, efforts to describe/analyze the intona-
fully between encoding and decoding mechanisms, tion of actual emotional expressions have been
how the sender encodes both push and pull effects limited by the use of simplified descriptors, such as
for a certain affect, and what processes are used to measures of overall pitch level, pitch range, or
decode the production influenced by emotional overall rise/fall of pitch contours. This line of re-
processes. The ideal way to understand these two search established that a number of acoustic fea-
different processes is to study them jointly as sug- tures — such as F0 mean or range, intensity mean
gested by the Brunswikian lens model (Brunswik, or range, and speech rate — vary continuously
1956). with emotional arousal (a review of this work is
In the following sections, we present three dif- described in Scherer, 2003). It is far less clear as to
ferent approaches to study and understand the what extent specific F0 contours can be associated
encoding and decoding processes at different with different emotions, especially independently
levels, using the conceptual distinctions explained of linguistic content. To examine this issue, quan-
in the above paragraphs. The first approach tifiable and comparable descriptions of F0 con-
is focused on the question of the specificity of tours are needed. The study we describe used
contour intonation for different emotions. The a simple procedure to stylize F0 contours for emo-
second exemplifies a study addressing the decoding tional expressions.
240

Fig. 4. Stylization example for an instance of (low aroused) ‘‘happiness’’ with sequence 1 (‘‘hät san dig prong nju ven tsi’’).

The stylization (see Fig. 4) was applied to correspond respectively to the minimum, maxi-
144 emotional expressions (sampled from a larger mum, and minimum of the F0 excursion for the
set of emotional expressions described in detail by first operationally defined ‘‘accent’’ of each se-
Banse and Scherer, 1996). Expressions produced by quence. Those local minima and maxima are
nine actors who pronounced two sequences of measured for the syllables ‘‘san dig’’ in sequence
seven syllables (1. ‘‘hät san dig prong nju ven tsi’’; 1 and for the syllables ‘‘gött laich’’ in sequence 2.
2. ‘‘fi gött laich jean kill gos terr’’) and expressed Points five (‘‘2 min1’’), six (‘‘2max’’), and seven
eight emotions were used in this study. Two in- (‘‘2 min2’’) correspond respectively to the mini-
stances of ‘‘fear,’’ ‘‘happiness,’’ ‘‘anger,’’ and ‘‘sad- mum, maximum, and minimum of the F0 excur-
ness,’’ with ‘‘low arousal’’ (labeled: ‘‘anxiety,’’ sion for the second operationally defined ‘‘accent’’
‘‘happiness,’’ ‘‘cold anger,’’ and ‘‘sadness’’) of each sequence. They are measured for the syl-
and ‘‘high arousal’’ (labeled: ‘‘panic fear,’’ ‘‘ela- lables ‘‘prong nju ven’’ and ‘‘jean kill gos.’’ Points
tion,’’ ‘‘hot anger,’’ ‘‘despair’’) were included in this eight (‘‘3 min’’), nine (‘‘3max’’), and ten (‘‘final’’)
study. correspond to the final ‘‘accent’’ of each sequence:
Ten key points were identified for each F0 con- The local minimum, maximum, and minimum for
tour. The first point (‘‘start’’) corresponds to the the syllables ‘‘tsi’’ and ‘‘ter.’’ Fig. 4 shows an il-
first F0 point detected for the first voiced section in lustration of this stylization for a happy expression
each expression. This point is measured on (first utterance).
the syllable ‘‘hät’’ in sequence 1 and on the sylla- The pattern represented in Figure 4 — two ‘‘ac-
ble ‘‘fi’’ in sequence 2. The second (‘‘1 min1’’), cents’’ (sequences of local F0 min1–max–min2)
third (‘‘1max’’), and fourth points (‘‘1 min2’’) followed by a final fall — was the most frequent
241

pattern for the 144 expressions submitted to this such as F0 mean. Looking at the residual variance
analysis. The count of F0 ‘‘rises’’ (local ‘‘min1’’ after regressing F0 mean (computed for each ex-
followed by ‘‘max’’), ‘‘falls’’ (local ‘‘max’’ followed pression) on the points represented in Fig. 5, there
by ‘‘min2’’), and ‘‘accents’’ (‘‘min1’’ followed by remains only a slight effect of expressed emotion
‘‘max’’ followed by ‘‘min2’’) for the first accented on points ‘‘2max’’ and ‘‘final.’’ The second max-
part, the second accented part, and the final syl- imum tends to be higher for recordings expressing
lable was not affected by the expressed emotions elation, hot anger, and cold anger than for re-
but varied for different speakers and for the two cordings expressing other emotions. The final F0
sequences of syllables that they pronounced. value tends to be relatively lower for hot anger and
In order to control for differences in F0 level cold anger than for other emotions. Figure 5 fur-
between speakers, a ‘‘baseline’’ value had to be ther shows that the range of F0 fluctuations is af-
defined for each speaker. An average F0 value was fected by emotional arousal and also that F0 range
computed from 112 emotional expressions (includ- (expressed in a linear scale) is on average larger for
ing the 16 expressions used in this study) produced portrayals with high arousal (and high F0 level)
by each speaker. Fig. 5 shows the differences in than for portrayals with low arousal (and low F0
hertz (averaged across speakers and sequences of level). It is likely that both the level and range of
syllables) between the observed F0 points in each F0 are enhanced in portrayals with high arousal as
expression and the speaker baseline value for each a consequence of increased vocal effort in those
expressed emotion. portrayals.
Figure 5 shows that F0 level is affected by emo- The results reported above show that in this
tional arousal. The F0 points for emotions with study only the overall level of the F0 contours was
low arousal (such as sadness, happiness, and anx- affected by expressed emotions and determined the
iety) are generally lower than the F0 points for emotion inferences of the judges in a powerful and
emotions with high arousal (despair, elation, panic statistically significant fashion. As could be ex-
fear, and hot anger). The description of the dif- pected from frequently replicated results in the lit-
ferent points in the contour does not appear to add erature, the height of F0 is likely to be reliably
much information to an overall measure of F0, interpreted as indicative of differential activation

Fig. 5. Average F0 values by portrayed emotion. Note: The number of observations varies from 18 (for ‘‘start’’ with hot anger, cold
anger, and elation; for ‘‘1max’’ with cold anger and panic fear) to 7 (for ‘‘final’’ with sadness). It should be noted also that there is a
sizable amount of variance around the average values shown for all measurement points.
242

or arousal. These results do not encourage the relevant parameters. Preliminary work using this
notion that there are emotion-specific intonation technique has shown the importance of voice qual-
contours. However, some of the detailed results ity in decoding intonation (Bänziger et al., 2004).
suggest that aspects of contour shape (such as
height of selected accents and final F0 movement)
may well differentially affect emotion inferences. The neural dynamics of intonation perception
However, it seems unlikely that such features will
have a discrete, iconic meaning with respect to The second study, which highlights the question
emotional content. It seems reasonable to assume related to the time course of decoding emotional
that, although the communicative value of F0 level prosody, was conducted by Grandjean et al. (in
may follow a covariation model, the interpretation preparation). The main goal of this experiment
of various features of F0 contour shape seems to was to address the timing related to the perception
be best described by a configuration model. Con- of emotional prosody, linguistic pragmatic ac-
cretely, contour shape, or certain central features cents, and phonemic identification. Using EEG
thereof, may acquire emotional meaning only in and spatio-temporal analyses of the electrical
specific linguistic and pragmalinguistic contexts brain signals, Grandjean et al. (2002) showed that
(including phonetic, syntactic, and semantic fea- different patterns of activation are related to spe-
tures, as well as normative expectations). Further- cific decoding processes in emotional prosody
more, the role of F0 contour may vary depending identification compared with pragmatic and pho-
on the complexity of the respective emotion and its nemic identifications. Three simple French words
dependence on a sociocultural context. Thus, one were used (‘‘ballon,’’ ‘‘talon,’’ and ‘‘vallon’’), with
would expect covariation effects for simple, uni- the F0 contour being systematically manipulated
versally shared emotions that are closely tied to using Mbrola synthesis (Dutoit et al., 1996) to
biological needs and configuration effects for com- produce happiness, sadness, and neutral emotion
plex emotions and affective attitudes that are expressions, as well as affirmative and interroga-
determined by socioculturally variable values and tive utterance types (see Fig. 6).
symbolic meaning. During EEG recording, the participants had to
To summarize, the results indicate that there are identify emotional prosody, linguistic prosody, and
no specific contours or different shapes related to phonemic differences within three different coun-
the emotions studied (Bänziger and Scherer, 2005). terbalanced blocks. Time slots for occurrences of
There are strong differences in the F0 level for the specific topographical brain maps obtained by clus-
different kinds of emotions related to the underly- ter analyses (Lehmann, 1987; see also Michel et al.,
ing arousal intensity, which is well known from past 2004) on grand average of event-related potentials
research (see Fig. 5). However, in this study, for the (ERPs) are different for the three recognition tasks.
first time, we have evidence that the accent struc- The results highlight specific processes related to
ture, even with meaningless speech, with an impor- emotional and semantic prosody identification
tant corpus, and spoken by nine different actors, is compared with phonemic identification (see Fig. 7).
very similar for different emotions. The only single Specifically, the first three ERP electrical brain
effect is related to elation, with a disproportional maps (C1, C2, and C3 maps on the Fig. 7) are
rise on the second accent (see Fig. 5). These results common to the different experimental conditions.
indicate that the accent structure alone does not Between 250 and 300 ms to 400 ms, specific
show much specificity: There may be slight effects processes occurred for emotional prosodic identi-
for the height of the secondary accent. There could fication and semantic linguistic identification, dem-
be emotion-specific effects in contour aspects that onstrating the involvement of different underlying
are time critical: Lengthening–shortening in partic- neural networks subserving these different mental
ular segments, upswing–downswing acceleration. processes. In fact, the statistical analyses show
To study these aspects systematically, we need specificity of the maps for both the emotional
further research using the synthetic variation of the prosody and the linguistic pragmatic conditions,
243

Fig. 6. Examples of pitch analyses and sonograms for the French utterance ‘‘ballon’’ for the different experimental conditions used in
the EEG study.

Fig. 7. Occurrence of the different topographical brain maps over time (1000 ms after the onset of the stimulus) for the three
experimental conditions (semantic and emotional prosody, and phonemic identifications). The different colors correspond to the
different brain electrical maps obtained from the grand average ERPs. The maps are represented on the global field power (GFP). Note
the specific map related to emotional identification characterized by a right anterior positivity (E map) and the specific map for the
semantic prosody condition with a large central negativity (S map).
244

when compared with the two other conditions, re- involved in different functions at the prosodic level
spectively (Grandjean et al., 2002). These results with vocal synthesis. In contrast to fMRI tech-
indicate that specific neural circuits are involved in niques, EEG methods allow the study of not only
the recognition of emotional prosody compared the interactions of different brain areas in prosodic
with linguistic and phonemic identification tasks. perception, but also the timing of these processes
A right anterior positivity was measured on the to identify the brain structures involved in prosody
scalp; this result is compatible with a previous perception.
fMRI study demonstrating anterior activations in
right dorsolateral and orbitofrontal regions during
emotional identification compared with phonetic Impact of attention on decoding of emotional
identifications of the same stimuli (Wildgruber et prosody
al., 2005). The involvement of the left part of the
frontal region was highlighted in another fMRI Another fundamental issue concerns the ability of
study when the participants had to identify lin- humans to rely on emotion to adapt or cope with
guistic information compared with emotional particularly relevant events or stimuli (Sander et
prosodic information (Wildgruber et al., 2004). al., 2003, 2005). Emotional prosody often serves
However, the two temporal parts of the hemi- the function of social communication and indeed
spheres are differentially involved in different sub- has an impact on behavioral level related to indi-
processes that contribute to the recognition of the vidual differences; that is, attention toward angry
emotional content of a word or a sentence (see voices increases neuronal activity in orbitofrontal
Schirmer and Kotz, 2006). For instance, different regions related to interindividual sensitivity of
brain networks process temporal information com- punishment (Sander et al., 2005). To detect emo-
pared with spectral information, respectively, in the tional signals in the environment, which is poten-
left temporal versus the right temporal parts of the tially relevant for survival, the CNS of organisms
brain (Zatorre and Belin, 2001). seems to be able to reorient attention via reflexive
The specific electrical map related to the recog- mechanisms, even if the voluntary attention
nition of emotional prosody in this EEG experi- processes are occupied with another specific task
ment cannot be explained solely by the fact that the (Vuilleumier, 2005). This reflexive process has been
intonation contour was modified, because we also extensively studied in the visual domain (Vuilleu-
used different F0 contours for the linguistic prag- mier et al., 2001; Pourtois et al., 2004) but only
matic condition (interrogation and affirmative con- rarely in the auditory domain (Mitchell et al.,
tours). Moreover, the same stimuli were used in the 2003; Wambacq et al., 2004). The last research
phonemic identification condition, demonstrating example in this chapter addresses this question
that this specific emotional prosody map is not re- through a brain imaging study of emotionally an-
lated to the differences of basic acoustical features gry prosody compared with neutral prosody in a
but rather related to the type of the participant’s dichotic listening study (see Fig. 8). In two experi-
recognition task. This study underlines the possi- ments using fMRI techniques, Grandjean et al.
bility of using speech synthesis to modify system- (2005) demonstrated an increase of neuronal ac-
atically acoustic features of emotional prosody, tivities in the bilateral superior temporal sulcus
inducing different types of categorization processes (STS), known to be sensitive for human voices
related to the participant’s tasks. In the future, this (Belin et al., 2000), when exposed to angry pros-
type of paradigm could allow researchers interested ody compared with neutral prosody (controlling
in the understanding of perception of emotional for signal amplitude level and envelope, as well as
prosody to study the integration of different sub- F0 level; see Fig. 8).
processes contributing to the subjective perception This increase of STS activity occurred even
of intonation in emotional processes. when the angry emotional prosody was not the
Further studies are needed to manipulate focus of attention in a dichotic listening paradigm,
systematically the different acoustical dimensions indicating possible reflexive mechanisms related to
245

Fig. 8. Attention and emotional prosody were manipulated in two fMRI experiments. (a) Dichotic listening paradigm allowing the
presentation of different vocalizations at the left and right ears and the manipulation of spatial attention toward the right or the left
side. (b) Cerebral activations of the right hemisphere. An increase of neuronal activity for angry relative to neutral speech prosody was
found in the right STS (red, Po0.001). An anterior region of right STS was modulated by spatial attention directed to the left relative
to the right ear (green, Po0.005). These modulations of activations by emotion and attention occurred within voice-selective areas
(blue line). (c) Right STS activation in Experiment 1. Blood Oxygen Level-Dependent (BOLD) responses were increased for angry
compared with neutral speech. (d) The same cluster in the right STS as in Experiment 2. Activation occurred only in response to vocal
stimuli and not synthetic sounds.

anger prosody. Thus, emotional prosody seems to by taking into account the specificity of acoustic
be able to induce an increase of neuronal activities signals not only in perception, but also in pro-
even when the respective emotional-relevant event is duction. Moreover, the different subprocesses in-
not the focus of voluntary attention, as previously volved in these two mechanisms, such as temporal
demonstrated in the visual domain (for a review, see unfolding of decoding or the differential effects of
Vuilleumier, 2005). Thus, the human auditory sys- push and pull effects, during encoding proc-
tem interacting with other brain areas such as the esses should be manipulated systematically in fu-
amygdala would be able to allocate attentional re- ture research to allow better differentiation of
sources to compute the relevant information, mod- the determinants of these phenomena. Future
ifying attention allocation (Sander et al., 2005). research in this field should systematically manip-
ulate different acoustical parameters (i.e., the
Conclusion acoustic features underlying prosody properly
speaking and vocal quality in a more general
The research examples described above highlight sense). In addition, the current research shows the
the importance of investigating emotional prosody utility of combining behavioral and neuroscience
246

research in trying to disentangle the complex ganization of Psychophysiology (IOP), August 2002,
system of intonational structure in a dually coded Montreal, Canada.
Grandjean, D., Sander, D., Pourtois, G., Schwartz, S., Seghier,
system, subject to psychobiological push and
M., Scherer, K.R. and Vuilleumier, P. (2005) The voices of
sociocultural pull. wrath: brain responses to angry prosody in meaningless
speech. Nat. Neurosci., 8: 145–146.
Ishii, K., Reyes, J.A. and Kitayama, S. (2003) Spontaneous
Abbreviations attention to word content versus emotional tone: differences
among three cultures. Psychol. Sci., 14: 39–46.
CNS central nervous system Lehmann, D. (1987) Principles of spatial analyses. In: Gevins,
A.S. and Rémond, A. (Eds.) Handbook of Electroencepha-
EEG electroencephalography
lography and Clinical Neurophysiology, Vol. 1. Methods of
ERPs event-related potentials Analyses of Brain Electrical and Magnetic Signals, Elsevier,
F0 fundamental frequency Amsterdam, pp. 309–354.
fMRI functional magnetic resonance Michel, C.M., Murray, M.M., Lantz, G., Gonzalez, S., Spinelli,
imagery L. and Grave de Peralta, R. (2004) EEG source imaging.
Clin. Neurophysiol., 115: 2195–2222.
max maximum
Mitchell, R.L., Elliott, R., Barry, M., Cruttenden, A. and
min minimum Woodruff, P.W. (2003) The neural response to emotional
ms milliseconds prosody, as revealed by functional magnetic resonance im-
STS superior temporal sulcus aging. Neuropsychologia, 41: 1410–1421.
Morton, E.S. (1982) Grading, discreteness, redundancy, and
motivation-structural rules. In: Kroodsma, D.E., Miller,
E.H. and Ouellet, H. (Eds.), Acoustic Communication in
Birds. Academic Press, New York, pp. 182–212.
References Pakosz, M. (1983) Attitudinal judgments in intonation: some
evidence for a theory. J. Psycholinguist. Res., 12: 311–326.
Banse, R. and Scherer, KR. (1996) Acoustic profiles in vocal Pourtois, G., Grandjean, D., Sander, D. and Vuilleumier, P.
(2004) Electrophysiological correlates of rapid spatial orient-
emotion expression. J. Pers. Soc. Psychol., 70: 614–636.
Bänziger, T. and Scherer, K.R. (2005) The role of intonation in ing towards fearful faces. Cereb. Cortex, 14: 619–633.
emotional expressions. Speech Commun., 46: 252–267. Salzmann, Z. (1993) Language, Culture and Society: An Intro-
Belin, P., Zatorre, R.J., Lafaille, P., Ahad, P. and Pike, B. duction to Linguistic Anthropology (3rd ed). Westview Press,
Boulder, CO.
(2000) Voice-selective areas in human auditory cortex. Na-
ture, 403: 309–312. Sander, D., Grafman, J. and Zalla, T. (2003) The human am-
Brunswik, E. (1956) Perception and The Representative Design ygdala: an evolved system for relevance detection. Rev. Ne-
of Psychological Experiments (2nd ed). University of Cali- urosci., 14: 303–316.
Sander, D., Grandjean, D., Pourtois, G., Schwartz, S., Seghier,
fornia Press, Berkeley, CA.
Darwin, C. (1998) The Expression of the Emotions in Man and M., Scherer, K.R. and Vuilleumier, P. (2005) Emotion and
Animals. John Murray, London. (Reprinted with introduc- attention interactions in social cognition: brain regions in-
volved in processing anger prosody. Neuroimage, 28:
tion, afterword, and commentary by P. Ekman, Ed.). Oxford
University Press, New York. (Original work published 1872) 848–858.
Dutoit, T., Pagel, V., Pierret, N., Bataille, F. and Van Der Sander, D., Grandjean, D. and Scherer, K.R. (2005) A systems
Vrecken, O. (1996) The MBROLA project: towards a set of approach to appraisal mechanisms in emotion. Neural Net-
works, 18: 317–352.
high-quality speech synthesizers free of use for non-commer-
cial purposes. Proc. ICSLP’96, 3: 1393–1396. Scherer, K.R. (1985) Vocal affect signalling: a comparative ap-
Fonagy, I. and Magdics, K. (1963) Emotional patterns in in- proach. In: Rosenblatt, J., Beer, C., Busnel, M.-C. and Slater,
tonation and music. Z Phonet., 16: 293–326. P.J.B. (Eds.) Advances in the Study of Behavior, Vol. 15.
Academic Press, New York, pp. 189–244.
Ghazanfar, A.A. (2005) The evolution of speech reading. In-
ternational Conference on Cognitive Neuroscience 9. Cuba. Scherer, K.R. (1986) Vocal affect expression: a review and a
Grandjean, D., Ducommun, C. and Scherer, K.R. (in prepa- model for future research. Psychol. Bull., 99: 143–165.
ration) Neural signatures of processing emotional and lin- Scherer, K.R. (2001) Appraisal considered as a process of
multi-level sequential checking. In: Scherer, K.R., Schorr, A.
guistic-pragmatic prosody compared to phonetic-semantic
word identification: an ERP study. and Johnstone, T. (Eds.), Appraisal Processes in Emotion:
Grandjean, D., Ducommun, C., Bernard, P.-J. and Scherer, Theory, Methods, Research. Oxford University Press, New
York and Oxford, pp. 92–120.
K.R. (2002) Comparison of cerebral activation patterns in
identifying affective prosody, semantic prosody, and pho- Scherer, K.R. (2003) Vocal communication of emotion: a re-
neme differences. Poster presented at the International Or- view of research paradigms. Speech Commun., 40: 227–256.
247

Scherer, K.R., Johnstone, T. and Klasmeyer, G. (2003) Vocal an event-related potentials study. Neuroreport, 15:
expression of emotion. In: Davidson, R.J., Scherer, K.R. and 555–559.
Goldsmith, H. (Eds.), Handbook of the Affective Sciences. Wildgruber, D., Hertrich, I., Riecker, A., Erb, M., Anders,
Oxford University Press, New York and Oxford, pp. S., Grodd, W. and Ackermann, H. (2004) Distinct frontal
433–456. regions subserve evaluation of linguistic and emotio
Scherer, K.R., Ladd, D.R. and Silverman, K.E.A. (1984) Vocal nal aspects of speech intonation. Cereb. Cortex, 14:
cues to speaker affect: testing two models. J. Acoust. Soc. 1384–1389.
Am., 76: 1346–1356. Wildgruber, D., Riecker, A., Hertrich, I., Erb, M., Grodd, W.,
Schirmer, A. and Kotz, S.A. (2006) Beyond the right hemi- Ethofer, T. and Ackermann, H. (2005) Identification of emo-
sphere: brain mechanisms mediating vocal emotional tional intonation evaluated by fMRI. Neuroimage, 15:
processing. Trends in Cognitive Sciences, 10: 24–30. 1233–1241.
Vuilleumier, P. (2005) How brains beware: neural mechanisms Wundt, W. (1900) Völkerpsychologie. Eine Untersuchung der
of emotional attention. Trends Cogn. Sci., 9: 585–594. Entwicklungsgesetze von Sprache, Mythos und Sitte. Band I.
Vuilleumier, P., Armony, J.L., Driver, J. and Dolan, R.J. (2001) Die Sprache. [Cultural Psychology: A Study of the Develop-
Effects of attention and emotion on face processing in the mental Laws of Language, Myth, and Customs. Vol. 1. Lan-
human brain: an event-related fMRI study. Neuron, 30: guage]. Kröner, Leipzig.
829–841. Zatorre, R.J. and Belin, P. (2001) Spectral and temporal
Wambacq, I.J., Shea-Miller, K.J. and Abubakr, A. (2004) Non- processing in human auditory cortex. Cereb. Cortex, 11:
voluntary and voluntary processing of emotional prosody: 946–953.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 13

Cerebral processing of linguistic and emotional


prosody: fMRI studies

D. Wildgruber1,2,, H. Ackermann3, B. Kreifelts1 and T. Ethofer1,2

1
Department of Psychiatry, University of Tübingen, Osianderstr. 24, 72076 Tübingen, Germany
2
Section MR of CNS, Department of Neuroradiology, University of Tübingen, 72076 Tübingen, Germany
3
Department of General Neurology, Hertie Institute for Clinical Brain Research, University of Tübingen,
Hoppe-Seyler-Str. 3, 72076 Tübingen, Germany

Abstract: During acoustic communication in humans, information about a speaker’s emotional state is
predominantly conveyed by modulation of the tone of voice (emotional or affective prosody). Based on
lesion data, a right hemisphere superiority for cerebral processing of emotional prosody has been assumed.
However, the available clinical studies do not yet provide a coherent picture with respect to interhem-
ispheric lateralization effects of prosody recognition and intrahemispheric localization of the respective
brain regions. To further delineate the cerebral network engaged in the perception of emotional tone, a
series of experiments was carried out based upon functional magnetic resonance imaging (fMRI). The
findings obtained from these investigations allow for the separation of three successive processing stages
during recognition of emotional prosody: (1) extraction of suprasegmental acoustic information predom-
inantly subserved by right-sided primary and higher order acoustic regions; (2) representation of mean-
ingful suprasegmental acoustic sequences within posterior aspects of the right superior temporal sulcus; (3)
explicit evaluation of emotional prosody at the level of the bilateral inferior frontal cortex. Moreover,
implicit processing of affective intonation seems to be bound to subcortical regions mediating automatic
induction of specific emotional reactions such as activation of the amygdala in response to fearful stimuli.
As concerns lower level processing of the underlying suprasegmental acoustic cues, linguistic and emotional
prosody seem to share the same right hemisphere neural resources. Explicit judgment of linguistic aspects of
speech prosody, however, appears to be linked to left-sided language areas whereas bilateral orbitofrontal
cortex has been found involved in explicit evaluation of emotional prosody. These differences in hemi-
spheric lateralization effects might explain that specific impairments in nonverbal emotional communi-
cation subsequent to focal brain lesions are relatively rare clinical observations as compared to the more
frequent aphasic disorders.

Keywords: affect; communication; emotion; fMRI; intonation; language; lateralization; prosody

Introduction words we use. Rather, in numerous situations it


seems to be much more important how we utter
During social interactions among humans, transfer them (Mehrabian, 1972). Emotional states, atti-
of information does not depend only upon the tudes (e.g., sympathy, dominance, politeness), and
intentions often are predominantly expressed by
Corresponding author. Tel. +49-7071-298-6543; Fax: +49- the modulation of the tone of voice (emotional or
7071-29-4141; E-mail: dirk.wildgruber@med.uni-tuebingen.de affective prosody). For example, if your head of

DOI: 10.1016/S0079-6123(06)56013-3 249


250

department comes around and says with an angry several neuroimaging studies reported rightward
intonation ‘‘I have just been reading your report. lateralization of hemodynamic activation within
We have to talk about it right now,’’ you will cer- temporal regions (Buchanan et al., 2000; Wild-
tainly get a fairly different impression of his inten- gruber et al., 2002, 2005; Kotz et al., 2003; Mitc-
tions as if he would produce the same sentences in hell, Elliot, Barry, Cruttenden, & Woodruff, 2003;
a friendly and happy manner. As concerns the cer- Grandjean et al., 2005) and revealed additional —
ebral correlates of prosody processing, observa- partially bilateral — responses within the frontal
tions in patients suffering from focal brain lesions cortex (George et al., 1996; Imaizumi et al., 1997;
indicate that the well-established left hemisphere Buchanan et al., 2000; Wildgruber et al., 2002,
dominance for language comprehension does not 2004, 2005; Kotz et al., 2003), the anterior insula
extend to the perception of emotional tone (Hugh- (Imaizumi et al., 1997; Wildgruber et al., 2002,
ling-Jackson, 1879; Pell and Baum, 1997a,b; Sch- 2004), and the basal ganglia (Kotz et al., 2003)
mitt, Hartje, & Williams, 1997; Baum and Pell, during recognition of emotional intonation. The
1999; Borod et al., 2001, 2002; Adolphs, 2002; considerable differences in lateralization and local-
Charbonneau, Scherzer, Aspirot, & Cohen, 2003; ization of the relevant lesion sites as well as hemo-
Wildgruber and Ackermann, 2003; Ackermann, dynamic activation spots, however, do not yet
Hertrich, Grodd, & Wildgruber, 2004). According allow for an indisputable determination of the
to an early neuroanatomical model proposed by neural substrates of prosody processing. Presuma-
Ross (1981), prosodic information is encoded bly, the discrepancies of the available data are due
within distinct right-sided perisylvian regions that to differences in the methods used such as stimulus
are organized in complete analogy to the left-sided selection, task and control conditions. In order to
language areas. Expression of emotional prosody, further clarify to what extent specific neural struc-
thus, is believed to depend upon the Broca’s homo- tures subserve different facets of the comprehen-
logue within the right inferior frontal cortex, sion of emotional prosody, our research group
whereas comprehension of intonational informa- conducted a variety of experiments based on func-
tion is presumed to be bound to the right superior tional magnetic resonance imaging (fMRI), a tech-
temporal region (Wernicke’s homologue). How- nique that can be used for the noninvasive
ever, the empirical evidence for this model was evaluation of task-related hemodynamic cerebral
based on a few case reports only, and more sys- responses at a high spatial (ca. 0.5 mm; Menon and
tematic investigations yielded rather discrepant re- Goodyear, 1999) and moderate temporal (o1 s;
sults. The majority of lesion studies seem to be Wildgruber, Erb, Klose, & Grodd, 1997) resolu-
compatible with the assumption that the right tion. Specifically, these studies were designed to
hemisphere posterior perisylvian cortex is highly delineate the neural substrates underlying distinct
important for the comprehension of speech melody facets of prosody processing: (a) extraction of sup-
(Heilman et al., 1975, 1984; Darby, 1993; Stark- rasegmental acoustic information, (b) representa-
stein, Federoff, Price, Leiguarda, & Robinson, tion of meaningful prosodic sequences, (c) explicit
1994; Adolphs, Tranel, & Damasio, 2001; Borod et judgment of emotional as compared to linguistic
al., 2002). However, various clinical examinations information, (d) connectivity between the neural
indicate a widespread network of — partially bi- structures involved, and (e) implicit processing of
lateral — cerebral regions including the frontal emotional prosody.
cortex (Hornack et al., 1996, 2003; Breitenstein
et al., 1998; Rolls, 1999; Adolphs, Damasio, &
Tranel, 2002) and the basal ganglia (Cancellier and Extraction of suprasegmental acoustic information
Kertesz, 1990; Weddel, 1994; Peper and Irle, 1997;
Breitenstein et al., 1998; Breitenstein, Van At the perceptual level, emotional tone is charac-
Lancker, Daum, & Waters, 2001; Pell and terized by the modulation of loudness (acoustic
Leonard, 2003) to contribute to the processing of correlate: sound intensity), pitch (fundamental fre-
emotional intonation. In line with these findings, quency variation), speech rhythm (duration of
251

syllables and pauses), and voice quality or timbre corresponding regions within the left hemi-
(distribution of spectral energy) across utterances sphere.
(Lehiste, 1970; Ackermann et al., 1993; Murray
and Arnott, 1993; Banse and Scherer, 1996; Cutler, The first experiment encompassed a simple pas-
Dahan, & Donselaar, 1997; Bachorowski and sive listening condition. Trains of noise bursts
Owren, 2003; Scherer, Johnstone, & Klasmeyer, (clicks) were presented at different rates (2.0, 2.5,
2003; Sidtis and Van-Lancker-Sidtis, 2003). These 3.0, 4.0, 5.0, 6.0 Hz) to eight healthy right-handed
suprasegmental features are imposed upon the se- subjects (four males and four females, aged 19–32
quence of speech sounds (segmental structure) of years) during fMRI measurements. The clicks had
verbal utterances. According to the acoustic later- been produced originally by striking a pen against
alization hypothesis (Fig. 1a), the encoding of a table. Each acoustic sequence of a given click
suprasegmental parameters of the speech signal rate had a duration of 6 s. Altogether, 90 trains (6
(rather slow shifts 4100 ms) is predominantly rates  15 repetitions) were presented in pseudo-
bound to right hemisphere structures whereas randomized order. During passive listening to
rapid transitions (o50 ms), contributing to the these stimuli, significant hemodynamic responses
differentiation of the various speech sounds at the across all different presentation rates emerged
segmental level (i.e., phonemes, syllables), are within the superior temporal gyrus of both sides,
mainly processed within contralateral areas (Van right hemisphere putamen, and the tectum. More-
Lancker and Sidtis, 1992; Belin et al., 1998; Ivry over, parametric analysis revealed lateralized rate-
and Robertson, 1998; Zatorre and Belin, 2001; dependent responses within the anterior insular
Zatorre, 2001; Zatorre et al., 2002; Meyer, Alter, cortex. During presentation of the click trains at
Friederici, Lohmann, & von Cramon, 2002; Po- slow rates, the right anterior insula showed the
eppel et al., 2004). These acoustic laterality effects highest activation levels. Furthermore, the hemo-
have been supposed to explain the differential dynamic responses of this region displayed a de-
hemispheric dominance patterns of language (left cline of amplitude in parallel with an increase of
hemisphere) and music processing (right hemi- stimulation frequency. By contrast, an opposite
sphere) (Wildgruber, et al., 1996, 1998, 2001, 2003; relationship emerged within the left anterior insu-
Belin et al., 1998; Ivry and Robertson, 1998; lar cortex (Ackermann et al., 2001). This double
Zatorre et al., 2002; Hugdahl and Davidson, 2003; dissociation of rate-response functions between the
Poeppel, 2004; Ackermann et al., 2006). In order two hemispheres is in a very good accordance with
to further separate the neural structures subserving the acoustic lateralization hypothesis (Fig. 1b).
the extraction of basic acoustic properties of Seventeen healthy volunteers (8 males, 9 fe-
speech prosody from those which respond to the males, aged 18–31 years) participated in a second
conveyed emotional ‘‘meaning’’, a series of fMRI experiment that investigated discrimination of du-
experiments was conducted. More specifically, the ration and pitch values at different levels of diffi-
following hypotheses were explored: culty. Complex sounds characterized by four
formant frequencies (500, 1500, 2500, 3500 Hz),
(a) Lateralization of hemodynamic responses manipulated either in duration (100–400 ms) or in
during passive listening to trains of noise fundamental frequency (100–200 Hz, realized by
bursts depends upon stimulus frequency. rhythmic intensity fluctuations throughout the sig-
(b) Extraction of specific acoustic parameters nal), served as stimuli. Sequences of two signals
(signal duration, fundamental frequency) is were presented to both ears each, and subjects ei-
associated with different activation patterns ther had to detect the longer duration (duration
at the level of primary and higher order task) or the higher pitch (pitch task), respectively.
acoustic regions. The behavioral data showed comparable hit scores
(c) Expressiveness of emotional prosody en- (mean values about 75%) with increasing accuracy
hances the hemodynamic responses of voice- rates in correlation to rising physical difference
sensitive areas within the right as compared to between the two acoustic signals for both, pitch
252

and duration discrimination (Fig. 1c). As com- cortices as well as the working memory network.
pared to baseline at rest, both tasks yielded bilat- A lateralization analysis, i.e., comparison of each
eral activation of frontal, temporal and parietal hemisphere with the contralateral side on a voxel-
regions including primary and secondary acoustic by-voxel basis, revealed, however, lateralization
253

effects toward the left side within insular and models as nuisance variables. During both tasks, a
temporal cortex during both tasks. Even more linear correlation between hemodynamic responses
noteworthy, a parametric analysis of hemodynam- and prosodic emotional expressiveness emerged
ic responses showed an increase of activation within the middle part of bilateral superior tempo-
within the right temporal cortex in parallel with ral sulcus (mid-STS). Responses of right hemisphere
the differences in sound properties of the stimulus mid-STS showed higher amplitudes, larger exten-
pairs (Fig. 1c). This positive linear relationship sion, and a stronger dependency on emotional in-
emerged both during the duration and the pitch tensity than those of the contralateral side (Fig. 1d).
task. Moreover, a comparison with the contralat- Similar response patterns were found both for ex-
eral hemisphere revealed significant lateralization plicit and implicit processing of emotional prosody
effects of the parametric responses toward the (Ethofer et al., 2006c). These observations support
right superior temporal sulcus during discrimina- the assumption that the mid-STS region contributes
tion of stimulus duration (Reiterer et al., 2005). to the encoding of emotionally salient acoustic
Slowly changing and highly different acoustic stimuli independent from task-related attentional
stimuli, thus, seem to be predominantly processed modulation (Grandjean et al., 2005).
within the right hemisphere whereas detection of In summary, these findings, related to the acous-
rapid changes or rather slight signal differences tic level of prosody processing, indicate extraction
might be linked to the left hemisphere. of suprasegmental acoustic information to be pre-
The findings of these first two experiments dominantly subserved within right-sided primary
indicate that differences in basic acoustic properties and higher order acoustic brain regions including
have a strong impact on brain activation patterns. mid-STS and anterior insula.
In a third study, 12 healthy right-handed subjects
(7 males, 5 females, aged 19–29 years) were asked to
judge in two separate sessions the emotional valence Representation of meaningful prosodic sequences
of either word content or prosody of altogether 162
German adjectives spoken in a happy, angry, or According to the neuroanatomical model pro-
neutral tone. Intonations of these different emo- posed by Elliot Ross, the Wernicke’s homologue
tional categories differ in various acoustic proper- region bound to the posterior aspects of right
ties (Banse and Scherer, 1996). To disambiguate hemisphere superior temporal gyrus represents the
more specific effects of emotional expressiveness key area for the comprehension of prosodic se-
from extraction of low-level acoustic parameters, quences (Ross, 1981). An important role of the
mean and variation of sound intensity and funda- right posterior perisylvian cortex for comprehen-
mental frequency were included in the statistical sion of speech melody has been confirmed in

Fig. 1. (a) According to the acoustic lateralization hypothesis, rapid changes of acoustic parameters (o50 ms) are predominantly
processed within the left whereas slow variations (4100 ms) are mainly encoded within the right hemisphere. (b) Parametric responses
during passive listening to trains of noise bursts: hemodynamic responses characterized by positive linear (red), negative linear (green),
or nonlinear (blue) rate-response functions. Activation clusters are displayed on transverse sections of the averaged anatomical
reference images (R ¼ right, L ¼ left). The relationship between signal intensity (in arbitrary units) and rate of acoustic stimulation
was determined within the right (green) and left (blue) insular cortex (see Ackerman et al., 2001). (c) Discrimination of sound duration:
pairs of complex acoustic signals that varied in duration (100–400 ms) were presented to healthy subjects. Accuracy rates demonstrate
increasing deviance in time to be correlated with higher performance scores. Parametric effects: significantly activated areas as a
function of linear increase with task performance emerged within the right MTG/STG during duration discrimination. Laterality
analysis: voxelwise comparison of the hemispheres revealed a significantly activated cluster within the left STG for the parametric effect
of duration discrimination (see Reiterer et al., 2005). (d) Parametric effects of prosodic emotional intensity. Conjunction of regions
showing a linear relationship between hemodynamic responses and prosodic emotional intensity during both implicit and explicit
processing of emotional prosody. Beta estimates (mean7standard error) corresponding to distinct intensity steps of emotional in-
tonations have been plotted for the most significant voxel of the cluster in the right and left STS during implicit (red) and explicit
(green) processing of emotional prosody (see Ethofer et al., 2006c).
254

various clinical examinations (Heilman, Scholes, examinations failed to show any interactions be-
& Watson, 1975, 1984; Darby, 1993; Starkstein et tween hemispheric lateralization and emotional
al., 1994; Borod et al., 2002). In some studies on valence (Pell, 1998; Baum and Pell, 1999; Borod et
the comprehension of emotional information, al., 2002; Kucharska-Pietura et al., 2003). Consid-
however, the valence of emotional expression has ering functional imaging data, however, distinct
been reported to influence lateralization of cere- cerebral activation patterns bound to specific emo-
bral responses (Canli et al., 1998; Davidson, Ab- tional categories such as disgust, anger, fear, or
ercrombie, Nitschke, & Putnam, 1999; Murphy, sadness have been observed during perception of
Nimmo-Smith, & Lawrence, 2003). According to facial emotional expressions (Sprengelmeyer, Ra-
the valence hypothesis, rightward lateralization of usch, Eysel, & Przuntek, 1998; Kesler-West et al.,
prosody processing only holds true for negative 2001; Phan, Wager, Tayler, & Liberzon, 2002;
emotions, whereas comprehension of happy stim- Murphy et al., 2003). Several studies have corrob-
uli is ascribed to the left hemisphere (Fig. 2). As orated the notion that responses of the amygdalae
concerns speech intonation, several clinical are specifically related to facial expressions of fear

Fig. 2. (a) According to the valence hypothesis, positive emotional information (i.e., happy expressions) is processed within the left
negative emotional information (expressions of fear, anger, disgust or sadness) within the right hemisphere. (b) Significant hemo-
dynamic responses during identification of emotional intonation as compared to vowel identification are superimposed upon the
cortical surface of a template brain and upon an axial slice at the level of the highest activated voxels within the activation clusters. The
emotional task yielded specific activation within the right STS (BA 22/42) and the right inferior frontal cortex (BA 45/47). Analysis of
valence effects, however, revealed no differences of cerebral responses depending upon valence or specific emotional categories (see
Wildgruber et al., 2005).
255

(Morris et al., 1996, 1998; Adolphs, 2002; Phan and they were provided with a list of possible re-
et al., 2002) whereas facial expressions of disgust sponse alternatives prior to testing. Since both
seem to elicit activation of the anterior insula tasks require evaluation of completely identical
(Phillips et al., 1998; Sprengelmeyer et al., 1998; acoustic stimuli and involve very similar response
Calder et al., 2000; Phan et al. 2002; Wicker et al., mechanisms, comparison of the respective hemo-
2003). Fear-specific responses of the amygdalae dynamic activation patterns should allow for the
have also been reported in association with vocal separation of task-specific cerebral responses inde-
emotional expressions (Phillips et al., 1998; Morris, pendently of stimulus characteristics and unspecific
Scott, & Dolan, 1999) whereas the predicted dis- task components. In order to delineate cerebral
gust-related activation of the anterior insula has structures contributing to the recognition of emo-
not been observed in a prior PET experiment (Phil- tional prosody independent of specific emotional
lips et al., 1998). It is unsettled, thus, to which categories, responses during the identification of
extent lateralization and exact localization of cer- emotional prosody across all emotional categories
ebral activation during comprehension of emo- were compared to the phonetic control condition.
tional prosody is linked to specific emotional To disentangle patterns of cerebral activation re-
categories. lated to comprehension of specific emotional cat-
Based on the aforementioned clinical and egories, each emotional category was compared
neuroimaging studies, presumably, there are cere- against the others. The main goal of the study,
bral regions, including the right posterior temporal thus, was to evaluate the following two hypotheses:
cortex, that contribute to comprehension of
emotional prosody independent of any specific (a) A network of right-hemisphere areas includ-
emotional content. Other regions, including ing the posterior temporal cortex supports
the amygdala and anterior insula, are selectively identification of affective intonation inde-
linked to comprehension of specific emotional cat- pendent of specific emotional information
egories. In order to separate these components, 100 conveyed.
short German declarative sentences with emotion- (b) Perception of different emotional categories
ally neutral content (such as ‘‘Der Gast hat sich für is associated with specific brain regions, i.e.,
Donnerstag ein Zimmer reserviert’’ [The visitor re- response localization varies with emotion
served a room for Thursday], ‘‘Die Anrufe werden type. Specifically, fear- specific responses are
automatisch beantwortet’’ [Phonecalls are answered linked to the amygdalae and disgust-specific
automatically]) were randomly ascribed to one of responses to the anterior insula.
five different target emotions (happiness, anger,
fear, sadness, or disgust). A professional actress During the fMRI experiment, subjects correctly
and an actor produced these test materials express- identified the emotional tone at a slightly lower
ing the respective emotion by modulation of affec- rate (mean: 75.277.9%) as compared to the vowel
tive intonation. Verbal utterances were presented detection task (mean: 83.477.0%, p o0.05). The
to 10 healthy subjects (5 males, 5 females, age: accuracy scores for happy (90%), angry (82%),
21–33 years) under two different task conditions and sad (84%) expressions reached comparable
during fMRI. As an identification task, subjects levels whereas fearful (51%) and disgusted (57%)
were asked to name the emotion expressed by the expressions were identified at significantly lower
tone of voice whereas the control condition (pho- rates (po0.05). These differences in performance
netic task) required the detection of the vowel fol- are in good accordance with prior observations
lowing the first /a/ in each sentence. Similarly to the and might be related to differences in recogniza-
emotion recognition task, vowel identification bility of the acoustic cues of the various emotions
also included a forced choice selection from five (Banse and Scherer, 1996). Response times for
alternatives, i.e., the vowels /a/, /e/, /i/, /o/, /u/. the emotional task (mean: 4.370.9 s) showed no
Under both conditions, participants were asked to significant differences as compared to the phonetic
give a verbal response as quickly as possible task (mean: 4.171.0 s) indicating comparable
256

levels of task difficulty. Cerebral responses ob- within these areas should be considered a second
tained during both tasks, as compared to the rest step of prosody processing. A further experiment
condition, yielded a bilateral network of hemody- was designed in order to evaluate the contribution
namic activation at the level of cortical and sub- of posterior STS and inferior frontal cortex to the
cortical regions including frontal, temporal and processing of emotional prosody as compared to
parietal cortex, thalamus, and cerebellum. To evaluation of linguistic prosody.
identify brain regions specifically contributing to
the encoding of emotional intonation, the respec-
tive activation patterns were directly compared to Explicit judgment of emotional prosody
the responses obtained during phonetic processing
of the identical acoustic stimuli (Wildgruber et al., As concerns its communicative functions, speech
2005). Using this approach, responses within two prosody serves a variety of different linguistic as
activation clusters, localized within the right pos- well as emotional purposes (Ackermann et al.,
terior superior temporal sulcus (BA 22/42) and the 1993, 2004; Baum and Pell, 1999). Among others,
right inferior frontal cortex (BA 45/47), could it is used to specify linguistic information at the
be assigned to recognition of emotional prosody word (content vs. content) and sentence level
(Fig. 2b). No significant impact of emotional va- (question vs. statement intonation: ‘‘It is new?’’
lence or specific emotional categories on the dis- vs. ‘‘It is new!’’; location of sentence focus: ‘‘he
tribution of brain activation could be observed. wrote this letter ‘‘vs. ‘‘he wrote this letter’’), and
Therefore, the results of the current study do not conveys information about a speaker’s personal-
support, in line with prior functional imaging ity, attitude (i.e., dominance, submissiveness, po-
(Buchanan et al., 2000; Wildgruber et al., 2002; liteness, etc.), and emotional state (Fig. 3). Based
Kotz et al., 2003; Mitchell et al., 2003) and recent on lesion studies, the functional lateralization hy-
lesion studies (Pell, 1998; Baum and Pell, 1999; pothesis proposes linguistic prosody to be proc-
Borod et al., 2002; Kucharska-Pietura et al., 2003), essed within the left hemisphere, whereas
the hypothesis of valence-specific lateralization emotional tone is bound to contralateral cerebral
effects during processing of emotional intonation. structures (Van Lancker, 1980; Heilman et al.,
The observed hemodynamic responses, however, 1984; Behrens, 1985; Emmorey, 1987; Pell and
indicate a task-dependent and stimulus-independ- Baum, 1997a; Borod et al., 1998, 2002; Geigen-
ent contribution of the right posterior STS (BA 22/ berger and Ziegler, 2001; Schirmer, Alter, Kotz, &
42) and the right inferior frontal cortex (BA 45/47) Friederici, 2001; Charbonneau et al., 2003). In or-
to the processing of suprasegmental acoustic infor- der to disentangle the functional and the acoustic
mation irrespective of specific emotional categories. level of prosody processing, sentences varying in
We assume, therefore, that the representation of linguistic accentuation (sentence focus) as well as
meaningful suprasegmental acoustic sequences emotional expressiveness were generated by

Fig. 3. (a) According to the functional lateralization hypothesis linguistic prosody is processed within the left emotional prosody is
bound to the right hemisphere. (b) Variation of linguistic (left) and emotional intonation (right). The German sentence ‘‘Der Schal ist
in der Truhe’’ (the scarf is in the chest) was digitally resynthesized with various pitch contours. Five different patterns of sentence focus
were realized by a stepwise increase of the fundamental frequency on the final word (left). The stress accentuation ranged between an
utterance clearly focused on the second word (solid line) and one that is focused on the final word (dotted line). For each of these
synthetic sentences, five variations of emotional expressiveness were generated by manipulation of the pitch range across the whole
utterance (right). Sentences with broader pitch ranges are perceived as being more excited. As shown for the middle contour (red), the
realization of linguistic accents remains constant during manipulation of emotional expressiveness. The sentences of each stimulus pair
differed in relative focus accentuation as well as in emotional intensity. (c) Significantly activated regions, identified by task com-
parisons, superimposed upon the cortical surface of a template brain and upon an axial slice at the level of the highest activated voxels
within each activation cluster: The emotional task (upper row) yielded significant responses within the bilateral orbitobasal frontal
cortex (BA 11/47), whereas activation of the left inferior frontal gyrus (BA 44/45) emerged during discrimination of linguistic prosody
(lower row) (see Wildgruber et al., 2004).
257

systematic manipulations of the fundamental fre- this utterance represents an answer to the question
quency contour of the simple declarative German ‘‘What is in the chest?’’. Shifting the accent to the
sentence ‘‘Der Schal ist in der Truhe’’ (The scarf is final word, the sentence provides information
in the chest). With its focus on the second word, about where the scarf is. This prosodic distinction
258

is realized by distinct pitch patterns characterized patterns of linguistic accentuation and emotional
by F0-peaks on the accented syllables (Cutler expressiveness at similar levels of accuracy (lin-
et al., 1997). As a first step, a series of five F0 guistic discrimination: 82%714%, emotional dis-
contours was generated extending from a clear-cut crimination 78%711%). Therefore, a comparable
focus on the second to an accent on the final word level of difficulty for both tasks can be assumed.
(Fig. 3b). On the basis of each of these five focus As compared to the baseline at rest, both condi-
patterns, second, five additional variations were tions yielded bilateral hemodynamic responses
generated differing in pitch range across the whole within supplementary motor area, anterior cingu-
sentence. These global variations are perceived as late gyrus, superior temporal gyrus, frontal op-
modulations of emotional expressiveness. Sen- erculum, anterior insula, thalamus, and
tences with broader F0 range clearly sound more cerebellum. Responses within the dorsolateral
excited (Banse and Scherer, 1996; Pihan et al., frontal cortex (BA 9/45/46) showed lateralization
1997). Ten healthy right-handed participants (6 effects toward the right side during both tasks
males, 4 females, age: 20–35 years) were asked to (Wildgruber et al., 2004). In order to identify brain
perform two different discrimination tasks during regions specifically contributing to the processing
pairwise presentation of these acoustic stimuli. In of linguistic or emotional intonation, the respec-
two different sessions of the experiment they had tive activation patterns were directly compared
to answer one of the following questions: (a) with each other.
‘‘Which of the two sentences is better suited as a During the linguistic task, significantly stronger
response to the question: Where is the scarf?’’ activation was observed within the left inferior
(discrimination of linguistic prosody) and (b) frontal gyrus (BA 44/45 ¼ Broca’s area). By con-
‘‘Which of the two sentences sounds more ex- trast, the affective condition yielded significant bi-
cited?’’ (discrimination of emotional expressive- lateral hemodynamic responses within orbitofrontal
ness). Since both conditions require the evaluation cortex (BA 11/47) as compared to the linguistic task
of completely identical acoustic signals, the com- (Fig. 3c). Comprehension of linguistic prosody re-
parison of hemodynamic responses obtained dur- quires analysis of the lexical, semantic, and syntac-
ing the two different runs allows for the separation tic aspects of pitch modulation patterns. Activation
of task-specific responses independent of stimulus of left inferior frontal cortex (Broca’s area) con-
characteristics. This experiment was primarily de- comitant with the discrimination of linguistic ac-
signed to explore the following two alternative hy- cents indicates that at least some of these operations
potheses: might be housed within the anterior perisylvian
language areas. In line with this assumption, native
(a) Lateralization effects during prosody
speakers of Thai, a tone language, showed activa-
processing are strongly bound to acoustic
tion of the left inferior frontal region during dis-
properties of the relevant speech signal:
crimination of linguistically relevant pitch patterns
Since comprehension of linguistic as well as
in Thai words. This activity was absent in English-
emotional prosody relies upon the extrac-
speaking subjects listening to identical stimuli
tion of suprasegmental features, a rightward
(Gandour, Wong, & Hutchins, 1998). Moreover,
lateralization must be expected during both
clinical observations support the assumption of a
conditions (acoustic lateralization hypothe-
specific contribution of the left hemisphere to
sis).
the comprehension of linguistic aspects of intona-
(b) Linguistic prosody is processed within left-
tion. For example, Heilman et al. (1984) found pa-
sided speech areas, whereas comprehension
tients suffering from focal left-sided brain lesions to
of emotional prosody must be expected to be
produce significantly more errors in a linguistic
bound to the right hemisphere (functional
prosody identification task as compared to the rec-
lateralization hypothesis).
ognition of affective intonation, whereas damage to
The obtained behavioral data clearly show that the right hemisphere was associated with a similar
the participants were able to discriminate the profile of deficits in both tasks. Furthermore,
259

Emmorey (1987) observed impaired discrimination behavior have been observed (Levin, Eisenberg,
of stress contrasts between noun compounds and & Benton, 1991; Blair and Cipolatti, 2000;
noun phrases after damage to the left hemisphere Wildgruber et al., 2000), resulting, conceivably,
whereas patients with right-sided lesions performed from compromised associations between actual
as well as normal control subjects. Predominant environmental stimuli with emotional memory
disturbance of linguistic prosody comprehension traces.
concomitant with relatively preserved processing of In conclusion, hemispheric specialization for
emotional intonation in patients with damage to the higher level processing of intonation contours has
left hemisphere has also been reported by Pell and been found to depend, at least partially, upon the
Baum (1997a) as well as Geigenberger and Ziegler functional role of the respective acoustic signals
(2001). within the communication process: Comprehen-
Discrimination of emotional expressiveness sion of linguistic aspects of speech melody relies
yielded a significant increase of hemodynamic re- predominantly upon left-sided perisylvian lan-
sponses within bilateral orbitofrontal cortex (BA guage areas, whereas the evaluation of emotional
11/47) as compared to the linguistic task indicat- signals, independent of modality and emotion
ing, thus, a specific contribution of this region to type, is bound to bilateral orbitofrontal regions.
the evaluation of emotional aspects of verbal ut- As a third step of prosody processing, thus,
terances conveyed by the tone of speech. On the explicit evaluation of emotional prosody seems to
basis of neuroanatomical considerations, e.g., re- be associated with bilateral inferior aspects of
ciprocal fiber connections to sensory cortices and frontal cortex including the orbitobasal surface
limbic regions, this region might serve as a subst- (BA 47/11).
rate for the judgment of emotional stimuli inde-
pendent of the stimulus modality (Price, 1999).
Accordingly, activation of the orbitobasal frontal Connectivity within the prosody network
cortex has been observed in preceding functional
imaging studies during perception of emotional So far, three successive steps of prosody processing
intonation (George et al., 1996; Wildgruber et al., have been identified: (1) extraction of supraseg-
2002), emotional facial expressions (Blair, Morris, mental acoustic information, (2) representation of
Frith, Perret, & Dolan, 1999; Nakamura et al., suprasegmental sequences, and (3) explicit judg-
1999), and affective gustatory judgments ( Small ment of emotional information. As concerns the
et al., 2001). Moreover, patients suffering from respective neuroanatomical correlates, extraction
unilateral focal damage to this area displayed im- of suprasegmental acoustic information seems to
paired identification of emotional face and voice be predominantly bound to the right primary and
expressions whereas performance in nonemotional secondary auditory regions. Presumably, the rele-
control tasks (i.e., discrimination of unfamiliar vant acoustic information is transferred from these
voices and recognition of environmental sounds) regions via direct fiber connections to an area
was found uncompromised (Hornak, Rolls, & within the posterior superior temporal sulcus
Wade, 1996; Hornak et al., 2003; Rolls, 1999). (post-STS) subserving the representation of mean-
These observations, in line with the results of the ingful intonational sequences. In case of explicit
present study, support the assumption that orbito- judgment of emotional prosody, a further tempo-
frontal areas contribute to the explicit evaluation rofrontal passage of information must be assumed
of emotional information conveyed by different accounting for the observed activation of bilateral
communicational channels. Blair and Cipolatti inferior frontal cortex during this task. It should
supposed this region to be critically involved in be emphasized, furthermore, that converging re-
building associations between the perceived sults from lesion studies (Hornak et al., 1996,
emotional signals and an emotional episodic mem- 2003; Ross, Thompson, & Yenkosky, 1997) and
ory. In patients suffering from lesions of orbito- functional imaging examinations (Imaizumi et al.,
frontal cortex, pronounced abnormalities of social 1997; Pihan, Altenmüller, Hertrich, & Ackermann,
260

2000; Wildgruber et al., 2002, 2004) suggest a con- of different brain regions, and (3) modulation of
tribution of these areas to the processing of emo- this coupling by experimental factors (for meth-
tional prosody, and an intact transcallosal odological details see Ethofer et al., 2006b). Using
communication of information has been assumed this technique, the following hypotheses were eval-
to be a prerequisite for comprehension of emo- uated:
tional prosody (Ross et al., 1997). It is unclear,
however, whether this cooperation of the two (a) Within the network of regions character-
hemispheres is based on a sequence of processing ized by task-dependent activation, the post-
steps or if both frontal lobes receive the respective STS serves as input region (receiving input
information independently via parallel connec- from primary and secondary acoustic re-
tions from the right posterior temporal cortex. In gions).
order to investigate the connectivity architecture (b) The frontal lobes, consecutively, receive
of the cerebral network involved in the processing their input from the post-STS. Moreover,
of emotional prosody, a further experiment was it was assessed whether both frontal lobes
carried out. Twenty-four healthy right-handed subserve two successive processing steps or
subjects (11 males, 13 females, mean age 24.4 receive their information independently
years) underwent event-related fMRI measure- from the right post-STS via parallel path-
ments while rating the emotional valence of either ways.
prosody or semantics of 162 binaurally presented
emotional adjectives (54  neutral, 54  positive, Conventional analysis of the fMRI data yielded,
54  negative content) spoken in happy, neutral, in very good accordance with prior investigations
or angry intonation by six professional actors (3 (Wildgruber et al., 2004, 2005), activation within
females/3 males). The adjectives were selected the right posterior STS and bilateral inferior fron-
from a sample of 500 adjectives on the basis of tal cortices during evaluation of emotional pros-
ratings obtained from 45 healthy German native ody. Subsequent determination of functional
speakers (see Kissler et al., this volume) along the connectivity revealed that the activation cluster
dimensions of valence and arousal on a nine-point within the right post-STS represents the most
self-assessment manikin scale (SAM, Bradley and likely input region into this task-specific network.
Lang, 1994). The stimuli comprised 54 highly This finding is in agreement with the assumption
arousing positive (mean arousal rating 44, mean that this region subserves representation of supra-
valence rating o4, e.g., ‘‘verführerisch’’ ¼ allur- segmental sequences and receives direct input from
ing), 54 highly arousing negative (mean arousal primary and secondary acoustic regions. To inves-
44, mean valence rating 46, e.g., ‘‘pan- tigate the intrinsic connectivity pattern within the
isch’’ ¼ panic), and 54 low-arousing neutral (mean network, dynamic causal models assuming paral-
arousal rating o4, mean valence rating between 4 lel, serial, or fully bidirectional connectivity pat-
and 6, e.g., ‘‘breit’’ ¼ broad). During separate terns were compared. The model based upon
functional imaging sessions, subjects had been parallel projections from the posterior STS to the
asked to judge either the valence of emotional frontal cortical regions turned out to be signifi-
word content or the valence of emotional prosody cantly superior to both serial models as well as the
on the nine-point SAM scale. Both the order of model with bilaterally connected brain regions
within-session stimulus presentation and the se- (Fig. 4a). In a post hoc analysis, an attempt was
quence of sessions were pseudorandomized across made to optimize this parallel pathway model
subjects. To assess functional connectivity of ac- by adding either unidirectional or bidirectional
tivated regions, the novel technique of dynamic connections between the two frontal regions or
causal modeling (Friston et al., 2003) was applied adding unilateral or bidirectional backward pro-
to the data. This approach allows inferences on (1) jections from the frontal areas to the right poste-
the parameters representing influence of experi- rior STS. The original parallel pathway model
mentally designed inputs, (2) the intrinsic coupling again was found to be significantly superior to all
261

Fig. 4. (a) To evaluate the intrinsic connectivity of regions contributing to the processing of emotional prosody, four different models
were compared. (Model 1) Parallel transmission from the right post-STS to both frontal regions. (Model 2) Successive conductance
from post-STS to right IFG and further on to the left IFG. (Model 3) Serial conductance from post-STS to left IFG and right IFG.
(Model A) Fully connected bidirectional flow of information. Based upon a prior analysis, in all these models external inputs were
specified to enter the network via the right post-STS. Dynamic causal modeling revealed a statistical superiority of the parallel
processing model (Model 1) as compared to all other models (Ethofer et al., 2006b). (b) Based on the these findings it is assumed that
explicit judgment of emotional prosody is carried out in at least three successive steps: (1) extraction of suprasegmental information
bound to predominantly right-sided primary and secondary acoustic regions, (2) representation of meaningful suprasegmental se-
quences within the right post-STS, and (3) explicit emotional judgment of acoustic information within the bilateral inferior frontal
cortices.
262

alternative models. These results provide further consequence, implicit transmission of emotional in-
empirical support for the hypothesis that process- formation by the induction of physiological emo-
ing of emotional prosody is carried out in three tional reactions, e.g., changes of heart rate and skin
successive steps: (1) extraction of suprasegmental conductance, might be linked to emotion-specific
acoustic information bound to predominantly subcortical regions, whereas the explicit evaluation
right-sided primary and higher order acoustic of emotional signals based on the retrieval of in-
regions, (2) representation of meaningful supra- formation from emotional memory appears to be
segmental sequences within the right post-STS, and processed within bilateral inferior frontal areas,
(3) explicit emotional judgment of acoustic infor- irrespective of emotion type and valence of the
mation within the bilateral inferior frontal cortices stimuli. In order to evaluate the neural basis of im-
(Fig. 4b). plicit processing of emotional prosody, a cross-
modal interaction experiment was conducted (for
methodological issues of cross-modal interaction
Implicit processing of emotional prosody experiments see Ethofer et al., this volume). This
experiment was designed to test the following two
During everyday interactions among humans, predictions:c
as a rule, the emotional connotations of commu-
nicative signals are not explicitly evaluated on a (a) Simultaneous presentation of emotional faces
quantitative scale. Rather, highly automatized un- and emotional prosody induces distinct in-
derstanding of the emotional information con- teraction effects: explicit judgment of facial
veyed by facial expressions, speech prosody, expressions is influenced by implicit process-
gestures, or the propositional content of verbal ing of unattended emotional prosody.
utterances seems to be much more important. (b) The impact of an unattended fearful tone of
A variety of empirical data indicate different cer- speech on explicit judgment of emotional
ebral pathways to be involved in explicit and im- faces is associated with activation of the
plicit processing of emotional signals (LeDoux, amygdala.
1996; Anderson and Phelps, 1998; Adolphs and
Tranel, 1999; Critchley, 2000; Adolphs et al., During this experiment, images of facial expres-
2002). As concerns the hemodynamic responses sions taken from the Ekman and Friesen battery
bound to specific emotional categories, a selective (Ekman and Friesen, 1976) were presented to 12
contribution of the amygdala to recognition of healthy right-handed subjects (7 males, 5 females,
fearful voices has been assumed on the basis of age: 19–29 years). Using digital morphing tech-
lesion data (Scott, Young, Calder, & Hellawell, niques, a series of visual stimuli was generated ex-
1997) and prior PET studies (Phillips et al., 1998; tending in facial expression from 100% fear to
Morris et al., 1999). Furthermore, a specific con- 100% happiness in incremental steps of 25% (Per-
tribution of the anterior insula and the basal gan- ret et al., 1994). In one run of the experiment, the
glia to the perception of vocal expressions of facial expressions were shown in isolation, and in
disgust has been predicted based on clinical find- another trial they were combined with acoustic
ings (Pell and Leonhard, 2003) and functional im- stimuli, i.e., short declarative sentences spoken in a
aging experiments during processing of facial fearful or happy tone by two professional actors
expressions (Sprengelmeyer et al., 1998; Phan et (one male, one female). In both of these runs, par-
al., 2002; Wicker et al., 2003). Responses of the ticipants were instructed to rate the emotional va-
amygdalae have been observed to depend on im- lence of the displayed facial expressions. A third
plicit processing of emotional signals, e.g., during run of the experiment required explicit judgment
passive listening tasks, whereas explicit judgments of emotional prosody. The behavioral results show
of emotional expressions were shown to result in that subjects rated fearful and neutral facial ex-
deactivation of this region (Morris et al., 1999; pressions as being more fearful when presented
Critchley et al., 2000; Adolphs 2002). As a concomitant with a fearfully spoken sentence as
263

Fig. 5. (a) Implicit impact of fearful prosody on judgment of emotional faces: (left) valence rating of facial expressions (mean7stand-
ard error) presented without acoustic stimuli (white bars) and in combination with fearful prosody (gray): Evaluation of facial
expressions in the presence of a fearful voice as compared to a happy intonation yielded significant activation in the right fusiform
gyrus (upper right). Analysis of cross-modal impact of fearful voices revealed significant correlations between individual behavioral
changes and hemodynamic responses in the left amygdala (see Ethofer et al., 2006a). (b) Cross-modal integration of emotional
communicative signals: (1) Extraction of different communicative signals (prosody, facial expressions, word content) is subserved by
the respective modality-specific primary cortices. (2) More complex features of these signals are processed within modality specific
secondary regions. (3) As a third step, explicit emotional judgments based on evaluation of associations with episodic emotional
memory seem to be linked to the bilateral inferior frontal cortex. This region is assumed to be involved in cross-modal integration
during explicit evaluation. On the other hand, emotional signals can yield an automatic (implicit) induction of emotional physiological
reaction (e.g., variation of heart rate and skin conductance) that is linked to specific subcortical regions. Presumably, both neural
pathways are interconnected at various levels.
264

compared to the no-voice condition. By contrast, Cross-modal integration of emotional


no significant shifts in interpretation occurred dur- communicative signals
ing presentation of happy expressions (Fig. 5a).
Thus, this experimental paradigm might provide a Emotional information may be conveyed via differ-
means for quantitative measurements of the im- ent communicative channels, e.g., prosodic features
plicit impact of emotional prosody on the judg- of the acoustic speech signal, facial expressions,
ment of facial expressions (de Gelder and and propositional content of verbal utterances.
Vroomen, 2000). A comparison of happy and Based on the findings presented here, several suc-
fearful intonations during explicit judgment of cessive steps during cross-modal integration of
prosody (unimodal auditory session) did not re- emotional signals can be separated and assigned to
veal any significant differences of the hemody- distinct cerebral correlates: (1) extraction of com-
namic cerebral responses. As concerns implicit municative signals is subserved by the respective
processing of emotional prosody, however, the modality-specific primary cortices, (2) modality-
middle section of the right fusiform gyrus showed specific higher order regions process emotional in-
a significantly stronger activation when facial ex- formation (e.g., prosody ¼ right STS, facial ex-
pressions were displayed in the presence of a fear- pressions ¼ fusiform face area, propositional
ful voice as compared to happy intonation. This meaning ¼ left posterior STG), (3) explicit emo-
region has been named the fusiform face area, be- tional judgments, presumably involving evalu-
cause it has been found crucial for the processing ation of associations with episodic emotional
of faces in clinical and experimental studies (Puce, memory, were found to be linked to bilateral or-
Allison, Gore, & McCarthy, 1995; Kanwisher bitofrontal cortex. Implicit processing of emotional
et al., 1997; Barton et al., 2002). Moreover, this signals, however, seems to rely on alternative path-
region shows stronger activation to emotional as ways including emotion-specific subcortical regions
compared to neutral faces (Morris et al., 1998) and involved in automatic physiological reaction (e.g.,
seems to respond particularly to stimuli signaling variation of heart rate and skin conductance). It
danger (Surguladze et al., 2003). The increased has been demonstrated, that both pathways of
hemodynamic responses within the fusiform gyrus emotion processing influence the behavior of the
in presence of an auditory expression of threat organism and that unattended processing of emo-
might reflect enhanced alertness for detection of tional information may interact with attended eval-
the respective visual cues, giving rise to shifts in the uation of emotional communicational signals
interpretation of facial expressions. Moreover, (Fig. 5b). Future research will be required, how-
comparison of hemodynamic responses with the ever, to further clarify the neuroanatomical basis of
individual explicit ratings of emotional facial ex- interaction effects between implicit and explicit
pressions in presence of unattended fearful pros- stimulus processing and integration of emotional
ody revealed a significant correlation within the signals conveyed by various means of commu-
basolateral part of the left amygdala extending nication.
into the periamygdaloid cortex. This finding indi-
cates the impact of voice on the processing of faces
to be mediated via these anterior temporal struc-
tures. In line with this assumption, the amygdala Abbreviations
has been observed to modulate neuronal activity in
brain regions subserving visual processing (Morris BA Brodmann area
et al., 1998; Davis and Whalen, 2001; Vuilleumier, fMRI functional magnetic resonance
Richardson, Armony, Driver, & Dolan, 2004), and imaging
it has been suggested the left-sided nuclei integrate IFC inferior frontal cortex
audiovisual fear-related emotional information IFG inferior frontal gyrus
into a common percept (Dolan, Morris, & De mid-STS middle part of the superior tem-
Gelder, 2001). poral sulcus
265

MTG middle temporal gyrus Barton, J.J.S., Press, D.Z., Keenan, J.P. and O’Connor, M.
post-STS posterior part of the superior (2002) Lesions of the fusiform face area impair perception of
facial configuration in prosopagnosia. Neurology, 58: 71–78.
temporal sulcus
Baum, S.R. and Pell, M.D. (1999) The neural basis of prosody:
STG superior temporal gyrus insights from lesion studies and neuroimaging. Aphasiology,
STS superior temporal sulcus 13: 581–608.
Behrens, S.J. (1985) The perception of stress and lateralization
of prosody. Brain Lang., 26: 332–348.
Belin, P., Zilbovicius, M., Crozier, S., Thivard, L., Fontaine,
Acknowledgments A., Masure, M.C. and Samson, Y. (1998) Lateralization of
speech and auditory temporal processing. J. Cogn. Neurosci.,
10: 536–540.
The reported studies were supported by the Junior Blair, R.J.R. and Cipolotti, L. (2000) Impaired social response
Science Program of the Heidelberger Academy of reversal. Brain, 123: 1122–1141.
Sciences and Humanities and the German Research Blair, R.J.R., Morris, J.S., Frith, C.D., Perret, D.I. and Dolan,
Foundation (DFG WI 2101 and SFB 550 B10). R.J. (1999) Dissociable neural responses to facial expressions
of sadness and anger. Brain, 122: 883–893.
Borod, J.C., Bloom, R.L., Brickman, A.M., Nakhutina, L.
and Curko, E.A. (2002) Emotional processing deficits in indi-
References viduals with unilateral brain damage. Appl. Neuropsychol., 9:
23–36.
Ackermann, H., Hertrich, I., Grodd, W. and Wildgruber, D. Borod, J.C., Obler, L.K., Erhan, H.M., Grunwald, I.S., Cicero,
(2004) Das Hören von Gefühlen: funktionell-neuroanato- B.A., Welkowitz, J., Santschi, C., Agosti, R.M. and Whalen,
mische Grundlagen der Verarbeitung affektiver prosodie. J.R. (1998) Right hemisphere emotional perception: evidence
Aktuelle Neurol., 31: 449–460. across multiple channels. Neuropsychology, 12: 446–458.
Ackermann, H., Hertrich, I. and Ziegler, W. (1993) Prosodische Borod, J.C., Zgaljardic, D., Tabert, M.H. and Koff, E. (2001).
störungen bei neurologischen erkrankungen: eine lit- Asymmetries of emotional perception and expression in nor-
eraturübersicht. Fortschr. Neurol. Psychiatr., 61: 241–253. mal adults. In: Gainotti G. (Ed.), Handbook of Neuropsy-
Ackermann, H., Riecker, A., Grodd, W. and Wildgruber, D. chology, 2nd Edition, Vol. 5. Elsevier Science, Amsterdam,
(2001) Rate-dependent activation of a prefrontal-insular-cer- pp. 181–205.
ebellar network during passive listening to trains of click Bradley, M.M. and Lang, P.J. (1994) Measuring emotion: the
stimuli: an fMRI study. NeuroReport, 18: 4087–4092. self-assessment manikin and the semantic differential. J. Be-
Ackermann, H., Riecker, A. and Wildgruber, D. (2006) Cer- hav. Ther. Exp. Psychiatry., 25: 49–59.
ebral correlates of singing capabilities in humans: clinical Breitenstein, C., Daum, I. and Ackermann, H. (1998) Emo-
observations, experimental-behavioural studies, and func- tional processing following cortical and subcortical brain
tional imaging data. In: Altenmüller, E., Kesselring, J. and damage: contribution of the fronto-striatal circuitry. Behav.
Wiesendanger, M. (Eds.), Music, Motor Control, and the Neurol., 11: 29–42.
Brain. Oxford University Press, Oxford, pp. 205–221. Breitenstein, C., Van Lancker, D., Daum, I. and Waters, C.H.
Adolphs, R. (2002) Neural systems for recognizing emotion. (2001) Impaired perception of vocal emotions in parkinson’s
Curr. Opin. Neurobiol., 12: 169–177. disease: influence of speech time processing and executive
Adolphs, R., Damasio, H. and Tranel, D. (2002) Neural sys- functioning. Brain Cogn., 45: 277–314.
tems for recognition of emotional prosody: a 3-D lesion Buchanan, T.W., Lutz, K., Mirzazade, S., Specht, K., Shah,
study. Emotion, 2: 23–51. N.J., Zilles, K. and Jäncke, L. (2000) Recognition of
Adolphs, R. and Tranel, D. (1999) Intact recognition of emotional prosody and verbal components of spoken lan-
emotional prosody following amygdala damage. Neuro- guage: an fMRI study. Cogn. Brain Res., 9: 227–238.
psychologia, 37: 1285–1292. Calder, A.J., Keane, J., Manes, F., Antoun, N. and Young,
Adolphs, R., Tranel, D. and Damasio, H. (2001) Emotion rec- A.W. (2000) Impaired recognition and experience of disgust
ognition from faces and prosody following temporal lobec- following brain injury. Nat. Neurosci., 3: 1077–1078.
tomy. Neuropsychology, 15: 396–404. Cancelliere, A.E. and Kertesz, A. (1990) Lesion localization in
Anderson, A.K. and Phelps, E.A. (1998) Intact recognition of acquired deficits of emotional expression and comprehension.
vocal expressions of fear following bilateral lesion of the Brain Cogn., 13: 133–147.
human amygdala. Neuroreport, 9: 3607–3613. Canli, T., Desmond, J.E., Zhao, Z., Glover, G. and Gabrieli,
Bachorowski, J.O. and Owren, M.J. (2003) Sounds of emotion: J.D. (1998) Hemispheric asymmetry for the emotional stimuli
production and perception of affect-related vocal acoustics. detected with fMRI. Neuroreport, 9: 3233–3239.
Ann. NY Acad. Sci., 1000: 244–265. Charbonneau, S., Scherzer, B.P., Aspirot, D. and Cohen, H.
Banse, R. and Scherer, K.R. (1996) Acoustic profiles in vocal (2003) Perception and production of facial and prosodic emo-
emotion expression. J. Pers. Soc. Psychol., 70: 614–636. tions by chronic CVA patients. Neuropsychologia, 41: 605–613.
266

Critchley, H., Daly, E., Phillips, M., Brammer, M., Bullmore, Heilman, K.M., Scholes, R. and Watson, R.T. (1975) Auditory
E., Williams, S., Van Amelsvoort, T., Robertson, D., David, affective agnosia: disturbed comprehension of affective
A. and Murphy, D. (2000) Explicit and implicit neural mech- speech. J. Neurol. Neurosurg. Psychiatry, 38: 69–72.
anisms for processing of social information from facial ex- Hornak, J., Bramham, J., Rolls, E.T., Morris, R.G., O’Doherty,
pressions: a functional magnetic resonance imaging study. J., Bullock, P.R. and Polkey, C.E. (2003) Changes in emotion
Hum. Brain Mapp., 9: 93–105. after circumscribed surgical lesions of the orbitofrontal and
Cutler, A., Dahan, D. and Donselaar, W. (1997) Prosody in the cingulate cortices. Brain, 126: 1691–1712.
comprehension of spoken language: a literature review. Lang. Hornak, J., Rolls, E.T. and Wade, D. (1996) Face and voice
Speech, 40: 141–201. expression identification in patients with emotional and be-
Darby, D.G. (1993) Sensory aprosodia: a clinical clue to lesions havioral changes following ventral frontal lobe damage. Ne-
of the inferior division of the right middle cerebral artery? uropsychologia, 34: 247–261.
Neurology, 43: 567–572. Hugdahl, K. and Davidson, R.J. (2003) The Asymmetrical
Davidson, R.J., Abercrombie, H., Nitschke, J.B. and Putnam, Brain. MIT Press, Cambridge, London.
K. (1999) Regional brain function, emotion and disorders of Hughling-Jackson (1879). On affections of speech from disease
emotion. Curr. Opin. Neurobiol., 9: 228–234. of the brain (reprint from Brain 1879). Brain (1915), 38
Davis, M. and Whalen, P.J. (2001) The amygdala: vigilance and 107–129.
emotion. Mol. Psychiatry, 6: 13–34. Imaizumi, S., Mori, K., Kiritani, S., Kawashima, R., Sugiura,
de Gelder, B. and Vroomen, J. (2000) The perception of emo- M., Fukuda, H., Itoh, K., Kato, T., Nakamura, A., Hatano,
tions by ear and eye. Cogn. Emotion, 14: 289–311. K., Kojima, S. and Nakamura, K. (1997) Vocal identification
Dolan, R.J., Morris, J.S. and De Gelder, B. (2001) Crossmodal of speaker and emotion activates different brain regions.
binding of fear in voice and face. Proc. Natl. Acad. Sci. USA, NeuroReport, 8: 2809–2812.
98: 10006–10010. Ivry, R.B. and Robertson, L.C. (1998) The Two Sides of Per-
Ekman, P. and Friesen, W. (1976) Pictures of Facial Affect. ception. MIT Press, Cambridge, MA.
Consulting Psychologists Press, Palo Alto. Kanwisher, N., McDermott, J. and Chun, M.M. (1997)
Emmorey, K.D. (1987) The neurological substrates for pro- The fusiform face area: a module in human extrastriate cor-
sodic aspects of speech. Brain Lang., 30: 305–329. tex spezialized for face perception. J. Neurosci., 17:
Ethofer, T., Anders, S., Erb, M., Droll, C., Royen, L., Saur, R., 4302–4311.
Reiterer, S., Grodd, W. and Wildgruber, D. (2006a). Impact Kesler-West, M.L., Andersen, A.H., Smith, C.D., Avison, M.J.,
of voice on emotional judgement of faces: an event-related Davis, C.E., Kryscio, R.J. and Blonder, L.X. (2001) Neural
fMRI study. Hum. Brain Mapp (in press). substrates of facial emotion processing using fMRI. Cogn.
Ethofer, T., Anders, S., Erb, M., Herbert, C., Wiethoff, S., Brain Res., 11: 213–226.
Kissler, J., Grodd, W. and Wildgruber, D. (2006b) Cerebral Kotz, S.A., Meyer, M., Alter, K., Besson, M., von Cramon,
pathways in processing of emotional prosody: a dynamic D.Y. and Friederici, A.D. (2003) On the lateralization of
causal modelling study. NeuroImage, 30: 580–587. emotional prosody: an event-related functional MR investi-
Ethofer, T., Erb, M., Anders, S., Wiethoff, S., Herbert, C., gation. Brain Lang., 86: 366–376.
Saur, R., Grodd, W. and Wildgruber, D. (2006c) Effects of Kucharska-Pietura, K., Phillips, M.L., Gernand, W. and Dav-
prosodic emotional intensity on activation of associative au- id, A.S. (2003) Perception of emotions from faces and voices
ditory cortex. NeuroReport, 17: 249–253. following unilateral brain damage. Neuropsychologia, 41:
Friston, K.J., Harrison, L. and Penny, W. (2003) Dynamic 1082–1090.
causal modeling. NeuroImage, 19: 1273–1302. LeDoux, J. (1996) The Emotional Brain. Simon & Schuster,
Gandour, J., Wong, D. and Hutchins, G. (1998) Pitch process- New York.
ing in the human brain is influenced by language experience. Lehiste, I. (1970) Suprasegmentals. MIT Press, Cambridge,
NeuroReport, 9: 2115–2119. MA.
Geigenberger, A. and Ziegler, W. (2001) Receptive prosodic Levin, H.S., Eisenberg, H.M. and Benton, A.L. (1991) Frontal
processing in aphasia. Aphasiology, 15: 1169–1188. lobe function and dysfunction. Oxford University Press, New
George, M.S., Parekh, P.I., Rosinsky, N., Ketter, T.A., Kimbr- York, pp. 318–338.
ell, T.A., Heilman, K.M., Herscovitch, P. and Post, R.M. Mehrabian, A. (1972) Nonverbal Communication. Albine-At-
(1996) Understanding emotional prosody activates right herton, Chicago.
hemisphere regions. Arch. Neurol., 53: 665–670. Menon, R.S. and Goodyear, B.G. (1999) Submillimeter func-
Grandjean, D., Sander, D., Pourtois, G., Schwartz, S., Seghier, tional localization in human striate cortex using BOLD con-
M.L., Scherer, K.R. and Vuilleumier, P. (2005) The voices of trast at 4 Tesla: implications for the vascular point-spread
wrath: brain responses to angry prosody in meaningless function. Magn. Reson. Med., 41: 230–235.
speech. Nat. Neurosci., 8: 145–146. Meyer, M., Alter, K., Friederici, A.D., Lohmann, G. and von
Heilman, K.M., Bowers, D., Speedie, L. and Coslett, H.B. Cramon, D.Y. (2002) FMRI reveals brain regions mediating
(1984) Comprehension of affective and nonaffective prosody. slow prosodic modulations in spoken sentences. Hum. Brain
Neurology, 34: 917–921. Mapp., 17: 73–88.
267

Mitchell, R.L.C., Elliot, R., Barry, M., Cruttenden, A. and Pihan, H., Altenmüller, E., Hertrich, I. and Ackermann,
Woodruff, P.W.R. (2003) The neural response to emotional H. (2000) Cortical activation patterns of affective speech
prosody, as revealed by functional magnetic resonance im- processing depend on concurrent demands on the subvocal
aging. Neuropsychologia, 41: 1410–1421. rehearsal system: a DC-potential study. Brain, 123:
Morris, J.S., Friston, K.J., Büchel, C., Frith, C.D., Young, 2338–2349.
A.W., Calder, A.J. and Dolan, R.J. (1998) A neuromodulary Poeppel, D., Guillemin, A., Thompson, J., Fritz, J., Bavelier, D.
role for the human amygdala in processing emotional facial and Braun, A. (2004) Auditory lexical decision, categorical
expressions. Brain, 121: 47–57. perception, and FM direction discrimination differentially
Morris, J.S., Frith, C.D., Perrett, D.I., Rowland, D., Young, engage left and right auditory cortex. Neuropsychologia, 42:
A.W., Calder, A.J. and Dolan, R.J. (1996) A differential 183–200.
neural response in the human amygdala to fearful and happy Price, J.L. (1999) Prefrontal cortical network related to visceral
facial expressions. Nature, 383: 812–815. function and mood. Ann. NY Acad. Sci., 877: 383–396.
Morris, J.S., Scott, S.K. and Dolan, R.J. (1999) Saying it with Puce, A., Allison, T., Gore, J. and McCarthy, G. (1995) Face-
feelings: neural responses to emotional vocalizations. Ne- sensitive regions in human extrastriate cortex studied by
uropsychologia, 37: 1155–1163. functional MRI. J. Neurophysiol., 74: 1192–1199.
Murphy, F.C., Nimmo-Smith, I. and Lawrence, A.D. (2003) Reiterer, S.M., Erb, M., Droll, C.D., Anders, S., Ethofer,
Functional neuroanatomy of emotions: a meta-analysis. T., Grodd, W. and Wildgruber, D. (2005) Impact of task
Cogn. Affect. Behav. Neurosci., 3: 207–233. difficulty on lateralization of pitch and duration discrimina-
Murray, I.R. and Arnott, J.L. (1993) Toward the simulation of tion. NeuroReport, 16: 239–242.
emotion in synthetic speech: a review of the literature on Rolls, E.T. (1999) The functions of the orbito-frontal cortex.
human vocal emotion. J. Acoust. Soc. Am., 93: 1097–1108. Neurocase, 5: 301–312.
Nakamura, K., Kawashima, R., Ito, K., Sugiura, M., Kato, T., Ross, E.D. (1981) The aprosodias: functional-anatomic organ-
Nakamura, A., Hatano, K., Nagumo, S., Kubota, K., ization of the affective components of language in the right
Fukuda, H. and Kojima, S. (1999) Activation of the right hemisphere. Arch. Neurol., 38: 561–569.
inferior frontal cortex during assessment of facial emotion. Ross, E.D., Thompson, R.D. and Yenkosky, J. (1997) Later-
J. Neurophysiol., 82: 1610–1614. alization of affective prosody in the brain and the callosal
Pell, M.D. (1998) Recognition of prosody following unilateral integration of hemispheric language functions. Brain Lang.,
brain lesions: influence of functional and structural attributes 56: 27–54.
of prosodic contours. Neuropsychologia, 36: 701–715. Scherer, K.R., Johnstone, T. and Klasmeyer, G. (2003) Vocal
Pell, M.D. and Baum, S.R. (1997a) The ability to perceive and expression of emotion. In: R.J. Davidson K.R. Scherer H.H.
comprehend intonation in linguistic and affective contexts by Goldsmith (Eds.), Handbook of Affective Sciences. Oxford,
brain-damaged adults. Brain Lang., 57: 80–99. New York, pp. 433–456.
Pell, M.D. and Baum, S.R. (1997b) Unilateral brain damage, Schirmer, A., Alter, K., Kotz, S. and Friederici, A.D. (2001)
prosodic comprehension deficits, and the acoustic cues to Lateralization of prosody during language production: a le-
prosody. Brain Lang., 57: 195–214. sion study. Brain Lang., 76: 1–17.
Pell, M.D. and Leonard, C.L. (2003) Processing emotional tone Schmitt, J.J., Hartje, W. and Williams, K. (1997) Hemispheric
from speech in Parkinson’s disease: a role for the basal gan- asymmetry in the recognition of conditional attitude con-
glia. Cogn. Affect. Behav. Neurosci., 3: 275–288. veyed by facial expression, prosody and propositional speech.
Peper, M. and Irle, E. (1997) Categorical and dimensional de- Cortex, 33: 65–81.
coding of emotional intonations in patients with focal brain Scott, S.K., Young, A.W., Calder, A.J. and Hellawell, D.J.
lesions. Brain Lang., 58: 233–264. (1997) Impaired auditory recognition of fear and anger fol-
Perrett, I., May, K.A. and Yoshikawa, S. (1994) Facial shape lowing bilateral amygdala lesions. Nature, 385: 254–275.
and judgements of female attractiveness. Nature, 368: Sidtis, J.J. and Van-Lancker-Sidtis, D. (2003) A neurobehavioral
239–242. approach to dysprosody. Semin. Speech Lang., 24: 93–105.
Phan, K.L., Wager, T., Tayler, S.F. and Liberzon, I. (2002) Small, D.M., Zatorre, R.J., Dagher, A., Evans, A.C. and Jones-
Functional neuroanatomy of emotion: a meta-analysis of Gotman, M. (2001) Changes in brain activity related to eat-
emotion activation studies in PET and fMRI. NeuroImage, ing chocolate: from pleasure to aversion. Brain, 124:
16: 331–348. 1720–1733.
Phillips, M.L., Young, A.W., Scott, S.K., Calder, A.J., Andrew, Sprengelmeyer, R., Rausch, M., Eysel, U.T. and Przuntek, H.
C., Giampietro, V., Williams, S.C.R., Bullmore, E.T., Bra- (1998) Neural structures associated with recognition of facial
mmer, M. and Gray, J.A. (1998) Neural responses to facial expressions of basic emotions. Proc. R. Soc. Lond. B Biol.
and vocal expressions of fear and disgust. Proc. R. Soc. Sci., 265: 1927–1931.
Lond., 265: 1809–1817. Starkstein, S.E., Federoff, J.P., Price, T.R., Leiguarda, R.C.
Pihan, H., Altenmüller, E. and Ackermann, H. (1997) The cor- and Robinson, R.G. (1994) Neuropsychological and neuro-
tical processing of perceived emotion: a DC-potential study radiologic correlates of emotional prosody comprehension.
on affective speech prosody. NeuroReport, 8: 623–627. Neurology, 44: 515–522.
268

Surguladze, S.A., Brammer, M.J., Young, A.W., Andrew, C., Wildgruber, D., Ackermann, H., Klose, U., Kardatzki, B. and
Travis, M.J., Williams, S.C.R. and Phillips, M.L. (2003) A Grodd, W. (1998) Hemispheric lateralization of speech pro-
preferential increase in the extrastriate response to signals of duction and singing at the level of the motor cortex in fMRI.
danger. NeuroImage, 19: 1317–1328. In: Ziegler, W. and Deger, K. (Eds.), Clinical Phonetics and
Van Lancker, D. (1980) Cerebral lateralization of pitch cues in Linguistics. Whurr, London, pp. 238–243.
the linguistic signal. Int. J. Hum. Commun., 13: 227–277. Wildgruber, D., Erb, M., Klose, U. and Grodd, W. (1997) Se-
Van Lancker, D. and Sidtis, J.J. (1992) The identification of quential activation of supplementary motor area and primary
affective-prosodic stimuli by left- and right-hemisphere-dam- motor cortex during self-paced finger movement in human
aged subjects: all errors are not created equal. J. Speech Hear. evaluated by functional MRI. Neurosci. Lett., 127: 161–164.
Res., 35: 963–970. Wildgruber, D., Hertrich, I., Riecker, A., Erb, M., Anders, S.,
Vuilleumier, P., Richardson, M.P., Armony, J.L., Driver, J. and Grodd, W. and Ackermann, H. (2004) Distinct frontal re-
Dolan, R.J. (2004) Distant influences of amygdala lesion on gions subserve evaluation of linguistic and affective aspects of
visual cortical activation during emotional face processing. intonation. Cereb. Cortex, 14: 1384–1389.
Nat. Neurosci., 7: 1271–1278. Wildgruber, D., Kischka, U., FaXbender, K. and Ettlin, T.
Weddell, R. (1994) Effects of subcortical lesion site on human (2000) The Frontal Lobe Score: evaluation of its clinical va-
emotional behaviour. Brain Cogn., 25: 161–193. lidity. Clin. Rehabil., 14: 272–278.
Wicker, B., Keysers, C., Plailly, J., Royet, J.P., Gallese, V. and Wildgruber, D., Pihan, H., Ackermann, H., Erb, M. and
Rizzolatti, G. (2003) Both of us disgusted in My insula: the Grodd, W. (2002) Dynamic brain activation during process-
common neural basis of seeing and feeling disgust. Neuron, ing of emotional intonation: influence of acoustic parameters,
40: 655–664. emotional valence and sex. NeuroImage, 15: 856–869.
Wildgruber, D. and Ackermann, H. (2003) Aphasie. In: Brandt, Wildgruber, D., Riecker, A., Hertrich, I., Erb, M., Grodd, W.,
T., Dichgans, J. and Diener, H.C. (Eds.), Therapie und Ethofer, T. and Ackermann, H. (2005) Identification of emo-
Verlauf neurologischer Erkrankungen. Kohlhammer, Stutt- tional intonation evaluated by fMRI. NeuroImage, 24:
gart, pp. 267–277. 1233–1241.
Wildgruber, D., Ackermann, H. and Grodd, W. (2001) Differ- Zatorre, R.J. (2001) Neural specializations for tonal processing.
ential contributions of motor cortex, basal ganglia and cer- Ann. NY Acad. Sci., 930: 193–210.
ebellum to speech motor control: effects of syllable repetition Zatorre, R.J. and Belin, P. (2001) Spectral and temporal
rate evaluated by fMRI. NeuroImage, 13: 101–109. processing in human auditory cortex. Cereb. Cortex, 11:
Wildgruber, D., Ackermann, H., Klose, U., Kardatzki, B. and 946–953.
Grodd, W. (1996) Functional lateralization of speech pro- Zatorre, R.J., Belin, P. and Penhune, V. (2002) Structure and
duction at primary motor cortex: a fMRI study. NeuroRe- function of auditory cortex: music and speech. Trends Cogn.
port, 7: 2791–2795. Sci., 6: 37–46.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 14

Affective and linguistic processing of speech prosody:


DC potential studies

Hans Pihan1,2,

1
Department of Neurology, Schulthess Klinik, 8008 Zurich, Switzerland
2
Department of Neurology, Inselspital, University of Bern, 3010 Bern, Switzerland

Abstract: Speech melody or prosody subserves linguistic, emotional, and pragmatic functions in speech
communication. Prosodic perception is based on the decoding of acoustic cues with a predominant function
of frequency-related information perceived as speaker’s pitch. Evaluation of prosodic meaning is a cog-
nitive function implemented in cortical and subcortical networks that generate continuously updated
affective or linguistic speaker impressions. Various brain-imaging methods allow delineation of neural
structures involved in prosody processing. In contrast to functional magnetic resonance imaging tech-
niques, DC (direct current, slow) components of the EEG directly measure cortical activation without
temporal delay. Activation patterns obtained with this method are highly task specific and intraindividually
reproducible. Studies presented here investigated the topography of prosodic stimulus processing in de-
pendence on acoustic stimulus structure and linguistic or affective task demands, respectively. Data ob-
tained from measuring DC potentials demonstrated that the right hemisphere has a predominant role in
processing emotions from the tone of voice, irrespective of emotional valence. However, right hemisphere
involvement is modulated by diverse speech and language-related conditions that are associated with a left
hemisphere participation in prosody processing. The degree of left hemisphere involvement depends on
several factors such as (i) articulatory demands on the perceiver of prosody (possibly, also the poser), (ii) a
relative left hemisphere specialization in processing temporal cues mediating prosodic meaning, and (iii) the
propensity of prosody to act on the segment level in order to modulate word or sentence meaning. The
specific role of top-down effects in terms of either linguistically or affectively oriented attention on lat-
eralization of stimulus processing is not clear and requires further investigations.

Keywords: DC potentials; slow potentials; prosody; hemisphere specialization; emotion; acoustics; digital
resynthesis; nonverbal communication

Biological theories of prosody processing frequency-related information such as pitch, (ii)


temporal information such as speech rate or
Prosody or speech melody subserves linguistic speech rhythm, and (iii) loudness. Misclassifica-
functions in speech communication, such as ques- tions of affective prosodic impressions are based
tion versus statement intonation, and mediates on certain similarities between emotions that result
emotional speaker states. Prosodic perception is from similar quantitative expression of acoustic
based on decoding of acoustic cues mediating (i) cues. For example, happiness and despair are more
likely to be confused than happiness and sadness
Corresponding author. Tel.: +41-44–385-7171; Fax: +41-44- because both emotions are characterized by high
385-7538; E-mail: hans_pihan@freesurf.ch values of pitch. These observations gave rise to

DOI: 10.1016/S0079-6123(06)56014-5 269


270

dimensional models that conceptualize emotional linguistic information (Van Lancker, 1980). For
perception as a result of multidimensional evalua- example, the difference of ‘ ‘hot dog’ versus
tion (Frijda, 1969). Typical dimensions used are ‘hot‘dog’ mediated by prosody would require LH
activation (e.g., highly aroused vs. calm) and mechanisms for perception whereas question and
valence (e.g., pleasant vs. unpleasant). Acoustic statement intonations (‘hot ‘dog?’ vs. ‘hot ‘dog!’)
correlates of these dimensions have been character- would predominantly involve the RH. Support for
ized (Frick, 1985; Ladd et al., 1985; Bergmann et this hypothesis has been found in lesion studies
al., 1988; Tischer, 1993; Banse and Scherer, 1996). and in recent imaging experiments (Pell and
Fundamental frequency (F0), for example, is the Baum, 1997b; Wildgruber et al., 2004; Wong
lowest frequency present in sound during phona- et al., 2004).
tion. It represents a physical correlate of perceived The differential specialization hypothesis sug-
pitch and indicates speaker activation. Voice dy- gests preferential RH processing of pitch patterns
namics can be characterized by measurement of irrespective of linguistic or emotional function and
fundamental frequency variation (F0-variability, LH specialization in high frequent temporal anal-
F0-range). Its expression in the extreme either re- ysis relevant for decoding speech sounds. This
flects a melodic speech expressing, for example, cortical asymmetry is assumed to result from the
happiness or a flat and poorly modulated tone of need to optimize processing of the acoustic envi-
voice typically mediating the impression of sadness. ronment in both temporal and frequency domains
Temporal parameters such as the variation of vowel (Zatorre et al., 2002). Accordingly, a right-later-
length also indicate dynamic voice properties either alized activation was observed when pitch patterns
typifying happiness (high variation) or amplifying were processed (Zatorre et al., 1992). Patients who
the impression of sadness (low variation). underwent resection of parts of the primary and
Perception studies on affective prosody further secondary acoustic cortex in the right temporal
demonstrated that judgments on dimensional cat- lobe were found to have elevated thresholds when
egories are based on different time scales. For ex- judging the direction of pitch change (Johnsrude
ample, during perception of a sentence recognition et al., 2000). In contrast, a predominant left hem-
of speaker activity was reliably completed after the isphere involvement was found when rapid fre-
first two words, whereas attribution of valence quency changes such as fast formant transitions,
continued to develop throughout the whole utter- which promote consonant detection, are processed
ance (Tischer, 1993). Thus, evaluation of prosodic in the auditory cortex (Zatorre et al., 1992; Belin
meaning can be considered a temporally extended et al., 1998; see review in Tervaniemi and Hugdahl,
mental process, which integrates different dimen- 2003).
sional aspects of intonation into a continuously A related version of the differential specializa-
updated impression. tion hypothesis, the acoustic cue hypothesis, sug-
In this review three relevant theories on prosody gests a functional dichotomy between the right and
processing are addressed which discuss different left hemisphere related to pitch (RH) and temporal
bottom-up versus top-down mechanisms of pros- processing (LH) of complex auditory stimuli. Pa-
ody processing. The functional lateralization hy- tients with RH lesions were found to make no use
pothesis proposes left hemisphere (LH) speciali- of F0-variability, but rather relied on duration
zation for linguistic intonation and right hemi- cues to assess affective prosody (Van Lancker and
sphere (RH) mediation of melodic aspects of Sidtis, 1992). Robin and co-worker (Robin et al.,
affective speech. This specialization is considered 1990) observed right temporoparietal lesions to
relative rather than absolute and is believed to de- disrupt the discrimination of tones, but not the
pend on the potency of intonation to vary seman- perception of time patterns, while lesions in the
tic understanding of speech content. Van Lancker homologous regions of the LH had opposite
suggested that the degree of LH involvement in effects. In accordance, a study using positron
prosodic processing depends on the extent to emission tomography (PET) found complemen-
which prosodic cues interact with segmental tary LH and RH auditory cortex activation in
271

response to temporal and spectral variation of two stimulus (S1) that was followed by an imperative
tones (Zatorre and Belin, 2001). These data sup- stimulus (S2) initiating a motor response. The neg-
port the idea of a pitch module located in the RH, ative DC potential occurring within S1–S2 intervals
representing a hard-wired right hemisphere spe- of up to 2 s was termed contingent negative vari-
cialization for nonverbal acoustic communication. ation (CNV). Later investigators using longer
As a consequence, the expected contribution of the S1–S2 intervals and focusing on cognitive, motor
RH to the processing of affective speech prosody preparatory (e.g., the ‘Bereitschaftspotential’), and
might reflect the discrimination of pitch contours motivational processes used the term slow cortical
rather than the evaluation of emotional signifi- potential or DC potential. For a review of this
cance of verbal utterances. method see Rockstroh et al. (1989) or Altenmüller
and Gerloff (1998).

DC potentials and prosodic evaluation


Processing of affective prosody mediated by pitch
In speech communication, mental activity is main- and temporal cues
tained over seconds when changes of speaker emo-
tion or variations of sentence meaning must be Affective prosodic discrimination: predominant RH
inferred from speech melody. These cognitive proc- involvement
esses strongly involve cortical networks. Sustained
cortical activation can directly be assessed by meas- There is clinical and experimental evidence that the
uring DC (direct current, slow) components of the RH has an advantage over the left in extracting
EEG signal. They presumably are generated by pitch information from complex auditory stimuli.
sustained excitatory input to apical dendrites of Patient with RH lesions were found to make no use
cortical pyramidal cells. Cortical activation patterns of F0-variability, but rather relied on duration cues
obtained with this method are task specific and in- to assess affective prosody (Van Lancker and Sidtis,
traindividually reproducible (Altenmüller, 1986; 1992). Furthermore, pitch analysis processors seem
Lang et al., 1988; Altenmüller et al., 1993; Pihan to be located within the right temporal lobe. Robin
et al., 1997, 2000). Microelectrode recordings dem- (Robin et al., 1990) observed that right temporo-
onstrated a close relationship between local field parietal lesions disrupted the discrimination of
potentials generated in particular from cells geo- tones but not the perception of time patterns, while
metrically arranged in the cortex, such as pyramidal lesions in the homologous regions of the LH had
cells, and changes of the blood oxygen level de- opposite effects. Patients who underwent resection
pendent (BOLD) signal (for a review see Logothetis of parts of the primary and secondary acoustic
and Pfeuffer, 2004). Results from these combined cortex in the right temporal lobe were found to
imaging and intracortical recordings support the have elevated thresholds when judging the direction
assumption that local intracortical processing is re- of pitch change (Johnsrude et al., 2000). An imag-
flected by both an extracranial recordable slow po- ing study using PET demonstrates a right-laterali-
tential and a correlating BOLD response. zed activation when pitch patterns were processed
DC potential investigations in affective prosody (Zatorre et al., 1992). Conceivably, therefore, the
processing presented here were designed in analogy expected contribution of the RH to the processing
to the ‘two stimulus paradigm’: Two stimuli (S1 of affective speech prosody reflects the discrimina-
and S2) are presented in succession with discrimin- tion of pitch contours rather than the evaluation of
able acoustic features, separated by a short time the emotional meaning of an utterance.
interval. Perception of S1 prepares subjects for per- In order to investigate specific contributions of
ception of a deviant acoustic structure of S2. A each hemisphere to prosody processing, control
verbal response is required after presentation of S2. tasks and/or appropriate control stimuli are needed
The original experiments were performed by Walter in order to evaluate bottom-up effects of acoustic
(Walter et al., 1964) who presented a warning cue processing and top-down effects of emotional
272

or linguistic engagement. All the more so as re- intense’. In sadly intonated pairs, the utterance
search in this field faces a corpus of experimental with longer syllable duration was expected to be
data which could not clearly support a differential labelled as ‘sadder’.
LH and RH involvement in prosody processing Sixteen healthy, right-handed subjects partici-
(Schlanger et al., 1976; Pell and Baum, 1997a, b; pated. During stimulus presentation DC potentials
Stiller et al., 1997; Imaizumi et al., 1998; Pell, 1998, were recorded from 26 nonpolarizable AgCl elec-
2006; Kotz et al., 2003; Doherty et al., 2004). trodes. The recordings were averaged and analysed
The first experiment reported here was designed relating the DC amplitudes during the presenta-
to test bottom-up effects of physical stimulus char- tion periods to a baseline taken from a 1.5 s pre-
acteristics according to the acoustic cue hypothesis. stimulus period. As shown in Fig. 2, activation of
Stimuli were presented as sequences of two succes- the second analysis interval was considerably
sive utterances, each of them with identical word- higher when compared to the first one reflecting
ing. The paired stimuli represented variants of the increasing cognitive demands on cognition and
same prosodic category (happy, sad, or neutral) working memory during stimulus perception.
differing perceptually in emotional intensity or Behavioural data demonstrated that vocally por-
arousal. Effects of varying intensity or arousal were trayed emotions were well recognized. Only few
created by digital manipulation of either pitch sentences with happy or sad tone were labelled
range or duration of stressed vowels. Pitch contours ‘neutral’ (Fig. 3). Discrimination of intensity based
of the original utterances were extended or reduced on variations of pitch range or duration of stressed
relative to the sentence-final F0-level being kept vowels yielded less consistent results. Few subjects
constant. Fig. 1(a) explicates the procedure and rated expressiveness at random or contrary to the
presents the average of maximum F0-range and its predictions made. We suggested that this behav-
standard deviation for all resynthesized stimuli. A ioural variability was driven by the inherent ambi-
copy of the original utterances was used in which guity of terms like ‘happier’ or ‘sadder’ with respect
emotional intensity was altered by varying vowel to dimensional specifications: For example, an ut-
duration of stressed syllables. From each original terance can be perceived as sadder, if it expresses
sentence, variants with short, middle, or long du- increasing passivity with low speaker activation
ration of stressed vowels were obtained by either (depression) or, in case one thinks of despair, if high
cutting out or doubling single pitch periods of the speaker arousal and activity is signalled. Even hap-
acoustic signal. Fig. 1(b) exemplifies the time ma- pier might equally denote a more extrovert, highly
nipulation and its effect on total stimulus duration. excited state (e.g. enthusiasm) as well as a rather
Stimulus pairs were presented in pseudoran- relaxed, self-content attitude such as placidity. In
domized order. As compared to the first sentence order to avoid the directing attention of the subjects
the second item of each pair varied either in pitch towards a pure evaluation of the activity dimension
range or in duration of stressed vowels. Wording and, possibly, towards a decision strategy based on
and emotional intonation were identical. At per- physical stimulus characteristics, response terms did
ceptual evaluation, the pitch- and time-manipu- not further specify the underlying emotion.
lated variants of happy and sad utterances differed Processing of pitch- and time-manipulated sen-
in the degree of perceived intensity of the respec- tence pairs with either happy or sad intonation
tive emotion. In contrast, sentences with neutral yielded a similar pattern of DC potentials. In each
intonation sounded more or less aroused. Healthy condition, cortical activation was significantly lat-
subjects specified the emotional category of each eralized towards the RH. Besides right-frontal ar-
stimulus pair (happy, sad, or neutral) and indi- eas, high activation was also observed over right-
cated whether the first or the second stimulus was central electrode positions (Fig. 4). These areas
more intense (emotional intonations) or higher were described to receive projections from the au-
aroused (neutral items). An answer was considered ditory cortex in the temporal plane (Keidel, 1971).
‘correct’ if the sentence with broader F0-range or Results of this experiment supported neuropsycho-
shorter vowel duration was recognized as ‘more logical assumptions of an RH dominance in
273

Fig. 1. (a) Acoustic signal (upper panel) and three synthetic pitch contours (lower panel) of the test sentence ‘Sie gab nach, und
verflogen war der ganze Unmut’, produced with neutral intonation. Voiced signal portions are indicated by bars below the acoustic
signal, numbers indicate peak and endpoint frequencies in Hz. In the lower panel, F0-range of the medial pitch contour is slightly
increased as compared to the original utterance. All original utterances were subjected to the same F0-range manipulation. Averages of
F0-range (highest minus lowest value) over four test sentences of each emotional and synthetic category are listed beneath (standard
deviation in parenthesis). (b) Durational variants of the vowel /o/ of the stressed syllable in ‘verflogen’. In the upper acoustic signal,
vowel duration has been reduced to 143 ms by cutting out single pitch periods; the middle and lower signals were lengthened to 195 and
232 ms, respectively, by doubling single pitch periods. Accented vowels of each test sentence were shortened/lengthened accordingly
constituting a corpus of 4  3 durational variants (short-, medial- or long-vowel duration) within each emotional category. Under-
neath, averages of total stimulus duration are listed (standard deviation in parenthesis). From Pihan et al. (2000).
274

Fig. 2. Time course of a single trial and averaged DC potential (electrode position FC4). Mean amplitudes within periods 1 and 2
provided the basis for data normalization and statistical evaluation. Evoked potentials within the first second of stimulus presentation
might partly reflect unspecific activation and orientation responses, and, therefore, were not analysed. The variable end of each
utterances lies within the light part of the grey bar. From Pihan et al. (2000).

Fig. 3. Percentage of correct answers across the subject group with respect to the identification of ‘emotional category’ (left panel) and
the evaluation of ‘expressiveness’ (right panel). F0 varied stimulus conditions are indicated by the suffix ‘-P’, duration-manipulated
conditions are marked ‘-T’. From Pihan et al. (1997).

emotion processing (Borod, 1992). As for prosody emotional intensity was only observed in neutral
processing, acoustic stimulus processing in the tem- stimulus pairs. Since the subjects correctly identified
poral plane as well as cognitive-affective evaluation those items as neutral utterances, the observed RH
were lateralized to the RH. No support was found lateralization during processing of pitch-manipu-
for the acoustic cue hypothesis. A differential effect lated neutral sentences cannot be explained by an
of pitch- and temporal-based manipulation of RH activation of stored emotional representations.
275

Fig. 4. (a) Grand average data are plotted into a black-and-white map with grey levels representing mean amplitude values of the
second presentation period. Highest amplitude values were recorded at right-frontal and right-central electrode positions. (b) Plot of
normalized amplitudes of all six stimulus conditions. Only neutrally intonated sentences differed significantly with respect to potentials’
amplitudes (po0.001) and distribution (po0.001) as revealed by contrast analysis. (c) Average data map from the processing of F0 and
duration varied pairs of sentences (second presentation period). From Pihan et al. (1997).

Psychoacoustic studies demonstrated that mean spectral components. An event-related design such
F0 and F0-range modulate emotional intensity as as ours in which subjects are not aware of the rel-
continuous variables. High values correspond to evant acoustic cue may provoke an activation of
high arousal, low values indicate low speaker ac- all mechanisms competent for prosodic evaluation.
tivity (Ladd et al., 1985; Bergmann et al., 1988). In Thus, possible lateralization effects of temporal
consideration of these observations our activation and spectral cue processing may have been
patterns suggest that evaluation of prosodic infor- masked.
mation as performed by the RH might extend be- Similar results were found in a functional mag-
yond specific emotions such as joy and sadness to a netic resonance imaging (fMRI) study that used the
broader range of a speaker’s inner states including same experimental design and identical stimuli. As
the level of arousal. This interpretation supports shown in Fig. 5, the temporal pattern of hemody-
earlier suggestions saying that components of emo- namic responses revealed increasing lateralization to
tional behaviour such as psychophysical activation the right hemisphere during processing of the sec-
and autonomic responses are mediated predomi- ond sentence. In addition, activation of mesial re-
nantly by RH mechanisms (for further information gions of both hemispheres, including anterior and
on this topic see Gainotti et al., 1993; Bradley and posterior cingulate gyrus, as well as supplementary
Lang, 2000). motor areas could be demonstrated. Thus, fMRI
A corresponding right-lateralized activation pat- yielded the same pattern of right-lateralized activa-
tern was observed for stimuli with pitch and tem- tion with frontal preponderance that was found in
poral variation of intensity. However, these the DC experiment. These results support the sug-
findings do not exclude a differential LH and gestion that blood oxygen level dependent (BOLD)
RH involvement in pitch and temporal cue signals from the cortex and DC potentials arise from
processing. Prosodic perception comprises com- similar electrochemical processes (see Logothetis
plex acoustic analyses including temporal and and Pfeuffer, 2004).
276

Fig. 5. Temporal sequence of fMRI activation across all subjects and stimuli conditions. Presentation of the stimuli, first sentence
indicated in yellow and second sentence in red, is shown at the left of the time axis. The assumed period of corresponding fMRI
activation under consideration of the haemodynamic delay of approximately 3–6 s is shown at the right side of the time axis. Areas
showing significant activation during the successive time intervals as compared to the baseline periods are projected on the lateral (left)
and mesial (right) surface of the template brain. The pictogram at the very right shows the corresponding contrast function for each
temporal interval. Successively analysed images are indicated with an arrow, baseline images as black bars and stimulation periods in
upper row (SPM96, n ¼ 12, po0.05 corrected, Z43.72, k48) From Wildgruber et al. (2002).

Inner speech: differential RH and LH involvement stable in all subjects but one. This test person
in prosody processing showed a bilateral activation with left-frontal pre-
ponderance. He indicated that he had repeated the
Activation patterns described above demonstrat- stimuli in parallel with the ongoing presentation
ing a predominant involvement of the RH were using inner speech. The test person assumed that
277

this manoeuvre had facilitated stimulus discrimi- pitch- and time-related information. A generally
nation. Although it is known that lateralization higher cortical activation was elicited in response
effects can be masked if linguistic operations are to time-varied as compared to pitch-manipulated
concurrently performed with nonlinguistic tasks sentences, whereas the reverse pattern was ob-
(Tompkins and Flowers, 1985), we were surprised served during discrimination of the same stimuli
that a comparatively easy and quickly automated (Figs. 6(b) and (c) — lower panel). Behaviourally,
task such as inner speech would abolish lateral- there was no significant change in recognition ac-
ization effects of affective prosody discrimination. curacy as compared to discrimination without in-
Possibly, the strategy of this exceptional test per- ner speech. We suggested that the pronounced left
son and the bilateral activation measured indi- hemisphere activation was not attributable to a
cated a left hemisphere involvement in prosody repeated low-level linguistic performance such as
processing. ‘inner speech’. The following line of argumenta-
In order to evaluate the function of inner speech tion was developed:
on affective prosody perception, a second group of The cognitive function of repeating words ‘in
eight healthy right-handed subjects performed in- our head’ has been conceptualized in terms of a
ner speech in addition to the discrimination task. phonological loop (Baddeley, 1995). It involves
They were instructed to repeat the wording of the two components, a memory store retaining pho-
utterance they were attending to in parallel to the nological information and an articulatory control
presentation. As only four semantically different process. Paulesu and co-workers (Paulesu et al.,
sentences were used this task was immediately 1993) localized the phonological store to the left
learned and completed without effort. supramarginal gyrus and the subvocal rehearsal
Under additional inner speech demands a bal- mechanism to Broca’s area. In our study, brain
anced RH/LH activation pattern arose with left- potentials during perception of the second sen-
frontal preponderance and reverse responses to tence were analysed. Besides inner speech there

Fig. 6. Grand average data of activation amplitudes; (a) spontaneous discrimination; (b) inner speech; (c) plot of normalized am-
plitudes averaged over all subjects including values of the second presentation/ analysis periods at frontal and temporoparietal
electrode positions. Upper panel: Mean activation of LH and RH during spontaneous discrimination (white rhombi) and inner speech
(black rhombi). Lower panel: Mean activation (average over LH and RH electrode values) during discrimination of F0 and duration
varied stimuli during spontaneous discrimination (inner speech: no) and inner speech (inner speech: yes). From Pihan et al. (2000).
278

was no need for further linguistic stimulus analysis represented by neural networks which provide an
nor for a phonological store during perception of interpretation of the acoustic structure in terms of
the second sentence. Consequently, differential a motor programme. Other input channels such as
brain responses towards frequency- and time-re- the visual do not seem to have the same coupling
lated information as shown in Fig. 6(c) were prob- strength to verbal output. Shaffer (1975), for ex-
ably not associated with phonological processing. ample, showed that auditory-vocal shadowing of
As compared to spontaneous discrimination, pro- continuous speech can be successfully combined
sodic evaluation under additional inner speech de- with visually controlled copy-typing. In contrast,
mands was paralleled by a shift of cortical copy-typing of continuous speech was not com-
involvement from the right hemisphere to a bilat- patible with a vocal reproduction of a written text,
eral processing with left-frontal preponderance. even for highly skilled audio-typists.
Stimulus pairs carrying temporal cues elicited The RH hypothesis of emotion processing
higher activation when compared to pitch-varied would suggest an independent activation of corre-
sentences. As duration of stressed vowels was al- sponding left- and right-sided neural networks in-
tered relative to their absolute length, the manip- volved in inner speech (which was performed
ulations did not change sentence rhythm or stress without effort) and prosodic affective evaluation
pattern. Its only function was to signal differences (which was difficult as shown in Fig. 3). As a con-
in prosodic intensity. We suggested that inner sequence, the left-frontal preponderance of brain
speech was associated with a bilateral prosodic activation observed in the inner speech experiment
processing and a differential weighting of acoustic does not support the right hemisphere hypothesis.
parameters used for evaluation of emotional con- Data presented suggest a bilateral processing of
tent: A relative increase of LH activation mediated prosody with LH involvement being dependent on
by temporal cues and a reduced RH processing of articulatory demands. They also indicate a possi-
frequency-related information. ble articulatory representation of suprasegmental
Results from the inner speech experiment are in speech characteristics in the left hemisphere. If this
line with data from lesion studies which found that is corroborated in future experiments, some results
RH damaged patients preferentially relied on du- of lesion studies may require reinterpretation. For
ration cues to make affective judgements whereas example, LH damaged patients have been shown
LH damaged subjects appeared to make use of F0- to improve in repetition and discrimination of
information (Van Lancker and Sidtis, 1992). Left affective prosody under reduced verbal-articula-
temporo-parietal lesions impaired the ability of tory load (Ross et al., 1997). This was taken as
gap detection and pattern perception when se- support for right hemisphere dominance for pro-
quences of tones were presented while frequency sodic processing. However, improvement may also
perception was completely preserved. Lesions of have occurred as a result of decoupling of acoustic
homologous regions of the RH had opposite input and verbal output channel in the LH, facil-
effects with normal processing of temporal infor- itating stimulus processing in the nonaffected RH.
mation (Robin et al., 1990). Patients with LH or RH lesions might both be
The left-frontal activation during inner speech impaired in prosodic evaluation: The LH group
was driven by an inherent functional coupling of by a disturbed use of temporal information and
acoustic input and verbal output channels. This the RH group from an impaired processing of
mechanism provides motor control information in frequency patterns. However, as shown by Van
early phase of speech analysis enabling subjects to Lancker (1992), preferential use of either param-
shadow (i.e. to repeat as fast as possible) speech eter by LH or RH damaged patients was not
segments with latencies at the lower limit for ex- sufficient for a good performance. This indicates
ecution of motor gestures (Porter and Lubker, that pitch range and duration parameters alone
1980). Shadowing studies indicated parallel proc- are not fully functional in mediating prosodic
esses of auditory analysis and generation of an impressions or in determining differential RH and
articulatory gesture. Perceived speech is rapidly LH involvement.
279

Biological speech perception/production models processing. As described in the first section of this
that incorporate emotional communication are chapter, the degree of LH involvement in prosodic
facing the same problem from different perspec- processing was suggested to depend on the extent
tives: The mechanisms of integration/decomposi- to which prosodic cues interact with segmental
tion of suprasegmental and segmental aspects of linguistic information (Van Lancker, 1980).
speech into/from articulatory-phonatory pro- In certain instances statements and questions
grammes remain unclear. These processes must are differentiated by pitch contours diverging over
be expected close to neural structures in the frontal large sentence segments. A study investigating the
(and maybe parietal) lobe which are functionally interdependence of sentence focus location and
connected to articulophonatory programmes. On speech intonation demonstrated that in utterances
the level of speech perception, an acoustic-motor with an initial focus, a clear divergence of question
interface system in the left hemisphere was sug- and statement intonations occurred after the initial
gested by Hickok and Poeppel, (2000). However, word (Eady and Cooper, 1986). In statements, the
besides phoneme processing, speech communica- F0 value dropped to a low number for the re-
tion extends to suprasegmental aspects such as mainder of the sentence, whereas in questions, F0
emotional intonation. The fact that we are able to showed rather high and increasing values for sub-
speak or to sing a given text and manage to do this sequent words. These findings were replicated by
in a happy or sad tone of voice is presumably ac- Pell (2001) who demonstrated a stable pattern of
complished by a differential contribution of each fall/rise of F0 contour for statements and ques-
hemisphere. As for speaking and singing, a recent tions in utterance with an initial sentence focus
fMRI study demonstrated differential lateralizat- and various affective intonations.
ion effects during covert performance and outlined These examples demonstrate that an intonational
motor cortex and anterior insula as functionally feature comprising a whole utterance (here F0 con-
involved areas (Riecker et al., 2000). Further ex- tour) may also appear in a linguistic context and in
periments are needed in order to investigate the parallel to other general acoustic characteristics in-
functional significance and the neural substrate of dicating speaker affect. Thus, the separation of
articulatory representation of suprasegmental as- intonational features into focal versus global related
pects of perceived speech. to linguistic versus affective perception appears in-
adequate as it does not consider overlapping lin-
guistic and emotive functions of intonation in
Processing of linguistic prosody mediated by pitch speech communication. Conceivably, emotional in-
direction tonations influence the linguistic perception of pitch
direction, for example, the discrimination of ques-
Speech communication presumably comprises tion versus statement intonation in the tone of voice.
both a phylogenetically young linguistic system Investigation of this topic promises to yield further
for the exchange of prepositional information and insight into the neural implementation of prosody
an older system for regulation of social behaviour processing. Spontaneous discrimination of affective
by means of emotional communication (Arndt and prosody has been shown to provoke a strong right-
Janney, 1991). Affective and linguistic prosodic lateralized activation. Acoustic cue hypotheses as
information is mediated by the tone of voice via well as differential specialization theory suggest a
different topological features of the acoustic sig- right hemisphere processing of pitch contour medi-
nal: a global pattern of pitch movement related to ating linguistic information (statement or question
emotive expressions and local characteristics cor- intonation). Emotional effects of intonation contour
responding to question versus statement intona- may or may not augment right hemisphere involve-
tion (in general at sentence-final location) or to ment. A bilateral or left-lateralized activation pat-
contrastive stress patterns (for example, ‘hot ‘dog’ tern would indicate participation of the left
vs. ‘hotdog’). These topological features are as- hemisphere and would support the functionalist hy-
sumed to determine lateralization of prosody pothesis. The following study was investigating the
280

cortical activation during linguistic perception of recorded from 64 nonpolarizable AgCl electrodes.
pitch direction of sentences that carried a neutral or The recordings were averaged and analysed relating
emotional intonation (happy, fearful). A total of the DC amplitudes during the presentation periods
three semantically different utterances were used. to a baseline taken from a 1.2 s prestimulus period.
They were recorded from two opera singers with Behavioural data showed that subjects were able
intonational focus on the first word according to to perceive intonational question and statement
Eady and Cooper (1986) and Pell (2001). Stimulus contrasts from pitch direction. The highest dis-
pairs were assembled which differed exclusively with crimination rates were observed for the neutrally
respect to pitch course direction. Acoustic differ- intoned sentence pairs. Fearful utterances were less
ences were created by means of digital resynthesis of consistently recognized, and items with happy in-
the original utterances. In order to create perceptual tonation had the lowest number of correct re-
contrasts of statement versus question intonations, sponses. Acoustic stimulus analysis suggested a
F0 values of successive voiced sentence segments perceptual interaction of pitch variability associ-
were changed either by a stepwise increase or de- ated with the type of emotional intonation and
crease of fundamental frequency values (Fig. 7). The pitch direction, indicating intonational question/
relative rise or fall of pitch contour as compared to statement contrasts. Higher pitch variability was
sentence initial floor values was not significantly paralleled by lower linguistic recognition accuracy:
different between fearful, neutral and happy items. Neutral sentences were easier to discriminate than
Sixteen healthy subjects listened to pairs of fearful items and happy stimuli were the most
sentences with happy, fearful or neutral speech difficult to discriminate (Fig. 8).
melody. Stimulus pairs were presented in pseudo- Analysis of DC potentials during processing of
randomized order. As compared to the first sen- the first utterance revealed a significant main effect
tence the second item of each pair varied in pitch for emotion (Fig. 9). Highest cortical activation
direction constituting a contrast up versus down or was observed when stimuli with happy intonation
up versus horizontal, respectively. Wording and were discriminated. Processing of neutral sentence
emotional intonation were identical. At perceptual pairs was associated with lowest brain activation.
evaluation, the sentence with continuously rising No significant difference was observed between the
contour was expected to represent an intonational left and right hemisphere. A post-hoc analysis
question. Subjects were asked to identify this item. performed on frontal, central and temporoparietal
During stimulus presentation DC potentials were regions yielded significant effects of emotion over

Fig. 7. Digital resynthesis of F0 direction. Acoustic signal (upper panel) and three synthetic pitch contours (lower panel) of the test
sentence ‘She did not believe the entire story’ produced with neutral intonation (female voice). Voiced signal portions are indicated by
bars under the corresponding acoustic signal. Numbers at the end of pitch contours indicate sentence final F0 differences (in Hz).
281

Fig. 8. Left and middle: Acoustic stimulus analyses. Bars indicate normalized mean values for pitch (mean F0) and pitch variability
(mean F0 SD). Utterances with neutral intonation were taken as a reference in order to control for gender related speaker effects. Both
acoustic parameters yielded significant effects of EMOTION in an ANOVA performed on ranked values (both, po0.01); right:
Accuracy of linguistic discrimination. Mean value of % correct answers plus standard deviations.

central and temporoparietal regions as indicated in


Fig. 9. Again, no significant difference in activa-
tion amplitude was observed when left and right
hemisphere regions were compared.
Previous research from our lab using DC poten-
tials had corroborated RH processing of affective
prosody when subjects evaluated valence and emo-
tional intensity (Pihan et al., 1997). In the current
study, stimulus processing lateralized to the right
hemisphere was not observed. In line with other
studies our data reflect bihemispheric involvement
(Heilman et al., 1975, 1984; Bryan, 1989; Cancelliere
and Kertesz, 1990; Pell and Baum, 1997b; Pihan et
al., 1997, 2000; Wildgruber et al., 2002; Kotz et al.,
2003; Pell, 2005). Although task demands required
focusing on linguistic aspects of intonation, a sig-
nificant activation main effect of emotional intona-
tion was observed which corresponded to the
behavioural results reported above. The data indi-
cated a bihemispheric competence in pitch pattern
processing and speak against a strong version of the
acoustic cue or differential specialization hypothesis.
Post-hoc analysis revealed that the main effect of
emotion primarily arose from central and temporo- Fig. 9. Cortical activation plotted separately for happy, neutral
and fearful intonations (normalized values) at frontal (left),
parietal regions. What might be the function of a
central (middle) and temporoparietal (right) electrode positions
posterior neural network when concurrent linguistic as indicated in the maps. Post-hoc analysis of the emotion main
and emotional pitch effects are processed? In the effect investigated in the respective regions: (n.s.) ¼ nonsignifi-
present study subjects were required to make cant; (*) ¼ significant.
282

judgements on linguistic functions of pitch direction. References


Presumably, interference occurred from emotional
pitch effects. It can be assumed that posterior left Altenmüller, E. (1986) Hirnelektrische Korrelate der cerebralen
and right hemisphere networks monitor task-rele- Musikverarbeitung beim Menschen. Eur. Arch. Psychiat.
Neurol. Sci., 235: 342–354.
vant linguistic (or emotive) pitch effects. The main
Altenmüller, E. and Gerloff, Ch. (1998) Psychophysiology and
effect of emotion observed in this study may have the EEG. In: Niedermeyer, E. and Lopes da Silva, F. (Eds.),
resulted from differential degrees of task difficulty Electroencephalography. Williams and Wilkins, Baltimore,
when performing the discrimination task. Posterior pp. 637–655.
brain regions getting differentially involved in lin- Altenmüller, E., Kriechbaum, W., Helber, U., Moini, S., Di-
guistic versus affective intonation processing were chgans, J. and Petersen, D. (1993) Cortical DC-potentials in
identification of the language-dominant hemisphere: linguis-
recently outlined in an fMRI study (Gandour et al., tic and clinical aspects. Acta Neurochir., 56: 20–33.
2003). However, the specific contribution of left Arndt, H. and Janney, R.W. (1991) Verbal, prosodic, and
hemisphere regions to linguistic processing of pitch kinesic emotive contrasts in speech. J. Pragma., 15: 521–549.
direction needs further clarification. Baddeley, A. (1995) Working memory. In: Gazzaniga, M.S.
(Ed.), The Cognitive Neuroscience. MIT, Baltimore, pp.
755–764.
Banse, R. and Scherer, K.R. (1996) Acoustic profiles in vocal
Conclusion emotion expression. J. Pers. Soc. Psychol., 70: 614–636.
Belin, P., Zilbovicius, M., Crozier, S., Thivard, L., Fontaine,
Studies on perception of affective prosody using A., Masure, M.C. and Samson, Y. (1998) Lateralization of
speech and auditory temporal processing. J. Cogn. Neurosci.,
DC potentials corroborated that the RH has a
10: 536–540.
predominant role in processing emotions from the Bergmann, G., Goldbeck, T. and Scherer, K.R. (1988) Emo-
tone of voice, irrespective of emotional valence. tionale eindruckswirkung von prosodischen sprechmerk-
However, the right hemisphere holds a merely rel- malen. Z. Exp. Angew. Psychol., 35: 167–200.
ative dominance, both for processing of F0 and for Borod, J.C. (1992) Interhemispheric and intrahemispheric con-
evaluation of emotional significance of sensory in- trol of emotion: A focus on unilateral brain damage. J. Con-
sult. Clin. Psychol., 60: 339–348.
put. The role of the left hemisphere is at least Bradley, M.M. and Lang, P.J. (2000) Measuring emotion: Be-
complementary for the analysis of intonation. The havior, feeling, and physiology. In: Lane, R.D. and Nadel, L.
degree of LH involvement seems to be determined (Eds.), Cognitive Neuroscience of Emotion. Oxford Univer-
by a number of speech- and language-related con- sity Press, New York, pp. 242–276.
ditions such as: articulatory demands on the per- Bryan, K.L. (1989) Language prosody and the right hemi-
sphere. Aphasiology, 3: 285–299.
ceiver of prosody (possibly also the poser); the Cancelliere, A.E.B. and Kertesz, A. (1990) Lesion localization
power of temporal cues mediating prosodic mean- in aquired deficits of emotional expression and comprehen-
ing; the propensity of prosody to act on the seg- sion. Brain Cogn., 13: 133–147.
ment level in order to modulate word or sentence Doherty, C.P., West, W.C., Dilley, L.C., Shattuck-Hufnagel, S.
and Caplan, D. (2004) Question/statement judgments: an
meaning. In addition, posterior neural networks in
fMRI study of intonation processing. Hum. Brain Mapp., 23:
the left and right hemisphere may monitor task- 85–98.
dependent linguistic (or emotive) pitch effects. The Eady, S.J. and Cooper, W.E. (1986) Speech intonation and fo-
specific contribution of the LH to linguistic cus location in matched statements and questions. J. Acoust.
processing of pitch direction is not clear and re- Soc. Am., 80: 402–415.
quires further investigations. Frick, R.W. (1985) Communicating emotion: The role of pro-
sodic features. Psychol. Bull., 97: 412–429.
Frijda, N.H. (1969) Recognition of emotion. In: Berkowitz, L.
(Ed.), Advances in Experimental Social Psychology. Aca-
Abbreviations demic Press, New York, pp. 167–223.
Gainotti, G., Caltagirone, C. and Zoccolotti, P. (1993) Left/
right and cortical/subcortical dichotomies in the neuropsy-
DC direct current
chological study of human emotions. Cogn. Emotion, 7:
F0 fundamental frequency 71–93.
LH left hemisphere Gandour, J., Wong, D., Dzemidzic, M., Lowe, M., Tong, Y.
RH right hemisphere and Li, X. (2003) A cross-linguistic fMRI study of perception
283

of intonation and emotion in Chinese. Hum. Brain Mapp., processing depend on concurrent demands on the subvocal
18: 149–157. rehearsal system: a DC-potential study. Brain, 123:
Heilman, K.M., Bowers, D., Speedie, L. and Coslett, H.B. 2338–2349.
(1984) Comprehension of affective and nonaffective prosody. Porter, R.J. and Lubker, J.F. (1980) Rapid reproduction of
Neurology, 34: 917–921. vowel-vowel sequences: evidence for a fast and direct acous-
Heilman, K.M., Scholes, R. and Watson, R.D. (1975) Auditory tic-motoric linkage in speech. J. Speech Hear. Res., 23:
affective agnosia. Disturbed comprehension of affective 576–592.
speech. J. Neurol. Neurosurg. Psychiatry, 38: 69–72. Riecker, A., Ackermann, H., Wildgruber, D., Dogil, G. and
Hickok, G. and Poeppel, D. (2000) Towards a functional ne- Grodd, W. (2000) Opposite hemispheric lateralization effects
uroanatomy of speech perception. Trends Cogn. Sci., 4: during speaking and singing at motor cortex, insula and cer-
131–138. ebellum. NeuroReport, 11: 1997–2000.
Imaizumi, S., Mori, K., Kiritani, S., Hosoi, H. and Tonoike, M. Robin, D.A., Tranel, D. and Damasio, H. (1990) Auditory
(1998) Task-dependent laterality for cue decoding during perception of temporal and spectral events in patients
spoken language processing. NeuroReport, 9: 899–903. with focal left and right cerebral lesions. Brain Lang, 39:
Johnsrude, I.S., Penhune, V.B. and Zatorre, R.J. (2000) Func- 539–555.
tional specificity in the right human auditory cortex for per- Rockstroh, B., Elbert, T., Canavan, A., Lutzenberger, W. and
ceiving pitch direction. Brain, 123: 155–163. Birbaumer, N. (1989) Slow Cortical Potentials and Behav-
Keidel, W.D. (1971) DC-potentials in the auditory evoked re- iour. Urban & Schwarzenberg, Munich.
sponse in man. Acta. Otolaryngol., 71: 242–248. Ross, E.D., Thompson, R.D. and Yenkosky, J. (1997) Later-
Kotz, S.A., Meyer, M., Alter, K., Besson, M., von Cramon, alization of affective prosody in brain and the callosal inte-
D.Y. and Friederici, A.D. (2003) On the lateralization of gration of hemisphere language functions. Brain Lang, 56:
emotional prosody: an event-related functional MR investi- 27–54.
gation. Brain Lang, 86: 366–376. Schlanger, B.B., Schlanger, P. and Gerstmann, L.J. (1976) The
Ladd, D.R., Silverman, K.E.A., Tolkmitt, F., Bergmann, G. perception of emotionally toned sentences by right-hemi-
and Scherer, K.R. (1985) Evidence for the independent func- sphere damaged and aphasic subjects. Brain Lang, 3:
tion of intonation contour type, voice quality, and F0 range 396–403.
in signaling speaker affect. J. Acoust. Soc. Am., 78: 435–444. Shaffer, L.H. (1975) Multiple attention in continuous verbal
Lang, W., Lang, M., Podreka, I., Steiner, M., Uhl, F., Suess, tasks. In: Rabbit, P.M.A. and Dornic, S. (Eds.), Attention
E., Muller, C. and Deecke, L. (1988) DC-potential shifts and and Performance V. Academic Press, New York.
regional cerebral blood flow reveal frontal cortex involve- Stiller, D., Gaschler-Markefski, B., Baumgart, F., Schindler, F.,
ment in human visuomotor learning. Exp. Brain Res., 71: Tempelmann, C., Heinze, H.J. and Scheich, H. (1997) Lat-
353–364. eralized processing of speech prosodies in the temporal cor-
Logothetis, N.K. and Pfeuffer, J. (2004) On the nature of the tex: a 3-T functional magnetic resonance imaging study.
BOLD fMRI contrast mechanism. Magn. Reson. Imag., 22: MAGMA, 5: 275–284.
1517–1531. Tervaniemi, M. and Hugdahl, K. (2003) Lateralization of au-
Paulesu, E., Frith, C.D. and Frackowiak, R.S.J. (1993) The ditory-cortex functions. Brain Res. Brain Res. Rev., 43:
neural correlates of the verbal component of working mem- 231–246.
ory. Nature, 362: 342–345. Tischer, B. (1993) Äusserungsinterne Änderungen des emotion-
Pell, M.D. (1998) Recognition of prosody following unilateral alen eindrucks mündlicher sprache: dimensionen und akus-
brain lesion: influence of functional and structural attributes tische korrelate der eindruckswirkung. Z. Exp. Angew.
of prosodic contours. Neuropsychologia, 36: 701–715. Psychol., XL, 4: 644–675.
Pell, M.D. (2001) Influence of emotion and focus location on Tompkins, C.A. and Flowers, C.R. (1985) Perception of emo-
prosody in matched statements and questions. J. Acoust. Soc. tional intonation by brain-damaged adults: the influence of
Am., 109: 1668–1680. task processing levels. J. Speech Hear. Res., 28: 527–538.
Pell, M.D. (2006) Cerebral mechanisms for understanding emo- Van Lancker, D. (1980) Cerebral lateralization of pitch cues in
tional prosody in speech. Brain Lang, 96: 221–234. the linguistic signal: papers in linguistic. Int. J. Hum. Co-
Pell, M.D. and Baum, S.R. (1997a) The ability to perceive and mmun., 13: 200–277.
comprehend intonation in linguistic and affective contexts by Van Lancker, D. and Sidtis, J.J. (1992) The identification of
brain-damaged adults. Brain Lang, 57: 80–99. affective-prosodic stimuli by left- and right-hemisphere-dam-
Pell, M.D. and Baum, S.R. (1997b) Unilateral brain damage, aged subjects: all errors are not created equal. J. Speech Hear.
prosodic comprehension deficits, and the acoustic cues to Res., 35: 963–970.
prosody. Brain Lang, 57: 195–214. Walter, W.G., Cooper, R., Aldridge, V.J., McCallum, W.C.
Pihan, H., Ackermann, H. and Altenmüller, E. (1997) The cor- and Winter, A.L. (1964) The contingent negative variation.
tical processing of perceived emotion: A DC-potential study An electrical sign of significance of association in the human
on affective speech prosody. NeuroReport, 8: 623–627. brain. Nature, 203: 380–384.
Pihan, H., Altenmüller, E., Hertrich, I. and Ackermann, H. Wildgruber, D., Hertrich, I., Riecker, A., Erb, M., Anders, S.,
(2000) Cortical activation patterns of affective speech Grodd, W. and Ackermann, H. (2004) Distinct frontal
284

regions subserve evaluation of linguistic and emotional as- Zatorre, R.J. and Belin, P. (2001) Spectral and temporal
pects of speech intonation. Cereb. Cortex, 14: 1384–1389. processing in human auditory cortex. Cereb. Cortex, 11:
Wildgruber, D., Pihan, H., Ackermann, H., Erb, M. and 946–953.
Grodd, W. (2002) Dynamic brain activation during process- Zatorre, R.J., Belin, P. and Penhune, V.B. (2002) Structure and
ing of emotional intonation: influence of acoustic parameters, function of auditory cortex: music and speech. Trends Cogn.
emotional valence, and sex. Neuroimage, 15: 856–869. Sci., 6: 37–46.
Wong, P.C., Parsons, L.M., Martinez, M. and Diehl, R.L. (2004) Zatorre, R.J., Evans, A.C., Meyer, E. and Gjedde, A. (1992)
The role of the insular cortex in pitch pattern perception: the Lateralization of phonetic and pitch discrimination in speech
effect of linguistic contexts. J. Neurosci., 24: 9153–9160. processing. Science, 256: 846–849.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 15

Lateralization of emotional prosody in the brain: an


overview and synopsis on the impact of study design

Sonja A. Kotz1,2,, Martin Meyer3 and Silke Paulmann1

1
Max Planck Institute for Human Cognitive and Brain Sciences, Stephanstrasse 1a, 04103 Leipzig, Germany
2
Day Care Clinic of Cognitive Neurology, University of Leipzig, Leipzig, Germany
3
Department of Neuropsychology, Institute for Psychology, University of Zürich, Zürich, Switzerland

Abstract: Recently, research on the lateralization of linguistic and nonlinguistic (emotional) prosody has
experienced a revival. However, both neuroimaging and patient evidence do not draw a coherent picture
substantiating right-hemispheric lateralization of prosody and emotional prosody in particular. The current
overview summarizes positions and data on the lateralization of emotion and emotional prosodic process-
ing in the brain and proposes that: (1) the realization of emotional prosodic processing in the brain is based
on differentially lateralized subprocesses and (2) methodological factors can influence the lateralization of
emotional prosody in neuroimaging investigations. Latter evidence reveals that emotional valence effects
are strongly right lateralized in studies using compact blocked presentation of emotional stimuli. In con-
trast, data obtained from event-related studies are indicative of bilateral or left-accented lateralization of
emotional prosodic valence. These findings suggest a strong interaction between language and emotional
prosodic processing.

Keywords: emotion; prosody; lateralization; fMRI; design; patients

Emotion and its supporting brain network the reinforcer is positive or negative. This classi-
fication model includes emotions associated with
Previous research postulates that emotions are the omission or termination of reinforcers and ex-
processed in a complex brain network including plains an emotion along a valence (positive vs.
the orbitofrontal cortex, the amygdala, the ante- negative) and an emotional intensity dimension
rior cingulate cortex, and the temporal and sub- (high vs. low) shaping the assumption that nega-
cortical structures (Davidson, 2000). However, not tive and positive emotions may be processed in
every study of emotional processing reports acti- partially different neural systems.
vation of all these brain regions. Here, differenti- Pharmacological investigations in the late sixties
ating accounts on emotion processing in the brain initially coined the valence hypothesis (Rosadini and
may shed some light on such differences. Rossi, 1967). However, in particular, the studies by
Rolls (1999), in a dimensional account of emo- Davidson (1992) shaped a lateralization hypothesis
tion, classified emotions according to their differ- of emotion processing on the basis of the valence
ent reinforcement contingencies and to whether dimension of emotion. The valence hypothesis
conceptualizes that emotion processing is anchored
in both hemispheres, but each hemisphere is special-
Corresponding author. E-mail: kotz@cbs.mpg.de ized for one valence. Several authors proposed that

DOI: 10.1016/S0079-6123(06)56015-7 285


286

the left hemisphere regulates positive emotions, Predicated on his investigations, Ross coined the
while the right hemisphere drives negative emotions notion of ‘‘aprosodia’’ encompassing all sorts of
(Robinson and Starkstein, 1989; Davidson, 1992; disorders of affective language (equivalent to the
Gur et al., 1994). It has also been noted in a range of term ‘‘aphasia’’ covering various disturbances of
patient investigations that next to cortical represen- propositional language). To further substantiate
tation of emotion, additional brain structures such this view, a subsequent publication by Ross et al.
as fronto-stiatal areas are involved in emotion (1981) reported the case study of a patient suffer-
processing (Sackheim et al., 1982; Morris et al., ing from right frontoparietal opercular damage
1996; Paradiso et al., 1999). However, lateralization who displayed impaired expression of all altera-
of function within these areas has not yet been fully tions in prosody, which signifies emotional speech.
addressed. On the basis of these clinical studies, Ross con-
In contrast to the valence hypothesis, the right cluded that the right hemisphere is dominant for
hemisphere hypothesis posits that the right hemi- organizing the affective-prosodic components of
sphere dominates the left hemisphere for all emo- language and gestural behavior (Ross, 1985).
tion processing. For example, early behavioral However, all strong lateralization hypotheses
studies reported enhanced left side emotional ex- have been challenged (Caltagirone et al., 1989;
pressiveness of the face (Sackheim et al., 1978). Kowner, 1995). In addition, several variants of the
Strauss and Moscovitch (1981) reported superior right hemisphere hypothesis have emerged. Some
identification of emotional facial expression when authors put forward that the right hemisphere en-
faces were presented in the left rather than the right gages in the perception and expression of emotion,
visual hemifield. These results were substantiated by but does not support the experience of an emotion
evidence of Bowers et al. (1985) reporting that the (Adolphs et al., 1999). Murphy et al. (2003) noted
identification of faces was disrupted after right a return in current emotion research to the concept
hemisphere rather than left hemisphere strokes. of individual neural systems coding distinct di-
Lastly, data from split-brain patients showed a re- mensions of emotions rather than the concept of
sponse to emotional stimuli but no capacity to ver- an integrated neural emotional system.
bally describe emotional stimuli (Gazzaniga, 1988).
In terms of prosody, as another multipurpose Tuning into emotional tone — the lateralization of
means of expression, clinical studies also but- emotional prosody
tressed the notion that affective components of
language are a dominant function of the right Motivated by the proposed lateralization hypothe-
hemisphere. The first clinical study to formally ses for general emotion processing, lateralization
address the neural organization of emotional hypotheses for emotional prosodic processing were
speech tested the prediction that damage to the developed. Taking both, voice and acoustic prop-
right hemisphere seriously disrupts the compre- erties, as well as valence into consideration, three
hension of affective aspects of speech (Heilman main hypotheses are proposed. Before each of these
et al., 1975). They observed that right brain-dam- hypotheses is discussed in turn, a brief discussion
aged patients were severely impaired when asked into acoustic voice properties will be presented.
to recognize emotions inserted into linguistically
neutral statements. In a series of seminal publica-
tions, Ross provided observations supporting the Voice and acoustic properties of emotion
view that the ‘‘functional-anatomic organization
in the right hemisphere mirrors that of the pro- The German saying ‘‘Der Ton macht die Musik,’’
positional language in the left hemisphere.’’ (Ross, literally means the tone defines how one under-
1981, p. 561). Examinations of patients suffering stands an expression. Thus, this saying often serves
from focal right hemisphere lesions indicated de- as an excellent introduction to prosody for a non-
ficient receptive and expressive prosody as well as linguist. Interestingly, the saying suits the purpose
impaired comprehension of emotional gestures. of emotional communication as well as it points to
287

the social relevance of appropriately using emo- Lateralization hypotheses — clinical evidence
tional prosodic cues in daily communication. For
example, we use a variety of emotional tones that Going beyond the acoustic characteristics of emo-
give the correct meaning to what we are saying. A tional vocalization and its potential brain basis in
simple example is the difference between the state- the limbic system and beyond, the last 30 years of
ment: ‘‘Tim is crazy’’ spoken in a positive or neg- emotional prosodic research reflect a quest in
ative tone of voice. Dependent on the acoustic search of cortical representation of emotion vo-
modulation of a voice articulating this sentence, calization. Comparable to the positions put for-
the interpretation of the utterance changes in that ward in emotion research, three main hypotheses
it could be evaluated as spoken in happy or neg- have been formulated. These hypotheses are the
ative tone. Modulating the emotional expressions result of clinical research and recently neuroimag-
of speech is thus dependent on physical regulation ing research on emotional prosody and will be
of the voice. According to Scherer (1989), this en- discussed in turn.
tails three physiological processes, i.e., respiration The right hemisphere hypothesis is based on pri-
that provides flow of air, phonation that trans- marily receptive but also expressive, clinical stud-
forms flow of air from lungs into sounds, and ies. In accord with the proposal by Ross (1981),
movement of articulators (jaw, tongue, and palate) the right hemisphere exclusively operates prosodic
that modulate speech sounds. While not investi- cues. He stipulated that the right inferior frontal
gating lateralization of vocalization per se, Scherer cortex serves expressive prosody, whereas the right
(1989; but see Belin et al., 2000 for a more recent posterior superior temporal cortex mediates re-
opinion on voice processing in general) proposes ceptive prosody. Even though Ross’ model was
that the processes regulating emotional vocaliza- based on parsimonious empirical data even to date
tion are controlled by the limbic system — a cir- some authors claim that both, linguistic and non-
cuitry composed of cortical and subcortical areas, linguistic (emotional) prosody are processed in the
which are supposed to support emotional proc- right hemisphere (Bryan, 1989; Dykstra et al.,
esses. The effects of emotional arousal on speech 1995). However, other studies that have investi-
production are primarily due to tonic activation of gated both types of prosody only show a right
the autonomic and somatic nervous system. Pred- hemisphere preference for emotional prosody
icated on the assumption that each emotion has its (Blonder et al., 1991; Borod, 1993; Starkstein
own acoustic profile, Banse and Scherer (1996) et al., 1994; but see Weintraub et al., 1981; Bradvik
analyzed the acoustic parameters of vocal cues to et al., 1991 for linguistic prosody). Early patient
define acoustic profiles of vocal emotional expres- evidence from identification tasks (Heilman et al.,
sion, as modulation of vocal cues is a good indi- 1975), discrimination tasks (Tucker et al., 1977),
cator of physiological arousal. For example, the and recognition tasks (Bowers et al., 1987) indi-
vocalization of anger reveals higher fundamental cates that right temporoparietal lesions result in
frequency than the vocalization of sadness. Fur- emotional prosodic deficits. A recent case report
thermore, intensity measurements reveal louder using intrasurgical electrocortical stimulation de-
vocalizations for happy than for sad utterances. scribes a particular sensitivity of the right fronto-
However, the issue of an acoustic description of central operculum for prosody (Montavont et al.,
emotion is not trivial. The particular quality of 2005). However, there is ample clinical evidence
prosodic emotions cannot solely be defined acous- that challenges the right hemisphere hypothesis for
tically. In addition, it has been demonstrated that emotional prosodic processing (Van Lancker and
decoding emotional information does not imper- Sidtis, 1992; Darby, 1993).
atively involve limbic regions. It rather appears Accordingly, the functional hypothesis claims that,
that individuals are capable of properly evaluating dependent on linguistic load, prosodic processing is
the emotional color of a spoken utterance with lateralized either to the left or to the right hemi-
only frontotemporal regions being recruited (Kotz sphere (Van Lancker, 1980). For example, data from
et al., 2003a). sentence-level linguistic prosodic processing show
288

selective influence of the left hemisphere (Van This short survey of empirical clinical evidence
Lancker, 1980; Emmorey, 1987). Thus, the relation and the developed lateralization hypotheses on
of lateralization and prosodic processing forms a emotional prosodic processing clearly show that
continuum. The more linguistic emphasis on the clinical research of emotional and linguistic pros-
task, the more pronounced is the left hemisphere ody provides very little convergent evidence that
involvement. In turn, one could speculate that the prosody is solely processed in the right hemisphere
smaller the linguistic load, the stronger the right (also see Baum and Pell, 1999 for similar conclu-
hemisphere involvement. That is, if the task empha- sions). Plausible factors driving such divergence in
sizes emotion, such as in emotional categorization, clinical research may result from (1) variable lesion
right hemisphere dominance predominates (Bowers locations and lesion size, (2) acute or chronic state
et al., 1987). As a consequence, lateralization of of testing, (3) secondary deficits such as depression
function could be influenced by attention to task and neglect, (4) stimulus characteristics (Van
over the course of an experiment. Lancker, 1980), (5) ill-defined concepts, and (6)
Van Lancker and Sidtis (1992) as well as Zatorre task complexity. For example, Tompkins and
et al. (2002) put forward a more detailed func- Flowers (1985) showed that task complexity in
tional lateralization hypothesis. The so-called pa- an emotional prosodic judgment correlated with
rameter dependence hypothesis states that prosody lateralization. The more complex a task was, the
perception is ruled by acoustic parameters such as more the left-hemispheric involvement during
pitch, duration, and intensity. It was found that emotional prosodic processing.
pitch is preferably processed in the right hemi- On the basis of these data, the following ques-
sphere, whereas duration, rhythm, and intensity tions remain to be solved in neuroimaging inves-
are primarily processed in the left hemisphere (see tigations of emotional prosodic processing in
Van Lancker and Sidtis, 1992; Sidtis and Van healthy participants: (1) What is the functional
Lancker-Sidtis, 2003). Such a hypothesis also calls specification of the right hemisphere, potentially
into question whether lateralization due to acous- relevant for prosodic processing in general (both,
tic properties is prosody specific (linguistic or non- linguistic and nonlinguistic), and (2) What sub-
linguistic; see Zatorre, 1988; Ouellette and Baum, stantiates a bilateral prosodic network also includ-
1993) or linguistic at all. Studies using nonlinguis- ing subcortical and basal temporal structures?
tic tasks revealed a left hemisphere superiority for
processing of time structure cues (Carmon and
Nachshon, 1971; Robinson and Starkstein, 1990), Lateralization hypotheses — neuroimaging evidence
and a right hemisphere preference for pitch
processing (Robin et al., 1990; Sidtis and Feld- Next to the question of interhemispheric special-
mann, 1990; Zatorre et al., 1994). ization of emotional prosodic processing, the in-
Lastly, the valence hypothesis of emotional pro- trahemispheric specialization is a critical factor in
sodic processing was put forward stating that lat- our understanding of the brain bases of emotional
eralization of emotional prosody can be influenced prosodic processing. As mentioned earlier, the
by specific emotional intonation. This hypothesis perception of emotional prosody may rely on a
is based on the clinical evidence on depression. distributed brain network. On the basis of patient
Davidson et al. (1999) reported that patients with data, right posterior temporal (parietal) brain
a depressive mood process positive emotions in the regions still seem to be the prime candidate
left hemisphere, but negative emotions in the right for emotional prosodic processing. However, ne-
hemisphere. However, Ross et al. (1981, 1997), as uroimaging studies, comparing neutral against
well as Pell and Baum (1997) reported data speak- emotional prosody, report bilateral activation of
ing against the valence-dependent lateralization of the posterior part of the superior temporal cortex
emotional prosody. According to these authors, (Phillips et al., 1998; Kotz et al., 2003a; Grandjean
there is no straightforward clinical evidence sup- et al., 2005; but see Wildgruber et al., 2002,
porting the valence hypothesis. 2004, 2005; Mitchell et al., 2003). As discussed
289

elsewhere (Schirmer and Kotz, 2006), latter acti- One function or multiple processing steps?
vation overlaps with areas reported for voice rec-
ognition (Belin et al., 2004) suggesting that Recently, we argued that defining emotional pros-
listeners potentially engage attention toward emo- ody as a multistep process rather than a holistic
tional tone (as realized by voice) rather than neu- concept would render lateralization of emotional
tral tone. prosody a less controversial issue (for more elab-
Next to the involvement of posterior temporal orate information please refer to Schirmer and
areas, some patient evidence links right frontopa- Kotz, 2006). By dividing emotional prosodic
rietal (Starkstein et al., 1994), right orbitofrontal processing into three subprocesses, that is, (1)
(Hornak et al., 1996), and right frontal brain re- acoustic analysis, (2) derivation of emotional
gions (Breitenstein et al., 1998) to the recognition meaning based on acoustic cues, and (3) evalua-
of emotional prosody. Significant blood flow tion processes two critical aspects influencing
changes of the right dorsal and ventral prefrontal lateralization are put forward such as, (1) lateral-
and orbitofrontal cortex have been reported dur- ization may vary as a function of the proposed
ing emotional prosodic judgment (George et al., subprocesses, and (2) lateralization can be sub-
1996; Morris et al., 1999; Gandour et al., 2003; but stantiated by a brain network supporting these
see Buchanan et al., 2000; Kotz et al., 2003a; subprocesses beyond the right hemisphere. Taking
Mitchell et al., 2003; Wildgruber et al., 2004, 2005 both, neuroimaging evidence with its high spatial
for bilateral activation). Furthermore, subcor- resolution and event-related brain potential (ERP)
tical structures, such as the basal ganglia, seem evidence with its high temporal resolution into
to be involved in processing emotional prosody consideration, we (Schirmer and Kotz, 2006) pro-
(Breitenstein et al., 1998, 2001; Pell and Leonard, posed the following: (1) acoustic analysis of tem-
2003; Sidtis and Van Lancker-Sidtis, 2003). To poral and frequency properties of a signal is
date, only three fMRI experiments report en- supported by left and right primary and second-
hanced activation of the basal ganglia during the ary auditory cortices, respectively (Zatorre and
processing of vocal verbal (Kotz et al., 2005) Belin, 2001; Liegeois-Chauvel et al., 2004; Zaehle
and nonverbal emotions (Morris et al., 1999; et al., 2004; but see Boemio et al., 2005 for a more
Wildgruber et al., 2002; Kotz et al., 2003a). fine-grained analysis of the lateralization of tem-
poral properties of the acoustic signal). As re-
ported in two recent fMRI studies, lateralization
Factors influencing the lateralization of emotional of basic acoustic information such as temporal
prosody information may occur when integration of such
information is in focus. Increase in left posterior
Given the short survey on clinical and neuroim- auditory cortex was reported as a consequence of a
aging evidence on the lateralization of emotional perceptual switch from nonspeech to speech
prosodic processing, one of the currently unsolved perception requiring integration of temporal in-
questions in the literature is: What drives lat- formation (Dehaene-Lambertz et al., 2005; Meyer
eralization of emotional prosody? As argued by et al., 2005). When frequency or intensity infor-
Davidson and Irwin (1999), divergent results could mation is discriminated (i.e., contrasting two
be the consequence of conceptual and methodo- tones), a more extensive right-lateralized superior
logical differences. Here, we focus on two possible temporal network including the posterior superior
factors, namely differentiation of emotional pro- temporal gyrus is recruited (Liebenthal et al.,
sodic processing and methodology, and substanti- 2003). Latter evidence points to the fact that the
ate the methodological factor by presenting some activation may not result from complex acoustic
of our recent data that contrast valence effects cue analysis per se, but may occur as a function of
under two different fMRI designs — event-related discrimination or contrast realized in an experi-
vs. blocked presentation (Donaldson and Buckner, mental setup. The role of task demands (i.e., al-
2001). location of attention) in this context awaits further
290

neuroimaging investigations. However, some re- least two of seven reports (George et al., 1996;
cent ERP evidence points to such a possibility Buchanan et al., 2000) used a blocked design re-
(Rinne et al., 2005). (2) Attributing emotional porting right hemisphere lateralization effects for
significance to a stimulus when comparing emo- emotional prosodic processing. Other reports
tional and neutral acoustic signals results in acti- (Wildgruber et al., 2002, 2004, 2005) describe right
vations along the bilateral posterior superior hemisphere lateralization of frontal/orbitofrontal
temporal cortex overlapping with areas identified brain regions for higher order judgment of emo-
as voice-specific (Belin et al., 2004). However, the tional prosody, but activation of temporal brain
functional specification of anterior and middle regions often is bilateral though with a right hemi-
portions of the superior temporal cortex related sphere dominance.
to emotional prosodic processing needs further In a recent study by Schäfer et al. (2005), the
investigations (but see Grandjean et al., 2005). authors pose the question whether stimulus ar-
(3) Dependent on the task utilized to test emo- rangement (blocked- and event-related designs)
tional prosodic processing, the right frontal cortex has a strong effect in the context of emotion
extending into the orbitofrontal cortex (Hornak et processing on the basis of the significant difference
al., 2003; Öngür et al., 2003) is recruited to explic- in detection power of the two design types. Their
itly evaluate the significance attributed to an emo- results show common activation between the de-
tional signal in the right posterior temporal cortex. sign types (bilateral amygdalae, left insula, and left
Taken together, starting to specify lateralization as orbitofrontal cortex), but distinct right orbitofron-
a function of level of processing allows to consol- tal and insula activation in the event-related de-
idate some of the controversial clinical and ne- sign. In addition, the extent of activation in all
uroimaging evidence reported above, even though brain areas was larger in the event-related design.
further evidence is needed to substantiate such a While the authors point out the advantages of an
working model of emotional prosodic processing. event-related design, it is also apparent that design
type can drive lateralization effects.
For this matter, we compared the results of a
Methodological factors — design and task previous experiment that had utilized a fully event-
related design (Kotz et al., 2003a) with results of a
We have recently argued that methodological fac- mixed design, that is, stimulus type (intelligible
tors may influence lateralization of emotional pro- emotional/nonemotional sentences and unintelli-
sodic processing (Kotz et al., 2003a, b). While it is gible [filtered] emotional/nonemotional sentences)
generally acknowledged that an event-related de- was presented in blocks, but within blocks, valence
sign is preferred over a blocked design, dependent (emotional vs. nonemotional sentences) was pre-
on the experimental realization, design per se may sented and analyzed in an event-related manner
critically affect lateralization of emotional pro- (Kotz et al., 2003b). As the same stimulus material
sodic processing. For example, it is well known and task (emotional prosodic categorization) was
that a blocked design provides enhanced statistical used in the two experiments, the effects of blocked-
power (Friston et al., 1999), if subtle differences presentation vs. fully event-related presentation
between conditions should be detected in the bold designs can be compared. We previously argued
signal as (1) artifacts are more easily detected, and (Kotz et al., 2003a) that the interleaving of stim-
(2) randomization and spacing of critical trials is ulus type in an event-related design might influ-
not necessary (Donaldson and Buckner, 2001). On ence bilateral and left-accented activation of an
the other hand, a blocked design suffers from emotional prosodic brain network. Given the fact
stimulus order predictability and does not allow to that the task, emotional prosodic categorization,
sort trial responses for individual response pat- engages the participants to verbally label prosodic
terns (Carter et al., 1998; Wagner et al., 1998) or as contours, the swift change from intelligible to un-
a function of stimulus characteristics (Pu et al., intelligible (filtered) emotional sentences may have
2001). Scanning the neuroimaging literature, at resulted in an effort to understand and categorize
291

unintelligible emotional sentences. In essence, refer to Kotz et al., 2003a). When visually compar-
as also argued by Vingerhoets et al. (2003), ing the valence effects (emotional [positive/negative]
verbally labeling emotional categories may pro- vs. neutral) of the respective blocked (Fig. 1) and
mote semantic processing and as a result enhance a fully event-related design (Fig. 2) design, present-
left-hemisphere effort over a right-hemisphere ing stimulus type in a blocked manner, while
analysis of emotional prosodic contours. Thus, keeping presentation within block event-related,
the evaluation of unintelligible emotional sen- results in a rightward shift of activation for both
tences may have become more effortful as a func- stimulus types.
tion of task. For that matter, both design and task Valence effects in the blocked-design revealed
can critically affect lateralization of emotional predominantly activation of the right anterior, mid-
prosodic processing. dle, and posterior superior temporal region (cover-
In a first step, we therefore manipulated the de- ing both, the superior temporal gyrus [STG] and
sign in a second fMRI investigation. The assump- superior temporal sulcus [STS]), as well as the right
tion of the follow-up study was that varying the middle frontal gyrus, right anterior insula, and right
design (from a fully event-related to a blocked- striatum in the intelligible sentence condition. Va-
presentation design), but keeping stimulus type, va- lence effects in the unintelligible sentence condition
lence, and task consistent, would critically test the are reflected in bilateral middle STG and right pos-
influence of design on the lateralization of function terior STS activation. In comparison to the results
(for procedure and recording specifications please of a fully event-related design (Kotz et al., 2003a),

Fig. 1. Displayed are in an axial view (nonradiological con- Fig. 2. Displayed are in an axial view (nonradiological con-
vention, left ¼ left; right ¼ right); the activation patterns for vention, left ¼ left; right ¼ right); the activation patterns for
intelligible emotional speech (left) and unintelligible emotional intelligible emotional speech (left) and unintelligible emotional
speech (right) for positive valence (top) and negative valence speech (right) for positive valence (top) and negative valence
(bottom) from the blocked presentation design. Functional ac- (bottom) from the event-related design. Functional activation
tivation was thresholded at ZX3.09 (uncorrected). was thresholded at ZX3.09 (uncorrected).
292

which resulted in strongly left-lateralized (though Blonder, L., Bowers, D. and Heilman, K. (1991) The role of the
bilateral) emotional valence effects for both stimu- right hemisphere in emotional communication. Brain, 114:
1115–1127.
lus types, the current results of the mixed presen-
Boemio, A., Fromm, S., Braun, A. and Poeppel, D. (2005) Hi-
tation design clarify the powerful effect of design on erarchical and asymmetric temporal sensitivity in human au-
the respective brain areas involved in emotional ditory cortices. Nat. Neurosci., 8: 389–395.
prosodic processing. Borod, J.C. (1993) Cerebral mechanisms underlying facial,
Lastly, Schäfer et al. (2005) discussed the po- prosodic and lexical emotional expression: a review of ne-
tential effects of task on the visual processing of uropsychological studies and methodological issues. Neuro-
psychology, 7: 445–493.
disgust and fear. Referring to an investigation by Bowers, D., Bauer, R.M., Coslett, H.M. and Heilman, K.M.
Taylor et al. (2003), the authors speculate that an (1985) Processing of face by patients with unilateral hemi-
active rather than a passive task can modulate spheric lesions: dissociations between judgments of facial
limbic activation with an active task reducing affect and facila identity. Brain Cogn., 4: 258–272.
Bowers, D., Coslett, H., Bauer, R.M., Speedie, L. and Heilman,
limbic activation. We have followed up this pos-
K. (1987) Comprehension of emotional prosody following
sibility and data from a nonverbal (unpublished unilateral hemispheric lesions: processing defect versus dis-
data) and verbal investigation (Kotz et al., 2005) traction defect. Neuropsychologia, 25: 317–328.
on emotion processing indicate that task affects Bradvik, B., Dravins, C., Holtas, S., Rosen, I., Ryding, E. and
both, activation of critical brain regions during the Ingvar, D. (1991) Disturbances of speech prosody following
respective emotional processing and lateralization right hemisphere infarcts. Acta Neurol. Scand., 84: 114–126.
Breitenstein, C., Daum, I. and Ackermann, H. (1998) Emo-
of these brain regions. tional processing following cortical and subcortical brain
Taken together, both factors evaluated here, damage: contribution of the frontostriatal circuitry. Behav.
namely differentiation of emotional prosody into Neurol., 11: 29–42.
subprocesses and the design type applied in fMRI Breitenstein, C., Van Lancker, D., Daum, I. and Waters, C.H.
investigations, have a major impact on our under- (2001) Impaired perception of vocal emotions in Parkinson’s
disease: influence of speech time processing and executive
standing of how emotional prosodic processing is functioning. Brain Cogn., 45: 277–314.
functionally anchored in the brain. Suffice to say, Bryan, K. (1989) Language prosody and the right hemisphere.
lateralization of function, and emotional prosody Aphasiology, 3: 285–299.
in particular, is not a simple concept of the right Buchanan, T., Lutz, K., Mirzazade, S., Specht, K., Shah, N.,
Zilles, K. and Jäncke, L. (2000) Recognition of emotional
hemisphere. Future clinical and neuroimaging in-
prosody and verbal components of spoken language: an
vestigations will considerably improve our knowl- fMRI study. Cogn. Brain Res., 9: 227–238.
edge on this function by keeping the critical factors Caltagirone, C., Ekman, P., Friesen, W., Gainotti, G., Ma-
discussed here in perspective. mmucari, A., Pizzamiglio, L. and Zoccolotti, P. (1989) Posed
emotional expression in unilateral brain damaged patients.
Cortex, 25: 653–663.
Carmon, A. and Nachshon, I. (1971) Effect of unilateral brain
References damage on perception of temporal order. Cortex, 7: 410–418.
Carter, C.S., Braver, T.S., Barch, D.M., Botvinick, M.M., Noll,
Adolphs, R., Russel, J. and Tranel, D. (1999) A role for the D.C. and Cohen, J.D. (1998) Anterior cingulate cortex, error
human amygdala in recognizing emotional arousal from un- detection and the on-line monitoring of performance. Sci-
pleasant stimuli. Psychol. Sci., 10: 167–171. ence, 280: 747–749.
Banse, R. and Scherer, K. (1996) Acoustic profiles in vocal Darby, D. (1993) Sensory aprosodia: a clinical clue to lesion of
emotion expression. J. Pers. Soc. Psychol., 3: 614–636. the inferior division of the right middle cerebral artery? Neu-
Baum, S. and Pell, M. (1999) The neural bases of prosody: rology, 43: 567–572.
insights from lesion studies and neuroimaging. Aphasiology, Davidson, R. (1992) Anterior cerebral asymmetry and the na-
13: 581–608. ture of emotion. Brain Cogn., 6: 245–268.
Belin, P., Zatorre, R.J., Lafaille, P., Ahad, P. and Pike, B. Davidson, R. (2000) Affective style, psychopathology, and re-
(2000) Voice-selective areas in human auditory cortex. Na- silience: brain mechanisms and plasticity. Am. Psychol., 55:
ture, 403: 309–312. 1196–1214.
Belin, P., Fecteau, S. and Bedard, C. (2004) Thinking the voice: Davidson, R., Abercrombie, H., Nitschke, J. and Putnam, K.
neural correlates of voice perception. Trends Cogn. Neuro- (1999) Regional brain function, emotion and disorders of
sci., 8: 129–135. emotion. Curr. Opin. Neurobiol., 9: 228–234.
293

Davidson, R.J. and Irwin, W. (1999) The functional neuroanat- Kotz, S.A., Paulmann, S. and Raettig, T. (2005) Varying task
omy of emotion and affective style. Trends Cogn. Sci., 3: 11–21. demands during the perception of emotional content: efMRI
Dehaene-Lambertz, G., Pallier, C., Serniclaes, W., Sprenger- evidence. J. Cogn. Neurosci., Suppl. S, 63.
Charolles, L., Jobert, A. and Dehaene, S. (2005) Neural cor- Kowner, R. (1995) Laterality in facial expressions and its effect
relates of switching from auditory to speech perection. Ne- on attributions of emotion and personality: a reconsidera-
uroimage, 24: 21–33. tion. Neuropsychologia, 33: 539–559.
Donaldson, D.I. and Buckner, R.L. (2001) Effective paradigm Liebenthal, E., Binder, J.R., Piorkowski, R.L. and Remez, R.E.
design. In: Matthews, P.M., Jezzard, P. and Evans, A.C. (2003) Short-term reorganization of auditory analysis induced
(Eds.), Functional Magnetic Resonance Imaging of the by phonetic experience. J. Cogn. Neurosci., 15: 549–558.
Brain: Methods for Neuroscience. Oxford University Press, Liegeois-Chauvel, C., Lorenzi, C., Trebuchon, A., Regis, J. and
Oxford, pp. 175–195. Chauvel, P. (2004) Temporal envelope processing in the human
Dykstra, K., Gandour, J. and Stark, R. (1995) Disruption of left and right auditory cortices. Cereb. Cortex, 14: 731–740.
prosody after frontal lobe seizures in the non-dominant Meyer, M., Zaehle, T., Gountouna, V.E., Barron, A., Jäncke,
hemisphere. Aphasiology, 9: 453–476. L. and Turk, A. (2005) Spectro-temporal processing during
Emmorey, K. (1987) The neurological substrates for prosodic speech perception involves left posterior auditory cortex.
aspects of speech. Brain Lang., 30: 305–320. Neuroreport, 19: 1985–1989.
Friston, K.J., Zarahn, E., Joseph, O., Henson, R.N.A. and Mitchell, R.L., Elliott, R., Barry, M., Cruttenden, A. and
Dale, A. (1999) Stochastic designs in event-related fMRI. Woodruff, P.W. (2003) The neural response to emotional
Neuroimage, 10: 609–619. prosody, as revealed by functional magnetic resonance im-
Gandour, J., Wong, D., Dzemidzic, M., Lowe, M., Tong, Y. aging. Neuropsychologia, 41: 1410–1421.
and Li, X. (2003) A cross-linguistic fMRI study of perception Montavont, A., Demarquay, G., Guenot, M., Isnard, J., Maugui-
of intonation and emotion in Chinese. Hum. Brain Mapp., ere, F. and Ryvlin, P. (2005) Ictal dysprosody and the role of the
18: 149–157. non-dominant frontal operculum. Epileptic Disord., 7: 193–197.
Gazzaniga, M. (1988) Brain modularity: towards a philosophy Morris, J., Frith, C., Perret, D., Rowland, D., Young, A.W.,
of conscious experience. In: Marcel, A. and Bisiach, E. Calder, A.J. and Dolan, R.J. (1996) A differential neural re-
(Eds.), Consciousness in Contemporary Science. Oxford sponse in the human amygdala to fearful and happy facial
University Press, Oxford, pp. 218–238. expressions. Nature, 383: 812–815.
George, M.S., Parekh, P.I., Rosinsky, N., Ketter, T.A., Kimbr- Morris, J., Scott, S. and Dolan, R. (1999) Saying it with feeling:
ell, T.A., Heilman, K.M., Herscovitch, P. and Post, R.M. neural responses to emotional vocalizations. Ne-
(1996) Understanding emotional prosody activates right uropsychologia, 37: 1155–1163.
hemisphere regions. Arch. Neurol., 53: 665–670. Murphy, F.C., Nimmo-Smith, I. and Lawrence, A. (2003)
Grandjean, D., Sander, D., Pourtois, G., Schwartz, S., Seghier, Functional neuroanatomy of emotions: a meta-analysis.
M.L., Scherer, K.R. and Vuilleumier, P. (2005) The voices of Cogn. Affect. Behav. Neurosci., 3: 207–233.
wrath: brain responses to angry prosody in meaningless Öngür, D., Ferry, A.T. and Price, J.L. (2003) Architectonic
speech. Nat. Neurosci., 8: 145–146. subdivision of the human orbital and medial prefrontal cor-
Gur, R., Skolnnick, B. and Gur, R. (1994) Effects of emotional tex. J. Comp. Neurol., 460: 425–449.
discrimination tasks on cerebral blood flow: regional activa- Ouellette, G. and Baum, S. (1993) Acoustic analysis of prosodic
tion and its relation to performance. Brain Cogn., 25: 271–286. cues in left and right-hemisphere-damaged patients. Aphasio-
Heilman, K., Scholes, R. and Watson, R. (1975) Auditory logy, 8: 257–283.
affective agnosia: disturbed comprehension of speech. J. Ne- Paradiso, S., Chemerinski, E., Yazici, K., Tartaro, A. and
urol. Neurosurg. Psychiatry, 38: 69–72. Robinson, R. (1999) Frontal lobe syndrome reassessed: com-
Hornak, J., Bramham, J., Rolls, E.T., Morris, R.G., O’Doh- parison of patients with lateral or medial frontal brain dam-
erty, J., Bullock, P.R. and Polkey, C.E. (2003) Changes in age. J. Neurol. Neurosurg. Psychiatry, 67: 664–667.
emotion after circumscribed surgical lesions of the orbito- Pell, M. and Baum, S. (1997) Unilateral brain damage, prosodic
frontal and cingulate cortices. Brain, 126: 1691–1712. comprehension deficits, and the acoustic cues to prosody.
Hornak, J., Rolls, E.T. and Wade, D. (1996) Face and voice Brain Lang., 57: 195–214.
expression identification in patients with emotional and be- Pell, M. and Leonard, C. (2003) Processing emotional tone
havioural changes following ventral frontal lobe damage. from speech in Parkinson’s disease: a role for the basal gan-
Neuropsychologia, 34: 247–261. glia. Cogn. Affect. Behav. Neurosci., 3: 275–288.
Kotz, S., Meyer, M., Alter, K., Besson, M., von Cramon, D.Y. Phillips, M., Young, A., Scott, S., Calder, A., Andrew, C.,
and Friederici, A.D. (2003a) On the lateralization of emo- Giampietro, V., Williams, S.C., Bullmore, E.T., Brammer,
tional prosody: an event-related functional MR investigation. M. and Gray, J.A. (1998) Neural responses to facial and
Brain Lang., 86: 366–376. vocal expressions of fear and disgust. Proc. R. Soc. Lond. B
Kotz, S., Meyer, M., Besson, M. and Friederici, A.D. (2003b) Biol. Sci., 265: 1809–1817.
On the lateralization of emotional prosody: effect of blocked Pu, Y., Liu, H.L., Spinks, J.A., Mahankali, S., Xiong, J., Feng,
versus event-related designs. J. Cogn. Neurosci. Suppl., 138. C.M., Tan, L.H., Fox, P.T. and Gao, J.H. (2001) Cerebral
294

hemodynamic response in Chinese (first) and English (sec- Strauss, E. and Moscovitch, M. (1981) Perception of facial ex-
ond) language processing revealed by event-related func- pressions. Brain Lang., 13: 308–332.
tional MRI. Magn. Reson. Imaging, 19: 643–647. Taylor, S.F., Phan, K.L., Decker, L.R. and Liberzon, I. (2003)
Rinne, T., Pekkola, J., Degerman, A., Autti, T., Jääskeläinen, Subjective rating of emotionally salient stimuli modulates
I.P., Sams, M. and Alho, K. (2005) Modulation of auditory limbic activity. Neuroimage, 18: 650–659.
cortex activation by sound presentation rate and attention. Tompkins, C. and Flowers, C. (1985) Perception of emotional
Hum. Brain Mapp., 26: 94–99. intonation by brain-damaged adults: the influence of task
Robin, D., Tranel, D. and Damasio, H. (1990) Auditory per- processing levels. J. Speech Hear. Res., 28: 527–538.
ception of temporal and spectral events in patients with focal Tucker, D., Watson, R. and Heilman, K. (1977) Discrimination
left and right cerebral lesions. Brain Lang., 39: 539–555. and evocation of affectively intoned speech in patients with
Robinson, R. and Starkstein, S. (1989) Mood disorders fol- right parietal disease. Neurology, 27: 947–950.
lowing stroke: new findings and future directions. J. Geriatr. Van Lancker, D. (1980) Cerebral lateralization of pitch cues in
Psychiatry Neurol., 22: 1–15. the linguistic signal. Int. J. Hum. Comm., 13: 101–109.
Robinson, R. and Starkstein, S. (1990) Current research in Van Lancker, D. and Sidtis, J. (1992) The identification of
affective disorders following stroke. J. Neuropsychiatry Clin. affective-prosodic stimuli by left- and right-hemisphere-dam-
Neurosci., 2: 1–14. aged subjects: all errors are not created equal. J. Speech Hear.
Rolls, E. (Ed.). (1999) The Brain and Emotion. Oxford Uni- Res., 35: 963–970.
versity Press, Oxford. Vingerhoets, G., Berckmoes, C. and Stoobant, N. (2003) Cer-
Ross, E.D. (1981) The aprosodias. Functional-anatomic or- ebral hemodynamics during discrimination of prosodic and
ganization of the affective components of language in the semantic emotions in speech studied by transcranial Doppler
right hemisphere. Arch. Neurol., 38: 561–569. ultrasonography. Neuropsychology, 17: 93–99.
Ross, E.D., Harney, J.H., de Lacoste-Utamsing, C. and Purdy, Wagner, A.D., Schacter, D.L., Rotte, M., Koutstaal, W., Maril,
P.D. (1981) How the brain integrates affective and propositio- A., Dale, A.M., Rosen, B.R. and Buckner, R.L. (1998) Build-
nal language into a unified behavioral function. Hyopthesis ing memories: remembering and forgetting of verbal experi-
based on clinicoanatomic evidence. Arch. Neurol., 38: 745–748. ences as predicted by brain activity. Science, 281: 1188–1191.
Ross, E.D., Thompson, R. and Yenkosky, J. (1997) Lateralizat- Weintraub, S., Mesulam, M. and Kramer, L. (1981) Distur-
ion of affective prosody in brain and the callosal integration of bances in prosody. A right-hemisphere contribution to lan-
hemispheric language functions. Brain Lang., 56: 27–54. guage. Arch. Neurol., 38: 742–744.
Rosadini, G. and Rossi, G.F. (1967) On the suggested cerebral Wildgruber, D., Hertrich, I., Riecker, A., Erb, M., Anders, S.,
dominance for consciousness. Brain, 90: 101–112. Grodd, W. and Ackermann, H. (2004) Distinct frontal re-
Sackheim, H., Greenberg, M., Weiman, A., Gur, R., Hunger- gions subserve evaluation of linguistic and emotional aspects
buhler, J., Geschwind, N. (1982) Hemispheric asymmetry in of speech intonation. Cereb. Cortex, 14: 1384–1389.
the expression of positive and negative emotions Neurologic Wildgruber, D., Pihan, H., Ackermann, H., Erb, M. and
evidence. Arch. Neurol., 39: 219–218. Grodd, W. (2002) Dynamic brain activation during process-
Sackheim, H., Gur, R. and Saucy, M. (1978) Emotions are ing of emotional intonation: influence of acoustic parameters,
expressed more intensely on the left side of the face. Science, emotional valence, and sex. Neuroimage, 15: 856–869.
202: 434–436. Wildgruber, D., Riecker, A., Hertrich, I., Erb, M., Grodd, W.,
Schäfer, A., Schienle, A. and Vaitl, D. (2005) Stimulus type and Ethofer, T. and Ackermann, H. (2005) Identification of emo-
design influence hemodynamic responses towards visual dis- tional intonation evaluated by fMRI. Neuroimage, 24:
gust and fear elicitors. Int. J. Psychophysiol., 57: 53–59. 1233–1241.
Scherer, K. (1989) Vocal correlates of emotional arousal and Zaehle, T., Wüstenberg, T., Meyer, M. and Jäncke, L. (2004)
affective disturbance. In: Wagner, H. and Manstead, A. Evidence for rapid auditory perception as the foundation of
(Eds.), Handbook of Social Physiology. Wiley, New York, speech processing: a sparse temporal sampling fMRI study.
pp. 165–197. Eur. J. Neurosci., 20: 2447–2456.
Schirmer, A. and Kotz, S.A. (2006) Beyond the right hemi- Zatorre, R. (1988) Pitch perception of complex tones and hu-
sphere: brain mechanisms mediating vocal emotional man temporal-lobe function. J. Acoust. Soc. Am., 84:
processing. Trends Cogn. Neurosci., 10: 24–30. 566–572.
Sidtis, J.J. and Feldmann, E. (1990) Transient ischemic attacks Zatorre, R.J. and Belin, P. (2001) Spectral and temporal
presenting with a loss of pitch perception. Cortex, 26: processing in human auditory cortex. Cereb. Cortex, 11:
469–471. 946–953.
Sidtis, J.J. and Van Lancker-Sidtis, D. (2003) A neurobehavioral Zatorre, R., Belin, P. and Penhune, V. (2002) Structure and
approach to dysprosody. Semin. Speech Lang., 24: 93–105. function of auditory cortex: music and speech. Trends Cogn.
Starkstein, S., Federoff, J., Price, T., Leiguarda, R. and Rob- Sci., 6: 37–46.
inson, R. (1994) Neuropsychological and neuroradiologic Zatorre, R., Evans, A. and Meyer, E. (1994) Neural mecha-
correlates of emotional prosody comprehension. Neurology, nisms underlying melodic perception and memory for pitch.
44: 515–522. J. Neurosci., 14: 1908–1919.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 16

Psychoacoustic studies on the processing of vocal


interjections: how to disentangle lexical and prosodic
information?

Susanne Dietrich1,, Hermann Ackermann2, Diana P. Szameitat1 and Kai Alter3

1
Max-Planck-Institute for Human Cognitive and Brain Sciences, Stephanstrasse 1a, 04103 Leipzig, Germany
2
Department of Neurology, Hertie Institute for Clinical Brain Research, University of Tübingen, Tübingen, Germany
3
School of Neurology, Neurobiology & Psychiatry, Newcastle upon Tyne, UK

Abstract: Both intonation (affective prosody) and lexical meaning of verbal utterances participate in the
vocal expression of a speaker’s emotional state, an important aspect of human communication. However, it
is still a matter of debate how the information of these two ‘channels’ is integrated during speech per-
ception. In order to further analyze the impact of affective prosody on lexical access, so-called interjections,
i.e., short verbal emotional utterances, were investigated. The results of a series of psychoacoustic studies
indicate the processing of emotional interjections to be mediated by a divided cognitive mechanism en-
compassing both lexical access and the encoding of prosodic data. Emotional interjections could be sep-
arated into elements with high- or low-lexical content. As concerns the former items, both prosodic and
propositional cues have a significant influence upon recognition rates, whereas the processing of the low-
lexical cognates rather solely depends upon prosodic information. Incongruencies between lexical and
prosodic data structures compromise stimulus identification. Thus, the analysis of utterances characterized
by a dissociation of the prosodic and lexical dimension revealed prosody to exert a stronger impact upon
listeners’ judgments than lexicality. Taken together, these findings indicate that both propositional and
prosodic speech components closely interact during speech perception.

Keywords: cognition; communication; emotion; language; recognition; semantics

Introduction As a prerequisite to the recognition of the emo-


tional meaning of a word, thus, propositional labels
Verbal utterances may convey information about a must refer to a knowledge-based semantic concept
speaker’s emotional state by the modulation of (lexical emotional meaning). These processes imply
intonational cues (affective prosody) as well as the existence of a mental lexicon ‘housing’ phonetic/
emotional word content (lexicality). Thus, both phonological patterns linked to emotional seman-
prosodic and lexical data structures must be ex- tics. A variety of functional imaging studies indicate
pected to closely interact during speech production. the ‘mental lexicon’ to be bound to specific cerebral
On the one hand, the sequential order of phonemes areas. It has been assumed that these regions act as
reflects the canonical structure of verbal utterances. an interface between word perception and long-
term memory representations of familiar words
Corresponding author. Tel.: +49-0341-9940268; (Wartburton et al., 1996; Wise et al., 2001). Evi-
E-mail: dietrich@cbs.mpg.de dence for the participation of specific cerebral areas

DOI: 10.1016/S0079-6123(06)56016-9 295


296

in representing sequences of phonemes comes, e.g., items? (influence of lexical semantics on recogni-
from cortical stimulation studies in patients who tion rates), and (c) Do incongruencies between
underwent surgical treatment for epilepsy (Anderson lexical and prosodic information, first, compro-
et al., 1999; Quigg and Fontain, 1999). On the other mise recognition rate as compared to congruent
hand, suprasegmental parameters of verbal utter- items and, second, disclose the relative impact of
ances such as pitch and loudness contours, syllable these two components upon listeners’ judgments?
durations, and various aspects of voice quality signal (interaction of prosodic and lexical information).
affective prosody of verbal utterances. These cues These investigations, thus, aimed at a further
are imposed upon the segmental structure of speech analysis of the perceptual and cognitive processes
utterances (Banse and Scherer, 1996; MacNeilage, underlying two of the most important aspects of
1998; Sidtis and Van-Lancker-Sidtis, 2003). Similar speech communication: prosody and lexicality. Par-
to lexical features, the recognition of the emotional ticipants were asked to judge the emotional meaning
meaning of an utterance might be based in these of auditory applied vocal interjections. The focus
instances upon the association of prosodic features was on so-called ‘basic emotions’ that can be easily
with a knowledge-based semantic concept (prosodic displayed by vocal expressions and reliably recog-
emotional meaning). nized from verbal utterances (Scherer and Oshinsky,
So far, it is unsettled whether the recognition of 1977). The following emotion categories were con-
emotional speech utterances involves a divided sidered for analysis: happiness, anger, disgust, and
cognitive mechanism encompassing both lexical sadness. Trained native speakers of German had
access and the encoding of prosodic information. produced these interjections with affective (happy,
For example, functional imaging studies revealed angry, disgusted, or sad) or neutral prosody.
different activation patterns in association with the
processing of lexical–semantic and prosodic data
structures (Wise et al., 2001; Friederici, 2002). Lexical meaning of verbal emotional interjections
These findings raise the question of how affective
prosodic information and lexicality influence each In order to separate segmental and suprasegmental
other during speech perception. An interaction levels of vocal affective expression, a series of in-
between both components could be a hint for sep- terjections of the German language was evaluated
arate processing of these two components within with respect to their emotional lexical meaning.
different cognitive systems. In order to address this Table 1 includes a list of items — spoken at a neu-
issue, emotional interjections, i.e., short utterances tral tone — that served as test materials. Under
often used during speech communication, were in- these conditions, affective prosodic cues are not
vestigated. These elements primarily reflect spon- available and, thus, only lexicality conveys emo-
taneous expressions of high emotional arousal tional meaning. Participants had to answer the
(Nübling, 2004), and their spectrum extends from question ‘Which emotion was expressed by the ut-
verbal utterances like ‘yuck’ (in German ‘igitt’) terance?’ and, thus, were instructed to focus on
and vowels of a distinct tone to nonverbal expres- lexical information ignoring prosody. Listeners
sions like bursts of laughter. This chapter reviews a rated the items either as neutral (‘neutral button’)
series of experiments conducted by our research or as emotionally meaningful (four ‘emotional but-
group to further elucidate perceptual and cognitive tons’). Thus, the probability to press an emotional
processing of affective vocal interjections. The fol- and not the neutral button amounts to 50%. Re-
lowing questions were addressed: (a) Can the sponses were coded in terms of the frequency (in
affective prosodic and the propositional compo- percent) of selecting any one of the emotional cat-
nents of interjections be separated and, thereby, egories (percentage of emotional judgments ¼
classified based on lexical emotional meaning? ‘emotional recognition rate’). Based on this
(lexical meaning of verbal emotional interjections), approach, the various vocal interjections spoken
(b) In how far depends the recognition of emo- at a neutral tone were found to differ in lexical
tional meaning upon the lexical ‘load’ of these emotional meaning (Fig. 1, Table 2), and the stimuli
297

Table 1. Items used in the study

Intonation Stimuli

Congruent Affective incongruent Neutrally spoken

Happy heissa [‘ha ısa], hurra [hu0 ra:], juhu a, achje, äh, auwei, bäh, e, herrje,
[ju’hu:], yippie [jıpi:], heida i, igitt, jeminee, o, oje, oweh, pfui,
[‘haıda], ja [ja:], aja [‘a:’ja:] u, uäh
Angry pha [pha], ahr [a:r], a [a:] e, ei, heida, heissa, hurra, i, juhu,
o, u, yippie
Disgusted igitt [i:gıth], pfui [pfui], i [i:], uäh a, o
[uæ:], bäh [bæ:], äh [æ:], e [e:], u
[u:h]
Sad auwei [auvaı:], oje [o’je:], oweh a, ahr, e, i, pah
[o:’ve:], achje [ax’je:], jeminee
[‘je:mıne:], herrje [her’je:], ach
[a:x], o [o:], u [u:], ei [a ı:]
Neutral a, ach, achje, äh, ahr, aja, auwei,
bäh, e, ei, heida, heissa, herrje,
hurra, i, igitt, ja, jeminee, juhu,
oje, oweh, pah, pfui, u, uäh, yippie

Three intonational conditions result in affective congruency or incongruency, and in neutrality.

could be assigned to two major categories, i.e., approach. Apart from two stimuli, the obtained
high-lexical (HLI) and low-lexical (LLI) interjec- highest percentage values achieved in all instances
tions. Thus, some interjections convey emotional the significance level considered. As expected, HLI
meaning even in the absence of relevant affective stimuli displayed an unambiguous lexical emo-
prosodic cues, i.e., exclusively based on their seg- tional meaning each. By contrast, LLI failed to
mental structure (HLI). Other items failed to show convey an unequivocal affective content. Never-
these effects, and in these instances emotional theless, some preferences in the attribution of lex-
meaning must be expected to depend primarily ical emotional meanings could be observed.
upon affective prosody (LLI). The findings of a recent behavioral study
A second step of analysis aimed at the determi- (Schröder, 2003) on interjections show some anal-
nation of the emotional category, i.e., happiness, ogies to the data of our study. In that report, the
anger, sadness, or disgust, displayed by the various term ‘transcription variability’ was introduced to
interjections. The probability to select one of the characterize the relationship between emotional ex-
four categories by chance amounts to 25%. For pression and the segmental structure of a stimulus,
each interjection, the percentage of the four differ- providing a criterion for the degree of convention-
ent responses was calculated and the category that ality. Thus, conventionality is considered to be
achieved the highest rank assumed to represent the based upon the canonical segmental structures, i.e.,
lexical emotional meaning of the respective item. If ‘lexicality’. In accordance with our investigation,
the largest percentage does not achieve the signifi- some of Schröder’s stimuli such as ‘yuck’ (in Ger-
cance level, the respective emotion category cannot man ‘igitt’) or an affective intonated vowel (‘o’)
be considered the lexical emotional meaning of show high (in the case of ‘igitt’) or low (in the case
that stimulus. Nevertheless, these data also had to of ‘o’) listening recognition rate. We found that the
be considered for analysis since otherwise it would vowel ‘a’ was found most frequently associated with
not have been possible to define response correct- anger (44%). However, this observation does not
ness in the following experiments. Table 2 and reach significance (t(15) ¼ 1.69, p ¼ 0.118). There-
Fig. 1 summarize the emotional meanings of fore, this item might allow for a broader range
the various interjections as determined by this of emotional interpretations such as happiness,
298

Fig. 1. Emotional recognition rates for neutrally spoken interjections in response to the task question ‘Which emotion is expressed by
the utterance?’. The figure illustrates how frequent an utterance was classified as an item with emotional lexical meaning. Gray symbols
were not significantly identified as emotional. Black symbols were significantly identified as emotional. Percentage values in brackets
indicate the frequency of choosing an emotional category.

disgust, and sadness. By contrast, vowel ‘u’ is as- recognition of vocally expressed emotions. Table 1
sociated with disgust and sadness to the same extent provides a list of interjections spoken with congru-
(50%). As a consequence, this stimulus is compat- ent prosodic modulation, e.g., the utterance ‘hurra’
ible with both emotional expressions. produced at a happy tone. In these instances, pro-
In summary, these studies indicate that at least a sodic and lexical emotional meanings converge.
subset of interjections conveys emotional meaning During auditory application of these items, partic-
both by lexical as well as by prosodic factors ipants had to answer the question ‘Which emotion
(HLI). Other utterances were found to predomi- was expressed?’ in the absence of any further in-
nantly rely on affective prosody as a means for the structions. Thus, listeners could base their judg-
display of emotional meaning (LLI) and, thus, ments either on prosodic or on lexical or on both
lack a significant contribution of lexical informa- types of cues. Again, participants had to assign one
tion in this regard. These data argue in favor of a of the four emotion categories considered, i.e., hap-
new classification scheme of interjections based on piness, anger, sadness, or disgust, to the verbal ut-
their lexical status. It could be documented, fur- terances. Responses were coded in terms of the
thermore, that a separation of affective prosodic frequency of correct ratings of the transmitted
and lexical dimensions is even possible in natural emotion (proportion of correct to incorrect answers
speech materials. Thus, verbal stimuli must not ¼ recognition rate in percent). HLI items yielded
necessarily be manipulated, e.g., by means of low- significantly higher recognition rates than their LLI
pass filtering, in order to remove lexical compo- cognates (Fig. 2; t(40) ¼ 3.81, po0.001).
nents of emotional meaning. Recognition rates were found to decline in the
absence of either one of the two speech compo-
nents ‘prosody’ or ‘lexicality’. The comparison of
Influence of lexical semantics on recognition rates HLI and LLI stimuli spoken with congruent affec-
tive prosody revealed better recognition rates for
A further experiment tried to assess the influence the items of high-lexical emotional meaning. In
of lexical meaning (HLI vs. LLI stimuli) on the these cases, both ‘channels’ of information are
Table 2. Recognition rates and identification of special emotional categories for neutrally spoken interjections in response to the task question ‘Which emotion is
expressed by the utterance?’

Stimulus Emotional identification Identification of special emotion

Elation Hot anger Disgust Sadness

t P P (%) SEM t P P (%) SEM t P P (%) SEM t P P (%) SEM t P P (%) SEM

Igitt 100 100


Pfui 100 100
I 100 100
Auwei 100 100
Oje 100 100
Oweh 100 100
Achje 100 100
Heissa 100 100
Hurra 100 100
Juhu 100 100
Yippie 100 100
Uäh 100 23.00 o0.001 97 3.13 7.00 o0.001 3 3.13
Bäh 100 3.00 o0.01 6 6.25 11.00 o0.001 94 6.25
Äh 15.00 o0.001 97 3.13 16.10 o0.001 94 4.27 4.39 o0.001 6 4.27
Jeminee 10.25 o0.001 94 4.27 3.00 o0.01 6 6.25 11.00 o0.001 94 6.25
Herrje 10.25 o0.001 94 4.27 1.73 0.104 13 7.22 8.66 o0.001 88 7.22
Heida 5.20 o0.001 88 7.22 100
Ach 2.09 0.054 69 8.98 2.50 o0.05 7 7.14 2.50 o0.05 7 7.14 6.26 o0.001 86 9.71
Pah 1.86 0.083 69 10.08 2.39 o0.05 58 13.69 1.26 0.23042 13.69
E 2.08 0.056 66 7.53 100
Ahr 1.17 0.261 63 10.70 6.20 o0.001 83 9.40 2.00 0.071 8 8.33 2.97 o0.05 8 5.62
O 1.00 0.333 59 9.38 0.54 0.598 19 10.65 2.25 o0.05 8 7.69 5.50 o0.001 4 3.85 3.32 o0.01 69 13.32
Ja 0.29 0.774 53 10.67 100
Aja 0.27 0.791 53 11.61 3.38 o0.01 70 13.33 0.38 0.716 30 13.33
A 0.20 0.846 52 10.53 0.74 0.476 33 11.32 1.69 0.118 44 11.53 3.03 o0.01 8 5.50 1.26 0.232 14 8.82
U 0.52 0.609 44 11.97 1.53 0.170 50 16.37 1.53 0.170 50 16.37
Ei 1.00 0.333 38 12.50 0.50 0.638 17 16.67 3.50 o0.05 83 16.67

Values of percentage as well as statistical parameters are given.


Abbreviations: t ¼ t-values of a one sample t-test, p ¼ significance level, P ¼ recognition rate, SEM ¼ error deviation.

299
300

Fig. 3. Recognition rates for interjections with congruent


Fig. 2. Recognition rates for interjections with congruent pros- (match) versus incongruent (mismatch) prosody. Values of per-
ody in response to the question task ‘Which emotion is ex- centage and error deviation are given for both lexical categories.
pressed by the speaker?’. Values of percentage and error
deviation are given for both lexical categories.
an adequate response. Recognition performance
available to the listeners. By contrast, evaluation was coded in terms of the frequency of correct rat-
of LLI utterances rather exclusively must rely on ings of the transmitted emotion (proportion of cor-
affective prosody. Thus, recognition of the emo- rect to incorrect answers ¼ recognition rate in
tional meaning of interjections appears to depend percent). As concerns affectively spoken interjec-
upon the range of available cues. In line with these tions with congruent emotional tone, a significantly
data, a variety of studies of animal communicative higher percentage of correct judgments emerged as
behavior suggest the increase of the amount of compared to affectively spoken stimuli with incon-
available information to enhance the probability gruent prosody (Fig. 3; for HLI: t(36) ¼ 4.63,
that a receiver makes a correct decision (Bradbury po0.001; for LLI: t(36) ¼ 3.64, po0.001).
and Vehrencamp, 1998). Comparison of affectively spoken stimuli with
congruent and incongruent prosody revealed bet-
ter recognition rates in the former case. HLI items
Interaction of prosodic and lexical information showed a stronger effect than LLI, most presum-
ably due to reduced lexicality of LLI stimuli. It can
In order to assess whether the processing of pro- be expected, therefore, that the processing of pros-
sodic and lexical information is bound to different ody and lexicality cues within the domain of emo-
mechanisms, stimuli spoken with congruent and tional interjections be bound to separate cognitive
incongruent prosody were compared to each other. mechanisms. However, since incongruencies be-
Given an impact of prosody on lexicality or vice tween the prosody and the lexicality of interjec-
versa, recognition rates must be expected to differ tions yield lower recognition rates, subsequent
between these two conditions. Table 1 summarizes higher order cognitive processes appear to match
the stimuli spoken with incongruent prosodic mod- both sources of information.
ulation. In this study, participants had to respond A further step of analysis addressed the question
to the question ‘Which emotion was expressed?’ in of whether the listeners’ judgments predominantly
the absence of any further instructions. As in the relied on prosodic or lexical information. Based on
preceding study, listeners, thus, could base their the sample of interjections with incongruent pros-
judgments either on prosodic or on lexical or on ody, lexical and prosodic decisions were separately
both types of cues. Again, participants had to as- analyzed and responses coded in terms of the fre-
sign one of the four emotion categories considered, quency of correct ratings of the transmitted emotion
i.e., happiness, anger, sadness, or disgust, to the (proportion of correct prosodic or lexical decisions
verbal utterances. In case of stimuli with incongru- to incorrect answers ¼ recognition rate in percent).
ent prosody, correct identification of either the lex- The recognition rates for HLI and LLI items asso-
ical or the prosodic emotional meaning counted as ciated with prosodic and lexical decisions were
301

analyzed in order to determine which of the two


processes was preferred. Prosodic decisions yielded
significantly higher recognition rates than the lexical
ones (Fig. 4; for HLI: t(36) ¼ 7.35, po0.001; for
LLI: t(36) ¼ 27.13, po0.001). As a consequence,
HLI items were more often correctly identified by
lexical decisions than their LLI counterparts (Fig. 4;
t(36) ¼ 6,42, po0.001). By contrast, the reversed
pattern was found in association with prosodic de-
cisions (Fig. 4; t(36) ¼ 4.42, po0.001).
Taken together, prosodic cues yielded more of- Fig. 4. Recognition rates for interjections with incongruent
ten correct judgments than lexical information. It prosody. Decisions guided by prosodic or lexical information
must be expected, therefore, that prosody plays a are illustrated separately. Values of percentage and error devi-
ation are given for both lexical categories.
dominant role during recognition of emotional
meaning. Other studies also reported that affective
prosodic meaning usually takes precedence if the utterances into different categories provided evi-
linguistic message is at odds with emotional into- dence for separate processing of prosodic and lexical
nation (Bolinger, 1972; Ackerman, 1983; Ross, elements during speech communication. Neverthe-
1997). In case of incongruently spoken utterances, less, both dimensions combined allow for a more
listeners have to focus either on the prosody or on reliable encoding of the emotional meaning of in-
the lexicality of verbal utterances. Conceivably, terjections than prosody on its own. Comparing
affective prosodic cues are easier to process and, both sources of information, prosody seems to have
therefore, more salient during speech perception. a more salient function in the transmission of emo-
Nevertheless, lexical content cannot be totally ig- tions. Taken together, these psychoacoustic data
nored even under these conditions since LLI suggest lexical and prosodic components to sepa-
showed higher recognition rates than HLI in as- rately contribute to the formulation of emotional
sociation with prosodic decisions. Since lexicality semantic information. In order to correctly perceive
seems to have a specific impact on prosody-based the affective message of verbal utterances, both
judgments, any mismatches must be expected to speech elements, however, have to be matched.
depend upon the lexical rather than the prosodic
informational component. Lexical emotional
Abbreviations
meaning is strongly associated with the segmental
structure of an utterance, a dimension difficult to
HLI high-lexical interjections
ignore. By contrast, prosody is bound to the sup-
LLI low-lexical interjections
rasegmental level of speech that can easily be rec-
ognized but does not as strictly dependent on the
segmental structure of verbal utterances. Thus, a Acknowledgments
more versatile use of prosodic patterns during
speech communication is quite conceivable. This study has been supported by the Max Planck
Institute for Human Cognitive and Brain Sciences,
Conclusions Leipzig, and the Department of General Neurol-
ogy, Hertie Institute for Clinical Brain Research,
Both affective prosody and lexical meaning partic- University of Tübingen, Germany.
ipate in the communication of a speaker’s emotional
states. The data reviewed here suggest that these two Appendix
dimensions of the emotional meaning of interjec-
tions are mediated by different cognitive mecha- Participants: All participants were native right-
nisms. Among others, the classification of these handed (Edinburgh Inventory; Oldfield, 1971)
302

German speakers (mean age ¼ 25.0 years, could make a short break upon their own request.
SD ¼ 3.72, range 18–45 years). None of them The three studies lasted about 45 min each.
had a history of any hearing disorders or neuro- Data analysis: One-sample t-tests were calcu-
logical diseases. They were all naive with regard to lated in order to determine whether the obtained
the purposes of the experiments. Sixteen volun- values differed significantly from chance level.
teers (8 females) participated in the first and 41 Furthermore, t-tests for dependent samples were
individuals (21 females) in the second experiment. performed to determine whether values of two
The third study encompassed 37 individuals (18 groups differed significantly from each other.
females). Participants were paid for participation,
and all of them had provided informed consent.
Recordings: Recordings were performed in a References
sound proven booth with a microphone (type Sen-
nheiser) positioned at a distance of 0.8 m to the Ackerman, B. (1983) Form and function in children’s under-
speaker’s mouth. Stimuli were recorded on a digital standing of ironic utterances. J. Exp. Child Psychol., 35:
DAT audiotape and digitized at 44.1 kHz with a 16- 487–508.
Anderson, J.M., Gilmore, R., Roper, S., Crosson, B., Bauer,
bit sampling rate. A sound editor (Cool Edit version R.M., Nadeau, S., Beversdorf, D.Q., Cibula, J., Rogish III,
2000, Syntrillium Software) was used for further M., Kortencamp, S., Hughes, J.D., Gonzalez Rothi, L.J. and
processing of the stimuli. In order to facilitate pro- Heilman, K.M. (1999) Conduction aphasia and the arcuate
duction of distinct affective tones, suitable frame fasciculus: a reexamination of the Wernicke-Gschwind
model. Brain Lang., 70: 1–12.
stories were presented to the speakers: (a) the
Banse, R. and Scherer, K.R. (1996) Acoustic profiles in vocal
speaker has just won a lot of money in a horse emotion expression. J. Pers. Soc. Psychol., 70: 614–636.
race (happiness); (b) the speaker is furious about a Bolinger, D. (1972) Intonation. Harmondsworth, UK, Penguin.
person he/she dislikes because this individual has Bradbury, J.W. and Vehrencamp, S.L. (1998). In: Principles of
broken once again a rule of proper conduct and Animal Communication. Sinauer Associates, Sunderland,
shows no sign of regret (anger); (c) the speaker re- MA, pp. 387–418.
Friederici, A.D. (2002) Towards a neural basis of auditory
gards moving maggots in a wound (disgust); (d) the sentence processing. Trends Cogn. Sci., 6: 78–84.
speaker has just heard about the death of a close MacNeilage, P.F. (1998) The frame/content theory of evolution
relative (sadness). During the recordings, speakers of speech production. Behav. Brain Sci., 21: 499–546.
silently read the frame story for a given emotional Nübling, D. (2004) Die prototypische interjektion: ein defini-
category and then produced the respective stimuli. tionsvorschlag. Z. Semiotik, 26(Band 1-2): 11–45.
Oldfield, R. (1971) The assessmentand analysis of handedness:
Procedure: Participants were seated in a sound- the Edinburgh inventory. Neuropsychologia, 9: 97–113.
proof booth in front of a computer screen with Quigg, M. and Fontain, N.B. (1999) Conduction aphasia elic-
loudspeakers placed to its left and right side, in- ited by stimulation of the left posterior superior temporal
structions about the experimental procedure and the gyrus. J. Neurol. Neurosurg. Psychatry, 66: 393–396.
Ross, E.D. (1997) Cortical representation of the emotions. In:
task to be performed being visually presented prior
Trimble, M.R. and Cummings, J.L. (Eds.), Contemporary
to each session. Participants had to listen to the in- Behavioral Neurology. Butterworth-Heinemann, Boston, pp.
terjections applied via loudspeakers and to identify 107–126.
the respective emotional category. After presentation Scherer, K.R. and Oshinsky, J.S. (1977) Cue utilization in emotion
of each item, the question to be answered appeared attribution from auditory stimuli. Motiv. Emot., 1: 331–346.
on the screen. Participants then had to respond as Schröder, M. (2003) Experimental study of affect bursts. Speech
Commun., 40: 99–116.
quickly and accurately as possible by pressing a but- Sidtis, J.J. and Van Lancker Sidtis, D. (2003) A neurobehavioral
ton on a response box. The time interval for the approach to dysprosody. Semin. Speech Lang., 24(2): 93–105.
reaction amounted to 5 s. Stimuli were presented in Wartburton, E., Wise, R.J., Price, C.J., Weiller, C., Hadar, U.,
pseudorandomized order, but balanced over the var- Ramsay, S. and Frackowiack, R.S. (1996) Noun and verb
retrieval by normal subjects: studies with PET. Brain, 119:
ious blocks with 50 items per block. Each stimulus
159–179.
was presented twice. Prior to the experiments, a Wise, R.J.S., Scott, S.K., Blank, S.C., Mummery, C.J., Mu-
short practice session including 20 stimuli was per- rphy, K. and Wartburton, E.A. (2001) Separate neural sub-
formed. Between successive blocks, participants systems within ‘Wernicke’s area’. Brain, 124: 83–95.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 17

Judging emotion and attitudes from prosody


following brain damage

Marc D. Pell

School of Communication Sciences and Disorders, McGill University,


1266 Ave. des Pins Ouest, Montréal, QC, H3G 1A8, Canada

Abstract: Research has long indicated a role for the right hemisphere in the decoding of basic emotions
from speech prosody, although there are few data on how the right hemisphere is implicated in processes
for understanding the emotive ‘‘attitudes’’ of a speaker from prosody. We describe recent clinical studies
that compared how well listeners with and without focal right hemisphere damage (RHD) understand
speaker attitudes such as ‘‘confidence’’ or ‘‘politeness,’’ which are signaled in large part by prosodic features
of an utterance. We found that RHD listeners as a group were abnormally sensitive to both the expressed
confidence and expressed politeness of speakers, and that these difficulties often correlated with impair-
ments for understanding basic emotions from prosody in many RHD individuals. Our data emphasize a
central role for the right hemisphere in the ability to appreciate emotions and speaker attitudes from
prosody, although the precise source of these social-pragmatic deficits may arise in different ways in the
context of right hemisphere compromise.

Keywords: interpersonal behavior; emotion; attitudes; prosody; right hemisphere; brain-damaged;


communication disorders; pragmatic language processing

Isolating prosodic functions in the brain ness, and temporal patterning), which dynamically
interact in the speech signal to convey intended
Researchers interested in the neurocognition of meanings to the hearer over various time intervals
prosody face a number of distinct challenges when (words, phrases) (Pell, 2001). Moreover, opera-
studying this communication channel. Speech tions for decoding prosodic distinctions in speech
prosody serves an array of functions that impart cooccur and are frequently interdependent with
meanings as a formal part of language, and si- those for processing finer segmental distinctions in
multaneously, which jointly refer to the emotions the verbal–semantic channel of language. These
and/or interpersonal stance (‘‘attitudes’’) of a variables, which emphasize the unique, multifac-
speaker within the communicative event (Sidtis eted status of prosody in the spectrum of human
and Van Lancker Sidtis, 2003; Grandjean et al., communication systems, have long-defied re-
this volume). The needs for linguistic and emotive searchers’ attempts to ‘‘isolate’’ effects due to
expression through prosody are achieved by ex- prosody from other sources of information in
ploiting a minimal and overlapping set of prosodic speech, although significant gains are now being
elements in speech (e.g., changes in pitch, loud- made (for reviews see Baum and Pell, 1999; Sidtis
and Van Lancker Sidtis, 2003).
Corresponding author. Tel.: +1-514–398-4133; Fax: +1–514- For instance, there is growing credence for the
398-8123; E-mail: marc.pell@mcgill.ca idea that understanding speech prosody engages

DOI: 10.1016/S0079-6123(06)56017-0 303


304

broadly distributed and bilateral networks in the recognition of emotional prosody in 9 right-hem-
brain (Mitchell et al., 2003; Pell and Leonard, isphere-damaged (RHD), 11 left-hemisphere-dam-
2003; Gandour et al., 2004). Asymmetries in net- aged (LHD), and 12 healthy control (HC) listeners
work functioning, when detected at specific stages (Pell, 2006). Each participant was required to dis-
of prosodic processing, are likely to reflect differ- criminate, identify, and rate expressions of five
ential sensitivity of the two hemispheres to be- basic emotions based on the prosodic features of
havioral, stimulus, and/or task-related variables ‘‘pseudo-utterances,’’ which contained no emo-
(Wildgruber et al., 2002; Kotz et al., 2003; Tong tionally relevant language cues (e.g., Someone mi-
et al., 2005; Pell, 2006). For example, it is increas- gged the pazing spoken in a ‘‘happy’’ or ‘‘sad’’
ingly apparent that processing prosody as a local tone). Participants were also required to identify
cue to linguistic–semantic structure, such as re- emotions from utterances with semantically bias-
trieving word meanings defined by tonal contrasts ing language content (e.g., I didn’t make the team
or changes in syllabic stress, favors mechanisms of spoken in a congruent ‘‘sad’’ tone). After compar-
the left cerebral hemisphere (Pell, 1998; Gandour ing the group patterns and inspecting individual
et al., 2004; Wildgruber et al., 2004; Tong et al., performance characteristics across our prosody
2005). This literature implies that when prosody tasks, we found that both RHD and LHD patients
acts as a linguistic device with conventionalized exhibited impairments in our ‘‘pure prosody’’
expression within the language system, prosodic tasks, which provided no semantic cues for iden-
attributes are treated like other acquired linguistic tifying the five emotions (Cancelliere and Kertesz,
elements that engage left hemisphere processing 1990; Starkstein et al., 1994; Ross et al., 1997; Pell,
mechanisms in a preferential manner (Hsieh et al., 1998). For individuals in the RHD group, these
2001). difficulties appeared to reflect a relatively pervasive
In contrast, the processing of less iconic pro- insensitivity to the emotional features of prosodic
sodic representations, which encode vocal expres- stimuli, whereas for individuals in the LHD group,
sions of emotion states such as ‘‘anger’’ or their problems appeared to stem to a greater extent
‘‘sadness’’ in speech, shows a distinct right-sided from difficulties interpreting prosody in the con-
bias in many lesion and neuroimaging studies text of concurrent language cues, especially when
(Pell, 1998; Wildgruber et al., 2002; Gandour semantically biasing utterances were presented.
et al., 2003; Pell, 2006). This relative right hemi- These findings served to reiterate that while un-
sphere dominance for decoding emotional prosody derstanding emotional prosody engages regions of
has been hypothetically traced to early stages for both hemispheres, the right hemisphere is proba-
structuring acoustic-perceptual features of the au- bly critical for retrieving the emotional details rep-
ditory input into an internal prosodic representa- resented by prosodic cues in speech prior to
tion over extended time frames (Poeppel, 2003) integrating this information with the meanings of
and/or to adjacent stages for evaluating the affec- language (Friederici and Alter, 2004; Pell, 2006).
tive significance of the event (Wildgruber et al., Thus, despite continued developments in the lit-
2004). However, it must be emphasized that sig- erature on prosody, there is repeated evidence of
nificant interhemispheric interactions are expected the left hemisphere’s superiority for processing
when processing emotional prosody in conjunc- prosody as language and of the right hemisphere’s
tion with language content (Plante et al., 2002; preferential involvement for processing prosody in
Vingerhoets et al., 2003; Schirmer et al., 2004; less-structured, emotive contexts. The prevalence
Hesling et al., 2005). Accordingly, the existence of of these findings justifies the opinion that hemi-
comparative mechanisms for resolving the emotive spheric sensitivities for prosody are partly directed
significance of prosody in reference to concurrent by the functional significance of prosodic cues in
semantic features of language has been highlighted speech (Baum and Pell, 1999; Gandour et al.,
in the latest research (Kotz et al., 2003; Pell, 2006). 2004). This functional hypothesis follows a model
In one of our recent studies, we looked at the proposed by Van Lancker (1980) who underscored
effects of task and stimulus parameters on the the operation of pitch in prosodic communication,
305

and her idea remains a viable tool for grossly what they are saying, if correctly interpreted by the
differentiating the respective roles of the two hem- hearer, will shape further aspects of the discourse
ispheres in the processing of speech prosody. event in distinctive ways (i.e., the hearer may in-
itiate actions based on the perceived certainty of
the information received). Alternatively, using
On the ‘‘attitudinal’’ functions of prosody prosody to determine levels of speaker self-identi-
fication or self-assertiveness toward the listener is
However, as was further characterized by Van centrally connected to interpreting how polite the
Lancker’s (1980) functional laterality continuum, speaker intends to be, a form of evaluation that
prosody fulfills a broader array of functions than is has major consequences on how discourse events
currently being investigated. Of principal interest, unfold.
prosody is known to assume a key role in commu- According to certain pragmatic descriptions of
nicating the interpersonal stance of the speaker in emotive communication, prosody and other cues
a variety of contexts, where these cues are rou- operate as various ‘‘emotive devices’’ in speech,
tinely understood by listeners as ‘‘speaker atti- which encode information along such dimensions
tudes’’ (Uldall, 1960; Brown et al., 1974; Ladd as evaluation (positive/negative), proximity (near/
et al., 1985). Accepting the view that functional far), and quantity (more/less), among others. These
properties of speech prosody govern patterns of devices allow humans to communicate a range of
laterality in a critical (although perhaps incomplete) emotive meanings and attitudes that correspond in
manner, an important question is: what neurocog- many ways to traditional psychological concepts
nitive mechanisms are engaged when prosody of ‘‘evaluation,’’ ‘‘potency,’’ and ‘‘activity’’ (Caffi
serves to mark the interpersonal or attitudinal and Janney, 1994). Other pragmatic frameworks
meanings of the speaker as its core function in emphasize the manner by which communicative
speech? For example, how are the cerebral hemi- strategies, such as changes in prosody, act to ‘‘at-
spheres specialized for structuring an internal rep- tenuate’’ or ‘‘boost’’ the illocutionary force of
resentation of prosody when it is used as a strategy speech acts that may be inherently positive or
for communicating the likely veracity of a state- negative in their impact on the listener, thereby
ment being uttered, the extent to which a speaker communicating the degree of belief, commitment,
wishes to affiliate with a particular idea or individ- or strength of feeling of the speaker’s intentions
ual being discussed, or the speaker’s intended po- (Holmes, 1984). These different concepts can be
liteness toward the hearer when making a request? used as a theoretical basis for investigating how
Empirical undertakings have largely failed to elu- ‘‘prosodic attitudes’’ are encoded and understood
cidate how the cerebral hemispheres respond to from speech in adults with and without unilateral
prosodic events that serve these interpersonal or brain damage.
attitudinal functions in speech. Given the evidence that right hemisphere re-
The ability to infer attitudes held by a speaker gions are highly sensitive to vocal cues of emotion
and to integrate these details with other sources of and other meanings of prosody not formally struc-
information during interpersonal events represents tured in language, one might reasonably predict
a vital area of pragmatic competence, which guides that the right hemisphere is centrally implicated
successful communication (Pakosz, 1983). From a when drawing conclusions about the emotive
social–pragmatic viewpoint, the attitudinal func- significance of prosodic attitudes. However, as
tions of prosody should be considered ‘‘emotive’’ noted earlier, this supposition is founded on little
(as opposed to emotional or linguistic) in nature empirical research and there are other variables to
because they encode various relational meanings, consider when studying the attitudinal func-
which are socially relevant to the speaker–hearer tions of prosody, which may dictate how cerebral
in the interpersonal context in which they appear mechanisms respond to events in this more het-
(Caffi and Janney, 1994). For example, prosodic erogeneous ‘‘functional category.’’ Whereas ‘‘lin-
cues that signal that a speaker is very confident in guistic’’ prosody is formally represented within the
306

language system and ‘‘emotional’’ prosody can at ‘‘truth value’’ to the listener, promote inferences
times be divorced from the propositional message about the level of speaker confidence in the in-
while retaining its meaning (e.g., Pell, 2005), the formation conveyed (Caffi and Janney, 1994).
ability to interpret speaker attitudes from prosody Prosodic features play an instrumental if not dom-
is usually informed by comparative relations in the inant role in how attributions of speaker confi-
significance of prosody, concurrent language fea- dence are made in the auditory modality (Brennan
tures, and extralinguistic parameters (e.g., situa- and Williams, 1995; Blanc and Dominey, 2003).
tional cues and/or existing knowledge of the Studies of young, healthy listeners indicate that
listener). In fact, many speaker attitudes are characteristic alterations in loudness, pitch (rising
achieved when speakers violate highly convention- or falling intonation contour), and the temporal
alized, cross-channel associations in the use of patterning of speech (e.g., pauses, speaking rate)
prosody and linguistic strategies, which acquire are all important for judging the degree of a
emotive meanings through association over time speaker’s confidence in an assertion being made; a
(Burgoon, 1993; Wichmann, 2002). high level of speaker confidence is identified
Thus, one must bear in mind that the ability to through increased loudness of voice, rapid rate of
understand speaker attitudes from prosody is speech, short and infrequent pauses, and a termi-
tightly intertwined with functional properties of nal fall in the intonation contour (Scherer et al.,
language (such as speech acts). As well, the pro- 1973; Kimble and Seidel, 1991; Barr, 2003). In
sodic cues that mark particular attitudes tend to contrast, low speaker confidence (i.e., doubt) cor-
coincide with specific linguistic strategies or de- responds with longer pre-speech delay, (filled)
vices that serve to elaborate the speaker’s inten- pauses, and a higher probability of rising intona-
tions. We attempted to control for these factors in tion or raised pitch (Smith and Clark, 1993;
a set of new patient studies that focused on the Brennan and Williams, 1995; Boltz, 2005). In ad-
possible contributions of the right hemisphere for dition to prosody, verbal ‘‘hedges’’ (e.g., I thinky,
processing two emotive attitudes signaled by pros- Probablyy, I’m surey) play a role in judging
ody: speaker confidence and speaker politeness. speaker confidence by attenuating or boosting the
Specifically, we sought to determine whether the perceived truth value of the utterance content
same RHD individuals who displayed impair- (Holmes, 1984; Caffi and Janney, 1994), although
ments in our experiment on emotional prosody it is unclear how much weight these verbal cues
(Pell, 2006) would also present with difficulties impose on listeners relative to prosody.
recognizing speaker attitudes under comparable Our first study of prosodic attitudes sought to
testing conditions, given the functional adjacency delineate whether RHD patients, many of whom
of ‘‘emotive’’ and ‘‘emotional’’ uses of prosody demonstrated impairments for recognizing emo-
(Van Lancker, 1980). In each of our studies of tional prosody (Pell, 2006), exhibit concurrent
speaker attitudes, we manipulated the emotive difficulties for recognizing speaker confidence in
value of prosody as well as linguistic cues to speech, especially prosodic markers of this atti-
achieve a fine-grained analysis of whether RHD tude. Each of our participants was asked to rate
listeners are insensitive to speaker attitudes on the the perceived level of speaker confidence when lis-
basis of a misuse of prosodic information, linguis- tening to simple assertions indicating the probable
tic cues, or both. location of an object. Given the possible role of
both prosody and verbal strategies for conveying
confidence, and the possibility that some RHD
Understanding prosody as a cue to speaker patients attend largely to verbal content when
confidence making complex interpretations based on redun-
dant cues (Bowers et al., 1987; Brownell et al.,
Verbal and prosodic elements that convey the rel- 1992), we evaluated our subjects in separate con-
ative commitment of a speaker to the propositio- ditions where utterances contained both prosodic
nal content of their utterances, or its probable and lexical choices for understanding speaker
307

confidence or prosodic cues alone. In the event Table 1, together with data on how the groups
that RHD patients display impairments for judg- performed in our study of emotional prosody.
ing speaker confidence as predicted, this design Each participant listened to a series of sentences
should inform whether failures to understand (6–11 syllables in length), which were constructed
speaker attitudes such as confidence are linked to to fit two distinct conditions for inferring speaker
prosodic impairments or to a more general misuse confidence. In the ‘‘linguistic’’ condition, stimuli
of speech-related cues for inferring speaker atti- were semantically informative statements (e.g.,
tudes in a normal manner. You turn left at the lights) that began with linguis-
tic phrases such as for sure, most likely, or perhaps
Study 1: evaluating speaker confidence following to convey a relatively high, moderate, or low de-
RHD gree of speaker confidence through linguistic and
prosodic cues of these utterances. In the ‘‘prosody’’
As reported in full by Pell (under review), we condition, comparable pseudo-utterances were
tested two groups of right-handed, English-speak- constructed to resemble the stimuli entered into
ing participants for whom we already had detailed the linguistic condition (e.g., You turn left at the
information on their ability to discriminate, cate- lights — You rint mig at the flugs). Each pseudo-
gorize, and rate basic emotions from prosody (Pell, utterance was produced to communicate a high,
2006). Our patient group consisted of nine right- moderate, and low degree of confidence by ma-
hemisphere-damaged (RHD) participants (four nipulating only prosodic features of the utterances.
males, five females, mean age ¼ 64.2 years) who Basic acoustic analyses of the utterances entered
had each suffered a single, thromboembolic event into each task were undertaken and these differ-
with anatomical lesion of the right cerebral hem- ences are summarized in Table 2. All stimuli were
isphere (see Pell, 2006 for details). All RHD pa- digitally recorded by four male speakers of English
tients were tested during the chronic stage of their and subjected to thorough pilot testing to establish
stroke (post-onset, range ¼ 2.0–11.9 years). Our the perceptual validity of each token prior to en-
control group consisted of eleven healthy control tering these materials in the patient study. Inde-
(HC) participants without neurological damage pendently for the ‘‘linguistic’’ and ‘‘prosody’’
(five males, six females, mean age ¼ 63.4 years), tasks, each participant was instructed to listen to
all but one of whom took part in our study of each sentence and then rate the degree of confi-
emotional prosody. All RHD and HC participants dence expressed by the speaker on a five-point
displayed good hearing and completed a battery of continuous scale, where ‘‘1’’ signifies the speaker
neuropsychological tests, which are summarized in was ‘‘not at all confident’’ and ‘‘5’’ signifies that

Table 1. Neuropsychological features of the healthy control (HC) and right-hemisphere-damaged (RHD) groups, including data on
the ability to identify basic emotions from prosody (Pell, 2006) (mean 7 standard deviation converted to percent correct)

Measure HC RHD

(n ¼ 11) (n ¼ 9)

Discriminating emotional prosody (/30) 79.079.6 68.5711.8


Identifying emotion prosody from pseudo-utterances (‘‘pure prosody’’ task, /40) 68.8724.0 43.9727.1
Identifying emotional prosody with congruent semantic cues (‘‘prosody-semantic’’ task, /40) 79.5724.1 65.0725.9
Identifying emotion from faces (/40) 86.379.1 68.6727.3
Identifying emotion from verbal scenarios (/10) 86.0712.0 64.0725.5
Mini-mental state exam (/30) — 90.777.2
Benton phoneme discrimination (/30) 92.176.7 82.6710.1
Benton face discrimination (/54) 87.277.1 77.876.5
Verbal working memory — words recalled (/42)a — 51.4711.1
a
Four of the RHD patients (R3, R7, R8, R9) did not complete this task.
308

Table 2. Acoustic measures of stimuli entered into the confidence study by task, according to levels of expressed confidence

Confidence level Mean f0 (Hz) f0 Range (Hz) Speech rate (seconds/syllable)

Linguistic condition High 110 58 0.19


Moderate 130 86 0.23
Low 133 87 0.27
Prosody condition High 116 62 0.23
Moderate 127 56 0.24
Low 136 103 0.33

the speaker was ‘‘very confident.’’ The numerical pants (with frequent ‘‘4’’ and ‘‘5’’ responses), but
ratings assigned were analyzed between groups and not to members of the patient group, yielding a
across experimental conditions. highly divergent pattern in this context. Nonethe-
When the mean ratings were examined for each less, the RHD patients displayed a certain capacity
group, it was shown that both the healthy and the to identify ‘‘high confidence’’ exemplars when lex-
RHD participants differentiated high, moderate, ical cues were present (see Fig. 1a).
and low confidence utterances by assigning pro-
gressively lower ratings to items within these cat-
egories for both the ‘‘linguistic’’ (HC ¼ 4.57, 3.03, Implications of the results on speaker confidence
1.68; RHD ¼ 4.01, 2.92, 2.40) and the ‘‘prosody’’
(HC ¼ 4.37, 3.01, 1.61; RHD ¼ 3.47, 2.47, 1.90) In general, we found that there was little difference
tasks. However, the ability of the RHD group to in how healthy participants rated speaker confi-
detect graded differences among the three confi- dence according to whether utterances contained
dence levels was noticeably reduced in range based lexical phrases (‘‘hedges’’) in addition to repre-
on the mean conditional ratings when compared to sentative prosodic features to confidence. In con-
the HC group, especially in the prosody task. trast, the RHD listeners appeared to rely more
When we examined the frequency distribution of strongly on lexical markers for inferring the extent
ratings assigned at each interval of the rating scale, of speaker confidence, with reduced capacity to
we found that members of the RHD and HC use prosodic cues alone to infer the extent of
groups responded in a distinct manner to speaker speaker confidence. For example, the RHD pa-
confidence cues, implying that the RHD patients tients tended to assign significantly higher ratings
were less sensitive to meaningful distinctions in the of speaker confidence in the linguistic vs. the pros-
degree of speaker confidence encoded by stimuli in ody condition, and exhibited a selective tendency
both our ‘‘linguistic’’ and ‘‘prosody’’ conditions for judging speakers as less or ‘‘not at all’’ con-
(see Figs. 1a, b). Closer inspection of these data fident on the basis of exposure to prosodic cues
revealed that the ability to interpret speaker con- alone when the two conditions were compared.
fidence from prosodic cues alone was especially These results argue that lexical hedges such as ‘‘I’m
problematic for the RHD group; as illustrated in sure’’ or ‘‘probably’’ assumed very little weight
Fig. 1b, there were marked qualitative differences when making attributions about speaker confi-
between the groups, which suggested that the dence for healthy listeners, but rather, that pro-
RHD patients frequently did not detect overt pro- sodic features were the decisive factor for rating
sodic indicators of ‘‘speaking with confidence,’’ speaker confidence in both conditions (Kimble and
such as reductions in pitch and a relatively fast Seidel, 1991; Brennan and Williams, 1995; Barr,
speech rate (Scherer et al., 1973; Kimble and Sei- 2003). Accordingly, it is likely that the observed
del, 1991; Pell, under review). This claim seems impairments for recognizing speaker confidence
particularly true when RHD patients rated ‘‘high’’ by our RHD patients revolved significantly
confidence utterances from prosody alone where around a failure to analyze prosody in both of
these cues were highly salient to healthy partici- our experimental conditions.
309

Fig. 1. The frequency of responses assigned at each interval of the five-point confidence scale by participants with right hemisphere
damage (RHD) and healthy controls (HC) in the linguistic and prosody conditions (frequency counts were converted to proportions).

In addition to demonstrating that RHD pa- fident attitudes following damage to the right
tients are less sensitive to prosodic attributes of hemisphere.
speaker confidence, our initial experiment revealed One limitation of our study of speaker confi-
that RHD individuals tend to be more attuned to dence was that we were only able to compare con-
lexical features when judging this attitude (in spite fidence judgments based on prosody alone to the
of the fact that lexical cues are often less inform- added (i.e., congruent) effects of lexical cues, which
ative to healthy listeners when inferring speaker proved to be relatively minimal in importance for
confidence). The idea that RHD patients fre- our stimuli. This precluded an analysis of many
quently accord greater weight to the value of lin- naturally occurring conditions in which listeners
guistic/verbal cues over prosody — perhaps as a must interpret speaker attitudes when the meaning
compensatory attempt to understand attitudinal of prosody and language content sometimes
meanings, which are encoded to a large extent conflict, such as for understanding the politeness
through prosody — has been described previously of a speaker who is making a request to the lis-
when RHD patients have performed other inter- tener. This motivated a second experiment that
pretive language tasks (Tompkins and Mateer, probed how RHD patients interpret speaker po-
1985; Bowers et al., 1987; Brownell et al., 1992; liteness under various stimulus conditions in which
Pell and Baum, 1997). For our experiment on prosodic strategies interact more extensively with
speaker confidence, this abnormal tendency to fo- linguistic strategies to indicate particular interpre-
cus on the language channel (to the extent possi- tations of this attitude, furnishing additional clues
ble) may have been directed by subtle impairments about the source of difficulties for processing
for processing the prosodic underpinnings of con- speaker attitudes in our RHD patients.
310

Understanding prosody as a cue to speaker tions where this attitude is marked by divergent
politeness cues in the prosody and language channels.

Emotive attributes of speech, which reflect the self-


assertiveness of the speaker vis-à-vis the listener, Study 2: evaluating speaker politeness following
are strongly associated with the notion of ‘‘polite- RHD
ness principles,’’ or the operation of ‘‘volitionality
devices’’ (Caffi and Janney, 1994). Politeness is For this study, we recruited six RHD patients and
conveyed linguistically through word selection ten HC subjects who originally participated in our
(e.g., please) and through choices in sentence studies of speaker confidence and emotional pros-
structure; in the latter case, perceived politeness ody. Our stimuli were English sentences phrased in
tends to increase in the face of ‘‘conventional in- the form of a command or request for a simple
directness,’’ for example, when a command is action to be performed by the listener, recorded by
framed as an optional request posed to the listener two female speakers. Again, distinct items were
(Clark and Schunk, 1980; Blum-Kulka, 1987; carefully prepared to enter into a ‘‘linguistic’’ task,
Brown and Levinson, 1987). At the level of pros- for eliciting politeness judgments based on com-
ody, politeness is communicated in large part bined prosody and lexical–semantic cues, and a
through conventionalized choices in intonational ‘‘prosody’’ task, for eliciting politeness judgments
phrasing; utterances with high/rising pitch tend to based on prosodic features alone.
be perceived as more polite than those with a ter- The linguistic stimuli were eight ‘‘stem’’ com-
minal falling contour (Loveday, 1981; Wichmann, mands (e.g., Do the dishes), which were linguisti-
2002; Culpeper et al., 2003), as are utterances cally modified to change the emotive force of the
produced with decreased loudness (Trees and command in four distinct ways (see Table 3 for a
Manusov, 1998; Culpeper et al., 2003). The idea summary). ‘‘Direct’’ utterances contained initial
that prosody and utterance type necessarily com- phrases, which linguistically boosted the negative
bine to signal speaker attitudes concerning the po- intent of the command, whereas ‘‘indirect,’’ ‘‘very
liteness of requests by each modifying (i.e., indirect,’’ and ‘‘please’’ utterances employed con-
attenuating or boosting) the negative, illocutio- ventional indirectness or explicit lexical markers to
nary force of the request to the listener is well ac- attenuate the negative force of the command. Each
cepted (Trees and Manusov, 1998; Wichmann, linguistic utterance type was then produced by the
2002; LaPlante and Ambady, 2003). actors in two prosodic modes: with a high/rising
Although research has examined how RHD pa- tone that tends to attenuate the imposition of re-
tients process the meaning of direct vs. indirect quests (i.e., be interpreted as polite); and a falling
requests (Hirst et al., 1984; Weylman et al., 1989; tone that tends to boost the negativity of the re-
Stemmer et al., 1994), the specific role of prosody quest (i.e., less polite) (Loveday, 1981; Culpeper
in judging the politeness of requests following et al., 2003). As demonstrated in Table 3, some of
RHD has not been carefully explored (Foldi, these cue combinations led to stimuli in which lin-
1987). In our follow-up study, we assessed whether guistic and prosodic conventions for understand-
listeners who took part in our investigations of ing speaker politeness were pragmatically opposed
emotional prosody and speaker confidence dis- (e.g., indirect language spoken with a low pros-
played normal sensitivity to speaker politeness ody). Stimuli presented in the prosody task were
based on the complex interplay of linguistic and comparable pseudo-utterances produced by the
prosodic strategies for understanding this attitude two actors in a ‘‘rising’’ and a ‘‘falling’’ tone as was
(or based on prosodic features alone). We antic- accomplished in the linguistic condition. Again,
ipated that RHD listeners would again experience basic acoustic differences in our stimuli were
difficulties using prosodic information to recognize explored, as reported in Table 4. Each RHD and
speaker attitudes of politeness but that these im- HC participant listened to all sentences in the lin-
pairments would be most pronounced in condi- guistic and prosody conditions and indicated on a
311

Table 3. Examples of stimuli presented in the experiment on speaker politeness by condition, according to whether language and
prosodic features served to ‘‘attenuate’’ or ‘‘boost’’ the negative impact of the request on the listener

Utterance type Example Prosody Emotive impact of utterance type/prosody

Linguistic condition Stem command Do the dishes High Boost/attenuate


Low Boost/boost
Direct You must do the dishes High Boost/attenuate
Low Boost/boost
Indirect Can you do the dishes High Attenuate/attenuate
Low Attenuate/boost
Very indirect Could I bother you to do the dishes High Attenuate/attenuate
Low Attenuate/ boost
Please Please do the dishes High Attenuate/attenuate
Low Attenuate/boost
Prosody condition Pseudo-utterance Gub the mooshes High –-/attenuate
Low –/boost

Table 4. Acoustic measures of stimuli entered into the politeness study by task, according to manipulations in utterance type and
prosody

Utterance type Mean f0 (Hz) f0 range (Hz) Speech rate (Seconds/


Syllable)

Falling Rising prosody Falling Rising Falling Rising


prosody prosody prosody prosody prosody

Linguistic condition Stem command 228 277 151 192 3.87 4.27
Direct 228 268 158 184 3.79 4.68
Indirect 216 258 139 221 4.63 4.74
Very indirect 229 270 218 312 4.32 4.70
Please 218 257 165 180 2.93 3.95
Prosody condition Pseudo-utterance 226 268 175 254 3.81 4.05

five-point (1–5) scale how ‘‘polite the speaker healthy listeners (Trees and Manusov, 1998; LaPl-
sounds,’’ where ‘‘1’’ represented ‘‘not-at-all polite’’ ante and Ambady, 2003). For example, when sen-
and ‘‘5’’ represented ‘‘very polite.’’ tences with overt linguistic mitigating strategies of
On the basis of the mean politeness ratings as- indirectness (Could I bother you to do the dishes)
signed by members of each group, we found that were produced in a corresponding high (polite)
sentences spoken in a high vs. a low prosody had tone, HC listeners perceived these stimuli as the
the largest influence on politeness ratings; for both most polite form of request, whereas the same
the HC and RHD listeners, sentences spoken with sentences produced in a low (impolite) tone were
a high prosody were always judged to be more rated as no more polite than listening to stem
polite than those spoken in a low prosody, in both commands (Do the dishes). However, we found no
the linguistic (HC: high ¼ 3.54, low ¼ 1.60; RHD: comparable evidence that the mean politeness rat-
high ¼ 3.40, low ¼ 2.06) and the prosody (HC: ings assigned by the RHD group were influenced
high ¼ 4.20, low ¼ 1.73; RHD: high ¼ 3.98, by the combined effects of prosody and utterance
low ¼ 2.13) tasks. As expected, the mean ratings type. Thus, although the RHD patients seemed
assigned by the healthy listeners in the linguistic capable of using gross differences in prosody to
condition indicated that prosodic factors inter- judge politeness, the subtle interactions of prosody
acted significantly with utterance type to deter- and utterance type, which served to alter the per-
mine the perceived degree of speaker politeness for ceived politeness of speakers for healthy listeners,
312

were recognized less frequently by members of the linguistic strategies were employed to render the
RHD group. request more polite in conjunction with a discon-
Given the interplay of utterance type and pros- firming, low, and impolite prosody (top right
ody for understanding speaker politeness, we con- graph in Fig. 2). In this context, RHD patients
ducted further analyses in our linguistic condition tended to rate these utterances as more polite than
to see whether the groups differed when levels of the HC listeners.
utterance type and prosody were defined strictly by
their emotive function as a device for ‘‘attenuat-
ing’’ or ‘‘boosting’’ the negativity of requests. Pre- Implications of the results on speaker politeness
vious research involving healthy adults implies
that unconventional associations of linguistic and In general, results of the politeness experiment in-
prosodic cues, which conflict in their emotive dicate that RHD patients were largely able to in-
function, are most likely to generate a negative terpret whether highly conventionalized utterance
(impolite) evaluation of the utterance (LaPlante types (Stemmer et al., 1994) or differences in into-
and Ambady, 2003). After analyzing the frequency national phrasing were acting to attenuate or boost
distribution of ratings recorded at each interval of the negative attitude of the speaker when the op-
the five-point politeness scale, we noted that the eration of cues in each channel was considered in-
two groups differed significantly only in the two dependently. However, the RHD group recognized
conditions in which prosody and language con- fewer meaningful distinctions among our stimulus
flicted in their emotive intent (see Fig. 2). Most conditions and demonstrated little sensitivity to the
notably, the RHD patients were relatively insen- interplay of prosodic and linguistic factors, which
sitive to the intended politeness of speakers when combine to signal speaker politeness. The fact that

Fig. 2. The frequency of responses assigned at each interval of the five-point politeness scale by participants with right hemisphere
damage (RHD) and healthy controls (HC) according to the combined function of prosody vs. language cues (frequency counts were
converted to proportions).
313

RHD patients were sensitive to broad differences communicative events (Brownell et al., 1992, 1997;
in prosody when making politeness ratings (espe- Cheang and Pell, 2006).
cially when pseudo-utterances were presented) im- Finally, in our politeness experiment we uncov-
plies that these listeners were able to judge attitudes ered further indications that RHD patients as-
that revolve around conventional, categorical signed greater weight to the functional significance
choices in contour type (Scherer et al., 1984; Wich- of language over prosodic content (Tompkins and
mann, 2002). These findings in our politeness ex- Mateer, 1985; Bowers et al., 1987; Pell and Baum,
periment contrast with arguments we arrived at 1997), although for speaker politeness the appar-
when studying speaker confidence, where there was ent ‘‘language focus’’ of our RHD patients did not
strong evidence that members of our RHD group reflect an incapacity to process relevant prosodic
could not meaningfully process representative pro- distinctions to this attitude. It is possible that
sodic cues to make appropriate decisions about many of the patients’ difficulties emerged at a stage
this speaker attitude. of retrieving learned associations between utter-
What posed greatest difficulties for RHD listen- ance type and intonation, which are routinely used
ers was to render interpretations of speaker po- by listeners to infer speaker politeness, particularly
liteness when these were signaled by functional in instances when violations of these expectations
discrepancies in the emotive force of linguistic occur and generate distinct attitudinal interpreta-
vs. prosodic cues of the stimuli. When conven- tions (Wichmann, 2002). More research will be
tional linguistic strategies for attenuating the im- needed to evaluate these initial hypotheses.
position of requests (e.g., Would you mindy) were
conjoined with conventional prosodic choices,
which helped boost the impoliteness (i.e., per- Individual profiles of impairment for understanding
ceived dominance) of the speaker, RHD patients emotions versus prosodic attitudes
were less sensitive to the intended politeness of
these utterances. These deficits were most evident One of the intended benefits of investigating right
in our condition, which resembled ‘‘mock polite- hemisphere contributions to speaker attitudes was
ness’’ (i.e., very indirect language + low prosody) to compare these findings to data on how the same
where RHD patients failed to detect the negative/ individuals recognized expressions of basic emo-
impolite attitude of the speaker, a situation that tions from prosody. Given the consistent deficits
resembles the case of sarcasm, which is commonly of our RHD group in each of these experiments
encountered in human discourse. (Pell, 2006; Pell, under review), our findings es-
Since RHD adults were fundamentally aware tablish a platform for predicting that many RHD
of how prosody functions to signal politeness patients who fail to comprehend emotional pros-
based on these features alone (prosody condition), ody also fail to normally appreciate many of the
it is likely that a major source of their errors in emotive attitudes conveyed by speakers at the level
the politeness task was the ability to engage in a of prosody. However, as emphasized in our pre-
comparative analysis of relevant cues across the vious research (Pell and Baum, 1997; Pell, 2006),
two speech channels (Foldi, 1987; Weylman et al., not all RHD individuals display receptive impair-
1989). Possibly, our RHD patients were less ments for emotional prosody and this was also
efficient at detecting violations in the emotive sig- true of our data on speaker attitudes.
nificance of language and prosody precluding in- As a supplementary measure, we compared in-
ferential processes, which would highlight the dividual performance features of the RHD pa-
intended (i.e., nonconventional) attitude of the tients in our confidence and politeness studies in
speaker in these situations. This explanation relation to their individual scores on tasks for un-
places the locus of their deficit in our politeness derstanding emotional prosody (Pell, 2006). We
experiment closer to the stage of applying and in- detected considerable overlap in the ability of in-
tegrating social and affective cues, which are rel- dividual RHD patients to judge speaker confi-
evant to understanding the intended meaning of dence and to judge vocal emotions in our ‘‘pure
314

prosody’’ identification task (Pell, 2006). Three prosodic defect in our RHD patients may have
RHD participants (R3, R5, R6) who were mark- contributed to difficulties in both the confidence
edly ‘‘impaired’’ in the recognition of speaker con- and emotional prosody tasks (Pell, 2006).
fidence, based on the expected range of their mean In contrast, our data on speaker politeness high-
confidence ratings, were also the most impaired lighted the interdependence of prosody and lan-
individuals for judging basic emotions from pro- guage content for understanding this attitude and
sodic cues alone (Pell, 2006). Participant R5 who established that the same RHD patients could suc-
had a large temporoparietal lesion and whose cessfully interpret categorical distinctions in into-
deficits for emotional prosody were most pro- nation contours for judging the polite or impolite
nounced was also most severely impaired for judg- attitude of speakers. At the level of prosody alone,
ing speaker confidence (assigning mean ratings of this task likely resembles the ability of many RHD
2.1, 2.2, and 2.2 for low, moderate, and high con- patients to recognize highly conventionalized, pho-
fidence utterances, respectively). nological pitch categories, which is known to be
Finally, it is noteworthy that two RHD patients intact in most RHD adults (Baum and Pell, 1999;
who were unimpaired in processing emotional Hsieh et al., 2001). However, RHD patients per-
prosody (R1, R7) also performed normally in the formed abnormally when prosody and language
confidence experiment, with mean conditional rat- choices combined in an unconventional (discordant)
ings approximating the HC group values. While manner to communicate politeness. This places the
attrition of our RHD group (including partici- likely source of impairment for understanding cer-
pants R5 and R6) precluded a similar analysis of tain attitudes such as politeness, not at the level of
individual data on speaker politeness with our prosody per se, but in resolving these attitudinal
emotional prosody results, it is nonetheless appar- meanings based on the comparative weight of pros-
ent that deficits for emotional prosody in our ody with other socially relevant cues such as utter-
RHD sample frequently correlated with deficits for ance type. This idea fits with a well-established
understanding speaker confidence, and perhaps, literature indicating that RHD patients do not suc-
other speaker attitudes pending more research that cessfully integrate all relevant sources of informa-
explores these issues in greater depth. tion when intended meanings are noncanonical or
nontransparent in form (for a review, see Brownell
et al., 1997; Martin and McDonald, 2003). Along
On the source of failures to appreciate prosodic similar lines, our data on speaker attitudes such as
attitudes following RHD politeness imply that RHD patients do not always
incorporate their evaluations of prosodic informa-
It would appear that for many RHD patients, a tion with the significance of language content,
failure to appreciate speaker attitudes may reflect which was often the primary focus of these listeners.
impairments at slightly different stages of process- Until further data are gathered to inform these
ing depending on the manner in which prosody, issues in a more detailed manner, one can argue that
language, and other variables which define speaker (1) when continuous changes in multiple prosodic
attitudes are assigned meaning by the interpersonal elements weigh heavily for assigning speaker atti-
context. In our experiment on speaker confidence, tudes (e.g., confidence), or (2) when prosody must
prosody was always the dominant cue for inferring be resolved with associated parameters of the ut-
this attitude, justifying our hypothesis that diffi- terance to mark noncanonical interpretations of the
culties relating specifically to prosody governed the expressed attitude (e.g., speaker politeness), RHD
RHD patients’ difficulties in that task. Like basic patients often fail to generate appropriate inferences
emotions, vocal features, which communicate dis- about the emotive intentions of the speaker. In
tinctions in confidence/doubt, are likely repre- either case, one must assume that these specific
sented by multiple prosodic elements with deficits are part of a wider array of ‘‘pragmatic
continuous expression in the speech signal (Scherer language deficits’’ exhibited by many RHD patients
et al., 1973, 1984). For this reason, an underlying who show marked difficulties in the generation
315

Acknowledgments
of inferences from socially and emotionally rele-
vant information in speech (Happe et al., 1999; see
The author thanks Tamara Long and Alana
Martin and McDonald, 2003 for a recent overview).
Pearlman for their helpful comments and assist-
It will be highly informative to see these findings
ance with data collected for the studies on speaker
extended to a broader array of interpersonal con-
texts that revolve around prosody usage, to a larger confidence and speaker politeness, respectively;
Elmira Chan and Nicole Hallonda Price for addi-
sample of RHD participants, and studied through
tional help with data collection, data organization,
other investigative approaches.
and manuscript preparation; the Canadian Insti-
Our data also call on researchers to explore how
tutes for Health Research (Institute of Aging) and
emotive features of prosody are interpreted by other
the Natural Sciences and Engineering Research
brain-damaged populations who exhibit deficits for
Council of Canada for generous operating sup-
emotional prosody, such as patients with acquired
port; and McGill University for investigator
left hemisphere lesions (Cancelliere and Kertesz,
1990; Van Lancker and Sidtis, 1992; Pell, 1998, support (William Dawson Chair).
2006; Adolphs et al., 2002; Charbonneau et al.,
2003) or Parkinson’s disease (Breitenstein et al., References
1998; Pell and Leonard, 2003). Although it is our
standpoint that emotional and emotive meanings of Adolphs, R., Damasio, H. and Tranel, D. (2002) Neural sys-
prosody rely on mandatory right hemisphere con- tems for recognition of emotional prosody: a 3-D lesion
study. Emotion, 2: 23–51.
tributions during speech comprehension, our find-
Barr, D.J. (2003) Paralinguistic correlates of conceptual struc-
ings are inconclusive regarding the probable input of ture. Psychonomic Bull. Rev., 10: 462–467.
left hemisphere (Kotz et al., 2003) and subcortical Baum, S.R. and Pell, M.D. (1999) The neural bases of prosody:
(Pell and Leonard, 2003) regions in this processing. insights from lesion studies and neuroimaging. Aphasiology,
Given the increasingly strong evidence of interhem- 13: 581–608.
ispheric cooperation in how emotional prosody is Blum-Kulka, S. (1987) Indirectness and politeness in requests:
same or different? J. Prag., 11: 131–146.
processed from speech (see Kotz et al., in this vol- Blanc, J.-M. and Dominey, P. (2003) Identification of prosodic
ume), coupled with the notion that emotive and in- attitudes by a temporal recurrent network. Cogn. Brain Res.,
terpersonal meanings of prosody bear an important 17: 693–699.
functional relationship to emotional prosody, one Boltz, M.G. (2005) Temporal dimensions of conversational in-
would predict that bilateral hemispheric mechanisms teraction: the role of response latencies and pauses in social
impression formation. J. Lang. Soc. Psychol., 24: 103–138.
are engaged when processing speaker attitudes em- Bowers, D., Coslett, H.B., Bauer, R.M., Speedie, L.J. and He-
bedded in spoken language content (Pell, 2006). ilman, K.M. (1987) Comprehension of emotional prosody
Through careful inspection of how speaker at- following unilateral hemispheric lesions: processing defect
titudes are dictated by the physical form of pros- versus distraction defect. Neuropsychologia, 25: 317–328.
Breitenstein, C., Daum, I. and Ackermann, H. (1998) Emo-
ody, the degree to which prosodic parameters are
tional processing following cortical and subcortical brain
conventionalized by the speaker–hearer, and the damage: contribution of the fronto-striatal circuitry. Behav.
relationship of the prosodic message to conjoined Neurol., 11: 29–42.
linguistic, contextual, and other relational varia- Brennan, S.E. and Williams, M. (1995) The feeling of another’s
bles, future research will undoubtedly culminate in knowing: prosody and filled pauses as cues to listeners about
increasingly finer descriptions of how speaker at- the metacognitive states of speakers. J. Mem. Lang., 34:
383–398.
titudes are processed in the human mind/brain. Brown, P. and Levinson, S. (1987) Politeness: Some Universals
in Language Usage. Cambridge University Press, Cambridge.
Brownell, H.H., Carroll, J.J., Rehak, A. and Wingfield,
Abbreviations A. (1992) The use of pronoun anaphora and speaker mood
in the interpretation of conversational utterances by right
hemisphere brain-damaged patients. Brain Lang., 43:
HC healthy control 121–147.
LHD left-hemisphere-damaged Brownell, H., Pincus, D., Blum, A., Rehak, A. and Winner,
RHD right-hemisphere-damaged. E. (1997) The effects of right-hemisphere brain damage on
316

patients’ use of terms of personal reference. Brain Lang., 57: Kotz, S., Meyer, M., Alter, K., Besson, M., Von Cramon,
60–79. Y. and Friederici, A. (2003) On the lateralization of emo-
Brown, B.L., Strong, W.J. and Rencher, A.C. (1974) Fifty-four tional prosody: an event-related functional MR investigation.
voices from two: the effects of simultaneous manipulations of Brain Lang., 86: 366–376.
rate, mean fundamental frequency, and variance of funda- Ladd, D.R., Silverman, K.E.A., Tolkmitt, F., Bergmann,
mental frequency on ratings of personality from speech. G. and Scherer, K.R. (1985) Evidence for the independent
J. Acoust. Soc. Am., 55: 313–318. function of intonation contour type, voice quality, and Fo
Burgoon, J.K. (1993) Interpersonal expectations, expectancy range in signaling speaker affect. J. Acoust. Soc. Am., 78:
violations, and emotional communication. J. Lang. Soc. 435–444.
Psychol., 12: 30–48. Laplante, D. and Ambady, N. (2003) On how things are said:
Caffi, C. and Janney, R.W. (1994) Toward a pragmatics of voice tone, voice intensity, verbal content and perceptions of
emotive communication. J. Prag., 22: 325–373. politeness. J. Lang. Soc. Psychol., 22: 434–441.
Cancelliere, A.E.B. and Kertesz, A. (1990) Lesion localization Loveday, L. (1981) Pitch, politeness and sexual role: an ex-
in acquired deficits of emotional expression and comprehen- ploratory investigation into the pitch correlates of English
sion. Brain Cogn, 13: 133–147. and Japanese politeness formulae. Lang. Speech, 24: 71–89.
Charbonneau, S., Scherzer, B.P., Aspirot, D. and Cohen, Martin, I. and McDonald, S. (2003) Weak coherence, no theory
H. (2003) Perception and production of facial and prosodic of mind, or executive dysfunction? Solving the puzzle of
emotions by chronic CVA patients. Neuropsychologia, 41: pragmatic language disorders. Brain Lang., 85: 451–466.
605–613. Mitchell, R.L.C., Elliott, R., Barry, M., Cruttenden, A. and
Cheang, H.S. and Pell, M.D. (2006) A study of humour and Woodruff, P.W.R. (2003) The neural response to emotional
communicative intention following right hemisphere stroke. prosody, as revealed by functional magnetic resonance im-
J. Clin. Linguist. Phonet., 20: 447–462. aging. Neuropsychologia, 41: 1410–1421.
Clark, H.H. and Schunk, D. (1980) Polite responses to polite Pakosz, M. (1983) Attitudinal judgments in intonation: some
requests. Cognition, 8: 111–143. evidence for a theory. J. Psycholinguist. Res., 12: 311–326.
Culpeper, J., Bousfield, D. and Wichmann, A. (2003) Impo- Pell, M.D. (1998) Recognition of prosody following unilateral
liteness revisited: with special reference to dynamic and pro- brain lesion: influence of functional and structural attributes
sodic aspects. J. Prag., 35: 1545–1579. of prosodic contours. Neuropsychologia, 36: 701–715.
Foldi, N.S. (1987) Appreciation of pragmatic interpretations of Pell, M.D. (2001) Influence of emotion and focus location on
indirect commands: comparison of right and left hemisphere prosody in matched statements and questions. J. Acoust. Soc.
brain-damaged patients. Brain Lang., 31: 88–108. Am., 109: 1668–1680.
Friederici, A. and Alter, K. (2004) Lateralization of auditory Pell, M.D. (2005) Prosody-face interactions in emotional
language functions: a dynamic dual pathway model. Brain processing as revealed by the facial affect decision task.
Lang., 89: 267–276. J. Nonver. Behav., 29: 193–215.
Gandour, J., Tong, Y., Wong, D., Talavage, T., Dzemidzic, Pell, M.D. (2006) Cerebral mechanisms for understanding emo-
M., Xu, Y., Li, X. and Lowe, M. (2004) Hemispheric roles in tional prosody in speech. Brain Lang., 96: 221–234.
the perception of speech prosody. NeuroImage, 23: 344–357. Pell, M.D. Reduced sensitivity to ‘‘prosodic attitudes’’ in adults
Gandour, J., Wong, D., Dzemidzic, M., Lowe, M., Tong, with focal right hemisphere brain damage. Brain Lang., in press.
Y. and Li, X. (2003) A cross-linguistic fMRI study of per- Pell, M.D. and Baum, S.R. (1997) The ability to perceive and
ception of intonation and emotion in Chinese. Hum. Brain comprehend intonation in linguistic and affective contexts by
Mapp., 18: 149–157. brain-damaged adults. Brain Lang., 57: 80–99.
Happe, F., Brownell, H. and Winner, E. (1999) Acquired theory Pell, M.D. and Leonard, C.L. (2003) Processing emotional tone
of mind impairments following stroke. Cognition, 70: from speech in Parkinson’s disease: a role for the basal gan-
211–240. glia. Cogn. Affect. Behav. Neurosci., 3: 275–288.
Hesling, I., Clement, S., Bordessoules, M. and Allard, M. Plante, E., Creusere, M. and Sabin, C. (2002) Dissociating
(2005) Cerebral mechanisms of prosodic integration: evi- sentential prosody from sentence processing: activation in-
dence from connected speech. NeuroImage, 24: 937–947. teracts with task demands. NeuroImage, 17: 401–410.
Hirst, W., Ledous, J. and Stein, S. (1984) Constraints on the Poeppel, D. (2003) The analysis of speech in different temporal
processing of indirect speech acts: evidence from aphasio- integration windows: cerebral lateralization as ‘asymmetric
logy. Brain Lang., 23: 26–33. sampling in time’. Speech Commun., 41: 245–255.
Holmes, J. (1984) Modifying illocutionary force. J. Prag., 8: Ross, E.D., Thompson, R.D. and Yenkosky, J. (1997) Later-
345–365. alization of affective prosody in brain and the collosal inte-
Hsieh, L., Gandour, J., Wong, D. and Hutchins, G. (2001) gration of hemispheric language functions. Brain Lang., 56:
Functional heterogeneity of inferior frontal gyrus is shaped 27–54.
by linguistic experience. Brain Lang., 76: 227–252. Scherer, K.R., Ladd, D.R. and Silverman, K. (1984) Vocal cues
Kimble, C.E. and Seidel, S. (1991) Vocal signs of confidence. to speaker affect: testing two models. J. Acoust. Soc. Am., 76:
J. Nonver. Behav., 15: 99–105. 1346–1356.
317

Scherer, K.R., London, H. and Wolf, J.J. (1973) The voice of Uldall, E. (1960) Attitudinal meanings conveyed by intonation
confidence: paralinguistic cues and audience evaluation. contours. Lang. Speech, 3: 223–234.
J. Res. Person., 7: 31–44. Van Lancker, D. (1980) Cerebral lateralization of pitch cues in
Schirmer, A., Zysset, S., Kotz, S. and Von Cramon, D.Y. the linguistic signal. Pap. Linguist., 13: 201–277.
(2004) Gender differences in the activation of inferior frontal Van Lancker, D. and Sidtis, J.J. (1992) The identification of
cortex during emotional speech perception. NeuroImage, 21: affective-prosodic stimuli by left- and right-hemisphere-dam-
1114–1123. aged subjects: all errors are not created equal. J. Speech
Sidtis, J. and Van Lancker Sidtis, D. (2003) A neurobehavioral Hearing Res., 35: 963–970.
approach to dysprosody. Semin. Speech Lang., 24: 93–105. Vingerhoets, G., Berckmoes and Stroobant, N. (2003) Cerebral
Smith, V.L. and Clark, H.H. (1993) On the course of answering hemodynamics during discrimination of prosodic and se-
questions. J. Mem. Lang., 32: 25–38. mantic emotion in speech studied by transcranial Doppler
Starkstein, S.E., Federoff, J.P., Price, T.R., Leiguarda, R.C. ultrasonography. Neuropsychology, 17: 93–99.
and Robinson, R.G. (1994) Neuropsychological and neuro- Weylman, S., Brownell, H., Roman, M. and Gardner, H. (1989)
radiologic correlates of emotional prosody comprehension. Appreciation of indirect requests by left- and right-brain-
Neurology, 44: 515–522. damaged patients: the effects of verbal context and conven-
Stemmer, B., Giroux, F. and Joanette, Y. (1994) Production tionality of wording. Brain Lang., 36: 580–591.
and evaluation of requests by right-hemisphere brain-dam- Wichmann, A. (2002) Attitudinal intonation and the inferential
aged individuals. Brain Lang., 47: 1–31. process. In: Bel, B., Marlien, I. (Eds.), Proceedings of Speech
Tompkins, C.A. and Mateer, C.A. (1985) Right hemisphere Prosody 2002 Conference, 11–13 April 2002. Laboratoire
appreciation of prosodic and linguistic indications of implicit Parole et Langage, Aix-en-Provence, France.
attitude. Brain Lang., 24: 185–203. Wildgruber, D., Hertrich, I., Riecker, A., Erb, M., Anders, S.,
Tong, Y., Gandour, J., Talavage, T., Wong, D., Dzemidzic, M., Grodd, W. and Ackermann, H. (2004) Distinct frontal re-
Xu, Y., Li, X. and Lowe, M. (2005) eural circuitry underlying gions subserve evaluation of linguistic and emotional aspects
sentence-level linguistic prosody. NeuroImage, 28: 417–428. of speech intonation. Cerebral Cortex, 14: 1384–1389.
Trees, A.R. and Manusov, V. (1998) Managing face concerns in Wildgruber, D., Pihan, H., Ackermann, H., Erb, M. and
criticism: integrating nonverbal behaviors as a dimension of Grodd, W. (2002) Dynamic brain activation during process-
politeness in female friendship dyads. Hum. Commun. Res., ing of emotional intonation: influence of acoustic parameters,
24: 564–583. emotional valence, and sex. NeuroImage, 15: 856–869.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 18

Processing of facial identity and expression: a


psychophysical, physiological, and computational
perspective

Adrian Schwaninger1,2,, Christian Wallraven1, Douglas W. Cunningham1 and


Sarah D. Chiller-Glaus2

1
Department of Bülthoff, Max Planck Institute for Biological Cybernetics, Spemannstr. 38, 72076 Tübingen, Germany
2
Department of Psychology, University of Zurich, Zurich, Switzerland

Abstract: A deeper understanding of how the brain processes visual information can be obtained by
comparing results from complementary fields such as psychophysics, physiology, and computer science. In
this chapter, empirical findings are reviewed with regard to the proposed mechanisms and representations
for processing identity and emotion in faces. Results from psychophysics clearly show that faces are
processed by analyzing component information (eyes, nose, mouth, etc.) and their spatial relationship
(configural information). Results from neuroscience indicate separate neural systems for recognition of
identity and facial expression. Computer science offers a deeper understanding of the required algorithms
and representations, and provides computational modeling of psychological and physiological accounts.
An interdisciplinary approach taking these different perspectives into account provides a promising basis
for better understanding and modeling of how the human brain processes visual information for recog-
nition of identity and emotion in faces.

Keywords: face recognition; facial expression; interdisciplinary approach; psychophysics of face processing;
computational modeling of face processing; face processing modules; component and configural processing

Introduction faces have not been seen for 50 years. Moreover,


people identify facial expressions very fast and
Everyday object recognition is usually a matter of even without awareness (see Leiberg and Anders,
discriminating between quite heterogeneous object this volume). These abilities seem to be remarkably
classes that differ with regard to their global shape, disrupted if faces are turned upside-down. Con-
parts, and other distinctive features such as color sider the pictures in Fig. 1. Although this woman is
or texture. Face recognition, in contrast, relies on a well-known celebrity, it is difficult to recognize
the discrimination of exemplars of a very homog- her from the inverted photographs. One might de-
enous category. According to Bahrick et al. (1975) tect certain differences between the two pictures
we are able to recognize familiar faces with an ac- despite the fact that both seem to have the same
curacy of 90% or more, even when some of these facial expression. Interestingly, after rotating this
page by 1801 so that the two faces are upright, one
can now easily identify the person depicted in these
Corresponding author. Tel.: +41-76-393-24-46; Fax: +49- pictures and grotesque differences in the facial ex-
7071-601-616; E-mail: adrian.schwaninger@tuebingen.mpg.de pression are revealed. This illusion was discovered

DOI: 10.1016/S0079-6123(06)56018-2 321


322

Fig. 1. Thatcher illusion. When the photographs are viewed upside-down (as above) it is more difficult to identify the person belonging
to the pictures and the facial expressions seem similar. When the pictures are viewed right side up, it is very easy to identify the person
depicted in these pictures and the face on the right appears highly grotesque.

by Thompson (1980). He used Margaret Thatch- the expression in either half of these composite
er’s face, which is why the illusion is known as the images. However, this effect diminished when
‘‘Thatcher illusion.’’ It was already well known by faces were misaligned or inverted, which parallels
painters and Gestalt psychologists that face the composite effect for facial identity by Young
processing is highly dependent on orientation et al. (1987). Interestingly, in an additional experi-
(e.g., Köhler, 1940). However, the finding that ment Calder et al. found evidence for the view that
upside-down faces are disproportionately more the composite effects for identity and expression
difficult to recognize than other inverted objects operate independently of one another.
has been referred to as the face inversion effect and These examples illustrate that information of
was first reported by Yin (1969). parts and spatial relations are somehow combined
Another interesting effect was discovered by in upright faces. In contrast, when faces are turned
Young et al. (1987). Composite faces were created upside-down it seems that only the local part-
by combining the top and bottom half of different based information is processed.
faces. Figure 2 shows an example. If the two halves In this chapter, we discuss the representations
were aligned and presented upright, a new face and processes used in recognition of identity and
resembling each of the two originals seemed to facial emotion. We follow a cognitive neuroscience
emerge. This made it difficult to identify the per- approach, by discussing the topic from a psycho-
sons shown in either half. If faces were inverted or physical, physiological, and computational per-
if the top and bottom halves were misaligned hor- spective. Psychophysics describes the relationship
izontally, then the two halves did not spontane- between stimuli in our external world and our
ously fuse to create a new face and the constituent internal representations. We first review the psy-
halves remained identifiable. chophysics literature on recognition of faces and
Calder et al. (2000) used the same technique to facial expressions. Because our goal is to gain a
investigate the processing of facial expressions. deeper understanding of how our brain produces
They prepared emotional face composites by behavior, we discuss possible neural substrates
aligning the top half of one expression (e.g., an- of the representations and processes identified in
ger) with the bottom half of another (e.g., happi- neuroscience. Computer science, the third perspec-
ness) from the same person. When the face tive, provides computational algorithms to solve
composites were aligned, a new facial expression certain recognition problems and the possibility of
emerged and participants were slower to identify biologically plausible computer models.
323

Fig. 2. Aligned and misaligned halves of different identities (Margaret Thatcher and Marilyn Monroe). When upright (as above), a
new identity seems to emerge from the aligned composites (left), which makes it more difficult to extract the original identities. This
does not occur for the misaligned composite face (right). When viewed upside-down, the two identities do not fuse to a new identity.

Psychophysical perspective recognize than in the whole face condition. How-


ever, when participants were trained to recognize
Recognition of identity inverted faces, scrambled faces, and houses no
such advantage of context was found. Tanaka and
Two main hypotheses have been proposed to ex- Farah concluded that face recognition relies
plain the recognition of identity in faces: the ho- mainly on holistic representations, in contrast to
listic hypothesis and the component-configural the recognition of objects. While the encoding and
hypothesis. According to the holistic hypothesis, matching of parts are assumed to be relatively
upright faces are stored as unparsed perceptual orientation invariant (see also Biederman, 1987),
wholes in which individual parts are not explicitly holistic processing is thought to be very sensitive
represented (Tanaka and Farah, 1991, 1993; Farah to orientation (see also Farah et al., 1995b; Bieder-
et al., 1995b). The main empirical evidence in fa- man and Kalocsai, 1997).
vor of this view is based on a paradigm by Tanaka The component-configural hypothesis is based
and Farah (1993). These authors argued that if on a qualitative distinction between component
face recognition relies on parsed representations, and configural information. The term component
then single parts of a face, such as nose, mouth, or (or part-based, piecemal, feature-based, featural)
eyes, should be easily recognized even if presented information refers to those elements of a face
in isolation. However, if faces are represented as that are perceived as parts of the whole (e.g., the
unparsed perceptual wholes (i.e., holistically), then eyes, mouth, nose, ears, chin, etc.). According to
recognition of the same isolated parts should be Bruce (1988), the term configural information (or
more difficult. In their experiments, participants configurational, spatial-relational, second-order
were shown a previously learned face together with relational information) refers to the ‘‘spatial inter-
a slightly different version in which one single part relationship of facial features’’ (p. 38). Examples
(e.g., nose or mouth) had been replaced. The task are the eye–mouth or intereye distance. Interest-
was to judge which face had been shown in the ingly, these distances are overestimated by
learning phase. The experiment was conducted in 30–40% (eye–mouth distance) and about 15% (in-
both a whole face condition and an isolated parts tereye distance) in face perception (Schwaninger
condition without facial context. In the isolated et al., 2003b). In practice, configural changes have
condition, face parts proved to be more difficult to been induced by altering the distance between
324

components or by rotating components (e.g., turn- second condition, the blur level was determined
ing the eyes and mouth upside-down within the that made the scrambled versions impossible to
facial context like in the Thatcher illusion de- recognize. This blur level was then applied to
scribed above). According to the component-con- whole faces in order to create configural versions
figural hypothesis, the processing of configural that by definition did not contain local featural
information is strongly impaired by inversion or information. These configural versions of previ-
plane rotation, whereas processing of component ously learned intact faces could be recognized re-
information is much less affected. There are now a liably. These results suggest that separate
large number of studies providing converging ev- representations exist for component and con-
idence in favor of this view (e.g., Sergent, 1984; figural information. Familiar face recognition
Rhodes et al., 1993; Searcy and Bartlett, 1996; was investigated in a second experiment by run-
Leder and Bruce, 1998; Schwaninger and Mast, ning the same conditions with participants who
2005; see Schwaninger et al., 2003a, for a review). knew the target faces (all distractor faces were un-
These studies changed component information by familiar to the participants). Component and con-
replacing components (e.g., eyes of one person figural recognition was better when the faces were
were replaced with the eyes of another person). familiar, but there was no qualitative shift in
Configural changes were induced by altering the processing strategy as indicated by the fact that
distance between components (e.g., larger or there was no interaction between familiarity and
smaller intereye distance). However, one possible condition (see Fig. 3).
caveat is that these types of manipulations often Schwaninger et al. (2002, 2003a) proposed a
change the holistic aspects of the face and they are model that allows integrating the holistic and
difficult to carry out selectively. For example, re- component-configural hypotheses. Pictorial as-
placing the nose (component change) might pects of a face are contained in the pictorial met-
change the distance between the contours of the ric input representation that corresponds to
nose and the mouth, which induces a configural activation of primary visual areas. On the basis
change (Leder and Bruce, 1998, 2000). Moving the of years of experience, neural networks are trained
eyes apart (configural change) can lead to an in- to extract specific information in order to activate
crease in size of the bridge of the nose, which is a component and configural representations in the
component change (see Leder et al., 2001). Such ventral visual stream. The output of these repre-
problems can be avoided by using scrambling and sentations converges towards the same identifica-
blurring procedures to reduce configural and com- tion units. These units are holistic in the sense that
ponent information independently (e.g., Sergent, they integrate component and configural informa-
1985; Davidoff and Donnelly, 1990; Collishaw and tion. Note that this concept of holistic differs from
Hole, 2000; Boutet et al., 2003). Schwaninger et al. the original definition of Tanaka and Farah (1993)
(2002) extended previous research by ensuring that and Farah et al. (1995b), which implies that faces
scrambling and blurring effectively eliminate con- are stored as perceptual wholes without explicit
figural and component information separately. representations of parts (component information).
Furthermore, in contrast to previous studies, The model by Schwaninger et al. (2002, 2003a)
Schwaninger et al. (2002) used the same faces in assumes that it is very difficult to mentally rotate a
separate experiments on unfamiliar and familiar face as a perceptual whole (Rock, 1973, 1974,
face recognition to avoid potential confounds with 1988; Schwaninger and Mast, 2005). When faces
familiarity. In an old–new recognition paradigm it are substantially rotated from upright, they have
was found that previously learned intact faces to be processed by matching parts, which explains
could be recognized even when they were scram- why information about their spatial relationship
bled into constituent parts. This result challenges (configural information) is hard to recover when
the assumption of purely holistic processing ac- faces are inverted (for a similar view see Valentine
cording to Farah et al. (1995b) and suggests that and Bruce, 1988). Since face recognition depends
components are encoded and stored explicitly. In a on detecting subtle differences in configural
325

Fig. 3. Recognition performance in unfamiliar and familiar face recognition across three different conditions at test. Scr: scrambled;
ScrBlr: scrambled and blurred; Blr: blurred. (Adapted with permission from Schwaninger et al., 2002.)

information, a large inversion effect is observed normal activation of component representations.


(Yin, 1969). Consistent with this view, Williams The unnatural spatial relationship (changed con-
et al. (2004) suggested that inverted faces are figural information) is hard to perceive due to ca-
initially processed by parts-based assessment be- pacity limitations of an orientation normalization
fore second-order relational processing is initiated. mechanism. As a consequence, the strange activa-
Sekuler et al. (2004) used response classification tion pattern of configural representations is re-
and found that the difference between the process- duced and the grotesque perception disappears.
ing of upright and inverted faces was of quanti- The composite face illusion can be explained on
tative, rather than of qualitative, nature, i.e., the basis of similar reasoning. Aligned upright face
information was extracted more efficiently from composites contain new configural information
upright faces than from inverted faces. This is also resulting in a new perceived identity. Inverting the
consistent with Schwaninger et al.’s model if one aligned composites reduces the availability of con-
assumes that configural processing is not abruptly figural information and it is easier to access the
but gradually impaired by rotation (Murray et al., two different face identification units on the basis
2000; Schwaninger and Mast, 2005) and integrated of component information alone.
with the output of component processing. Invert- Note that the model does not apply to process-
ing the eyes and mouth within an upright face re- ing component and configural information for
sults in a strange activation pattern of component gaze perception. As shown by Schwaninger et al.
and configural representations that appears gro- (2005), there is also an inversion effect on gaze
tesque when upright, but not when upside-down perception. However, this effect is not due to im-
(Thatcher illusion). This can be explained by the paired configural processing but due to orienta-
integrative model as follows: in an inverted tion-sensitive processing of local component
Thatcher face, the components themselves are in information in the eyes. This difference between
the correct orientation which results in a relatively inversion effects for recognition of identity versus
326

perceived eye gaze direction are consistent with Given the general preoccupation with the role of
separate functional systems for these tasks, which featural information in the recognition of facial
is consistent with physiological evidence discussed expressions, it should not be surprising that the
below. vast majority of descriptive systems and models of
In short, the model by Schwaninger et al. (2002, facial expressions are explicitly part-based (Frois-
2003a) allows the integration of the component Whittmann, 1930; Frijda and Philipszoon, 1963;
configural hypothesis and holistic aspects of Leventhal and Sharp, 1965; Ekman and Friesen,
face processing relevant to recognition of identity. 1978; Izard, 1979; Tronick et al., 1980; Essa
It can also explain striking perceptual effects such and Pentland, 1994; Ellison and Massaro, 1997).
as the Thatcher illusion and the composite face Perhaps the most widely used methods for para-
illusion. Most importantly, it provides an integra- metrizing the high-dimensional space of facial ex-
tive basis for understanding special characteristics pressions is the facial action coding system (or
of face recognition such as the specialization FACS; Ekman and Friesen, 1978), which segments
in upright faces and the sensitivity to configural the visible effects of facial muscle activity and
information. rigid head motion into ‘‘action units.’’ Combina-
tions of these action units can then be used to
Recognition of expressions describe different expressions. It is important to
note that FACS was designed as a system for
The structure and perception of facial expressions describing the elements of photographs of facial
have been subject to scientific examination since expressions. It is not a model of facial expression
at least Duchenne’s (1990) and Darwin’s (1872) processing and makes no claims about which
seminal work. The majority of these studies con- elements go together to produce different expres-
sisted of showing static photographs of expres- sions (Sayette et al., 2001).
sions to observers and examining the relationship Massaro and colleagues proposed a parts-based
between statically visible deformations of the model of perception (the fuzzy logical model of
facial surface and the judgments made by the ob- perception or FLMP) in which the features are
servers. It is, of course, clear that different independently processed and subsequently inte-
facial areas are important for the recognition of grated. The model makes specific claims about
different emotions (Hanawalt, 1944; Plutchik, how the featural information is processed and in-
1962; Nummenmma, 1964; Bassili, 1979; Cunning- tegrated, and thus makes clear predictions about
ham et al., 2005). For example, as mentioned the perception and categorization of facial expres-
above, Bassili (1979) used point-light faces to show sions. In one study, Ellison and Massaro (1997)
that the upper portions of the face are important used computer graphics animation techniques to
for some expressions, while the lower portions of produce static facial expressions where either (a)
the face are important for other expressions. the mouth shape was parametrically varied, (b) the
Facial features also play differentiated roles in eyebrow shape was parametrically varied, or (c)
other aspects of facial expression processing, such both were independently parametrically varied.
as the perception of sincerity. For example, ac- The faces were shown to a number of observers,
cording to Ekman and Friesen (1982), a true smile who were asked if the expression in the photo-
of enjoyment, which Ekman refers to as a graphs was that of happiness or anger. Ellison and
Duchenne smile, has a characteristic mouth shape Massaro found that both features (eyebrow posi-
as well as specific wrinkles around the eyes. tion and mouth position) affected the participants’
Faked expressions of enjoyment, in contrast, con- judgments, and that the influence of one feature
tain just the mouth information. Furthermore, was more prominent when the other feature was
Ekman and Friesen (1982) have shown that de- neutral or ambiguous. Moreover, the FLMP cap-
ceptive expressions of enjoyment appear to have tured patterns in the data better than either holistic
different temporal characteristics than spontane- models or a straight-forward additive model based
ous ones. on recognition rates of the individual features.
327

Elison and Massaro consequently claimed that can describe expressions in terms of their features
the perceptual system must be using featural infor- and configuration, and that dynamic information
mation in the recognition process and cannot be is represented in the human visual system and is
employing a purely holistic approach. These results important for various aspects of facial expression
are consistent with the finding that the aligned processing. Moreover, simple, purely holistic mod-
combination of two different emotions leads to els do not seem to describe the perception of facial
decreased recognition performance (Calder et al., expressions very well.
2000).
Just as is true for the processing of facial iden- Dynamic information
tity, the separate roles of component-configural
and holistic information have been discussed The vast majority of research on the perception of
within the context of facial expressions. There faces has tended to focus on the relatively stable
are at least two models that integrate holistic in- aspects of faces. This has consequently led to a
formation (Izard et al., 1983; White, 2000). White strong emphasis on static facial information (i.e.,
(2000) proposed a ‘‘hybrid model,’’ according to information that is available at any given instant,
which expression recognition is part-based on one such as eye color, distance between the eyes, etc.).
hand and holistic in the sense of undecomposed In general, however, human faces are not static
wholes on the other hand. entities. Humans are capable of moving their faces
Several researchers have examined the role of in a wide variety of ways, and they do so for an
temporal information in the perception and rec- astonishingly large number of reasons. Recent ad-
ognition of expressions (Bassili, 1978, 1979; Bruce, vances in technology, however, have allowed re-
1988; Edwards, 1998; Kamachi et al., 2001). searchers to begin examining the role of motion in
Kamachi et al., for example, manipulated the ve- face processing.
locity in which a neutral face turned into an emo- Before one can determine what types of motion
tional one. They found that happiness and surprise are used in the recognition of faces and facial ex-
were better recognized from fast sequences, sad- pressions (i.e., the dynamic features), one must
ness was better recognized from slow sequences, determine if motion plays any role at all. To this
and angriness was best recognized at medium end, it has been clearly established that dynamic
speed. This indicates that different expressions information can be used to recognize identity (Pike
seem to have a characteristic speed or rate of et al., 1997; Bruce et al., 1999, 2001; Lander et al.,
change. In an innovative study, Edwards (1998) 1999, 2001; Knappmeyer et al., 2003) and to rec-
presented participants with photographs of indi- ognize expressions (Bassili, 1978, 1979; Hump-
vidual frames from a video sequence of a dynamic hreys et al., 1993; Edwards, 1998; Kamachi et al.,
expression, in a scrambled order, and asked 2001; Cunningham et al., 2005; Wallraven et al.,
participants to place the photographs in the cor- 2005a). Overall, the positive influence of dynamic
rect order. Participants were remarkably accurate information is most evident when static informa-
in their reconstructions, showing a particularly tion is degraded.
strong sensitivity to the temporal characteristics in It is difficult, if not impossible, to present dy-
the early phases of an expression. Interestingly, namic information without static information.
participants performed better when asked to com- Thus, Pike et al. (1997) and Lander and colleagues
plete the task with extremely tight time constraints performed a series of control experiments to en-
than when given unlimited time, from which sure that the apparent advantage moving faces
Edwards concluded that conscious strategies are have over static faces is due to information that is
detrimental to this task. He further concluded solely available over time (i.e., dynamic informa-
that the results show that humans do encode and tion). One might, for example, describe dynamic
represent temporal information about expressions. sequences as a series of static snapshots. Under
In sum, it is clear that different expressions re- such a description, the advantage of dynamic
quire different features to be recognized, that one stimuli would not lie with dynamic information,
328

but with the fact that a video sequence has more conditions performing either specific expressions
static information (i.e., it has supplemental infor- (happy, sad, surprise, disgust, interest, fear, and
mation provided by the different views of the face). anger) or any facial motion the actor/actress de-
To test this hypothesis, Lander et al. (1999) asked sired. They kept their eyes closed during the re-
participants to identify a number of famous faces. cording sessions. Thus, in the final video sequence,
The faces were presented in three different for- all that was visible were the 100 white points. Each
mats. In one condition, participants saw a nine- participant saw a single display, either as a single
frame video sequence. In a second condition, static snapshot or a full video recording of one ex-
participants saw all nine frames at once, arranged pression, and was asked to describe what they saw.
in an ordered array. In the final condition, the The collection of points was recognized as being a
participants saw all nine frames at once in a jum- face more often in the dynamic conditions than in
bled array. Lander et al. found that the faces were the static conditions (73% vs. 22% of the time,
better recognized in the video condition than in the respectively). Additionally, the sequences were rec-
either of the two static conditions, and that the ognized as containing a face slightly more often for
performances in the two static conditions did not the free-form motion conditions than for the ex-
differ from one another. Thus, the reason why pression conditions (55% vs. 39%, on average, re-
video sequences are recognized better is not simply spectively). In a second experiment, the actors and
that they have more snapshots. actresses were again recorded while performing the
To test whether the advantage is due to motion various emotions. In the first set of recordings, they
in general or due to some specific type of motion, again wore the black makeup and white spots. The
Lander and Bruce (2000) and Pike et al. (1997) second set of recordings was made without make-
presented a video where the images were in a ran- up. Participants were asked to identify the expres-
dom order. Note that such sequences have infor- sion using a forced choice task. Overall, faces were
mation about the motion, but this motion is recognized more often in the full-face condition
random (and does not occur in nature). It was than in the dots-only condition (65% vs. 33% cor-
found that identity is more accurately recognized rect responses, respectively). Critically, the percent-
in normal sequences that in random sequences, age of correct responses in the point-light condition
implying that it is not just the presence of motion (33%) is significantly higher than expected by
that is important, but the specific, naturally oc- chance, suggesting that the temporal information is
curring motion that provides the advantage. Fur- sufficient to recognize expressions. Basilli (1979)
ther, it was found that reversing the direction of went on to examine the role of upper versus lower
motion (by playing the sequence backwards) de- internal facial motion for the recognition of ex-
creases recognition performance, suggesting that pressions and found that different facial areas were
the temporal direction of the motion trajectories is important for different expressions.
important (Lander and Bruce, 2000). Finally, by It remains unclear as exactly what the appro-
changing the speed of a motion sequence (e.g., by priate dynamic features are. One traditional way
playing parts or all of a video sequence too fast or of describing motion is to separate it into rigid and
slow), the researchers showed that the specific nonrigid motions (see, e.g., Gibson, 1957, 1966;
tempo and rhythm of motion is important for face Roack et al., 2003). Rigid face motion generally
recognition (Lander and Bruce, 2000). refers to the rotations and translations of the entire
In a perhaps more direct examination of the role head (such as the one which occurs when someone
of motion, Bassili (1978, 1979) used Johonnson nods his/her head). Nonrigid face motion, in con-
point-light faces as stimuli (see Johonnson, 1973, trast, generally refers to motion of the face itself,
for more information on point-light stimuli). More which consists mostly of nonlinear surface defor-
specifically, the face and neck of several actors and mations (e.g., lip motion, eyebrow motion). Most
actresses were painted black and then covered with naturally occurring face-related motion contains
approximately 100 white spots. These actors and both rigid and nonrigid motion. Indeed, it is very
actresses were then recorded under low light difficult for humans to produce facial (i.e.,
329

nonrigid) motion without moving their head (rigid specific area for the processing of faces exists in
motion), and vice versa. Thus, it should not be the human brain. Neuropsychological evidence
surprising that few studies have systematically ex- for specialization has been derived from proso-
amined the separate contributions of rigid and pagnosia, a deficit in face identification following
nonrigid face motions. Pike et al. (1997) conducted inferior occipitotemporal lesions (e.g., Damasio
one of the few studies to explicitly focus on the et al., 1982; for a review see DeRenzi, 1997).
contribution of rigid motion. They presented a 10 s There have been a few reports of prosopagnostic
clip of an individual rotating in a chair through a patients in which object recognition seemed
full 3601 (representing a simple change in relative to have remained intact (e.g., McNeil and
viewpoint). They found higher identity recognition Warrington, 1993; Farah et al., 1995a; Bentin
performance in dynamic conditions than in static et al., 1999). Prosopagnosia has been regarded as a
conditions. Christie and Bruce (1998), in contrast, face-specific deficit that does not necessarily
presented five frames of a person moving his/her reflect a general disorder in exemplar recognition
head up and down (e.g., representing social com- (e.g., Henke et al., 1998). Consistent with this
munication — a nod of agreement) and found no view, patients who suffered from associative
difference between static and dynamic conditions. object agnosia have been reported, while their
They suggest that the apparent conflict between face identification remained unaffected (e.g.,
the two studies comes from the type of rigid head Moscovitch et al., 1997). Such a double dissocia-
motion: viewpoint change versus social signal. tion between face and object recognition would
Munhall et al. (2004) focused explicitly on the role imply that the two abilities are functionally
of rigid head motion in communication and distinct and anatomically separable. However, on
showed in an elegant study that the specific pat- the basis of methodological concerns, some au-
tern of rigid head motion that accompanies speech thors have doubted whether face recognition can
can be used to disambiguate the speech signal really be dissociated from object recognition
when the audio is degraded. Hill and Johnson based on current literature on prosopagnosia
(2001) used facial animations to show that rigid (e.g., Gauthier et al., 1999a; see also Davidoff
head motion is more useful than nonrigid motion and Landis, 1990).
for identity recognition and that nonrigid motion Evidence for the uniqueness of face processing
was more useful than rigid motion in recognizing has also been derived from event-related potential
the gender of an individual. (ERP) and magnetoencephalographic (MEG)
In sum, it is clear that some form of facial in- studies. A response component called the N170
formation is available only over time, and that it (or M170 in MEG) occurring around 170 ms after
plays an important role in the recognition of iden- stimulus onset is usually twice as large for face
tity, expression, speech, and gender. Moreover, at stimuli when compared to other control stimuli
least several different types of motion seem to ex- such as hands, houses, or animals (e.g., Bentin et
ist, they play different roles, and a simple rigid/ al., 1996; Liu et al., 2002). However, the debate on
nonrigid dichotomy is neither sufficient nor ap- whether such activation is unique for faces or
propriate to describe these motions. Additional whether it represents effects of expertise that are
research is necessary to determine what the dy- not specific to face processing is still ongoing (for
namic features for face processing are. recent discussions see, e.g., Rossion et al., 2002;
Xu et al., 2005).
In functional brain imaging, several areas have
Physiological perspective been identified as being of special importance
for the processing of faces (see Haxby et al., 2000,
Face-selective areas — evidence from neuroscience for a review). These involve a region in the
lateral fusiform gyrus, the superior temporal
At least since the discovery of the face inversion sulcus (STS), and the ‘‘occipital face area’’
effect (Yin, 1969) it has been discussed whether a (OFA; Gauthier et al., 2000a). All areas have been
330

identified bilaterally, albeit with a somewhat found that only the FFA showed a difference in
stronger activation in the right hemisphere. The activity between upright and inverted faces.
face-selective area in the fusiform gyrus has been This can be interpreted as functional dissociation
referred to as the ‘‘fusiform face area’’ (FFA) by between FFA and the other cortical regions in-
Kanwisher et al. (1997). While FFA activation has volved in face processing. The authors also con-
been related to facial identity, the STS in humans cluded that the FFA appears to be the main
reacts particularly to changing aspects of faces neurological source for the behavioral face
with social value, such as expression, direction of inversion effect originally reported by Yin (1969).
gaze, and lip movement (e.g., Puce et al., 1998; The latter, however, is not exclusive to faces.
Hoffman and Haxby, 2000). In a recent functional In a behavioral study, Diamond and Carey
magnetic resonance imaging (fMRI) study using (1986) found comparable inversion effects for
adaptation (reduction of brain activity due to faces and side views of dogs when dog experts
repetitive stimulus presentation), Andrews and were tested. Subsequent behavioral and
Ewbank (2004) investigated differences in face imaging studies using recognition experiments
processing by the FFA versus the STS. Activity in with trained experts and artificial objects (‘‘Gree-
the FFA was reduced over time by stimuli having bles’’) as well as bird and car experts with bird and
the same identity. Adaptation was dependent on car images provided further evidence in favor
viewpoint but not on size changes. The STS of a process-specific interpretation rather than a
showed no adaptation to identity but an increased domain-specific interpretation (Gauthier et al.,
response when the same face was shown with a 1999b, 2000a). According to their view (‘‘expertise
different expression or from different viewpoints. hypothesis’’), FFA activity is related to the
These results suggest a relatively size-invariant identification of different classes of visual stimuli
neural representation in FFA for recognition of if they share the same basic configuration
facial identity, and a separate face-selective region and if substantial visual expertise has been
in STS involved in processing changeable aspects gained. The question on whether FFA activity is
of a face such as facial expression. OFA and in- domain or process specific is being debated
ferior occipital gyrus seem to be associated with since several years now. It is beyond the scope
early structural encoding processes; they are pri- of this chapter to review this ongoing debate but
marily sensitive to sensory attributes of faces for an update on the current status see, for exam-
(Rotshtein et al., 2005). Rossion et al. (2003) ob- ple, Downing et al. (2005), Xu (2005), Bukach
tained results in an fMRI study suggesting et al. (in press), Kanwisher and Yovel (in press).
that OFA and FFA might be functionally associ- Nevertheless, it should be noted that activation in
ated: PS, a patient suffering from severe pros- face-selective regions of the fusiform area is not
opagnosia due to lesions in the left middle exclusive to faces. Significant responses to other
fusiform gyrus and the right inferial occipital cor- categories of objects have been found in normal
tex, performed poorly in a face-matching task de- subjects, for example, for chairs, houses, and tools
spite normal activation of the intact right FFA. (Chao et al., 1999; Ishai et al., 1999, 2000; Haxby
Rossion et al. thus concluded that the FFA et al., 2001). Moreover, it has also been shown that
alone does not represent a fully functional mod- face-selective regions in the fusiform area can be
ule for face perception, but that for normal face modulated by attention, emotion, and visual im-
processing intact OFA and FFA in the right agery, in addition to modulation by expertise as
hemisphere with their re-entrant integration are mentioned above (e.g., O’Craven et al., 1999;
necessary. Yovel and Kanwisher (2005) came Vuilleumier et al., 2001; Ishai et al., 2002). In re-
to a different conclusion. They correlated the be- cent years, substantial progress has been made re-
havioral performance in a face-matching task of garding models on how different brain areas
upright and inverted faces with the neuronal re- interact in processing information contained in
sponses to upright and inverted faces in the faces. Three main accounts are summarized in the
three regions: FFA, STS, and OFA. It was following section.
331

Cognitive neuroscience models of face processing which are forwarded to the STS via the middle
temporal (MT) area; and person-specific motion
The model by Bruce and Young (1986) is one of (‘‘dynamic facial signatures’’). O’Toole et al. sug-
the most influential accounts in the psychological gest that the latter type of information is also
face processing literature. This framework pro- processed by the STS, representing an additional
poses parallel routes for recognizing facial identity, route for familiar face recognition. This model is
facial expression, and speech-related movements in accordance with the supplemental information
of the mouth. It is a rather functional account hypothesis that claims that facial motion consti-
since Bruce and Young did not provide specifics tutes additional information to static information.
regarding the neural implementation of their According to O’Toole et al., structure-from-mo-
model. The recent physiological framework pro- tion may also support face recognition by com-
posed by Haxby et al. (2000) is consistent with the munication between the ventral and the dorsal
general conception proposed by Bruce and streams. For instance, the structural representa-
Young. According to Haxby et al.’s model, the tion in FFA could be enhanced by input from MT.
visual system is hierarchically structured into a Thus, the model also integrates the representation
core and an extended system. The core system enhancement hypothesis.
comprises three bilateral regions in occipitotem- In a detailed review of psychological and neural
poral visual extrastriate cortex: inferior occipital mechanisms, Adolphs (2002) provides a descrip-
gyrus, lateral fusiform gyrus, and STS. Their func- tion of the processing of emotional facial expres-
tion is the visual analysis of faces. Early perception sions as a function of time. The initial stage
of facial features and early structural encoding provides automatic fast perceptual processing of
processes are mediated by processing in inferior highly salient stimuli (e.g., facial expressions of
occipital gyrus. The lateral fusiform gyrus anger and fear). This involves the superior col-
processes invariant aspects of faces as the basis liculus and pulvinar, as well as activation of the
for the perception of unique identity. Changeable amygdala. Cortical structures activated in this
properties such as eye gaze, expression, and lip stage are V1, V2, and other early visual cortices
movement are processed by STS. The representa- that receive input from the lateral geniculate nu-
tions of changeable and invariant aspects of faces cleus of the thalamus. Then, a more detailed struc-
are proposed to be independent of one another, tural representation of the face is constructed until
consistent with the Bruce and Young model. The about 170 ms. This processing stage involves the
extended system contains several regions involved fusiform gyrus and the superior temporal gyrus,
in other cognitive functions such as spatially which is consistent with Haxby et al.’s core system.
directed attention (intraparietal sulcus), prelexical Dynamic information in the stimulus would en-
speech perception (auditory cortex), emotion gage MT, middle superior temporal area, and
(amygdala, insula, limbic system), and personal posterior parietal visual cortices. Recognition
identity, name, and biographical information modules for detailed perception and emotional re-
(anterior temporal region). action involve Haxby et al.’s extended system. Af-
The model of Haxby et al. has been taken as a ter 300 ms conceptual knowledge of the emotion
framework for extension by O’Toole et al. (2002). signaled by the face is based on late processing in
By taking into account the importance of dynamic the fusiform and superior temporal gyri, orbito-
information in social communication, they frontal and somatosensory cortices, as well as ac-
further explain the processing of facial motion. tivation of the insula.
In their system, dynamic information is processed The assumption of separate processes for facial
by the dorsal stream of face recognition and static identity and facial expression is supported by a
information is processed by the ventral stream. number of studies. Neuropsychological evidence
Two different types of information are contained suggests a double dissociation; some patients show
in facial motion: social communication signals impairment in identity recognition but normal
such as gaze, expression, and lip movements, emotion recognition, and other patients show
332

intact identity recognition but impaired emotion general systems than damage (or impaired access)
recognition (for reviews see Damasio et al., 1982, to visual representations of facial expression. The
1990; Wacholtz, 1996; Adolphs, 2002). In a recent authors do not completely reject the dissociation
study, Winston et al. (2004) revealed dissociable of identity and expression, but they suggest that
neural representations of identity and expression the bifurcation takes place at a much later stage
using an fMRI adaptation paradigm. They found than that proposed by the model of Haxby et al.,
evidence for identity processing in fusiform cortex namely only after a common representational sys-
and posterior STS. Coding of emotional expres- tem. This alternative approach is supported by
sion was related to a more anterior region of STS. computational modeling studies using principal
Bobes et al. (2000) showed that emotion matching component analysis (PCA; see next section). A
resulted in a different ERP scalp topography com- critical problem of these approaches, however, is
pared to identity matching. In another ERP study, that they rely on a purely holistic processing strat-
Eimer and Holmes (2002) investigated possible egy of face stimuli, which in light of the previously
differences in the processing of neutral versus discussed behavioral evidence seems not plausible.
fearful facial stimuli. They found that the N170, As discussed in the previous section, there is a
which is related to structural encoding of the face growing number of studies in the psychophysics
in processing identity, did occur in both the neu- literature that clearly suggest an important role of
tral and the fearful conditions. This indicates that both component and configural information in
structural encoding is not affected by the presence face processing. This is supported by neurophys-
of emotional information and is also consistent iological studies. In general, it has been found that
with independent processing of facial expression cells responsive to facial identity are found in in-
and identity. However, results from other studies ferior temporal cortex while selectivity to facial
challenge the assumption of completely independ- expressions, viewing angle, and gaze direction can
ent systems. DeGelder et al. (2003) found that be found in STS (Hasselmo et al., 1989; Perret
subjects suffering from prosopagnosia performed et al., 1992). For some neurons, selectivity for
much better when faces showed emotions than particular features of the head and face, e.g. the
when they depicted a neutral expression. With eyes and mouth, has been revealed (Perret et al.,
normal subjects, the case was the opposite. DeG- 1982, 1987, 1992). Other groups of cells need the
elder et al. assume that the areas associated with simultaneous presentation of multiple parts of a
expression processing (amygdala, STS, parietal face, which is consistent with a more holistic type
cortex) have a modulatory role in face identifica- of processing (Perret and Oram, 1993; Wachsmuth
tion. Their findings challenge the notion that et al., 1994). Yamane et al. (1988) have discovered
different aspects of faces are processed independ- neurons that detect combinations of distances bet-
ently (assumption of dissociation) and only after ween facial parts, such as the eyes, mouth, eye-
structural encoding (assumption of hierarchical brows, and hair, which suggest sensitivity for the
processing). Calder and Young (2005) share a spatial relations between facial parts (configural
similar view. They argue that a successful proof of information).
the dissociation of identity and expression would Although they are derived from different phys-
require two types of empirical evidence. First, pa- iological studies, the three models by Haxby,
tients with prosopagnosia but without any impair- O’Toole et al., and Adolphs share many common
ment in facial expression recognition. Second, features. Nevertheless, it seems that some links to
intact processing of facial identity and impaired behavioral and physiological studies are not taken
recognition of emotion without impairment of up in these models. As discussed above, the con-
other emotional functions. On the basis of their cept of component and configural processing
review the authors conclude that such clear seems to be a prominent characteristic of face
patterns have not been revealed yet. The reported processing. The models, however, do not make this
selective disruption of facial expression recogni- processing step explicit by specifying at which
tion would rather reflect an impairment of more stage this information is extracted. Furthermore,
333

the distributed network of brain regions involved interaction in which the actions of the user have
in the processing of face stimuli has so far not been to be analyzed and recognized in a temporal
characterized in terms of the features that are context.
processed in each region — how does a face look As the previous sections have shown, however,
like for the amygdala, for example? Some of these apart from having commercial prospects, tech-
questions may be answered in connection with a niques developed by the computer vision commu-
closer look at the computational properties of face nity also have wide uses in cognitive research: by
recognition. In the next section, we therefore analyzing and parametrizing the high-dimensional
present a brief overview of computational models space of face appearances, for example, research-
of identity and expression recognition. ers gain access to a high-level, statistical descrip-
tion of the underlying visual data. This description
can then be used to design experiments in a well-
Computational perspective defined subspace of facial appearance (for a review
of face spaces see Valentine, 1995). A well-known
Since the advent of the field of computer vision example consists of the PCA of faces that defines
(Marr, 1982), face recognition has been and con- prototypes in a face space (Leopold et al., 2001).
tinues to be one of its best-researched topics with The same holds true in the case of facial expres-
hundreds of papers being published each year in sions as the amount of spatio-temporal data
conferences and journals. One reason for this in- quickly becomes prohibitive in order to conduct
tense interest in face recognition is certainly due to controlled experiments at a more abstract level
the growing range of commercial applications for that goes beyond mere pixels. A recent study that
computational face recognition systems — espe- has used advanced computer vision techniques to
cially in the areas of surveillance and biometrics, manipulate components of facial expressions is the
but increasingly also in other areas such as hu- study by Cunningham et al. (2005). In the follow-
man–computer interaction or multimedia applica- ing, we will briefly review the main advances and
tions. Despite these tremendous efforts, however, approaches in both the area of identity recognition
even today there exists no single computational and recognition of expressions (see Li and Jain,
system that is able to match human performance 2004, for further discussion).
— both in terms of recognition discriminability As a first observation, it is interesting to note
and in terms of generalization to new viewing that both identity and expression recognition in
conditions including changes in lighting, pose, computer vision follow the same basic structure: in
viewing distance, etc. It is especially this fact that the first step, the image is scanned in order to find
has led to a growing interest of the computer vi- a face — this stage is usually called face detection
sion community in understanding and applying and can also encompass other tasks such as esti-
the perceptual, cognitive, and neurophysiological mating the pose of the face. As a result of space
issues underlying human performance. Similar restrictions, we will not deal with face detection
statements could be made about the area of auto- explicitly — rather, the reader is referred to
matic recognition of facial expressions — the crit- Hjelmas and Low (2001). Interestingly, the topic
ical difference being that commercial interest in of face detection has received relatively little at-
such systems is less than in systems that can per- tention in cognitive research so far (see, e.g., Lewis
form person identification. Nevertheless, the area and Edmonds, 2003) and needs to be further ex-
of expression recognition continues to be a very plored. Following face detection, in a second step
active topic in computer vision because it deals the image area that comprises the face is further
with the temporal component of visual input: how analyzed to extract discriminative features ranging
the face moves and how computers might be able from holistic approaches using the pixels them-
to map the space of expressions are of interest for selves to more abstract approaches extracting the
computer vision researchers leading to potential facial components. Finally, the extracted features
applications in, for example, human–computer are compared to a database of stored identities or
334

expressions in order to recognize the person or In the early 1990s, Turk and Pentland (1991)
their expression. This comparison can be done by developed a holistic recognition system called ‘‘ei-
a range of different classification schemes from genfaces,’’ which used the full pixel information to
simple, winner-take-all strategies to highly com- construct an appearance-based low-dimensional
plex algorithms from machine learning. representation of faces. This approach proved to
Research in face recognition can be roughly di- be very influential for computer vision in general
vided into three areas following the type of infor- and inspired many subsequent recognition algo-
mation that is used to identify the person in the rithms. Its success is partially due to the fact that
feature extraction step: (1) Holistic approaches natural images contain many statistical redundan-
use the full image pixel information of the area cies. These can be exploited by algorithms such as
subtended by the face. (2) Feature-based ap- PCA by building lower-dimensional representa-
proaches try to extract more abstract information tions that capture the underlying information con-
from the face area ranging from high-contrast tained in, for example, the space of identities given
features to semantic facial features. (3) Hybrid a database of faces. The result of applying PCA to
systems combine these two approaches. such a database of faces is a number of eigenvec-
The earliest work in face recognition focused tors (the ‘‘eigenfaces’’) that encode the main sta-
almost exclusively on high-level, feature-based ap- tistical variations in the data. The first eigenvector
proaches. Starting in the 1970s, several systems is simply the average face and corresponds to the
were proposed which relied on extracting facial prototype face used in psychology. Recognition of
features (eyes, mouth, and nose) and in a second a new face image is done by projecting it into the
step calculated two-dimensional geometric prop- space spanned by the eigenvectors and looking for
erties of these features (Kanade, 1973). Although it the closest face in that space. This general idea of a
was shown that recognition using only geometric face space is shared by other algorithms such as
information (such as distances between the eyes, linear discriminant analysis (LDA; Belhumeur
the mouth, etc.) was computationally effective and et al., 1997), independent component analysis
efficient, the robust, automatic extraction of such (ICA; Bartlett et al., 2002), non-negative matrix
high-level facial features has proven to be very factorization (NMF; Lee and Seung, 1999), or
difficult under general viewing conditions (Brunelli support vector machines (SVMs; Phillips, 1999).
and Poggio, 1993). One of the most successful face The main difference between these algorithms lies
recognition systems based on local image infor- in the statistical description of the data as well as
mation therefore used much simpler features that in the metrics used to compare different elements
were supplemented by rich feature descriptors: in of the face space: PCA and LDA usually result in
the elastic bunch-graph matching approach, a face holistic descriptions of the data where every region
image is represented as a collection of nodes which of the face contributes to the final result; ICA and
are placed in a regular grid over the face. Each of NMF can yield more sparse descriptors with spa-
these nodes carries so-called ‘‘jets’’ of Gabor-filter tially localized responses; SVMs describe the space
responses, which are collected over various scales of face identities through difficult-to-recognize
and rotations. This representation is very compact face exemplars (the support vectors) rather than
yet has proven to be very discriminative, therefore through prototypical faces as PCA does. In terms
enabling good performance even under natural of metrics, there are several possibilities ranging
viewing conditions (Wiskott et al., 1997). It is from simple Euclidean distances to weighted
interesting to note this system’s similarity to the distances, which can take the statistical properties
human visual system (see Biederman and Kalocsai, of the face space into account.
1997) as the Gabor filters used closely resemble the The advantage of PCA (and other holistic ap-
receptive field structure found in the human cor- proaches) in particular is that it develops a gen-
tex. The advantage of such low-level features as erative model of facial appearance which enables
used also in later recognition systems lies in their it, for example, to reconstruct the appearance of a
conceptual simplicity and compactness. noisy or occluded input face. An extreme example
335

of this is the morphable model by Blanz and Vetter Tolerance to changes in viewing conditions: Ex-
(1999, 2003), which does not work on image pixels isting databases often contain only frontal im-
but works on three-dimensional (3D) data of laser ages or a single illumination. In the real world,
scans of faces. Because of their holistic nature, however, changes in illumination and pose are
however, all of these approaches require specially rather frequent — usually several changes occur
prepared training and testing data with very care- at the same time.
fully aligned faces in order to work optimally. Dealing with large amounts of training data:
Given the distinction between local and holistic Connected to the previous point, typical recog-
approaches, it seems natural to combine the two nition systems need massive amounts of train-
into hybrid recognition architectures. Eigenfaces ing data to learn facial appearance under
can of course be extended to ‘‘eigenfeatures’’ by various viewing conditions. While this is rem-
training facial features instead of whole images. iniscent of the extensive experience humans ac-
Indeed, such systems have been shown to work quire with faces, this represents a challenge for
much better under severe changes in the appear- the statistical algorithms as the relevant dis-
ance of the face such as due to occlusion by other criminative information has to be extracted
objects or make-up (see Swets and Weng, 1996). from the images.
Another system uses local information extracted Taking into account the context: Faces seldom
from the face to fit a holistic shape model to the appear without context — in this case context
face. For recognition, not only holistic informa- can mean a particular scene, evidence from
tion is used, but also local information from the other modalities, or the fact that faces are part
contour of the face (Cootes et al., 2001). Finally, in of the human body and therefore co-occur. This
a system proposed by Heisele et al. (2003), several information could be used not only for more
SVMs are trained to recognize facial features in an reliable face detection, but could also assist in
image, which are then combined into a configura- recognition tasks. Interestingly, context effects
tion of features by a higher-level classification are also not well studied in the behavioral lit-
scheme. Again, such a scheme has been shown to erature.
outperform other, purely holistic, approaches. Dealing with spatio-temporal data: Even though
Recently, several rigorous testing schemes have humans can recognize faces from still images,
been proposed to evaluate the various methods we generally experience the world dynamically
proposed by computer vision researchers: the (see Section ‘‘Dynamic information’’ above).
FERET (FacE REcognition Technology) Although first steps have been made (Li and
evaluations in 1994–1996 and three face recogni- Chellappa, 2001), the full exploitation of this
tion vendor tests in 2000, 2002, and 2005 (see fact remains to be done.
http://www.frvt.org). Although these evaluations Characterizing performance with respect to hu-
have shown that performance has increased mans: Although the claim that current systems
steadily over the past years, several key challenges do not yet reach human performance levels is
still need to be addressed before face recognit- certainly valid, there has been relatively little
ion systems can become as good as their human research on trying to relate computational and
counterpart: human performance in a systematic manner
(examples include the studies by Biederman and
Robust extraction of facial features: Although Kaloscai, 1997; O’Toole et al., 2000; Wallraven
face detection is possible with very high accu- et al., 2002, 2005b; Schwaninger et al., 2004).
racy, this accuracy is usually achieved by anal- Such information could be used to fine-tune
yzing large databases of face and nonface existing as well as develop novel approaches to
images to extract statistical descriptors. These face recognition.
statistical descriptors usually do not conform to
meaningful face components — components In addition to work on face recognition, con-
that humans rely on to detect faces. siderable attention has been devoted to the
336

automatic recognition of facial expressions. Early important to stress again that the co-activation of
work in this area has focused mainly on recogni- action units into complex expressions is external to
tion of the six prototypical or universal expres- FACS, making it a purely descriptive rather than
sions (these are angry, disgust, fear, happy, sad, an inferential system. In addition, the mapping
and surprised; see Ekman et al., 1969) whereas in from action units to complex expressions is am-
later work the focus has been to provide a more biguous. Nevertheless, most recent systems try to
fine-grained recognition and even interpretation of recognize action units from still images or image
core components of facial expressions. sequences — perhaps due to the fact that FACS is
As mentioned above, all computational systems one of the few parametric, high-level descriptions
follow the same basic structure of face detection, of facial motion. As a result of the highly localized
feature extraction, and classification. Moreover, structure of action units, these systems rely mostly
one can again divide the proposed systems into on local information and use a combination of
different categories on the basis of whether they low-level, appearance-based features and geomet-
use holistic information, local feature-based infor- ric facial components for increased robustness
mation, or a hybrid approach. An additional as- (Bartlett et al., 1999; Donato et al., 1999). One of
pect is whether the systems estimate the the most advanced recognition systems (Tian et
deformation of a neutral face for each image or al., 2001) uses a hybrid scheme combining both
whether they rely explicitly on motion to detect a low- and high-level local information in a neural
facial expression. network to recognize 16 action units from static
Systems based on holistic information have em- images with an accuracy of 96%.
ployed PCA (Calder et al., 2001) to recognize Similarly to face recognition systems there exist
static images, estimating dense optic flow to anal- several standard databases for benchmarking fa-
yze the deformation of the face in two dimensions cial expression algorithms (Ekman and Friesen,
(Bartlett et al., 1999), as well as 3D deformable 1978; Kanade et al., 2000) — so far, however, no
face models (DeCarlo and Metaxas, 1996). In con- comprehensive benchmark comparing the differ-
trast, systems based on local information rely on ent systems for facial expression analysis has been
analyzing regions in the face that are prone to developed.
changes under facial expressions. Such systems Interestingly, several of the systems discussed
have initially used tracking of 2D contours (Ter- here have been explicitly benchmarked against hu-
zopoulos and Waters, 1993) or high-contrast re- man performance — both in the case of prototyp-
gions (Rosenblum et al., 1996; Black and Yacoob, ical expressions (Calder et al., 2001) and in the case
1997), elastic bunch-graph matching with Gabor of action unit recognition (Bartlett et al., 1999;
filters (Zhang et al., 1998), as well higher-level 3D Tian et al., 2001). This shows that in the area of
face models (Gokturk et al., 2002). Despite good facial expression recognition, the coupling between
recognition results on a few existing test databases, psychological and computational research is much
however, these systems mostly focused on recog- closer than that seen in identity recognition — one
nition of the six prototypical expressions. There- of the reasons may be that expression analysis has
fore, they could not extract important dimensions drawn more heavily from results in psychological
such as the intensity of the recognized expressions research (Ekman et al., 1969; Ekman and Friesen,
— dimensions, which are critical for human–com- 1978).
puter interface (HCI) applications, for example. Finally, the following presents a list of key chal-
A very influential description of facial expres- lenges in the area of automatic recognition of fa-
sions is FACS developed by Ekman and Friesen in cial expressions that still need to be addressed to
the late 1960s and continuously improved in the design and implement robust systems with human-
following years. As mentioned above (see section like performance:
on recognition of expressions), FACS encodes
both anatomical muscle activations as well as so- Robust extraction of facial features: Even more
called miscellaneous actions in 44 action units. It is than in the case of identity recognition,
337

extraction of the exact shape of facial compo- and their spatial relationship (configural informa-
nents would enable to determine action units tion). The integrative model by Schwaninger et al.
very precisely. (2002, 2003a) provides a good basis for combining
Dealing with variations in appearance: In con- the component-configural hypothesis and holistic
trast to identity recognition, expressions need to aspects of face processing. According to the model,
be recognized despite changes in gender and component and configural information are first
identity (as well as of course additional param- analyzed separately and then integrated for recog-
eters such as pose, lighting, etc.). Current algo- nition. Rotating the faces in the plane results in a
rithms do not yet perform these generalizations strong impairment of configural processing, while
well enough. component processing is much less, if at all,
Going beyond FACS: Although FACS has re- affected by plane rotation. This could be due to
mained popular, it is still unclear whether it re- capacity limitations of an orientation normaliza-
ally is a useful basis for describing the space of tion mechanism such as mental rotation, which is
facial expressions — both from a human per- required in order to extract configural information
spective as well as from a statistical point of from the plane rotated or inverted faces. Because
view. Among the alternatives that have been adult face recognition relies more on the process-
proposed is FACS+, an encoding system for ing of configural information than basic level ob-
facial actions that seeks to determine well-de- ject recognition, a strong inversion effect is
fined actions as well as the mapping between obtained.
these actions and complex expressions (Essa Different facial areas and facial motions are im-
and Pentland, 1997). Another alternative is the portant for the recognition of different emotions.
MPEG4 face-coding scheme, a very generic face Most of the models on facial expression processing
animation framework based on movement of have stressed the importance of component infor-
keypoints on the face (e.g., Koenen, 2000; see mation while some models also integrate con-
also Section ‘‘Dynamic information’’ above). figural information (e.g., Izard et al., 1983; White,
Full spatio-temporal, high-level models for rec- 2000). As pointed out by Schwaninger et al. (2002,
ognition: Facial expressions are a highly efficient 2003a), a model which assumes separate process-
form of communication — the communicative ing of component and configural information be-
aspect is not yet exploited by current systems, fore integrating them can explain the effects of
which is partially due to the lack of explicit facial expression processing in the Thatcher illu-
models of how humans employ facial motions sion and the face composite illusion. In upright
to convey meaning. Such knowledge could, for faces, both component and configural information
example, prove useful as a high-level prior for can be processed. This results in a bizarre facial
automatic recognition of expressions. expression in the Thatcher illusion and in a new
More descriptive power: For humans, recogni- identity or facial expression when different face
tion of the expression is only the first step in a composites are aligned. Turning faces upside-
longer pipeline, which involves not only judg- down disrupts configural processing. As a conse-
ments of intensity, but also other interpretative quence, the bizarre facial expression in the
dimensions such as believability, naturalness, Thatcher illusion vanishes. Similarly, the absence
etc. Our behaviour may be more determined by of interference from configural information in in-
these dimensions rather than by the classifica- verted composites makes it easier to identify the
tion itself. different identities (Young et al., 1987) or emo-
tions (Calder et al., 2000).
The model by Bruce and Young (1986) proposes
Summary and conclusions separate parallel routes for recognizing facial iden-
tity, facial expression, and speech. Recent physio-
The review of psychophysical studies showed that logical models proposed by Haxby et al. (2000),
faces are processed in terms of their components O’Toole et al. (2002), and Adolphs (2002) are
338

consistent with this view. Although now much is model of component and configural processing in
known about the role and interaction of different identity recognition. On the basis of the model
brain areas in recognition of identity and expres- developed by Schwaninger et al. (2003a, 2004)
sion, the neuronal implementation of analyzing outlined above, they implemented the two routes
component and configural information and their of face processing using methods from computer
integration with motion information is less clear. vision. The building blocks of the face represen-
Interestingly, both identity and expression rec- tation consisted of local features that were ex-
ognition in computer vision follow the same basic tracted at salient image regions at different spatial
processing steps. Following face detection, in a frequency scales. The basic idea for the implemen-
second step the image area containing a face is tation of the two routes was that configural
processed to extract discriminative features, which processing should be based on a global, position-
are then compared to a database of stored iden- sensitive connection of these features, whereas
tities or expressions. The different recognition al- component processing should be local and posi-
gorithms can be distinguished on whether they use tion insensitive. Using these relatively simple in-
holistic information, local feature-based informa- gredients, they showed that the model could
tion, or a hybrid approach; the last two are usually capture the human performance pattern observed
more successful than the first one. One example in the psychophysical experiments. As can be seen
of a close connection between computational in Fig. 4, the results of the computational model
modeling and psychophysical research in this con- are very similar to the psychophysical results ob-
text is the set of studies by Wallraven et al. (2002, tained with humans in the experiment conducted
2005b) on the implementation and validation of a by Schwaninger et al. (2002) (see also Fig. 3).

Fig. 4. Human and computational (Comp) performance for the face recognition task of Schwaninger et al. (2002) for unfamiliar
(Unfam) and familiar (Fam) face recognition (adapted from Wallraven et al., 2005b). Performance is shown as area under the ROC-
curve (AUC). In order to determine the relative contributions of component and configural processing, participants had to recognize
previously learnt faces in either scrambled (Scr), blurred (Blr), or scrambled-blurred (ScrBlr) conditions (also see human data in Fig. 3).
The recognition performance of the computational model (Comp) was very similar to human performance. Moreover, the results
indicate that the two-route implementation of configural and component processing captures the relative contributions of either route
to recognition. The observed increase for familiar face recognition by humans could also be modeled with the computational system.
339

Moreover, combining the two routes resulted in Bartlett, M.S., Movellan, J.R. and Sejnowski, T.J. (2002) Face
increased recognition performance, which has im- recognition by independent component analysis. IEEE
Trans. Neural Networks, 13(6): 1450–1464.
plications in a computer vision context regarding
Bassili, J.N. (1978) Facial motion in the perception of faces and
recognition despite changes in viewpoint (see Wall- of emotional expression. J. Exp. Psychol.: Hum. Percep.
raven et al., 2005b). Perform., 4(3): 373–379.
In general, a closer coupling between psycho- Bassili, J. (1979) Emotion recognition: the role of facial motion
physical research, neuroscience, and computer vi- and the relative importance of upper and lower areas of the
sion would benefit all research areas by enabling a face. J. Pers. Soc. Psychol., 37: 2049–2059.
Belhumeur, P., Hespanha, J. and Kriegman, D. (1997) Eigen-
more advanced statistical analysis of the informa- faces vs. fisherfaces: recognition using class specific linear
tion necessary to recognize individuals and ex- projection. IEEE Trans. Pattern Anal.Mach. Intell., 19(7):
pressions as well as the development of better, 711–720.
perceptually motivated recognition algorithms Bentin, S., Allison, T., Puce, A., Perez, E. and McCarthy, G.
(1996) Electrophysiological studies of face perception in hu-
that are able to match human classification per-
man. J. Cogn. Neurosci., 8: 551–565.
formance. This will be necessary in order to better Bentin, S., Deouell, L.Y. and Soroker, N. (1999) Selective vis-
understand processing of component, configural, ual streaming in face recognition: evidence from develop-
and motion information and their integration for mental prosopagnosia. NeuroReport, 10: 823–827.
recognition of identity and facial expression. Biederman, I. (1987) Recognition-by-components: a theory of
human image understanding. Psychol. Rev., 94(2): 115–147.
Biederman, I. and Kalocsai, P. (1997) Neurocomputational
Abbreviations bases of object and face recognition. Philos. Trans. R. Soc.
Lond. Ser. B, 352: 1203–1219.
Black, M. and Yacoob, Y. (1997) Recognizing facial expres-
FACS facial action coding system sions in image sequences using local parameterized models of
FERET FacE REcognition Technology image motion. Int. J. Comput. Vis., 25(1): 23–48.
FFA fusiform face area Blanz, V. and Vetter, T. (1999) A morphable model for the
FLMP fuzzy logical model of perception synthesis of 3D faces. SIGGRAPH’99, Conference Proceed-
ings, Los Angeles, CA, USA, pp. 187–194.
HCI human–computer interface Blanz, V. and Vetter, T. (1993) Face recognition based on fit-
ICA independent component analysis ting a 3D morphable model. IEEE Trans. Pattern Anal.
LDA linear discriminant analysis Mach. Intell., 25: 1063–1074.
MT middle temporal Bobes, M.A., Martı́n, M., Olivares, E. and Valdés-Sosa, M.
NMF non-negative matrix factorization (2000) Different scalp topography of brain potentials related
to expression and identity matching of faces. Cogn. Brain
OFA occipital face area Res., 9: 249–260.
STS superior temporal sulcus Boutet, I., Collin, C. and Faubert, J. (2003) Configural face
SVM support vector machine encoding and special frequency information. Percept. Psy-
chophys., 65(7): 1087–1093.
Bruce, V. (1988) Recognising Faces. Lawrence Erlbaum Asso-
ciates, Hillsdale, NJ.
Bruce, V., Henderson, Z., Greenwood, K., Hancock, P.J.B.,
References Burton, A.M. and Miller, P. (1999) Verification of face iden-
tities from images captured on video. J. Exp. Psychol. Appl.,
Adolphs, R. (2002) Recognizing emotion from facial expres- 5(4): 339–360.
sions: psychological and neurological mechanisms. Behav. Bruce, V., Henderson, Z., Newman, C. and Burton, M.A.
Cogn. Neurosci. Rev., 1(1): 21–61. (2001) Matching identities of familiar and unfamiliar faces
Andrews, T.J. and Ewbank, M.P. (2004) Distinct representa- caught on CCTV images. J. Exp. Psychol.: Appl., 7: 207–218.
tions for facial identity and changeable aspects of faces in Bruce, V. and Young, A. (1986) Understanding face recogni-
human temporal lobe. NeuroImage, 23: 905–913. tion. Brit. J. Psychol., 77: 305–327.
Bahrick, H.P., Bahrick, P.O. and Wittlinger, R.P. (1975) Fifty Brunelli, R. and Poggio, T. (1993) Face recognition: features
years of memory for names and faces: a cross-sectional ap- versus templates. IEEE Trans. Pattern Anal. Mach. Intell.,
proach. J. Exp. Psychol.: Gen., 104: 54–75. 15(10): 1042–1052.
Bartlett, M.S., Hager, J.C., Ekman, P. and Sejnowski, T.J. Bukach, C.M., Gauthier, I. and Tarr, M.J. (in press) Beyond
(1999) Measuring facial expressions by computer image anal- faces and modularity: the power of an expertise framework.
ysis. Psychophysiology, 36: 253–263. Trends Cogn. Sci., 10(4): 159–166.
340

Calder, A.J., Burton, A.M., Miller, P., Young, A.W. and Aka- Duchenne, B. (1990) The Mechanism of Human Facial Ex-
matsu, S. (2001) A principal component analysis of facial pression or an Electro-Physiological Analysis of the Expres-
expressions. Vis. Res., 41: 1179–1208. sion of the Emotions. Cambridge University Press, New
Calder, A.J. and Young, A.W. (2005) Understanding the rec- York.
ognition of facial identity and facial expression. Nat. Rev. Edwards, K. (1998) The face of time: temporal cues in facial
Neurosci., 6: 641–651. expressions of emotion. Psychol. Sci., 9: 270–276.
Calder, A.J., Young, A.W., Keane, J. and Deane, M. (2000) Eimer, M. and Holmes, A. (2002) An ERP study on the time
Configural information in facial expression perception. J. course of emotional face processing. Cogn. Neurosci. Ne-
Exp. Psychol., 26(2): 527–551. uropsychol., 13(4): 427–431.
Chao, L.L., Haxby, J.V. and Martin, A. (1999) Attribute-based Ekman, P. and Friesen, W.F. (1978) Facial Action Coding
neural substrates in temporal cortex for perceiving and System. Consulting Psychologists Press, Palo Alto.
knowing about objects. Nat. Neurosci., 2: 913–919. Ekman, P. and Friesen, W.V. (1982) Felt, false, and miserable
Christie, F. and Bruce, V. (1998) The role of dynamic infor- smiles. J. Nonverb. Behav., 6: 238–252.
mation in the recognition of unfamiliar faces. Mem. Cogn., Ekman, P., Hager, J., Methvin, C. and Irwin, W. (1969) Ek-
26(4): 780–790. man-Hager Facial Action Exemplars. Human Interaction
Collishaw, S.M. and Hole, G.J. (2000) Featural and configu- Laboratory, University of California, San Francisco.
rational processes in the recognition of faces of different fa- Ellison, J.W. and Massaro, D.W. (1997) Featural evaluation,
miliarity. Perception, 29: 893–910. integration, and judgment of facial affect. J. Exp. Psychol.:
Cootes, T., Edwards, G. and Taylor, C. (2001) Active appear- Hum. Percept. Perform., 23(1): 213–226.
ance models. IEEE Trans. Pattern Anal. Mach. Intell., 23: Essa, I. and Pentland, A. (1994) A vision system for observing
681–685. and extracting facial action parameters. Proceedings of the
Cunningham, D., Kleiner, M., Wallraven, C. and Bülthoff, H. International Conference on Computer Vision and Pattern
(2005) Manipulating video sequences to determine the com- Recognition (CVPR’94), Seattle, WA, USA, pp. 76–83.
ponents of conversational facial expressions. ACM Trans. Essa, I. and Pentland, A. (1997) Coding, analysis, interpretation
App. Percept., 2(3): 251–269. and recognition of facial expressions. IEEE Trans. Pattern
Damasio, A.R., Damasio, H. and Van Hoesen, G.W. (1982) Anal. Mach. Intell., 19: 757–763.
Prosopagnosia: anatomic bases and behavioral mechanisms. Farah, M.J., Levinson, K.L. and Klein, K.L. (1995a) Face
Neurology, 32: 331–341. perception and within-category discrimination in pros-
Damasio, A.R., Tranel, D. and Damasio, H. (1990) Face opagnosia. Neuropsychologia, 33: 661–674.
agnosia and the neural substrates of memory. Annu. Rev. Farah, M.J., Tanaka, J.W. and Drain, H.M. (1995b) What
Neurosci., 13: 89–109. causes the face inversion effect? J. Exp. Psychol.: Hum. Per-
Darwin, C. (1872) The Expression of the Emotions in Man and cept. Perform., 21(3): 628–634.
Animals. John Murray, London. Frijda, N.H. and Philipszoon, E. (1963) Dimensions of recog-
Davidoff, J. and Donnelly, N. (1990) Object superiority: a com- nition of emotion. J. Abnorm. Soc. Psychol., 66: 45–51.
parison of complete and part probes. Acta Psychol., 73: 225–243. Frois-Wittmann, J. (1930) The judgment of facial expression. J.
Davidoff, J. and Landis, T. (1990) Recognition of unfamiliar Exp. Psychol., 13: 113–151.
faces in prosopagnosia. Neuropsychologia, 28: 1143–1161. Gauthier, I., Behrmann, M. and Tarr, M.J. (1999a) Can face
DeCarlo, D. and Metaxas, D., 1996. The integration of optical recognition be dissociated from object recognition? J. Cogn.
flow and deformable models with applications to human face Neurosci., 11: 349–370.
shape and motion estimation. Proceedings of the Interna- Gauthier, I., Skudlarski, P., Gore, J.C. and Anderson, A.W.
tional Conference on Computer Vision and Pattern Recog- (2000a) Expertise for cars and birds recruits brain areas in-
nition (CVPR ‘96), San Francisco, CA, USA, pp. 231–238. volved in face recognition. Nat. Neurosci., 3: 191–197.
DeGelder, B., Frissen, I., Barton, J. and Hadjikhani, N. (2003) Gauthier, I., Tarr, M.J., Anderson, A.W., Skudlarski, P. and
A modulatory role for facial expressions in prosopagnosia. Gore, J.C. (1999b) Activation of the middle fusiform area
Proc. Natl. Acad. Sci. USA, 100(22): 13105–13110. increases with expertise in recognizing novel objects. Nat.
DeRenzi, E. (1997) Prosopagnosia. In: Feinberg, T.E. and Far- Neurosci., 6: 568–573.
ah, M.J. (Eds.), Behavioral Neurology and Neuropsycholo- Gauthier, I., Tarr, M.J., Moylan, J., Skudlarski, P., Gore, J.C.
gy. McGraw-Hill, New York, pp. 245–256. and Anderson, A.W. (2000b) The fusiform ‘‘face area’’ is part
Diamond, R. and Carey, S. (1986) Why faces are and are not of a network that processes faces at the individual level. J.
special: an effect of expertise. J. Exp. Psychol.: Gen., 115(2): Cogn. Neurosci., 12(3): 495–504.
107–117. Gibson, J.J. (1957) Optical motions and transformations as
Donato, G., Bartlett, S., Hager, J., Ekman, P. and Sejnowski, stimuli for visual perception. Psychol. Rev., 64: 228–295.
T. (1999) Classifying facial actions. IEEE Trans. Pattern Gibson, J.J. (1966) The Senses Considered as Perceptual Sys-
Anal. Mach. Intell., 21(10): 974–989. tems. Houghton Mifflin, Boston, MA.
Downing, P.E., Chan, A.W., Peelen, M.V., Dodds, C.M. and Gokturk, S., Tomasi, C., Girod, B. and Bouguet, J. (2002)
Kanwisher, N. (2005). Domain specificity in visual cortex. Model-based face tracking for view-independent facial ex-
Cereb. Cortex (Dec 7, electronic publication, ahead of print). pression recognition. In: Fifth IEEE International Conference
341

on Automatic Face and Gesture Recognition, Washington, Kanade, T. (1973) Computer Recognition of Human Faces.
D.C., USA, pp. 287–293. Basel and Stuttgart, Birkhauser.
Hanawalt, N. (1944) The role of the upper and lower parts of Kanade, T., Cohn, J.F. and Tian, Y. (2000) Comprehensive
the face as the basis for judging facial expressions: II. In database for facial expression analysis. Proceedings of the 4th
posed expressions and ‘‘candid camera’’ pictures. J. Gen. International Conference on Automatic Face and Gesture
Psychol., 31: 23–36. Recognition, Grenoble, France, pp. 46–53.
Hasselmo, M.E., Rolls, E.T. and Baylis, C.G. (1989) The role of Kanwisher, N., McDermott, J. and Chun, M.M. (1997) The
expression and identity in the face-selective responses of neu- fusiform face area: a module in human extrastriate cortex
rons in the temporal visual cortex of the monkey. Exp. Brain specialized for face perception. J. Neurosci., 17: 4302–4311.
Res., 32: 203–218. Kanwisher, N. and Yovel, G. (in press) The fusiform face area:
Haxby, J.V., Gobbini, M.I., Furey, M.L., Ishai, A., Schouten, a cortical region specialized for the perception of faces. Phi-
J.L. and Pietrini, P. (2001) Distributed and overlapping rep- los. Trans. R. Soc. Lond. Ser. B.
resentations of faces and objects in ventral temporal cortex. Knappmeyer, B., Thornton, I.M. and Bülthoff, H.H. (2003)
Science, 293: 2425–2430. The use of facial motion and facial form during the process-
Haxby, J.V., Hoffman, E.A. and Gobbini, M.I. (2000) The ing of identity. Vis. Res., 43: 1921–1936.
distributed human neural system for face perception. Trends Koenen, R. (2000) Mpeg-4 Project Overview, International Or-
Cogn. Sci., 4(6): 223–233. ganization for Standardization, ISO/IEC JTC1/SC29/WG11.
Heisele, B., Ho, P., Wu, J. and Poggio, T. (2003) Face recog- Köhler, W. (1940) Dynamics in Psychology. Liveright, New
nition: comparing component-based and global approaches. York.
Comput. Vis. Image Understanding, 91(1–2): 6–21. Lander, K. and Bruce, V. (2000) Recognizing famous faces:
Henke, K., Schweinberger, S.R., Grigo, A., Klos, T. and Som- exploring the benefits of facial motion. Ecol. Psychol., 12(4):
mer, W. (1998) Specificity of face recognition: recognition of 259–272.
exemplars of non-face objects in prosopagnosia. Cortex, Lander, K., Bruce, V. and Hill, H. (2001) Evaluating the effec-
34(2): 289–296. tiveness of pixelation and blurring on masking the identity of
Hill, H. and Johnson, A. (2001) Categorization and identity familiar faces. Appl. Cogn. Psychol., 15: 101–116.
from the biological motion of faces. Curr. Biol., 11: 880–885. Lander, K., Christie, F. and Bruce, V. (1999) The role of
Hjelmas, E. and Low, B. (2001) Face detection: a survey. Com- movement in the recognition of famous faces. Mem. Cogn.,
put. Vis. Image Understanding, 83: 236–274. 27(6): 974–985.
Hoffman, E. and Haxby, J. (2000) Distinct representations of Leder, H. and Bruce, V. (1998) Local and relational aspects of
eye gaze and identity in the distributed human neural system face distinctiveness. Quart. J. Exp. Psychol., 51A(3): 449–473.
for face perception. Nat. Neurosci., 3: 80–84. Leder, H. and Bruce, V. (2000) When inverted faces are rec-
Humphreys, G., Donnelly, N. and Riddoch, M. (1993) Expres- ognized: the role of configural information in face recogni-
sion is computed separately from facial identity, and is com- tion. Quart. J. Exp. Psychol., 53A(2): 513–536.
puted separately for moving and static faces: neuro- Leder, H., Candrian, G., Huber, O. and Bruce, V. (2001) Con-
psychological evidence. Neuropsychologia, 31: 173–181. figural features in the context of upright and inverted faces.
Ishai, A., Haxby, J.V. and Ungerleider, L.G. (2002) Visual im- Perception, 30: 73–83.
agery of famous faces: effects of memory and attention re- Lee, D. and Seung, H. (1999) Learning the parts of objects by
vealed by fMRI. NeuroImage, 17: 1729–1741. non-negative matrix factorization. Nature, 401: 788–791.
Ishai, A., Ungerleider, L.G., Martin, A. and Haxby, J.V. (2000) Leopold, D.A., O’Toole, A., Vetter, T. and Blanz, V. (2001)
The representation of objects in the human occipital and Prototype-referenced shape encoding revealed by high-level
temporal cortex. J. Cogn. Neurosci., 12: 35–51. after effects. Nat. Neurosci., 4: 89–94.
Ishai, A., Ungerleider, L.G., Martin, A., Schouten, J.L. and Leventhal, H. and Sharp, E. (1965) Facial expression as indi-
Haxby, J.V. (1999) Distributed representation of objects in cators of distress. In: Tomkins, S.S. and Izard, C.E. (Eds.),
the human ventral visual pathway. Proc. Natl. Acad. Sci. Affect, Cognition and Personality: empirical Studies.
USA, 96: 9379–9384. Springer, New York, pp. 296–318.
Izard, C.E. (1979) The maximally discriminative facial move- Lewis, M.B. and Edmonds, A.J. (2003) Face detection: map-
ment coding system (MAX). Unpublished manuscript. ping human performance. Perception, 32(8): 903–920.
(Available from Instructional Resource Center, University Li, B. and Chellappa, R. (2001) Face verification through track-
of Delaware, Newark, DE.) ing facial features. J. Op. Soc. Am. A, 18(12): 2969–2981.
Izard, C.E., Dougherty, L.M. and Hembree, E.A. (1983) A Li, S. and Jain, A. (Eds.). (2004) Handbook of Face Recog-
system for identifying affect expressions by holistic judg- nition. Springer, New York.
ments. Unpublished manuscript, University of Delaware. Liu, J., Harris, A. and Kanwisher, N. (2002) Stages of process-
Johonnson, G. (1973) Visual perception of biological motion and ing in face perception: an MEG study. Nat. Neurosci., 5:
a model for its analysis. Percept. Psychophys., 14: 201–211. 910–916.
Kamachi, M., Bruce, V., Mukaida, S., Gyoba, J., Yoshikawa, Marr, D. (1982) Vision. Freeman Publishers, San Francisco.
S. and Akamatsu, S. (2001) Dynamic properties influence the McNeil, J.E. and Warrington, E.K. (1993) Prosopognosia: a
perception of facial expressions. Perception, 30: 875–887. face-specific disorder. Quart. J. Exp. Psychol., 46A: 1–10.
342

Moscovitch, M., Winocur, G. and Behrmann, M. (1997) What Rosenblum, M., Yacoob, Y. and Davis, L. (1996) Human ex-
is special about face recognition? Nineteen experiments on a pression recognition from motion using a radial basis func-
person with visual object agnosia and dyslexia but normal tion network architecture. IEEE Trans. Neural Networks,
face recognition. J. Cogn. Neurosci., 9: 555–604. 7(5): 1121–1138.
Munhall, K.G., Jones, J.A., Callan, D.E., Kuratate, T. and Rossion, B., Caldara, R., Seghier, M., Schuller, A.M., Laze-
Vatikiotis-Bateson, E. (2004) Visual prosody and speech in- yras, F. and Mayer, E. (2003) A network of occipito-tempo-
telligibility: head movement improves auditory speech per- ral face-sensitive areas besides the right middle fusiform
ception. Psychol. Sci., 15(2): 133–137. gyrus is necessary for normal face processing. Brain, 126:
Murray, J.E., Yong, E. and Rhodes, G. (2000) Revisiting the 2381–2395.
perception of upside-down faces. Psychol. Sci., 11: 498–502. Rossion, B., Curran, T. and Gauthier, I. (2002) A defense of the
Nummenmma, T. (1964) The language of the face. In: subordinate-level expertise account for the N170 component.
Jyvaskyla Studies in Education, Psychology, and Social Re- Cognition, 85: 189–196.
search. Jyvaskyla, Finland. Rotshtein, P., Henson, R.N.A., Treves, A., Driver, J. and
O’Craven, K.M., Downing, P.E. and Kanwisher, N. (1999) Donlan, R.J. (2005) Morphing Marilyn into Maggie disso-
FMRI evidence for objects as the units of attentional selec- ciates physical identity face representations in the brain. Nat.
tion. Nature, 401: 584–587. Neurosci., 8(1): 107–113.
O’Toole, A.J., Phillips, P.J., Cheng, Y., Ross, B. and Wild, Sayette, M.A., Cohn, J.F., Wertz, J.M., Perrott, M.A. and
H.A. (2000) Face recognition algorithms as models of human Dominic, J. (2001) A psychometric evaluation of the facial
face processing. Proceedings of the 4th IEEE International action coding system for assessing spontaneous expression. J.
Conference on Automatic Face and Gesture Recognition, Nonverb. Behav., 25: 167–186.
Grenoble, France. Schwaninger, A., Carbon, C.C. and Leder, H. (2003a) Expert
O’Toole, A.J., Roark, D.A. and Abdi, H. (2002) Recognizing face processing: specialization and constraints. In: Schwarzer,
moving faces: a psychological and neural synthesis. Trends G. and Leder, H. (Eds.), Development of Face Processing.
Cogn. Sci., 6(6): 261–266. Göttingen, Hogrefe, pp. 81–97.
Perret, D.I., Hietanen, J.K., Oram, M.W. and Benson, P.J. (1992) Schwaninger, A., Lobmaier, J. and Collishaw, S.M. (2002)
Organization and functions of cells in the macaque temporal Component and configural information in face recognition
cortex. Philos. Trans. R. Soc. Lond. Ser. B, 335: 23–50. (Lectures Notes). Comput. Sci., 2525: 643–650.
Perret, D.I., Mistlin, A.J. and Chitty, A.J. (1987) Visual neu- Schwaninger, A., Lobmaier, J. and Fischer, M. (2005) The in-
rones responsive to faces. Trends Neurosci., 10: 358–364. version effect on gaze perception is due to component infor-
Perret, D.I. and Oram, M.W. (1993) Image Vis. Comput., 11: mation. Exp. Brain Res., 167: 49–55.
317–333. Schwaninger, A. and Mast, F. (2005) The face inversion effect
Perret, D.I., Rolls, E.T. and Caan, W. (1982) Visual neurons can be explained by capacity limitations of an orientation
responsive to faces in the monkey temporal cortex. Exp. normalization mechanism. Jpn. Psychol. Res., 47(3): 216–222.
Brain Res., 47: 329–342. Schwaninger, A., Ryf, S. and Hofer, F. (2003b) Configural in-
Phillips, P.J. (1999) Support vector machines applied to face formation is processed differently in perception and recog-
recognition. Adv. Neural Inform. Process. Systems, 11: nition of faces. Vis. Res., 43: 1501–1505.
803–809. Schwaninger, A., Wallraven, W. and Bülthoff, H.H. (2004)
Pike, G., Kemp, R., Towell, N. and Phillips, K. (1997) Rec- Computational modeling of face recognition based on psy-
ognizing moving faces: the relative contribution of motion chophysical experiments. Swiss J. Psychol., 63(3): 207–215.
and perspective view information. Vis. Cogn., 4: 409–437. Searcy, J.H. and Bartlett, J.C. (1996) Inversion and processing
Plutchik, R. (1962) The Emotions: Facts, Theories, and A New of component and spacial-relational information in faces. J.
Model. Random House, New York. Exp. Psychol.: Hum. Percept. Perform., 22(4): 904–915.
Puce, A., Allison, T., Bentin, S., Gore, J.C. and McCarthy, G. Sekuler, A.B., Gaspar, C.M., Gold, J.M. and Bennett, P.J.
(1998) Temporal cortex activation in humans viewing eye and (2004) Inversion leads to quantitative, not qualitative,
mouth movements. J. Neurosci., 18: 2188–2199. changes in face processing. Curr. Biol., 14(5): 391–396.
Rhodes, G., Brake, S. and Atkinson, A.P. (1993) What’s lost in Sergent, J. (1984) An investigation into component and con-
inverted faces? Cognition, 47: 25–57. figural processes underlying face perception. Br. J. Psychol.,
Roack, D.A., Barrett, S.E., Spence, M., Abdi, H. and O’Toole, 75: 221–242.
A.J. (2003) Memory for moving faces: psychological and Sergent, J. (1985) Influence of task and input factors on hem-
neural perspectives on the role of motion in face recognition. ispheric involvement in face processing. J. Exp. Psychol.:
Behav. Cogn. Neurosci. Rev., 2(1): 15–46. Hum. Percept. Perform., 11(6): 846–861.
Rock, I. (1973) Orientation and Form. Academic Press, New Swets, D. and Weng, J. (1996) Using discriminant eigenfeatures
York. for image retrieval. IEEE Trans. Pattern Anal. Mach. Intell.,
Rock, I. (1974) The perception of disoriented figures. Sci. Am., 18: 831–836.
230: 78–85. Tanaka, J.W. and Farah, M.J. (1991) Second-order relational
Rock, I. (1988) On Thompson’s inverted-face phenomenon properties and the inversion effect: testing a theory of face
(Research Note). Perception, 17: 815–817. perception. Percept. Psychophys., 50(4): 367–372.
343

Tanaka, J.W. and Farah, M.J. (1993) Parts and wholes in face Wallraven, C., Schwaninger, A. and Bülthoff, H.H. (2005b)
recognition. Quart. J. Exp. Psychol., 46A(2): 225–245. Learning from humans: computational modeling of face rec-
Terzopoulos, D. and Waters, K. (1993) Analysis and synthesis ognition. Network: Comput. Neural Syst, 16(4): 401–418.
of facial image sequences using physical and anatomical Wallraven, C., Schwaninger, A., Schuhmacher, S. and Bülthoff,
models. IEEE Trans. Pattern Anal. Mach. Intell., 15: H.H. (2002) View-based recognition of faces in man and
569–579. machine: re-visiting inter-extra-ortho (Lectures Notes). Com-
Thompson, P. (1980) Margaret Thatcher: a new illusion. Per- put. Sci., 2525: 651–660.
ception, 9: 483–484. White, M. (2000) Parts and wholes in expression recognition.
Tian, Y., Kanade, T. and Cohn, J. (2001) Recognizing action Cogn. Emotion, 14(1): 39–60.
units for facial expression analysis. IEEE Trans. Pattern Williams, M.A., Moss, S.A. and Bradshaw, J.L. (2004) A
Anal. Mach. Intell., 23(2): 97–115. unique look at face processing: the impact of masked faces on
Tronick, E., Als, H. and Brazelton, T.B. (1980) Monadic the processing of facial features. Cognition, 91: 155–172.
phases: a structural descriptive analysis of infant-mother Winston, J.S., Henson, R.N.A., Fine-Goulden, M.R. and Do-
face-to-face interaction. Merrill-Palmer Quart. Behav. Dev., lan, R.J. (2004) fMRI-adaption reveals dissociable neural
26: 3–24. representations if identity and expression in face perception.
Turk, M. and Pentland, A. (1991) Eigenfaces for recognition. J. J. Neurophysiol., 92: 1830–1839.
Cogn. Neurosci., 3: 72–86. Wiskott, L., Fellous, J. and von der Malsburg, C. (1997) Face
Valentine, T. (1995) Cognitive and Computational Aspects of recognition by elastic bunch graph matching. IEEE Trans.
Face Recognition: Explorations in Face Space. Routledge, Pattern Anal. Mach. Intell., 19: 775–779.
London. Xu, Y. (2005) Revisiting the role of the fusiform face area in
Valentine, T. and Bruce, V. (1988) Mental rotation of faces. visual expertise. Cereb. Cortex, 15(8): 1234–1242.
Mem. Cogn., 16(6): 556–566. Xu, Y., Liu, J. and Kanwisher, N. (2005) The M170 is selective
Vuilleumier, P., Armony, J.L., Driver, J. and Dolan, R.J. (2001) for faces, not for expertise. Neuropsychology, 43: 588–597.
Effects of attention and emotion on face processing in the Yamane, S., Kaji, S. and Kawano, K. (1988) What facial fea-
human brain: an event-related fMRI study. Neuron, 30: tures activate face neurons in the inferotemporal cortex of the
829–841. monkey? Exp. Brain Res., 73: 209–214.
Wacholtz, E. (1996) Can we learn from the clinically significant Yin, R. (1969) Looking at upside-down faces. J. Exp. Psychol.,
face processing deficits, prosopagnosia and capgras delusion? 81(1): 141–145.
Neuropsychol. Rev., 6: 203–258. Young, A.W., Hellawell, D. and Hay, D.C. (1987) Configural
Wachsmuth, E., Oram, M.W. and Perret, D.I. (1994) Recog- information in face perception. Perception, 16: 747–759.
nition of objects and their component parts: responses of Yovel, G. and Kanwisher, N. (2005) The neural basis of the
single units in the temporal cortex of the macaque. Cereb. behavioral face-inversion effect. Curr. Biol., 15: 2256–2262.
Cortex, 4: 509–522. Zhang, Z., Lyons, M., Schuster, M. and Akamatsu, S. (1998)
Wallraven, C., Breidt, M., Cunningham, D.W. and Bülthoff, Comparison between geometry-based and Gabor-wavelets-
H.H. (2005a) Psychophysical evaluation of animated facial based facial expression recognition using multi-layer per-
expressions. Proceedings of the 2nd Symposium on Applied ceptron. Proceedings of the Third IEEE International Con-
Perception in Graphics and Visualization, A Coruña, Spain, ference on Automatic Face and Gesture Recognition, Nara,
pp. 17–24. Japan, pp. 454–459.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 19

Investigating audiovisual integration of emotional


signals in the human brain

Thomas Ethofer1,2,, Gilles Pourtois3 and Dirk Wildgruber2

1
Section of Experimental MR of the CNS, Department of Neuroradiology, Otfried-Müller-Str. 51, University of Tübingen,
72076 Tübingen, Germany
2
Department of General Psychiatry, University of Tübingen, Tübingen, Germany
3
Laboratory for Neurology and Imaging of Cognition, Departments of Neurology and Neurosciences, Centre Médical
Universitaire, University of Geneva, Geneva, Switzerland

Abstract: Humans can communicate their emotional state via facial expression and affective prosody. This
chapter reviews behavioural, neuroanatomical, electrophysiological and neuroimaging studies pertaining to
audiovisual integration of emotional communicative signals. Particular emphasis will be given to ne-
uroimaging studies using positron emission tomography (PET) or functional magnetic resonance imaging
(fMRI). Conjunction analyses, interaction analyses, correlation analyses between haemodynamic responses
and behavioural effects and connectivity analyses have been employed to analyse neuroimaging data. There
is no general agreement as to which of these approaches can be considered ‘‘optimal’’ to classify brain
regions as multisensory. We argue that these approaches provide complementing information as they assess
different aspects of multisensory integration of emotional information. Assets and drawbacks of the
different analysis types are discussed and demonstrated on the basis of one fMRI data set.

Keywords: conjunction analysis; connectivity analysis; correlation analysis; emotion; facial expression;
interaction analysis; multisensory; prosody

Behavioural studies perceptual sensitivity and shortened response late-


ncies on a behavioural level (Miller, 1982, 1986;
In natural environment, most events generate stimu- Schröger and Widmann, 1998). This is of particular
lation via multiple sensory channels. Integration of importance for perception of emotionally relevant
inputs from different modalities enables a unified information which can be simultaneously perceived
representation of the world and can provide infor- via the visual modality (e.g. facial expression, body
mation that is unavailable from any single modality postures and gestures) and the auditory modality
in isolation. A compelling example of such merging (e.g. affective prosody and propositional content). It
of information is the McGurk effect (McGurk has been demonstrated that congruency in infor-
and MacDonald, 1976) in which a heard syllable / mation expressed via facial expression and affective
ba/ and a seen syllable /ga/ are perceived as /da/. prosody facilitates behavioural reactions to such
Moreover, integration of information obtained emotion-laden stimuli (Massaro and Egan, 1996; de
from different modalities can result in enhanced Gelder and Vroomen, 2000; Dolan et al., 2001).
Furthermore, affective information obtained via
Corresponding author. Tel.: +49-7071-2987385; Fax: +49- one sense can alter information processing in an-
7071-294371; E-mail: thomas.ethofer@med.uni-tuebingen.de other, for example a facial expression is more likely

DOI: 10.1016/S0079-6123(06)56019-4 345


346

to be perceived as fearful if accompanied by a to the respective unimodal stimuli). Second, the


fearful (as opposed to a neutral) voice (Massaro and stronger these crossmodal interaction effects, the
Egan, 1996; Ethofer et al., 2006). Such crossmodal less effective are the unimodal stimuli in generating
biases occur even under the explicit instruction to a response in the multisensory cell (inverse effec-
ignore information conveyed in the concurrent tiveness rule). Third, spatially disparate crossmodal
channel (de Gelder and Vroomen, 2000; Ethofer cues result in pronounced response depression in
et al., 2006) and are unconstrained by the allocation multisensory cells (i.e. the response to a stimulus
of attentional resources (Vroomen et al., 2001). can be severely diminished by a spatially incongru-
These findings argue for a mandatory nature of ent stimulus from another modality).
processes underlying integration of facial and vocal A similar behaviour has been described for mul-
affective information. tisensory convergence sites in the cortex, such as
the banks of the STS (Bruce et al., 1981; Hikosaka
et al., 1988; Barraclough et al., 2005) and posterior
Neuroanatomical studies insula (Loe and Benevento, 1969; Fallon et al.,
1978). However, neurons in these cortical regions
In animal experiments, several areas with con- and the superior colliculus do not project to or
verging projections from visual and auditory cor- receive afferents from each other (Wallace et al.,
tices have been identified. Such convergence zones 1993) and show different sensitivity to spatial fac-
(Damasio, 1989) constitute candidate regions for tors (Stein and Wallace, 1996). Therefore, it is be-
mediation of audiovisual integration and crossmo- lieved that they fulfil different integrative functions
dal effects in humans (for a review, see Mesulam, (attention/orientating behaviour in the superior
1998; Driver and Spence, 2000; Calvert, 2001). colliculus vs. perceptual judgements in the cortex;
These regions include both cortical structures, Stein et al., 1996).
such as the banks of the superior temporal sulcus Animal experiments provided insight into possi-
(STS; Jones and Powell, 1970; Seltzer and Pandya, ble multisensory integration sites in the brain that
1978), the insula (Mesulam and Mufson, 1982) enable definition of regions of interest for analysis
and the orbitofrontal cortex (Jones and Powell, of human neuroimaging studies. With a typical
1970; Chavis and Pandya, 1976), as well as subcor- spatial resolution of 3  3  3 mm3, each data point
tical structures comprising the superior colliculus acquired in a neuroimaging experiment is attribut-
(Fries, 1984), claustrum (Pearson et al., 1982) and able to the averaged response of several millions of
several nuclei within the amygdala (Turner et al., neurons (Goldman-Racic, 1995). In view of the fact
1980; Murray and Mishkin, 1985; McDonald, that only about 25% of the neurons in cerebral
1998; Pitkänen, 2000) and thalamus (Mufson and multisensory integration sites are sensitive to stimuli
Mesulam, 1984). from more than one modality (Wallace et al., 1992),
At the single neuron level, the most intensively the effects elicited by multisensory integration proc-
studied of these convergence zones is the superior esses are expected to be small. Restricting the search
colliculus (Gordon, 1973; Meredith and Stein, 1983; region to small anatomical structures strongly im-
Peck, 1987; Wallace et al., 1993, 1996) which plays proves the sensitivity to identify such integration
a fundamental role in attention and orientation be- sites by reducing the problem of multiple compari-
haviour (for review, see Stein and Meredith, 1993). sons (Worsley et al., 1996).
On the basis of their detailed studies on multisen-
sory neurons in deep layers of the superior col-
liculus, Stein and colleagues (Stein and Meredith, Electrophysiological studies
1993) formulated a series of key ‘‘integration rules’’:
First, multimodal stimuli that occur in a close tem- Recording of electric brain responses over the
poral and spatial proximity elicit supra-additive re- human scalp (event-related potentials or ERPs)
sponses (i.e. the number of impulses to a bimodal has been primarily employed to investigate the
stimulus exceeds the arithmetic sum of impulses time course of crossmodal binding of affective
347

audiovisual information, given the high temporal of ERP data does not allow inference on which
resolution of this technique. De Gelder et al. brain regions are involved in integration of multi-
(1999) demonstrated that a facial expression with sensory information.
conflicting emotional information in relation to a
simultaneously presented affective voice evokes an
early mismatch-negativity response around 180 ms Neuroimaging studies
after its onset. These ERP findings indicate that
auditory processing is modulated by concurrent Positron emission tomography (PET) and func-
visual information. A subsequent ERP study from tional magnetic resonance imaging (fMRI) have
the same group demonstrated that the auditory been used to shed light on the functional neuro-
N1 component occurring around 110 ms after anatomy of multisensory integration of emotional
presentation of an affective voice is significantly information. However, definition of the appropri-
enhanced by an emotionally congruent facial ex- ate analysis in multimodal studies is not trivial and
pression. This effect occurs for upright but not for several approaches have been applied to model
inverted faces (Pourtois et al., 2000). Recognition audiovisual integration or crossmodal effects in
of emotional facial expressions is substantially the brain. These approaches include conjunction
hindered by face inversion (White, 1999). Thus, analyses, interaction analyses, correlation analyses
the observation that modulation of auditory ERP with effects observed on behavioural level and
components is restricted to upright faces suggests connectivity analyses. We demonstrate the appli-
that this effect is driven by the expressed facial cation of these approaches on the basis of one data
affect and is not attributable to low-level pictural set acquired in an event-related fMRI study con-
features of the visual stimuli. Finally, an analysis ducted to investigate the neuronal correlates un-
focused on the positive deflection following the derlying audiovisual integration of emotional
N1-P1 component around 220 ms poststimulus information from face and voice (Ethofer et al.,
revealed a shortened latency of this deflection in 2006). Twelve right-handed subjects participated
emotionally congruent as compared to incongru- in this experiment conducted at a 1.5 T scanner
ent audiovisual trials (Pourtois et al., 2002). These (Siemens VISION, Erlangen, Germany) compris-
faster ERP responses parallel behavioural effects ing two sessions with 36 visual (V) trials, two ses-
showing facilitated responses to affective prosody sions with 36 auditory (A) trials and two sessions
when presented simultaneously with a congruent with 36 audiovisual (AV) trials. In the visual ses-
versus an incongruent emotional facial expres- sions, every trial consisted of a presentation of a
sion (de Gelder and Vroomen, 2000). In summary, facial expression shown for 1 s. These visual stim-
electrophysiological studies on audiovisual inte- uli were obtained from the Ekman and Friesen
gration of emotional information indicate that series (1976) and comprised facial expressions
multisensory integration occurs at an early stage of ranging from neutral to 100% fear and from neu-
cerebral processing (i.e. around 110–220 ms post- tral to 100% happiness in incremental steps of
stimulus). The observation that crosstalk between 25% created by digital morphing techniques. In
the modalities takes place during the early per- the auditory sessions, sentences spoken by profes-
ceptual rather than during late decisional stages sional actors in an either happy or fearful voice
offers an explanation for the finding that cross- were presented. In the bimodal sessions, auditory
modal biases between the modalities occur irre- and visual stimuli were presented with facial ex-
spective of attentional resources (Vroomen et al., pressions being shown during the last second of
2001) and instructions to ignore a concurrent the spoken sentences. Subjects were instructed to
stimulus (de Gelder and Vroomen, 2000; Ethofer judge the emotional valence of the stimuli on
et al., 2006). Furthermore, these electrophysio- a nine-point self-assessment manikin scale (SAM;
logical findings point to neuronal structures that Bradley and Lang, 1994) by pressing buttons in
conduct early steps in the processing of external their right or left hand. In all the sessions, the
information. However, the low spatial resolution SAM scale was presented for 4 s 200 ms after
348

stimulus offset (see Fig. 1a). In unimodal sessions, motor cortices and cerebellum which are most
subjects rated the emotional valence of the pre- probably attributable to the motor responses made
sented stimuli (facial expressions or prosody). In the by the subjects to judge the stimuli. Furthermore,
bimodal sessions, subjects were instructed to judge dorsolateral prefrontal cortices of both hemispheres
the emotional valence of the facial expression and presumably subserving working memory processes
ignore the concomitant affective voice. fMRI data showed responses to stimulus presentation in both
were analysed using statistical parametric mapping unimodal and bimodal sessions.
software (SPM2, Welcome Department of Imaging
Neuroscience, London, UK, http://www.fil.ion.u-
cl.ac.uk/spm). Coordinates of activation clusters Conjunction analyses
are given in MNI space (Montreal Neurological
Institute; Collins et al., 1994). Main effects of pres- Conjunction analyses were originally introduced
entation of unimodal (A and V) and bimodal (AV) by Price and Friston (1997). The objective of
stimuli are shown in Fig. 1b. As expected, unimodal this approach is to test for commonalities in brain
presentation of either emotional facial expressions activation induced by different stimuli or tasks.
or affectively spoken sentences engaged bilateral Recently, conjunctions employed in analysis of
primary and higher-order visual or auditory corti- neuroimaging data have been revised and divided
ces, respectively, while in bimodal trials both visual into those that test for a global null hypothesis
and auditory cortices were found active. In all (H0: No effect in any of components; H1: Signi-
three conditions, activations were found in bilateral ficant effect in at least one of the components;

Fig. 1. Experimental design (a) and brain activation (b) for auditory (upper panels), visual (middle panels) and audiovisual trials
(lower panels) as compared to rest. Brain activations are thresholded at a height threshold of po0.001 (uncorrected) and corrected for
multiple comparisons at cluster level k450, po0.05 (corrected).
349

Friston et al., 1999, 2005) and those that test for fMRI study identifying audiovisual integration
a conjunction null hypothesis (H0: No effect in at sites of motion processing in lateral parietal cortex
least one of the components; H1: Significant effects (Lewis et al., 2000). In our study, the intersection
in all of the components; Nichols et al., 2005). A \ V revealed significant responses in candi-
Since only rejection of the conjunction null hy- date areas for audiovisual convergence, such as the
pothesis can be taken as evidence for a logical posterior thalamus extending into the superior
AND, all conjunction analyses reported here were colliculus and the right posterior temporooccipito-
carried out on the basis of a conjunction null parietal junction (see Fig. 2). However, areas pre-
hypothesis. sumably involved in the nonsensory components
An obvious property of multisensory neural of the task, such as the dorsolateral prefrontal
structures is their responsiveness to stimuli ob- cortex (working memory) and motor cortex, sup-
tained from more than one modality. Thus, a plementary motor area and cerebellum (motor re-
straightforward approach to locate brain regions sponses), also showed significant responses during
containing such structures is a conjunction anal- both unimodal sessions. These findings illustrate
ysis on responses to unimodal stimuli of both mo- that the results provided by the simple intersection
dalities (Unimodal 1 \ Unimodal 2). This A \ V cannot separate multimodal convergence
approach has been applied in a PET study dem- zones from unspecific activations attributable to
onstrating multisensory convergence zones in the nonsensory components of the task. Therefore,
left intraparietal sulcus for spatial attention to vi- results produced by such a conjunction do not
sion and touch (Macaluso et al., 2000) and an necessarily reflect multisensory convergence zones.

Fig. 2. Intersection of brain activation during unimodal auditory and visual stimulation (A \ V). Brain activations are thresholded at a
height threshold of po0.001 (uncorrected) and corrected for multiple comparisons at cluster level, k450, po0.05 (corrected).
350

Furthermore, brain regions responding exclusively conjunctions by correcting the search volume to
to bimodal stimulation or achieving suprathresh- small anatomical regions (small volume correction
old activations by supra-additive responses in- (SVC); Worsley et al., 1996). Definition of regions
duced by the simultaneous presence of stimuli of of interest in our analysis relied on knowledge in-
two modalities will be missed compromising the ferred from neuroanatomical studies and previous
sensitivity of this approach. neuroimaging studies and comprised the cortex
Both the lack of specificity for multimodal in- adjacent to the posterior STS (Jones and Powell,
tegration sites as well as the impaired sensitivity 1972; Seltzer and Pandya, 1978; Beauchamp et al.,
to detect regions responding exclusively to multi- 2004a, 2004b; van Atteveldt et al., 2004), orbito-
modal stimuli can be overcome by investigating frontal cortex (Jones and Powell, 1972; Chavis and
brain areas that show a significantly stronger re- Pandya, 1976), insular cortex (Mesulam and Muf-
sponse to bimodal stimuli than to unimodal stim- son, 1982), claustrum (Pearson et al., 1992; Olson
uli of both modalities. This can be achieved by et al., 2002), superior colliculus (Fries, 1984;
computing the conjunction [Bimodal – Unimodal Calvert et al., 2000), thalamus (Mufson and Me-
1] \ [Bimodal – Unimodal 2]. This analytic strat- sulam, 1984) and amygdala (Turner et al., 1980;
egy has been employed by Grefkes et al. (2002) to McDonald, 1998; Pitkänen, 2000; Dolan et al.,
investigate brain regions subserving crossmodal 2001). The conjunction analysis [AV – V] \ [AV –
transfer of visuotactile information and in a com- A] revealed activation clusters in bilateral poste-
parable way by Calvert et al. (1999) to detect neu- rior STS, right orbitofrontal cortex, bilateral pos-
ral structures involved in processing of audiovisual terior thalamus and right posterior insula/
speech. Recently, a more elaborated form of this claustrum (Fig. 3). The activation cluster in the
approach was used by Pourtois et al. (2005) in a left posterior STS was significant (SVC on a 6 mm
PET study on audiovisual integration of emotional radius spherical volume of interest centred at
information. In this study, the experimental design x ¼ –50, y ¼ –54, z ¼ 6, a priori coordinates
included two different AV conditions in which derived from Beauchamp et al., 2004b). This re-
subjects were instructed to judge either the infor- sult is in keeping with reports showing stronger
mation from the visual (AV(judge V)) or auditory responses in the posterior STS cortices during
channel (AV(judge A)). The conjunction of audiovisual presentation of objects (Beauchamp
[AV(judge A) – A(judge A)] \ [AV(judge V) – et al., 2004b), letters (van Atteveldt et al., 2004)
V(judge V)] was computed which represents a and speech (Wright et al., 2003; van Atteveldt
sound way to remove task-related brain responses. et al., 2004) than with unimodal presentation of
However, it should be noted that any conjunc- these stimuli. Thus, there is converging evidence
tion analysis based on a conjunction null hypoth- implicating posterior STS cortices in integration of
esis (Nichols et al., 2005) is a very conservative audiovisual stimuli irrespective of the type of in-
strategy which only gives an upper bound for the formation conveyed by these stimuli. Further-
false positive rate (Friston et al., 2005). While such more, a recent PET study (Pourtois et al., 2005)
conjunction analyses remain valid even in statis- demonstrated increased cerebral blood flow in the
tical worst-case scenarios (Nichols et al., 2005), left middle temporal gyrus during audiovisual
their conservativeness must be paid for by a loss of presentation of emotional information as com-
sensitivity (Friston et al., 2005). This is especially pared to isolated presentation of emotional faces
critical if two differential contrasts that are ex- or voices. However, it has to be noted that the
pected to yield small effects are submitted to such activation cluster in our study was localized more
an analysis. Accordingly, in none of the studies posterior and superior than the cluster described
that employed this approach (Calvert et al., 1999; by Pourtois et al. (2005). Differences in the imag-
Grefkes et al., 2002; Pourtois et al., 2005) brain ing modality and task instructions (gender differ-
activations were significant if corrected for multi- entiation in the PET study of Pourtois et al., 2005
ple comparisons across the whole brain. Therefore, as compared to rating of emotional information in
it is decisive to increase the sensitivity of such the fMRI study by Ethofer et al., 2006) might
351

Fig. 3. Conjunction analysis [AV – A] \ [AV – V] showing activations in (a) bilateral posterior superior temporal sulcus (x ¼ –54,
y ¼ –51, z ¼ 12, Z ¼ 2.90, k ¼ 51 and x ¼ 51, y ¼ –42, z ¼ 12, Z ¼ 2.79, k ¼ 119 for the left and right STS, respectively), right
orbitofrontal cortex (x ¼ 39, y ¼ 24, z ¼ –12, Z ¼ 2.69, k ¼ 64), (b) bilateral posterior thalamus (x ¼ –30, y ¼ –30, z ¼ 0, Z ¼ 2.82,
k ¼ 120 and x ¼ 12, y ¼ –24, z ¼ 9, Z ¼ 2.88, k ¼ 76 for the left and right thalamic cluster, respectively) and (c) right posterior insula/
claustrum (x ¼ 36, y ¼ –3, z ¼ 6, Z ¼ 2.45, k ¼ 124). Brain activations are thresholded at a height threshold of po0.05 (uncorrected).
(d) Event-related responses to unimodal auditory (red), unimodal visual (blue) and bimodal (magenta) stimuli in the left posterior STS.

constitute an explanation for the different locali- presentation of either happy voices or happy facial
zation of the activation clusters within the left expressions in the right posterior insula/claustrum
temporal lobe in the two studies. (x ¼ 39, y ¼ –6, z ¼ –3, Z ¼ 3.66, k ¼ 91, po0.05
Another promising approach employed by SVC corrected for right insula). The only brain
Pourtois et al. (2005) to investigate brain regions structure included in our regions of interest showing
subserving integration of audiovisual emotional enhanced responses to audiovisual fear as com-
information is to define separate conjunction anal- pared to unimodally presented fear [AVfear – Afear]
yses for specific emotions, such as [AVhappy – \ [AVfear – Vfear] was the right amygdala (x ¼ 27,
Ahappy] \ [AVhappy – Vhappy] or [AVfear – Afear] \ y ¼ –9, z ¼ –24, Z ¼ 2.00, k ¼ 29). However, this
[AVfear – Vfear]. The results of this analysis enable activation failed to reach significance within the
inference on the localization of brain regions search volume comprising the amygdala. In con-
showing stronger responses if a certain emotion clusion, the results from conjunction analyses of
is signalled via two modalities as compared to our experiment suggest that neocortical areas in vi-
unimodal presentation of this emotion in either cinity to the STS might be more generally con-
modality. We submitted the fMRI data set re- cerned with integration of audiovisual signals, while
ported here to an analogous conjunction analysis. phylogenetically older structures, such as the pos-
In this analysis, facial expressions were considered terior insula or the amygdala, show additive re-
to express happiness or fear, if they showed at least sponses if certain emotions are expressed in a
50% of the respective emotion. The conjunction congruent way via different sensory channels.
[AVhappy – Ahappy] \ [AVhappy – Vhappy] revealed However, one limitation of all analyses relying
stronger responses for bimodal presentation of on conjunctions of [AV – V] \ [AV – A] is that
happiness as compared to responses to unimodal they have the potential to detect brain regions in
352

which responses to information from auditory and motor responses) showing similar positive re-
visual channels sum up in a linear way and might sponses in both unimodal and the bimodal trials
therefore simply reflect areas in which both neu- will produce a negative interaction effect. Further-
rons responsive to unimodal auditory and unimo- more, brain areas which deactivate in a similar way
dal visual information coexist without the need of in all three conditions will produce a positive in-
multimodal integration in these areas (Calvert, teraction effect. Positive and negative interactions,
2001; Calvert and Thesen, 2004). as computed by AV – (A+V) are shown in Fig. 5a
in red and green, respectively. At first glance, the
finding of a positive interaction in the right inferior
Interaction analyses parietal cortex and the right orbitofrontal cortex is
an interesting result. Inspection of the event-related
Calvert and Thesen (2004) suggested that activat- responses during unimodal auditory, unimodal vis-
ions of multisensory integration sites should differ ual and bimodal trials, however, reveals that in
from the arithmetic sum of the respective activat- both these regions the haemodynamic response
ions to unimodal stimuli: If the response to a bi- decreases in all three conditions. The finding of
modal stimulus exceeds the sum of the unimodal unspecific deactivations to stimulus presentation in
responses [Bimodal4Unimodal 1+Unimodal 2], the right inferior parietal cortex is in agreement
then this is defined as positive interaction; while if with the view that this area belongs to the resting
the summed responses are greater than the bimo- state network and is tonically active in the baseline
dal response [BimodaloUnimodal 1+Unimodal state (Raichle et al., 2003; Fox et al., 2005). Thus,
2], then this is defined as negative interaction effect the positive response in this region as calculated by
(Calvert et al., 2000, 2001). Usually, the most effi- the interaction AV – (A+V) is caused by the fact
cient way to investigate interactions between two that the added deactivations during unimodal trials
factors is a 2  2 factorial design. Theoretically, exceed the deactivation during bimodal trials (see
such a 2  2 factorial design for investigating Fig. 4b). A more complex behaviour of the haemo-
interactions between two sensory modalities would dynamic response was found in the orbitofrontal
include one bimodal and two unimodal conditions, cortex showing a decrease of the blood oxygen level
in which the subject judges some aspect of the dependent (BOLD) response with varying delay
presented stimuli, and a control condition which followed by a positive response further complicat-
contains all components of the other conditions ing the interpretation of the positive interaction
(e.g. working memory, judgement and motor re- computed by AV – (A+V). Negative interactions
sponses), but no sensory stimulation. However, for were found in dorsolateral prefrontal areas, motor
all paradigms including a behavioural task, it is cortex and cerebellum. Event-related responses in
practically impossible to implement such a control the right dorsolateral prefrontal cortex (see Fig.
condition since it is impossible for the subject to 4d), however, show that the negative interaction in
judge a specific aspect (e.g. gender or conveyed this region is due to very similar responses in all
emotion) of a stimulus if it is not presented. three conditions. The vulnerability of the interac-
Therefore, all imaging studies investigating inter- tion AV – (A+V) to unspecific activations attribu-
actions between the modalities omitted this control table to the behavioural task resulting in negative
condition and simply compared haemodynamic interactions and unspecific deactivations of resting
responses obtained during bimodal stimulation to state network components producing positive in-
the sum of the responses of the unimodal condi- teractions demands caution in the application of
tions: Bimodal  [Unimodal 1+Unimodal 2]. this technique. Therefore, we suggest that interpre-
However, this omission of the control condition tation of both positive and negative interactions
has serious consequences for the interpretation of calculated by this approach should be based on
both positive and negative interactions. For ex- inspection of time series.
ample, brain regions involved in nonsensory com- Calvert et al. (2001) suggested that electrophysi-
ponents of the task (e.g. working memory and ological criteria for investigation of multimodal
353

Fig. 4. (a) Positive (red) and negative (green) interactions as calculated by AV– (A+V). Brain activations are thresholded at a height
threshold of po0.001 (uncorrected). Event-related responses to unimodal auditory (red), unimodal visual (blue) and bimodal (ma-
genta) stimuli in (b) the right lateral inferior parietal cortex (MNI coordinates: x ¼ 57; y ¼ –66; z ¼ 12, Z ¼ 4.81, k ¼ 36), (c) the right
orbitofrontal cortex (MNI coordinates: x ¼ 39; y ¼ 21; z ¼ –21, Z ¼ 5.23, k ¼ 46) and (d) the right dorsolateral prefrontal cortex
(MNI coordinates: x ¼ 45; y ¼ 6; z ¼ 21, Z ¼ 4.71, k ¼ 229).

integration should be applied to the BOLD effect. sponse depression: Bimodal(incongruent)o Max-
According to these criteria, cells subserving multi- imum(Unimodal 1, Unimodal 2)). Calvert et al.
modal integration show responses to congruent (2001) demonstrated that BOLD responses within
information obtained via several modalities that the superior colliculi fulfil these criteria showing
exceed the sum of responses to the respective supra-additive responses to audiovisual nonspeech
unimodal stimuli (supra-additivity: Bimodal(con- stimuli if they are presented in temporal synchrony
gruent)4Unimodal 1+Unimodal 2). Further- and corresponding response depressions if the
more, conflicting multimodal information results audiovisual stimuli are presented in an asynchro-
in response depression in which the response to nous manner. We investigated whether one of our
incongruent bimodal information is smaller than regions of interest shows a comparable behaviour
the stronger of the two unimodal responses (re- if congruence of audiovisual information is defined
354

by emotional content conveyed via affective (see Fig. 5b). No evidence for a depression of re-
prosody and facial expressions. To this end, we sponses below the level of unimodal responses was
compared responses to audiovisual trials with con- found. Instead, the results of this analysis suggest
gruent emotional information (e.g. showing at that the right posterior insula/claustrum also re-
least 50% of the respective emotion in the facial sponds, albeit to some lesser extent, to audiovisual
expression as expressed via emotional prosody) to information conveying conflicting emotional infor-
the sum of haemodynamic responses to unimodal mation. The observation that this region shows
visual and auditory trials (AV(congruent) – strongest haemodynamic responses if auditory and
(A+V), see Fig. 5a). Obviously, this analysis tech- visual information are simultaneously available is
nique is burdened with the same drawbacks as the concordant with results obtained from single-cell
simple interaction AV – (A+V) and not surpris- recordings of the posterior insula cortex (Loe and
ingly, positive interactions driven by similar de- Benevento, 1969). The finding that activation in
creases of the BOLD response in all three this region is stronger during congruent than dur-
conditions were found in the right lateral inferior ing conflicting audiovisual emotional information
parietal and orbitofrontal cortex. In addition, sig- is in keeping with observations from an fMRI
nificant supra-additive responses were also found experiment on synchronized and desynchronized
in the right posterior insula/claustrum (see Fig. audiovisual speech (Olson et al., 2002). In this ex-
5b). Inspection of event-related responses demon- periment, audiovisual stimuli presented in tempo-
strates that the interaction found in this region is ral synchrony resulted in stronger haemodynamic
of a completely different nature than those found responses in the left claustrum than asynchronally
in the parietal and orbitofrontal cortex showing presented stimuli. Olson et al. (2002) reasoned that
robust activation to congruent audiovisual trials audiovisual integration might be better explained by
and slightly negative responses to unimodal stim- a ‘‘communication relay’’ model in which subcor-
ulation in either modality (see Fig. 5b). To inves- tical areas, such as the claustrum, receive their in-
tigate whether the responses of this region fulfil formation directly from unimodal cortices than by a
the criteria of response depression to conflicting ‘‘site-specific’’ model assuming integration in multi-
emotional content as conveyed by voice and face, sensory brain regions, such as the posterior STS.
we plotted event-related responses to bimodal We argue that increased haemodynamic responses
trials with incongruent emotional information as found for temporally congruent information in

Fig. 5. (a) Positive interaction in the right posterior insula/claustrum (MNI coordinates: x ¼ 39; y ¼ –6; z ¼ –3; Z ¼ 4.05; k ¼ 14;
po0.05, SVC corrected) as calculated by AV(congruent) – (A+V). Brain activations are thresholded at a height threshold of po0.001
(uncorrected). (b) Event-related responses to unimodal auditory (red), unimodal visual (blue) and bimodal trials with congruent
emotional information (magenta) and incongruent information (cyan) have been plotted.
355

the study of Olson et al. (2002) and emotionally face-fearful voice (fF)) and for two conditions that
congruent information in our study might reflect a were conflicting information is expressed via the
more general role of the claustrum in determining two modalities (e.g. happy face-fearful voice (fH)
whether multisensory information matches or not. and fearful face-happy voice (hF)). The interaction
A possible problem with application of neuro- is then calculated by (hH – fH) – (hF – fF). For
physiological criteria to analyses of the BOLD interpretation of results calculated by this inter-
signal arises from the fact that the signal acquired action term it is worth noting that the interaction
in an fMRI is caused by haemodynamic pheno- of emotional information is mathematically equiv-
mena. While supra-additivity is a well-known alent to contrasting congruent with incongruent
electrophysiological characteristic of multimodal conditions: (hH+fF) – (fH+hF). This factorial
cells (Stein and Meredith, 1993), there is reason to design was used by Dolan et al. (2001) to inves-
doubt that supra-additive neuronal responses must tigate interactions between visually and auditorily
necessarily translate into supra-additive BOLD re- presented fear and happiness. In this study, par-
sponses. There is compelling evidence that the ticipants were instructed to categorize the pre-
BOLD response to two stimuli in temporal proxi- sented facial expression in either fearful or happy,
mity is overpredicted by simply adding the re- while ignoring the simultaneously presented emo-
sponses to the two stimuli presented in isolation tional voice. Behaviourally, the authors found a
(Friston et al., 1998; Mechelli et al., 2001). This facilitation of the emotion categorization as indi-
phenomenon has been named ‘‘haemodynamic re- cated by shortened response latencies for con-
fractoriness’’ and is specific for an fMRI since the gruent as compared to incongruent audiovisual
major part of this nonlinear behaviour arises from trials. On a cerebral level, a significantly stronger
the transformation of changes in regional cerebral activation of the left basolateral amygdala as
blood flow to the BOLD response (Friston et al., calculated by (hH – fH) – (hF – fF) was found.
1998; Mechelli et al., 2001). The fact that a pre- Inspection of parameter estimates of the four con-
ceding or simultaneously presented stimulus can ditions revealed that this interaction was mainly
attenuate the BOLD response to a second stimulus driven by an augmentation of haemodynamic re-
might compromise the sensitivity of an fMRI data sponses during the fear congruency condition (fF).
analysis in which responses during multimodal On the basis of their results, Dolan et al. (2001)
integration are expected to exceed the linear sum concluded that it is the left amygdala that sub-
of the responses to unimodal stimuli. It might serves crossmodal integration in fear processing.
therefore be useful to estimate the neuronal re- This interpretation is in line with observations
sponse from the BOLD response via a plausible from neuropsychological studies indicating that
biophysical model (Friston et al., 2000; Gitelman lesions of the amygdala can impair recognition of
et al., 2003) and search for regions in which the both fearful faces (Adolphs et al., 1994) and fear-
neuronal response exhibits supra-additive effects. ful voices (Scott et al., 1997) as well as neuro-
While construction of a full 2  2 factorial imaging studies of healthy subjects demonstrating
design for investigation of interaction between enhanced activation to fear signalled via the face
sensory modalities is compromised by the lack of (Breiter et al., 1996; Morris et al., 1996) and the
an appropriate control condition, investigating in- voice (Phillips et al., 1997). So far, little is known
teractions between emotions expressed via these on the neuronal substrates underlying integra-
modalities is not burdened with this problem. Fur- tion of audiovisual emotional information other
thermore, no unimodal trials are required in fac- than fear. The factorial design employed by Dolan
torial analyses designed to investigate interactions et al. (2001) has the potential to clarify whether
between emotional information expressed via two audiovisual integration of other basic emotions
different modalities. Instead, such designs include also occurs via the amygdala or is mediated by
only bimodal trials with two conditions in which different neuroanatomical structures. However, we
emotional information is presented in a congruent feel that interpretability of the results provided by
way (e.g. happy face-happy voice (hH) and fearful such factorial designs could be improved by using
356

neutral facial expressions (N) and intonations (n) in 1996; de Gelder and Vroomen, 2000) and it has
combination with facial expressions (E) and into- been demonstrated that fearful or neutral faces are
nations (e) of one emotion and then calculate the perceived as more fearful when accompanied by
interaction accordingly (e.g. (eE – nE) – (eN – nN)). a fearful voice (see Fig. 6a, Ethofer et al., 2006). To
examine which brain regions mediate this shift
in judgement of facial affect, we correlated the
Correlation analyses between brain activation and difference in brain responses to facial expressions
crossmodal behavioural effects in presence and absence of a fearful voice with
the difference of the subjects’ valence rating of the
On a behavioural level, crossmodal integration re- facial expressions in both conditions. A significant
sults in shortened response latencies to congruent correlation was found in the left basolateral am-
bimodal information (Miller et al., 1982; Schröger ygdala extending into the periamygdaloid cortex
and Widmann, 1998; de Gelder and Vroomen, (see Fig. 6b). This finding indicates that cogni-
2000). Furthermore, judgement of sensory infor- tive evaluation of emotional information signaling
mation from one modality can be influenced by threat or danger is modulated by amygdalar re-
information obtained from another modality sponses and is in agreement with the view that the
(McGurk and MacDonald, 1976). To investigate amygdala has a key integrative role in processing
which brain regions mediate the McGurk illusion, of emotional content, particularly when fear is
Jones and Callan (2003) employed the McGurk expressed across sensory channels (Dolan et al.,
paradigm in an fMRI experiment and correlated 2001). Response-related correlation analyses be-
haemodynamic responses with the influence of vis- tween brain activation and crossmodal behavioural
ual cues on judgement of auditory information. effects represent a useful approach to model sys-
Activity in the occipitotemporal junction was tems associated with the behavioural outcome of
found to correlate with the strength of the McGurk multisenory integration.
illusion, suggesting that modulation of responses
within this region might constitute the neural Connectivity analyses
substrate of the McGurk effect (Jones and Callan,
2003). Crossmodal biases also occur in perception All analysis techniques discussed so far are con-
of emotional information (Massaro and Egan, cerned with the localization of brain areas that

Fig. 6. (a) Valence ratings of facial expressions ranging from 0% to 100% fear in presence (red) and absence (blue) of a fearfully
spoken sentence. (b) Correlation analysis between crossmodal impact of fearful voices on judgement of faces and BOLD response
revealed a cluster within the left amygdala (MNI coordinates: x ¼ –24, y ¼ –6, z ¼ –24, Z ¼ 3.84, k ¼ 42, po0.05, SVC). (c) Event-
related responses of the left basolateral amygdala for trials in which faces were judged as being more fearful when presented together
with a fearful voice (red) and trials where no such shift in interpretation occurred (blue). Modified from Ethofer et al. (2006).
357

mediate a certain cognitive process of interest. between voice- and face-sensitive cortices, as de-
Recent developments (Friston et al., 2003; Gitel- scribed previously for speaker identification (von
man et al., 2003) in modeling effective connectivity Kriegstein et al., 2005), but rather via supramodal
between brain regions (i.e. the influence one neural relay areas, such as the amygdala. The amygdala is
system exerts over another, Friston et al., 1997) anatomically well positioned to provide such a
offer the opportunity to investigate neural inter- supramodal relay function (Murray and Mishkin,
actions between sensory systems. In an fMRI 1985) since it is bidirectionally connected with vis-
study on the neural substrates of speaker recogni- ual and auditory higher-order cortices (Pitkänen,
tion, von Kriegstein et al. (2005) found that fa- 2000). Furthermore, augmented effective connec-
miliar voices activate the fusiform face area (FFA, tivity between amygdala and fusiform cortex in
Kanwisher et al., 1997). A psychophysiological in- context of a fearful voice is in agreement with data
teraction (PPI) analysis (Friston et al., 1997) with obtained from lesion (Adolphs et al., 1994; Scott et
FFA activity as physiological and familiarity of al., 1997) and neuroimaging studies (Breiter et al.,
the voices as psychological factor revealed that this 1996; Morris et al., 1996; Phillips et al., 1997) im-
crossmodal activation of face-sensitive areas by an plicating the amygdala in fear processing, and re-
auditory stimulus was driven by activity of voice- sults from a PET experiment suggest that the
sensitive areas in the middle part of the STS. On amygdala exerts a top-down control on neural ac-
the basis of these results, von Kriegstein et al. tivity in extrastriate cortex (Morris et al., 1998).
(2005) suggested that assessment of person fami- Although anatomically segregated, voice- and
liarity does not necessarily engage a ‘‘supra- face-processing modules have to interact to form
modal cortical substrate’’ but can be the result of a unified percept of the emotional information ex-
a ‘‘supramodal process’’ constituting an enhanced pressed via different sensory channels. Analyses of
functional coupling of face- and voice-sensitive effective connectivity have the potential to inves-
cortices. In our study, significantly stronger acti- tigate the interaction between these modules and
vations were found in the right FFA (MNI co- could elucidate whether integration is achieved via
ordinates: x ¼ 24, y ¼ –69, z ¼ –15, Z ¼ 3.91, supramodal nodes or by direct coupling of face-
k ¼ 46) during judgement of facial expressions in and voice-sensitive cortices.
presence of fearful as compared to happy voices
(Ethofer et al., 2006). To elucidate which cortical
areas mediate this crossmodal effect of a fearful Conclusion
voice on processing within the fusiform face area,
a PPI analysis with activity of the right FFA as Integration of information from different sensory
physiological and emotion expressed via affective channels is a complex phenomenon and recruits
prosody (fear or happiness) as psychological factor several cerebral structures. Application of different
was carried out. Increased effective connectivity types of analyses aimed at identifying multisensory
between the right FFA and left basolateral am- integration sites to the fMRI data set presented
ygdala/periamygdaloid cortex (MNI coordinates: here revealed the cortex in the left posterior
x ¼ –18; y ¼ –12; z ¼ –30; Z ¼ 2.68; k ¼ 5, STS, the right posterior insula/claustrum and the
po0.05 small volume corrected for the amygdalar left amygdala as being implicated in audiovisual
cluster in which activity was correlated with be- integration.
havioural responses, see above) was found during The left posterior STS responded significantly
rating of facial expressions in presence of a fearful stronger to bimodal stimuli than to isolated pres-
as compared to a happy voice. Even at low thresh- entation of either faces or voices as determined by
olds (po0.05, uncorrected), no evidence for mod- the conjunction [AV – V] \ [AV – A]. Notably, the
ulation of responses in the right FFA by activity in left posterior STS did not show supra-additive
STS regions was found, suggesting that crossmo- BOLD responses as determined by the compar-
dal effects of emotional auditory information on ison of bimodal to the sum of unimodal responses
FFA responses is not mediated by direct coupling (AV – (A+V)). This lack of supra-additivity in
358

posterior STS regions in our study is in agreement threat or danger (Dolan et al., 2001). The right
with results obtained in previous fMRI experi- fusiform cortex showed stronger responses when
ments on audiovisual integration (Beauchamp facial expressions were rated in presence of a fear-
et al., 2004a, b; van Atteveldt et al., 2004). On ful as compared to a happy voice (Ethofer et al.,
the basis of the observation that the BOLD re- 2006). A psychophysiological interaction analysis
sponse in posterior STS to audiovisual stimuli revealed enhanced effective connectivity between
does not exceed the sum of responses to the re- the left amygdala and right fusiform cortex pro-
spective unimodal stimuli as expected from single- viding a possible neural basis for the observed be-
cell recordings (Bruce et al., 1981; Hikosaka et al., havioural effects.
1988; Barraclough et al., 2005), Beauchamp (2005) The aim of this chapter was to review the differ-
challenged the concept of supra-additivity as an ent methodological approaches to model multi-
appropriate criterion for definition of multisensory sensory integration in neuroimaging including con-
regions in neuroimaging. The ‘‘haemodynamic junction analyses, interaction analyses, correlation
refractoriness’’ (Friston et al., 1998) of BOLD analyses between brain responses and behavioural
responses to temporally proximate stimuli might effects and connectivity analyses. None of these
constitute a possible explanation for this discrep- approaches can be considered as the optimal
ancy between fMRI and electrophysiological data. method to clarify as to which brain structures par-
The right posterior insula/claustrum responded ticipate in multisensory integration. Rather, we
stronger to congruent audiovisual emotional would like to emphasize that each of these analyses
information than to unimodal information of both elucidates different aspects of the interplay of brain
modalities or to incongruent audiovisual informa- regions in integrational processes and thus pro-
tion. This stronger responsiveness to congruent vides complementing information.
than to conflicting audiovisual information is in
line with previous reports on the claustrum show- Abbreviations
ing stronger responses to synchronized than to
desynchronized audiovisual speech (Olson et al., A auditory
2002). These converging findings obtained across AV audiovisual
domains (temporal domain in the study of Olson BOLD blood oxygen level dependent
et al. (2002) and emotional domain in our study) EE bimodal trial with emotional
suggest a more general role of the claustrum in voice and emotional face
processes that determine whether information eN bimodal trial with emotional
gathered across different channels matches or voice and neutral face
not. Future studies relying on analyses of effec- ERP event-related potentials
tive connectivity should address the question fF bimodal trial with fearful voice
whether the claustrum receives its information di- and fearful face
rectly from unimodal cortices as suggested by the FFA fusiform face area
‘‘communication relay’’ model or via multimodal fH bimodal trial with fearful voice
cortices in posterior STS. and happy face.
Activation in the left amygdala was found to fMRI functional magnetic resonance
correlate with changes in rating of emotional facial imaging
expressions induced by a simultaneously presented H0 null hypothesis
fearful voice (Ethofer et al., 2006). Correlation of hF bimodal trial with happy voice
amygdalar activity with behavioural effects sug- and fearful face
gests that the amygdala modulates cognitive judge- hH bimodal trial with happy voice
ments. This finding is consistent with previous and happy face
suggestions implicating the amygdala in integration MNI Montreal Neurological Institute
of emotional information obtained from different nE bimodal trial with neutral voice
modalities, particularly if this information signals and emotional face
359

nN bimodal trial with neutral voice in sensory-specific cortices during crossmodal binding. Ne-
and neutral face uroReport, 10: 2619–2623.
PET positron emission tomography Calvert, G.A., Campbell, R. and Brammer, M. (2000) Evidence
from functional magnetic resonance imaging of crossmodal
PPI psychophysiological interaction binding in human heteromodal cortex. Curr. Biol., 10: 649–657.
SAM self-assessment manikin Calvert, G.A., Hansen, P.C., Iversen, S.D. and Brammer, M.J.
SPM statistical parametric mapping (2001) Detection of audio-visual integration in humans by
STS superior temporal sulcus application of electrophysiological criteria to the BOLD
SVC small volume correction effect. NeuroImage, 14: 427–438.
Calvert, G.A. and Thesen, T. (2004) Multisensory integration:
V visual methodological approaches and emerging principles in the
human brain. J. Physiol. Paris, 98: 191–205.
Chavis, D.A. and Pandya, D.N. (1976) Further observations on
Acknowledgments corticofrontal connections in the rhesus monkey. Brain Res.,
117: 369–386.
Collins, D.L., Neelin, P., Peters, T.M. and Evans, A.C. (1994)
This study was supported by the Deutsche Forsc- Automatic 3D intersubject registration of MR volumetric
hungsgemeinschaft (SFB 550) and by the Junior data in standardized Talairach space. J. Comput. Assist.
Science Programme of the Heildelberger Academy Tomogr., 18: 192–205.
of Sciences and Humanities. Damasio, A.R. (1989) Time-locke multiregional retroactiva-
tion: a systems-level proposal for the neural substrates of
recall and recognition. Cognition, 33: 25–62.
References de Gelder, B., Böcker, K.B.E., Tuomainen, J., Hensen, M.
and Vroomen, J. (1999) The combined perception of emo-
Adolphs, R., Tranel, D., Damasio, H. and Damasio, A. (1994) tion from voice and face: early interaction revealed by human
Impaired recognition of emotion in facial expressions fol- electric brain responses. Neurosci. Lett., 260: 133–136.
lowing bilateral damage to the human amygdala. Nature, de Gelder, B. and Vroomen, J. (2000) The perception of emo-
372: 669–672. tion by ear and eye. Cognit. Emotion, 14: 289–312.
Barraclough, N.F., Xiao, D., Baker, C.I., Oram, M.W. and Dolan, R.J., Morris, J. and de Gelder, B. (2001) Crossmodal
Perret, D.I. (2005) Integration of visual and auditory infor- binding of fear in voice and face. Proc. Natl. Acad. Sci., 98:
mation by superior temporal sulcus neurons responsive to the 10006–10010.
sight of actions. J. Cogn. Neurosci., 17: 377–391. Driver, J. and Spence, C. (2000) Multisensory perception: be-
Beauchamp, M.S. (2005) Statistical criteria in FMRI studies of yond modularity and convergence. Curr. Biol., 10: 731–735.
multisensory integration. Neuroinformatics, 3: 93–113. Ekman, P. and Friesen, W.V. (1976) Pictures of Facial Affect.
Beauchamp, M.S., Argall, B.D., Bodurka, J., Duyn, J.H. and Consulting Psychologists Press, Palo Alto.
Martin, A. (2004a) Unravelling multisensory integration: Ethofer, T., Anders, S., Erb, M., Droll, C., Royen, L., Saur, R.,
patchy organization within human STS multisensory cortex. Reiterer, S., Grodd, W. and Wildgruber, D. (2006). Impact of
Nat. Neurosci., 7: 1190–1192. voice on emotional judgment of faces: an event-related fMRI
Beauchamp, M.S., Lee, K.E., Argall, B.D. and Martin, A. study. Hum. Brain Mapp., In press. DOI: 10.1002/hbm.20212.
(2004b) Integration of auditory and visual information about Fallon, J.H., Benevento, L.A. and Loe, P.R. (1978) Frequency-
objects in superior temporal sulcus. Neuron, 41: 809–823. dependent inhibition to tones in neurons of cat insular cortex
Bradley, M.M. and Lang, P.J. (1994) Measuring emotion: The (AIV). Brain Res., 779: 314–319.
self-assessment manikin and the semantic differential. J. Be- Fox, M.D., Snyder, A.Z., Vincent, J.L., Corbetta, M., Van Es-
hav. Ther. Exp. Psychiatry, 225: 49–59. sen, D.C. and Raichle, M.E. (2005) The human brain is in-
Breiter, H.C., Etcoff, N.L., Whalen, P.J., Kennedy, W.A., Ra- trinsically organized into dynamic anticorrelated functional
uch, S.L., Buckner, R.L., Strauss, M.M., Hyman, S.E. and networks. Proc. Natl. Acad. Sci., 102: 9673–9678.
Rosen, B.R. (1996) Response and habituation of the human Fries, W. (1984) Cortical projections to the superior colliculus
amygdala during visual processing of facial expression. Neu- in the macaque monkey: a retrograde study using horseradish
ron, 2: 875–887. peroxidase. J. Comp. Neurol., 230: 55–76.
Bruce, C.J., Desimone, R. and Gross, C.G. (1981) Visual prop- Friston, K.J., Buechel, C., Fink, G.R., Morris, J., Rolls, E. and
erties of neurons in a polysensory area in superior temporal Dolan, R.J. (1997) Psychophysiological and modulatory in-
sulcus of the macaque. J. Neurophysiol., 46: 369–384. teractions in neuroimaging. NeuroImage, 6: 218–229.
Calvert, G.A. (2001) Crossmodal processing in the human Friston, K.J., Harrison, L. and Penny, W. (2003) Dynamic
brain: Insights from functional neuroimaging studies. Cereb. causal modelling. NeuroImage, 19: 1273–1302.
Cortex, 11: 1110–1123. Friston, K.J., Holmes, A.P., Price, C.J., Buechel, C. and Wo-
Calvert, G.A., Brammer, M.J., Bullmore, E.T., Campbell, R., rsley, K.J. (1999) Multisubject fMRI studies and conjunction
Iversen, S.D. and David, S.A. (1999) Response amplification analyses. NeuroImage, 10: 385–396.
360

Friston, K.J., Josephs, O., Rees, G. and Turner, R. (1998) Meredith, M.A. and Stein, B.E. (1983) Interactions among
Nonlinear event-related responses in fMRI. Magn. Res. converging sensory inputs in the superior colliculus. Science,
Med., 39: 41–52. 221: 389–391.
Friston, K.J., Mechelli, A., Turner, R. and Price, C.J. (2000) Mesulam, M.M. (1998) From sensation to cognition. Brain,
Nonlinear responses in fMRI: the balloon model, volterra 121: 1013–1052.
kernels, and other hemodynamics. NeuroImage, 12: 466–477. Mesulam, M.M. and Mufson, E.J. (1982) Insula of the old
Friston, K.J., Penny, W. and Glaser, D.E. (2005) Conjunction world monkey. III: efferent cortical output and comments on
revisited. NeuroImage, 25: 661–667. function. J. Comp. Neurol., 212: 38–52.
Gitelman, D.R., Penny, W.D., Ashburner, J. and Friston, K.J. Miller, J.O. (1982) Divided attention: evidence for coactivation
(2003) Modelling regional and psychophysiologic interac- with redundant signals. Cogn. Psychol., 14: 247–279.
tions in fMRI: the importance of hemodynamic deconvolu- Miller, J.O. (1986) Time course of coactivation in bimodal
tion. NeuroImage, 19: 200–207. divided attention. Percept. Psychophys., 40: 331–343.
Gordon, B.G. (1973) Receptive fields in the deep layers of the Morris, J.S., Friston, K.J., Buechel, C., Frith, C.D., Young,
cat superior colliculus. J. Neurophysiol., 36: 157–178. A.W., Calder, A.J. and Dolan, R.J. (1998) A neuromodula-
Goldman-Racic, P. (1995) Architecture of the prefrontal cortex tory role for the human amygdala in processing emotional
and the central executive. In: Grafman, J., Holyoak, K. and facial expressions. Brain, 121: 47–57.
Bollder, F. (Eds.), Structure and Functions of the Human Morris, J.S., Frith, C.D., Perrett, D.I., Rowland, D., Young,
Prefrontal Cortex. The New York Academy of Sciences, NY, A.W., Calder, A.J. and Dolan, R.J. (1996) A differential
USA, pp. 71–83. response in the human amygdala to fearful and happy facial
Grefkes, C., Weiss, P.H., Zilles, K. and Fink, G.R. (2002) expressions. Nature, 383: 812–815.
Crossmodal processing of object features in human anterior Mufson, E.J. and Mesulam, M.M. (1984) Thalamic connections
intraparietal cortex: an fMRI study implies equivalencies be- of the insula in the rhesus monkey and comments on the
tween humans and monkeys. Neuron, 35: 173–184. paralimbic connectivity of the medial pulvinar nucleus.
Hikosaka, K., Iwai, E., Saito, H. and Tanaka, K. (1988) J. Comp. Neurol., 227: 109–120.
Polysensory properties of neurons in the anterior bank of the Murray, E.A. and Mishkin, M. (1985) Amygdalactomy impairs
caudal superior temporal sulcus of the macaque monkey. J. crossmodal association in monkeys. Science, 228: 604–606.
Neurophysiol., 60: 1615–1637. Nichols, T., Brett, M., Andersson, J., Wager, T. and Poline,
Jones, J.A. and Callan, D.E. (2003) Brain activity during au- J.B. (2005) Valid conjunction inference with the minimum
diovisual speech perception: an fMRI study of the McGurk statistic. NeuroImage, 25: 653–660.
effect. NeuroReport, 14: 1129–1133. Olson, I.R., Gatenby, J.C. and Gore, J.C. (2002) A comparison
Jones, E.G. and Powell, T.P.S. (1970) An anatomical study of of bound and unbound audio–visual information processing
converging sensory pathways within the cerebral cortex of in the human cerebral cortex. Brain Res. Cogn. Brain Res.,
the monkey. Brain, 93: 793–820. 14: 129–138.
Kanwisher, N., Mc Dermott, J. and Chun, M.M. (1997) The Peck, C.K. (1987) Auditory interactions in cat’s superior col-
fusiform face area: a module in human extrastriate cortex liculus: their role in the control of gaze. Brain Res., 420:
specialized for face processing. J. Neurosci., 17: 4302–4311. 162–166.
Lewis, J.W., Beauchamp, M.S. and DeYoe, E.A. (2000) A Pearson, R.C., Brodal, P., Gatter, K.C. and Powell, T.P.
comparison of visual and auditory motion processing in hu- (1982) The organisation of the connections between the cor-
man cerebral cortex. Cereb. Cortex, 10: 873–888. tex and the claustrum in the monkey. Brain Res., 234:
Loe, P.R. and Benevento, L.A. (1969) Auditory–visual inter- 435–441.
action in single units in the orbito-insular cortex of the cat. Phillips, M.L., Young, A.W., Scott, S.K., Calder, A.J., Andrew,
Electroencephalogr. Clin. Neurophysiol., 26: 395–398. C., Giampietro, V., Williams, S.C., Bullmore, E.T., Brammer,
Macaluso, E., Frith, C.D. and Driver, J. (2000) Selective spatial M. and Gray, J.A. (1997) Neural responses to facial and vocal
attention in vision and touch: unimodal and multimodal expressions of fear and disgust. Proc. R. Soc. Lond. Ser. B,
mechanisms revealed by PET. J. Neurophysiol., 83: 3062–3075. 265: 1809–1817.
Massaro, D.W. and Egan, J.W. (1996) Perceptual recognition Pitkänen, A. (2000). Connectivity of the rat amygdaloid com-
of facial affect: cross-cultural comparisons. Mem. Cogn., 24: plex. In: The Amygdala: A Functional Analysis. Oxford
812–822. University Press, New York.
McDonald, A.J. (1998) Cortical pathways to the mammalian Pourtois, G., Debatisse, D., Despland, P.A. and de Gelder, B.
amygdala. Prog. Neurobiol., 55: 257–332. (2002) Facial expressions modulate the time course of long
McGurk, H. and MacDonald, J. (1976) Hearing lips and seeing latency auditory potentials. Cogn. Brain Res., 14: 99–105.
voices. Nature, 264: 746–748. Pourtois, G., de Gelder, B., Bol, A. and Crommelinck, M.
Mechelli, A., Price, C.J. and Friston, K.J. (2001) Nonlinear (2005) Perception of facial expression and voices and their
coupling between evoked rCBF and BOLD signals: a simu- combination in the human brain. Cortex, 41: 49–59.
lation study on hemodynamic responses. NeuroImage, 14: Pourtois, G., de Gelder, B., Vroomen, J., Rossion, B. and
862–872. Crommelinck, M. (2000) The time-course of intermodal
361

binding between seeing and hearing affective information. van Atteveldt, N., Formisano, E., Goebel, R. and Blomert,
NeuroReport, 11: 1329–1333. L. (2004) Integration of letters and speech sounds in the hu-
Price, C.J. and Friston, K.J. (1997) Cognitive conjunctions: a man brain. Neuron, 43: 271–282.
new approach to brain activation experiments. NeuroImage, von Kriegstein, K., Kleinschmidt, A., Sterzer, P. and Giraud,
5: 261–270. A.-L. (2005) Interaction of face and voice areas during
Raichle, M.E., MacLeod, A.M., Snyder, A.Z., Powers, W.J., speaker recognition. J. Cogn. Neurosci., 17: 367–376.
Gusnard, D.A. and Shulman, G.L. (2003) A default mode of Vroomen, J., Driver, J. and de Gelder, B. (2001) Is cross-modal
brain function. Proc. Natl. Acad. Sci., 98: 676–682. integration of emotional expressions independent of attent-
Schröger, E. and Widmann, A. (1998) Speeded responses to ional resources? Cogn. Affect. Behav. Neurosci., 1: 382–387.
audiovisual signal changes result from bimodal integration. Wallace, M.T., Meredith, M.A. and Stein, B.E. (1992) Integra-
Psychophysiology, 35: 755–759. tion of multiple sensory modalities in cat cortex. Exp. Brain.
Scott, S.K., Young, A.W., Calder, A.J., Hellawell, D.J., Aggl- Res., 91: 484–488.
eton, J.P. and Johnson, M. (1997) Impaired auditory recog- Wallace, M.T., Meredith, M.A. and Stein, B.E. (1993) Con-
nition of fear and anger following bilateral amygdala lesions. verging influences from visual, auditory, and somatosensory
Nature, 385: 254–257. cortices onto output neurons of the superior colliculus.
Seltzer, B. and Pandya, D.N. (1978) Afferent cortical connections and J. Neurophysiol., 69: 1797–1809.
architectonics of the superior temporal sulcus and surrounding cortex Wallace, M.T., Wilkinson, L.K. and Stein, B.E. (1996) Repre-
in the rhesus monkey. Brain Res., 149: 1–24. sentation and integration of multiple sensory inputs in pri-
Stein, B.E., London, N., Wilkinson, L.K. and Price, D.D. (1996) mate superior colliculus. J. Neurophysiol., 76: 1246–1266.
Enhancement of perceived visual intensity by auditory stimuli: White, M. (1999) Representation of facial expressions of emo-
a psychophysical analysis. J. Cogn. Neurosci., 8: 497–506. tion. Am. J. Psychol., 112: 371–381.
Stein, B.E. and Meredith, M.A. (1993) Merging of Senses. MIT Worsley, K.J., Marrett, S., Neelin, P., Vandal, A.C., Friston,
Press, Cambridge. K.J. and Evans, A.C. (1996) A unified statistical approach
Stein, B.E. and Wallace, M.T. (1996) Comparison of cross- for determining significant signals in images of cerebral ac-
modal integration in midbrain and cortex. Prog. Brain Res., tivation. Hum. Brain Mapp., 4: 58–73.
112: 289–299. Wright, T.M., Pelphrey, K.A., Allison, T., McKeown, M.J. and
Turner, B.H., Mishkin, M. and Knapp, M. (1980) Organization McCarthy, G. (2003) Polysensory interactions along lateral
of the amygdalopetal projections from modality-specific temporal regions evoked by audiovisual speech. Cereb. Cor-
cortical association areas. J. Comp. Neurol., 191: 515–543. tex,, 13: 34–43.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 20

Role of the amygdala in processing visual social


stimuli

Ralph Adolphs and Michael Spezio

Division of the Humanities and Social Sciences, HSS 228-77, California Institute of Technology, Pasadena,
CA 91125, USA

Abstract: We review the evidence implicating the amygdala as a critical component of a neural network of
social cognition, drawing especially on research involving the processing of faces and other visual social
stimuli. We argue that, although it is clear that social behavioral representations are not stored in the
amygdala, the most parsimonious interpretation of the data is that the amygdala plays a role in guiding
social behaviors on the basis of socioenvironmental context. Thus, it appears to be required for normal
social cognition. We propose that the amygdala plays this role by attentionally modulating several areas of
visual and somatosensory cortex that have been implicated in social cognition, and in helping to direct
overt visuospatial attention in face gaze. We also hypothesize that the amygdala exerts attentional mod-
ulation of simulation in somatosensory cortices such as supramarginal gyrus and insula. Finally, we argue
that the term emotion be broadened to include increased attention to bodily responses and their repre-
sentation in cortex.

Keywords: amygdala; face processing; simulation; lesion studies; social cognition; emotion

Introduction severely limit a primate’s range of social responses,


perhaps going so far as to eliminate some part or
The amygdala has long been implicated in primate all of the social repertoire altogether. More recent
social cognition and behavior, due primarily to the findings challenge the view that the amygdala is
well-known work by Kluver and Bucy (1939) and required for basic social behaviors. Yet the ques-
the studies by Kling and colleagues (Dicks et al., tion remains open whether the amygdala is a re-
1968; Kling, 1968, 1974; Kling et al., 1970, 1979; quired component for normal social cognition. For
Brothers et al., 1990). An influential view of the example, is the amygdala necessary for the normal
amygdala emerging from early studies of its func- information processing associated with an organ-
tion was that it acts as a generative locus of social ism’s evaluation of a visual social stimulus, such as
cognition and behavior, required to link the per- a facial expression (on which subsequent behaviors
ception of any stimuli to information about their could then be based)? We will see that an answer to
value to an organism (Weiskrantz, 1956). One this question depends on a new consideration of
interpretation of this view is that the amygdala is a evidence for the amygdala’s role. The view we will
primary source of social behavior, and the lack of present takes into consideration evidence regard-
a functioning amygdala would be expected to ing the amygdala’s role in modulating autonomic
arousal, new evidence regarding the amygdala’s
Corresponding author. Tel.: +1-626-395-4486; potential to affect visuospatial and visual object-
Fax: +1-626-793-8580; E-mail: radolphs@hss.caltech.edu based attention, and recent accounts that explain

DOI: 10.1016/S0079-6123(06)56020-0 363


364

social cognition in terms of simulation theory. rather than the placidity that tends to be seen in
These newer developments have posed something adults (Prather et al., 2001; Bauman et al., 2004b).
of a puzzle for older theories of the amygdala. We Furthermore, there are notable differences between
will review some older findings first, and the (the relatively small number of) different studies of
framework that was based on them. Then we will how amygdala lesions in monkeys affect basic so-
introduce the new findings and framework, and cial behaviors such as canonical facial expressions,
end by proposing a framework describing the amy- bodily postures (e.g., ‘‘present groom’’), and at-
gdala’s function in recognizing the social value of tachment behaviors. These differences between
stimuli. studies likely reflect effects of additional factors
such as lesion methodology and extent, lab- vs. wild
rearing, or the exact species used (Bachevalier and
Shifting views of the amygdala in social cognition Loveland, 2006). For example, there is evidence
suggesting that amygdala lesions profoundly impair
Studies of the primate amygdala began in the 1930s even basic social behaviors in monkeys (Bachevalier
with Kluver and Bucy’s well-known experiments in et al., 1999; Meunier et al., 1999; Meunier and
monkeys (Kluver and Bucy, 1937, 1997). Following Bachevalier, 2002). Amaral and colleagues, how-
large bitemporal lesions that encompassed the am- ever, found no impairment in basic social behaviors
ygdala, the animals came to exhibit a constellation following selective neurotoxic amygdala lesions,
of impairments in recognizing the emotional and though they did find impairments in the appropri-
social meaning of stimuli—the so-called ‘‘psychic ate deployment of these social behaviors (Bauman
blindness’’ of Kluver–Bucy syndrome. Notably, the et al., 2004a, b). Given this heterogeneity, it has
monkeys became exceptionally tame and placid, a been argued that the amygdala is not required for
behavioral abnormality that has been replicated to monkeys to show the full repertoire of social be-
some degree in later studies (Emery et al., 2001; haviors, because under some circumstances animals
Kalin et al., 2001, 2004; Izquierdo et al., 2005). The with complete bilateral amygdala lesions, neverthe-
animals also exhibited a variety of other unusual less, can show all the components of emotional and
behaviors, including hypermetamorphosis and social behaviors that normal monkeys would show
hypersexuality that have not been so reliably rep- (Amaral et al., 2003a), even if they are deployed
licated. Modern-day studies using selective neuro- abnormally (Bauman et al., 2004a).
toxins to lesion the amygdala, sparing surrounding The most parsimonious interpretation of the
tissues, not surprisingly provide a much more data thus far is rather that the amygdala plays a
muted and selective picture. Such selective lesions, role in guiding social behaviors on the basis of the
like the earlier lesions, do result in the lack of a socioenvironmental context with which an animal
normal ‘‘brake’’ on behavior, and the animals tend is faced. It is important to keep in mind that the
to approach objects and situations that normal socioenvironmental context is likely to include not
monkeys would avoid—they are also seldom re- only what is available to the monkey in the im-
garded as dominant by other monkeys (Meunier mediate circumstance, but also information that
et al., 1999; Emery et al., 2001). Yet the selective has been neurally encoded throughout develop-
amygdala lesions do not produce monkeys that ex- ment as the monkey has learned and adapted to
hibit the array of unusual behavior as Kluver and the surrounding world. Here we see a shift in
Bucy described. viewpoint from the amygdala as a structure that
The recent lesion studies in monkeys have also itself stores and activates patterns of basic social
begun to highlight how complex the role of the behaviors to one in which the amygdala plays an
amygdala in regulating social behavior is likely to influential role in the deployment of these be-
be. The consequences of amygdala lesions are quite haviors. This view predicts that primates lacking a
different depending on the age at which they are functional amygdala retain the ability to display
made, and infant monkeys with amygdala lesions the full range of basic social behaviors while being
actually show exaggerated social fear responses impaired in the appropriate context-dependent
365

deployment of these behaviors and of more com- (happiness, surprise, fear, anger, disgust, and sad-
plex social behaviors. The idea is similar to the ness), SM was insensitive to the intensity of the
difference between a novice chess player who emotion shown in fear, but not in other expressions
knows how each piece moves and even several (Fig. 1D) (Adolphs et al., 1994). The specificity to
useful openings and a grand master who can rap- fear was confirmed using morphs (linear blends)
idly choose the appropriate move among a myriad between the different emotions: the closer to the
of options. This shift in understanding of the amy- fear prototype the emotional expressions were, the
gdala’s role in social cognition to some extent more impaired SM’s recognition became. The im-
parallels debates regarding the amygdala’s role in pairment was all the more striking because she was
certain forms of memory. As with social behavior, able to recognize other kinds of information from
declarative memory is not stored in the amygdala fearful faces normally (such as judging their gender,
as such, but is influenced by the amygdala’s age, or identity), and because she was able to dis-
processing and projection into other structures, criminate the faces normally when presented pair-
such as the hippocampus in the case of memory wise (on same/different judgments). When shown
(McGaugh, 2004). As we will see below, this new an expression of fear and asked in an unconstrained
framework of the amygdala’s role in social cogni- task simply to name the emotion, she typically re-
tion is supported by a number of studies showing plied that she did not know what the emotion was.
that the amygdala influences the evaluation of If forced, she would often mistake fear for surprise,
stimuli in contributing to the perception, recogni- anger, or sadness (but never happiness).
tion, and judgment of socially relevant stimuli. SM’s impaired recognition of fear was followed
up in a series of studies that showed that SM does
have at least some components of the concept of
Impaired social cognition in humans following fear, because she can use the word relatively ap-
amygdala damage propriately in conversation, she believes she knows
what fear is, and she can in fact retrieve many facts
While the amygdala has been implicated in mon- about fear (such that being confronted with a bear
key social behavior for some time, it is only very would make one afraid, etc.) (Adolphs et al.,
recently that such a role has been established in 1995). While the amygdala’s role in recognizing
humans, and that detailed hypotheses have been fear in other sensory modalities remains unclear, in
investigated regarding the underlying mechanisms. SM’s case she was even able to recognize fear from
Here, we review evidence that the amygdala has a tone of voice (prosody) normally (Adolphs and
role in the recognition of emotion from faces, in Tranel, 1999). But she could not recognize it from
interpreting eye gaze, and in more complex social the face, nor could she generate an image of the
judgments in humans. Two early studies showed facial expression when given the name (e.g., when
that bilateral damage confined mainly to the am- asked to draw it) (Adolphs et al., 1995). The im-
ygdala resulted in a disproportionately severe im- paired recognition of fear thus seemed to be rela-
pairment in the ability to recognize fear from facial tively specific to the facial expression—the use of
expressions (Adolphs et al., 1994; Young et al., other visual information, such as context and body
1995). One patient in particular, SM, had damage posture, was less compromised. In fact, adding fa-
that was relatively restricted to the amygdala (Fig. cial expressions to scenes initially devoid of faces
1A–C), and an impairment that was very specific decreased the accuracy of emotion recognition in
to the recognition of fear (Adolphs et al., 2000). subjects with bilateral amygdala damage, whereas
SM’s lesion encompassed the entire amygdala bi- it increased it in healthy subjects (Adolphs and
laterally and extended also into anterior portions Tranel, 2003).
of the entorhinal cortex; there was no damage evi- Several studies have followed up these initial
dent to any other structures. findings. Other lesion studies have found impaired
When shown standardized emotional facial ex- recognition of fear from facial expressions following
pressions that depicted the six ‘‘basic’’ emotions bilateral amygdala damage (Calder et al., 1996;
366

Fig. 1. Bilateral amygdala lesions impair recognition of emotion from faces. There are two structural MRI slices (A) showing intact
hippocampus (B) and bilateral amygdala lesion (C) for SM. SM’s ratings of the degree to which a face expressed a particular emotion
are shown in (D) (SM: closed triangles; controls: open circles) and the correlation between SM’s ratings and normal ratings is shown in
(E) for each facial expression. From Adolphs et al. (1994).
367

Broks et al., 1998; Adolphs et al., 1999b), and func- with more selective amygdala lesions, and by using
tional imaging studies have found activation of the more sophisticated ways of assessing social be-
amygdala when subjects view fearful faces (Breiter havior (Emery and Amaral, 1999; Emery et al.,
et al., 1996; Morris et al., 1996; Whalen et al., 2001). 2001), and consistent findings have been shown
However, the findings are not as specific as in the now also in humans. We have found that the amy-
case of SM. Several lesion subjects with complete gdala is important for judging complex mental
bilateral amygdala damage were impaired also on states and social emotions from faces (Adolphs
emotions other than fear (always negatively va- et al., 2002), and for judging the trustworthiness of
lenced emotions) (Adolphs et al., 1999b), and in people from viewing their face (Adolphs et al.,
several cases their impairment in recognizing anger, 1998; Winston et al., 2002). Relatedly, the am-
disgust, or sadness was more severe than their im- ygdala shows differential habituation of activation
pairment in recognizing fear. Similarly, functional to faces of people of another race (Hart et al.,
imaging studies found activation of the amygdala 2000), and amygdala activation has been found to
to expressions other than fear, such as happiness correlate with race stereotypes of which the viewer
(Canli et al., 2002; Williams et al., 2005), surprise may be unaware (Phelps et al., 2000). On the basis
(Kim et al., 2004), sadness (Wang et al., 2005), and of these findings, some recent studies suggest a
anger (Whalen et al., 2001). general role for the amygdala in so-called ‘‘theory
A further function of the amygdala in processing of mind’’ abilities: the collection of abilities whereby
aspects of faces comes from studies of its role in we attribute internal mental states, intentions, de-
processing the eyes in a face. The eyes and their di- sires, and emotions to other people (Baron-Cohen
rection of gaze are key social signals in many species et al., 2000; Fine et al., 2001).
(Emery, 2000), especially apes and humans, whose Various theories have been put forth to account
white sclera makes the pupil more easily visible and for some of these findings, some proposing that the
permits better discrimination of gaze. Eyes signal amygdala is specialized for recognition of emotions
important information about emotional states, and that are high in arousal (Adolphs et al., 1999a), or
there is evidence from functional imaging studies that relate to withdrawal (Anderson et al., 2000), or
that at least some of this processing recruits the amy- that require disambiguation (Whalen, 1999). It is
gdala (Baron-Cohen et al., 1999; Kawashima et al., fair to say that, at present, there is no single ac-
1999; Wicker et al., 2003b). The amygdala’s involve- cepted scheme to explain which emotion categories
ment in processing gaze direction in emotional faces are affected by amygdala damage.
has been explored recently. It was found that direct These differences notwithstanding, we can iden-
gaze facilitated amygdala activation in response to tify a general framework for understanding the
approach-oriented emotions such as anger, whereas mechanisms by which the amygdala normally con-
averted gaze facilitated amygdala activation to tributes to emotion judgment and social cognition.
avoidance-oriented emotions such as fear (Adams The framework is built upon (1) recent work show-
and Kleck, 2003). Further, the amygdala has been ing (a) the amygdala’s ability to influence visual
found to be active during monitoring for direct gaze processing at early stages, and (b) the amygdala’s
(Hooker et al., 2003). role in influencing overt attention to the eyes in a
The amygdala’s role is not limited to making face; (2) the amygdala’s role in autonomic arousal;
judgments about basic emotions, but includes a role and (3) work implicating the pulvinar and Brod-
in making social judgments, as well. This function mann area 40 (SII) in the processing of affectively
was already suggested by earlier studies in nonhu- aversive visual stimuli. Each of these elements is
man primates (Kluver and Bucy, 1937; Rosvold supported by evidence from neuroanatomical stud-
et al., 1954; Brothers et al., 1990; Kling and Broth- ies of the internal and external connectivity of the
ers, 1992), which demonstrated impaired social be- amygdala. Our current neuroanatomical under-
havior following amygdala damage and amygdala standing of the amygdala, which consists of a
responses to complex social stimuli. They have been number of separate nuclei in primates (Price, 2003),
corroborated in recent times by studies in monkeys supports a scheme whereby faces are associated
368

with their emotional meaning in the lateral and higher order cortices in temporal lobe first encode
basolateral nuclei, in interaction with additional the visual properties of socially relevant stimuli,
brain structures such as orbitofrontal and medial and that this information is then subsequently
prefrontal cortices (Ghashghaei and Barbas, 2002). passed to neurons within the ventromedial pre-
This evaluation is conveyed to central and baso- frontal cortex and the amygdala that associate the
medial amygdala nuclei whose projections then in- visual percept with its emotional meaning.
fluence processing in visual cortex, processing that This standard view of a strong feedforward input
elicits autonomic and motor responses in the body to the amygdala, one in which visual cortices in the
(Price, 2003), and/or processing that involves so- temporal lobe comprise a series of visual processing
matosensory areas putatively involved in simula- stages the later components of which feed into the
tion-based transformations of the visual percept to amygdala, is now being modified. Accumulating ev-
an internal bodily representation (Gallese et al., idence strongly supports the notion that the amy-
2004; Rizzolatti and Craighero, 2004; see also gdala can directly influence visual processing, even
Keysers and Gazzola, this volume). We will con- at very early stages. Recent anatomical studies show
sider each of these aspects of the amygdala’s func- that the amygdala projects topographically to the
tion in social cognition. ventral visual stream, from rostral temporal cortical
area TE to caudal primary visual cortex (V1) (Am-
aral et al., 2003b). A majority of projections from
The amygdala influences early visual processing of the basal nucleus to V1 and TE colocalize with
faces and affective stimuli synaptophysin, suggesting that the amygdala can
exert direct influence on synaptic associations at
There is abundant data regarding the cortical multiple stages of primary and object-based visual
processing of faces, and such cortical processing processing (Freese and Amaral, 2005). Such direct
presumably can serve to provide highly processed influence on cortical visual processing may be a later
input to the amygdala. To briefly review this, evolutionary adaptation, as these anatomical pro-
functional magnetic resonance imaging (fMRI) jections have not been reported in rats and cats
studies have revealed an array of higher order vis- (Price, 2003). Through this architecture in primates,
ual cortical regions that are engaged in face the amygdala can link the perception of stimuli
processing, including the fusiform face area to an emotional response, and then subsequently
(FFA) in the fusiform gyrus, the face-sensitive modulate cognition on the basis of the value of the
area in the superior temporal sulcus (STS), and perceived stimulus (Amaral and Insausti, 1992;
superior and middle temporal gyrus (Kanwisher Adolphs, 2002). Thus, perception and evaluation
et al., 1997; McCarthy, 1999; Haxby et al., 2000). of faces are closely intertwined.
The STS in particular has been implicated in the Functional neuroimaging in humans indicates
detection of gaze direction in humans and nonhu- that these structural pathways from the amygdala to
man primates (Campbell et al., 1990; Puce et al., visual areas are put to use in social cognition, spe-
1998; Wicker et al., 1998, 2003b; Calder et al., cifically in the modulation of attention. Activation
2002; Hooker et al., 2003; Pourtois et al., 2004). A in the amygdala has been shown to predict extra-
distributed array of visually responsive regions in striate cortex activation specific to fearful facial ex-
the temporal lobe appears to encode classes of bi- pressions (Morris et al., 1998a). Lesions of the
ologically salient objects, notably faces and bodies, amygdala eliminate facial expression-specific acti-
in humans (Downing et al., 2001; Haxby et al., vations in occipital and fusiform cortices (Vuilleu-
2001; Spiridon and Kanwisher, 2002) as in mon- mier et al., 2004). Such findings are consistent with
keys (Pinsk et al., 2005). Regions in the superior the dependence of visual processing on prior amy-
temporal lobe appear specialized to process bio- gdala processing of visual information, in a manner
logical motion stimuli, such as point-light displays specific to the information’s associated value for the
of people (Haxby et al., 2000; Grossman and organism. Even more striking is evidence from sin-
Blake, 2002). It has been generally supposed that gle unit studies of face-selective neurons in TE and
369

STS of macaque monkeys (Sugase et al., 1999). consistent with evidence from neuroimaging of sub-
These neurons discriminate between faces and ob- liminal processing of faces in healthy controls
jects about 50 ms faster than they discriminate bet- (Morris et al., 1998b). Further evidence supporting
ween facial expressions, which is enough time for the involvement of a pulvinar–amygdala–infero-
the action of projections from the amygdala. Ex- temporal pathway in the rapid visual processing of
pression-dependent activity in these neurons occurs emotional stimuli comes from a patient who sus-
within 150 ms following stimulus onset, consistent tained a complete and focal loss of the left pulvinar
with the notion that input from the amygdala occurs (Ward et al., 2005). In a paradigm designed to
early in visual processing. measure how threatening images interfere with a
Clearly, rapid input of visual information to the goal-directed task, the patient’s behavior indicated
amygdala is required for the amygdala to exert an that the threatening images interfered with subse-
expression-dependent influence on the ventral vis- quent color identification of a single letter (‘‘O’’)
ual system prior to the observed expression-de- when the images were presented to the ipsilesional
pendent activity in the latter system. The medial field, but no interference was observed when the
nucleus of the pulvinar complex provides such a threatening images were presented to the contrales-
pathway, as it forms a strong projection to the lat- ional field. Interference by images in the contrales-
eral and basolateral nuclei of the amygdala in mac- ional field returned if they were displayed for a
aque monkeys (Jones and Burton, 1976; Aggleton relatively long time (600 ms vs. 300 ms). In light of
et al., 1980; Romanski et al., 1997). There is now the evidence presented here, it appears that the
evidence that these connections exist and are func- pulvinar–amygdala pathway is required for the ex-
tionally active in humans. In healthy controls, tremely rapid processing of threat, and is capable of
masked facial stimuli activate the amygdala in the using the results of this processing to influence vis-
absence of awareness (Ohman, 2005), together with ual perception in primary and higher visual cortices.
activation of the superior colliculus and the
pulvinar (Liddell et al., 2005). Functional connec-
tivity of the right amygdala with the right pulvinar The amygdala influences face gaze
and superior colliculus increases, and connectivity
with fusiform and orbitofrontal cortices decreases, In addition to influencing visual processing even at
during subliminal presentation of fear-conditioned very early stages, recent evidence suggests that the
faces (Morris et al., 1998b, 1999). The left amygdala amygdala affects face gaze in a surprisingly direct
shows no masking-dependent changes in connec- manner (Adolphs et al., 2005). This is consistent
tivity. A patient with blindsight (i.e., residual visual with the amygdala’s influence on visual processing
capacity without perceptual awareness) in the right and with previous work showing that the amygdala
cortical field nevertheless showed preserved ability affects visual and visuospatial attention. Lesions of
to guess correctly the identity of facial expressions the amygdala, particularly of the left amygdala, se-
presented to his blind hemifield (de Gelder et al., riously impair the attentional benefit in the percep-
1999). Both fearful and fear-conditioned faces pre- tion of aversive words during an attentional blink
sented to the blind hemifield increased the func- paradigm involving rapid stimulus presentation and
tional connectivity between the right amygdala, target detection (Anderson and Phelps, 2001). Emo-
superior colliculus, and posterior thalamus (i.e., the tional facial expressions and aversive nonfacial stim-
pulvinar) (Morris et al., 2001). A recent study of a uli overcome attentional deficits in patients showing
patient with total cortical blindness (i.e., destruction neglect due to right parietal lesions (Vuilleumier and
of bilateral visual cortices) found that the patient Schwartz, 2001a, b). It is likely that the latter finding
could correctly guess the facial expression of a dis- is the result of exogenous attentional modulation by
played face, but could not guess the identity of the amygdala in visual cortex and perhaps in visu-
other stimuli, i.e., emotional or not (Pegna et al., ally responsive prefrontal cortex. Recall that the
2005). The right but not the left amygdala in this amygdala is required for facial expression-specific
patient showed expression-dependent activation, activation of early visual cortex (Vuilleumier et al.,
370

2004), evidence that fits well within an understand- being required for basic social behaviors. SM
ing of the amygdala as part of an attentional net- clearly displayed direct eye gaze after being in-
work responsive to visual stimuli having value for structed to do so, and was even able to use the
an organism. information that direct eye gaze provided to fully
We have seen that the amygdala influences infor- recover her recognition of fear faces. An absence
mation processing in visual cortices and that it is of a functioning amygdala thus does not result in a
strongly implicated in attention to evaluatively sa- loss of the ability to engage in the social behavior
lient stimuli. It is possible, then, that the amygdala of direct eye gaze. However, the amygdala is
plays a role, via its projections to visual cortex par- required for the appropriate deployment of this
ticularly, in directing overt attention during the ex- social behavior via its processing of socioenviron-
ploration of a face in social judgment. Face gaze, mental context and its influence on visual attent-
that is, might be dependent on evaluative processing ional systems involved in social cognition. We will
within the amygdala. More specifically, in light of see this theme reappear in relation to the amy-
the evidence that the amygdala is sensitive to gaze gdala’s role in autonomic arousal to facial expres-
direction in a face, it is likely that an attentional role sions and in visuosomatosensoric processing of
for the amygdala would include directing gaze to the facial expressions of emotion.
eyes in a face. Indeed, a recent study of face gaze in
a patient with bilateral amygdala damage supports
this view (Adolphs et al., 2005). The amygdala mediates autonomic arousal elicited
The study tested a patient (SM) with complete by faces
and focal bilateral amygdala lesions during emotion
judgment, measuring both face gaze and the use of The human amygdala was originally thought to
facial information. To understand SM’s use of fa- have a key role in the generation of normal auto-
cial information during a simple emotion judgment, nomic responses associated with orienting and
the study used the Bubbles technique and compared arousal due to studies of amygdalectomized mon-
the result with those obtained from typical, age- keys (Bagshaw and Benzies, 1968; Bagshaw and
matched controls. SM displayed a marked reduc- Coppock, 1968; Pribram et al., 1979). Monkeys
tion impairment of the ability to use the eyes in a with bilateral amygdalectomies fail to produce the
face, compared to controls (Fig. 2A–D). Subse- expected changes in skin conductance response
quent investigation of SM’s face gaze using eye- (SCR), heart rate, and respiratory rate in response
tracking revealed a near absence of gaze to the eyes to irregularly repeated sounds, while ear move-
during visual exploration of faces (Fig. 2E–G). ments to the sounds are normal (Bagshaw and
When SM was instructed to look only at the eyes Benzies, 1968). Further, these animals show no
while making emotion judgments from faces (Fig. Pavlovian conditioning of SCR when a condi-
2H), performance in recognizing fear returned to tioned stimulus is paired with electrical stimulation
normal (Fig. 2I). Yet this remarkable recovery in to the skin (Bagshaw and Coppock, 1968; Pribram
emotion judgment was not sustained once SM went et al., 1979), although they do show normal
back to nondirected, free viewing of faces. These SCR in response to the unconditioned stimulus
results provide the first evidence showing a require- (Bagshaw and Coppock, 1968). However, in
ment for the amygdala in direct eye gaze, extending humans, amygdala lesions appear not to affect
our understanding of the amygdala’s influential orienting SCRs (Tranel and Damasio, 1989), while
role in visuospatial attention to faces during social severely impairing Pavlovian conditioning of
judgment. SCRs (Bechara et al., 1995). It is therefore the
In keeping with the new view of the amygdala linking of a conditioned stimulus with an uncon-
described in the section ‘‘Shifting views of the ditioned stimulus and its associated autonomic re-
amygdala in social cognition,’’ these findings sup- sponse that requires the amygdala, and not the
port the notion that the amygdala is a crucial generation of the autonomic response, associated
component of normal social cognition, while not with orienting or otherwise.
371

Fig. 2. Bilateral amygdala lesions impair the use of the eyes and gaze to the eyes during emotion judgment. Using the Bubbles method
(see Adolphs et al., 2005) to identify face areas used during emotion judgment, SM (B) differed from controls (A), such that controls
exhibited much greater use of the eyes (C) than SM, while SM did not rely more on any area of the face than did controls (D). While
looking at whole faces, SM exhibited abnormal face gaze (E), making far fewer fixations to the eyes than did controls. This was
observed across emotions (F) and across tasks (G; free viewing, emotion judgment, gender discrimination). When SM was instructed to
look at the eyes (I, ‘‘SM eyes’’) in a whole face, she could do this (H), resulting in a remarkable recovery in ability to recognize the facial
expression of fear (I).

Neuroimaging studies of classical conditioning in the amygdala, while they found increased acti-
are consistent with the work using the lesion vation in the amygdala only with conditioned SCR
method. In an analysis contrasting SCR+ trials (Williams et al., 2000; Knight et al., 2005). A study
with SCR- trials in an orienting paradigm, activat- of SCR in several different cognitive/behavioral
ions occurred in the hippocampus, anterior cingu- tasks found that SCR covaries with activation in
late, and ventromedial prefrontal cortex, but not ventromedial prefrontal cortex (Brodmann 10/32),
372

supramarginal gyrus (Brodmann 40), cingulomotor the basomedial nucleus to the insula influence
cortex (Brodmann 6/24), posterior cingulate cortex these nuclei that appear crucial to the generating
(Brodmann 23/30), right cerebellum, and thalamus of SCR (Amaral and Price, 1984).
(Patterson et al., 2002). More recently, in a study of
the neural systems underlying arousal and SCR
elicited by static images, the brain area that most The amygdala and simulation: somatosensory
closely associated with SCR variability was the attention as a component of emotional response to
ventromedial prefrontal cortex (Anders et al., faces in social judgment
2004). These findings are consistent with what was
seen using the lesion method and support the no- So far, we have seen that the amygdala acts to in-
tion that the amygdala is not required for SCR. fluence key components of object-based visual
Facial images are conditioned with an aversive processing, visuospatial attention, and autonomic
unconditioned stimulus; however, fMRI reveals responses during the processing of facial expres-
CS+ specific activations in the anterior cingulate, sions in social judgment. One important component
anterior insula, and bilateral amygdala (Buchel of emotion judgment not yet addressed in this
et al., 1998), suggesting a role for the amygdala in scheme is systems that have been implicated by
evaluative associations. Such influence of the amy- simulation theoretic approaches to social cognition
gdala in evaluative assessment is supported by (Gallese et al., 2004; Rizzolatti and Craighero,
evidence implicating the right amygdala in the 2004, see also Keysers and Gazzola, this volume).
generation of SCR in response to emotionally We will briefly outline a proposal for the amy-
arousing visual stimuli (Glascher and Adolphs, gdala’s action on this system, which we view as
2003). Lesions of the right temporal lobe and bi- primarily being one of somatosensory attention.
lateral temporal area, including lesions to the amy- Attentional modulation of the somatosensory cor-
gdala, impaired normal SCRs to nonfacial stimuli tices by the amygdala, we propose, involves several
that were emotionally arousing. Further, when aspects analogous to attentional modulation in
SCR is used to partition recorded amygdala acti- other contexts. First, the amygdala’s action could
vation in a fearful vs. neutral face contrast, no increase the sensitivity of somatosensory cortices to
expression-dependent difference in amygdala acti- the signals received from the body. Second, amy-
vation is seen unless an associated SCR is observed gdala inputs could enhance selectivity of inputs to,
(Williams et al., 2001). Again, this is consistent activity within, and outputs from somatosensory
with an evaluative function of the amygdala, this cortices. Finally, past associations established
time directly in relation to facial expressions of within these cortices may be reactivated so as to
emotion. facilitate neural traces having resulted from previ-
Here is another example of how a function that ous learning. In sum, we are extending the
was once held to be dependent directly on the amygdala’s role in emotional response from its im-
amygdala, namely SCR, is actually influenced portant and well-established role as facilitating
by the action of the amygdala without actually bodily responses to emotional stimuli (Damasio,
requiring the amygdala. Rather, the amygdala 1996) to a role in modulating the cortical processing
evaluates the socioenvironmental context and in- of those responses via somatosensory attentional
fluences the deployment of SCR in an appropriate mechanisms. This move implies that emotion may
manner, either for learning in a classical condi- be understood as being both increased autonomic
tioning paradigm or for normal evaluation of vis- responses (i.e., the ‘‘body’’ loop) and stored cortical
ual stimuli such as faces. It is likely that this action representations of those responses (the ‘‘as-if’’
is dependent on the central nucleus of the am- loop), as well as increased attention to those re-
ygdala (Price and Amaral, 1981; Price, 2003), sponses and their representation in cortex. On
though it is also possible that projections from the this view, emotion, at least in part, is attentional
basal nucleus of the amygdala to the ventromedial modulation of those neural systems dedicated
prefrontal cortex and cingulate cortex and from to processing somatosensory signals, serving to
373

establish the value of a particular socioenviron- monkey parietal cortex (Amaral and Price, 1984).
mental context for an organism. Moreover, the medial division of the pulvinar
Several lines of evidence now point to the in- projects strongly to area 7b along with other areas
volvement of somatosensory cortices in the judg- in the parietal cortex (Mesulam et al., 1977;
ment of emotion from faces. Bilateral lesions of Romanski et al., 1997), and it is known that the
the insula completely abolish the ability to judge central nucleus of the amygdala projects back to the
the emotion of disgust from static and dynamic medial pulvinar, the same nucleus that conveys rapid
facial displays (Calder et al., 2000; Adolphs et al., visual input to the amygdala (Price and Amaral,
2003). Such lesions also appear to abolish the 1981). It is plausible, then, that the amygdala acts on
ability to be disgusted by nonfood items that are SMG via the pulvinar, as well as by its direct pro-
widely recognized as disgusting. Consistent with jections. The amygdala also projects heavily into the
the lesion data, neuroimaging reveals activation of insula region (Amaral and Price, 1984), an area
the insula when observing dynamic displays of fa- strongly implicated in simulation-based processing
cial disgust (Wicker et al., 2003a). Lesions of the of facial emotion and in the representation of emo-
right somatosensory areas, particularly including tion (Adolphs et al., 2003; Wicker et al., 2003a). It is
the supramarginal gyrus (SMG; Brodmann 40), thus more likely in the case of the insula than in the
seriously impair judging emotion from faces case of the SMG that the amygdala acts directly via
(Adolphs et al., 1996, 2000) and from bodily mo- its many projections from basal and basomedial nu-
tion conveyed in point-light displays (Heberlein clei into this cortical region.
et al., 2004). Again, neuroimaging data are con- The proposal here regarding the amygdala’s pos-
sistent with the idea that SII is important for sible role in attentionally modulating somatosenso-
judging emotion from faces (Winston et al., 2003). ry cortices is consistent with what is established by
Looking at dynamic displays of smiling activates the evidence reviewed in the previous two sections.
areas such as SII in the right hemisphere, including It is not likely, in other words, that the amygdala
regions within the supramarginal gyrus and left itself is a locus of simulation. Rather, it is much
anterior insula, and these areas are also activated more plausible that it interprets the socioenviron-
when smiling (Hennenlotter et al., 2005). The ven- mental context and then affects simulation net-
tral amygdala is found to be activated only during works such as may inhabit the somatosensory
observation, however. cortices detailed here.
The pivotal role in judging facial emotion sus- Two brief points might be made before moving
pected to be played by the supramarginal gyrus is on to a summary of the current view of the amy-
intriguing in light of its evolutionary homology to gdala’s function in judging emotions from faces
area 7b in macaque monkeys (Rizzolatti and Craig- and other visual social stimuli. One is that evi-
hero, 2004). Area 7b is a cortical region with facial dence from neuroimaging experiments and single
haptic neurons whose haptic spatial receptive fields unit studies is required in order to test the frame-
and preferred directions of haptic stimulation over- work detailed here. Use of dynamic causal mode-
lap considerably with their visuospatial receptive ling or Granger causality, for example, in the
fields and preferred directions of movement in the analysis of fMRI data would help discern whether
visual field (Duhamel et al., 1998). Most impor- amygdala activation precedes and predicts activa-
tantly, neurons in 7b exhibit mirror neuron-like tion in SMG and insula. The second point is a
qualities in single unit recordings (Rizzolatti and more general one regarding the relation of amy-
Craighero, 2004). gdala activation to emotional experience and pa-
In the monkey, there are several neuroanatomical thology. It is simply this: the amygdala likely is not
pathways that could permit the amygdala to act on itself a generator of such experience, either in
somatosensory cortices such as SMG and insula in a healthy persons or in emotional disturbance, a
way similar to that described at the beginning of this view consistent with data from some amygdala
section. The basal and basomedial nuclei of the amy- lesion patients (Anderson and Phelps, 2002). In-
gdala project lightly and directly to area 7 of the stead, the amygdala helps to control attention
374

inward, i.e., toward the body and encoded emo- from object-selective visual cortex soon follows;
tional associations. Malfunction in these inward and this, together with input from other areas,
attention networks could very likely yield the kind leads to the generation of autonomic responses via
of negatively valent ideation and sensations often action by the central nucleus.
accompanying mental illness. Each of these steps casts the amygdala as an
important (attentional) modulator of neural sys-
The new model for how the amygdala contributes to tems, and a key aspect of the proposal here is the
the recognition of emotion from visual social stimuli amygdala’s influence on simulation systems. Im-
portantly, each element in this new framework of
We are now able to articulate a coherent view of the amygdala’s function is supported with empir-
the amygdala’s action in judging emotion from a ical data. Moreover, the connection between amy-
face. The story proceeds like this (Fig. 3): visual gdala processing and simulation networks is
input to the amygdala, which can occur very rap- supported by anatomical detail, though the func-
idly via the pulvinar, results in initial modulation tional relevance of this connectivity has yet to be
of subsequent visual inputs from visual cortex. clearly established. A more complete functional
Attentional modulation of somatosensory (i.e., understanding of this relationship is sure to come
putative simulation) cortex occurs so as to increase given the evident energy and productivity of re-
sensitivity to and selectivity for bodily responses search into these networks of social cognition.
and encoded emotional associations. Modulation
of temporal visual cortex by the amygdala may, Acknowledgments
via coarse visuospatial coding in these neurons,
influence the dorsal ‘‘where’’ stream so as to direct The authors thank Fred Gosselin and Dirk Neu-
visuospatial attention to emotionally salient fea- mann for helpful discussions. This work was sup-
tures (e.g., the eyes in a face). Richer visual input ported by grants from the National Institute of

Fig. 3. Schematic of the proposed action of the amygdala in attentionally modulating visual and somatosensory cortical areas either
directly or via projections to the pulvinar.
375

Mental Health, the Cure Autism New Foundation, Amaral, D.G., Bauman, M.D., Capitanio, J.P., Lavenex, P.,
and the Pfeiffer Research Foundation. Mason, W.A., Mauldin-Jourdain, M.L. and Mendoza, S.P.
(2003a) The amygdala: is it an essential component of the
neural network for social cognition? Neuropsychologia, 41:
517–522.
References Amaral, D.G., Behniea, H. and Kelly, J.L. (2003b) Topo-
graphic organization of projections from the amygdala to the
Adams Jr., R.B., Gordon, H.L., Baird, A.A., Ambady, N. and visual cortex in the macaque monkey. Neuroscience, 118:
Kleck, R.E. (2003) Effects of gaze on amygdala sensitivity to 1099–10120.
anger and fear faces. Science, 300: 1536. Amaral, D.G. and Insausti, R. (1992) Retrograde transport of
Adams Jr., R.B. and Kleck, R.E. (2003) Perceived gaze direc- D-[3 H]-aspartate injected into the monkey amygdaloid com-
tion and the processing of facial displays of emotion. Psychol. plex. Exp. Brain Res., 88: 375–388.
Sci., 14: 644–647. Amaral, D.G. and Price, J.L. (1984) Amygdalo-cortical pro-
Adolphs, R. (2002) Recognizing emotion from facial expres- jections in the monkey (Macaca fascicularis). J. Comp. Ne-
sions: psychological and neurological mechanisms. Behav. urol., 230: 465–496.
Cogn. Neurosci. Rev., 1: 21–61. Anders, S., Lotze, M., Erb, M., Grodd, W. and Birbaumer, N.
Adolphs, R., Baron-Cohen, S. and Tranel, D. (2002) Impaired (2004) Brain activity underlying emotional valence and
recognition of social emotions following amygdala damage. arousal: a response-related fMRI study. Hum. Brain Mapp.,
J. Cogn. Neurosci., 14: 1264–1274. 23: 200–209.
Adolphs, R., Damasio, H., Tranel, D., Cooper, G. and Dam- Anderson, A.K. and Phelps, E.A. (2001) Lesions of the human
asio, A.R. (2000) A role for somatosensory cortices in the amygdala impair enhanced perception of emotionally salient
visual recognition of emotion as revealed by three-dimen- events. Nature, 411: 305–309.
sional lesion mapping. J. Neurosci., 20: 2683–2690. Anderson, A.K. and Phelps, E.A. (2002) Is the human amy-
Adolphs, R., Damasio, H., Tranel, D. and Damasio, A.R. gdala critical for the subjective experience of emotion? Evi-
(1996) Cortical systems for the recognition of emotion in fa- dence of intact dispositional affect in patients with amygdala
cial expressions. J. Neurosci., 16: 7678–7687. lesions. J. Cogn. Neurosci., 14: 709–720.
Adolphs, R., Gosselin, F., Buchanan, T.W., Tranel, D., Schyns, Anderson, A.K., Spencer, D.D., Fulbright, R.K. and Phelps,
P. and Damasio, A.R. (2005) A mechanism for impaired fear E.A. (2000) Contribution of the anteromedial temporal lobes
recognition after amygdala damage. Nature, 433: 68–72. to the evaluation of facial emotion. Neuropsychology, 14:
Adolphs, R., Russell, J.A. and Tranel, D. (1999a) A role for the 526–536.
human amygdala in recognizing emotional arousal from un- Bachevalier, J., Beauregard, M. and Alvarado, M.C. (1999)
pleasant stimuli. Psychol. Sci., 10: 167–171. Long-term effects of neonatal damage to the hippocampal
Adolphs, R. and Tranel, D. (1999) Intact recognition of emo- formation and amygdaloid complex on object discrimination
tional prosody following amygdala damage. Neuro- and object recognition in rhesus monkeys (Macaca mulatta).
psychologia, 37: 1285–1292. Behav. Neurosci., 113: 1127–1151.
Adolphs, R. and Tranel, D. (2003) Amygdala damage impairs Bachevalier, J. and Loveland, K.A. (2006) The orbitofrontal-
emotion recognition from scenes only when they contain fa- amygdala circuit and self-regulation of social-emotional be-
cial expressions. Neuropsychologia, 41: 1281–1289. havior in autism. Neurosci. Biobehav. Rev., 30: 97–117.
Adolphs, R., Tranel, D. and Damasio, A.R. (1998) The human Bagshaw, M.H. and Benzies, S. (1968) Multiple measures of the
amygdala in social judgment. Nature, 393: 470–474. orienting reaction and their dissociation after amygdalec-
Adolphs, R., Tranel, D. and Damasio, A.R. (2003) Dissociable tomy in monkeys. Exp. Neurol., 20: 175–187.
neural systems for recognizing emotions. Brain Cogn., 52: Bagshaw, M.H. and Coppock, H.W. (1968) Galvanic skin re-
61–69. sponse conditioning deficit in amygdalectomized monkeys.
Adolphs, R., Tranel, D., Damasio, H. and Damasio, A. (1994) Exp. Neurol., 20: 188–196.
Impaired recognition of emotion in facial expressions fol- Baron-Cohen, S., Ring, H.A., Bullmore, E.T., Wheelwright, S.,
lowing bilateral damage to the human amygdala. Nature, Ashwin, C. and Williams, S.C. (2000) The amygdala theory
372: 669–672. of autism. Neurosci. Biobehav. Rev., 24: 355–364.
Adolphs, R., Tranel, D., Damasio, H. and Damasio, A.R. (1995) Baron-Cohen, S., Ring, H.A., Wheelwright, S., Bullmore, E.T.,
Fear and the human amygdala. J. Neurosci., 15: 5879–5891. Brammer, M.J., Simmons, A. and Williams, S.C. (1999) So-
Adolphs, R., Tranel, D., Hamann, S., Young, A.W., Calder, cial intelligence in the normal and autistic brain: an fMRI
A.J., Phelps, E.A., Anderson, A., Lee, G.P. and Damasio, study. Eur. J. Neurosci., 11: 1891–1898.
A.R. (1999b) Recognition of facial emotion in nine individ- Bauman, M.D., Lavenex, P., Mason, W.A., Capitanio, J.P. and
uals with bilateral amygdala damage. Neuropsychologia, 37: Amaral, D.G. (2004a) The development of mother-infant in-
1111–1117. teractions after neonatal amygdala lesions in rhesus mon-
Aggleton, A., Burton, M. and Passingham, R. (1980) Cortical keys. J. Neurosci., 24: 711–721.
and subcortical afferents to the amygdala of the rhesus mon- Bauman, M.D., Lavenex, P., Mason, W.A., Capitanio, J.P. and
key (Macaca mulatta). Brain Res., 190: 347–368. Amaral, D.G. (2004b) The development of social behavior
376

following neonatal amygdala lesions in rhesus monkeys. Emery, N.J. and Amaral, D.G. (1999) The role of the amygdala
J. Cogn. Neurosci., 16: 1388–1411. in primate social cognition. In: Lane R.D., NadelL. (Ed.),
Bechara, A., Tranel, D., Damasio, H., Adolphs, R., Rockland, Cognitive Neuroscience of Emotion. Oxford University
C. and Damasio, A.R. (1995) Double dissociation of condi- Press, Oxford.
tioning and declarative knowledge relative to the amygdala Emery, N.J., Capitanio, J.P., Mason, W.A., Machado, C.J., Men-
and hippocampus in humans. Science, 269: 1115–1118. doza, S.P. and Amaral, D.G. (2001) The effects of bilateral
Breiter, H.C., Etcoff, N.L., Whalen, P.J., Kennedy, W.A., Ra- lesions of the amygdala on dyadic social interactions in rhesus
uch, S.L., Buckner, R.L., Strauss, M.M., Hyman, S.E. and monkeys (Macaca mulatta). Behav. Neurosci., 115: 515–544.
Rosen, B.R. (1996) Response and habituation of the human Fine, C., Lumsden, J. and Blair, R.J. (2001) Dissociation be-
amygdala during visual processing of facial expression. Neu- tween ‘theory of mind’ and executive functions in a patient
ron, 17: 875–887. with early left amygdala damage. Brain, 124: 287–298.
Broks, P., Young, A.W., Maratos, E.J., Coffey, P.J., Calder, Freese, J.L. and Amaral, D.G. (2005) The organization of pro-
A.J., Isaac, C.L., Mayes, A.R., Hodges, J.R., Montaldi, D., jections from the amygdala to visual cortical areas TE and V1
Cezayirli, E., Roberts, N. and Hadley, D. (1998) Face in the macaque monkey. J. Comp. Neurol., 486: 295–317.
processing impairments after encephalitis: amygdala damage Gallese, V., Keysers, C. and Rizzolatti, G. (2004) A unifying
and recognition of fear. Neuropsychologia, 36: 59–70. view of the basis of social cognition. Trends Cogn. Sci., 8:
Brothers, L., Ring, B. and Kling, A. (1990) Response of neu- 396–403.
rons in the macaque amygdala to complex social stimuli. Ghashghaei, H.T. and Barbas, H. (2002) Pathways for emotion:
Behav. Brain Res., 41: 199–213. interactions of prefrontal and anterior temporal pathways in
Buchel, C., Morris, J., Dolan, R.J. and Friston, K.J. (1998) the amygdala of the rhesus monkey. Neuroscience, 115:
Brain systems mediating aversive conditioning: an event-re- 1261–1279.
lated fMRI study. Neuron, 20: 947–957. Glascher, J. and Adolphs, R. (2003) Processing of the arousal
Calder, A.J., Keane, J., Manes, F., Antoun, N. and Young, of subliminal and supraliminal emotional stimuli by the hu-
A.W. (2000) Impaired recognition and experience of disgust man amygdala. J. Neurosci., 23: 10274–10282.
following brain injury. Nat. Neurosci., 3: 1077–1078. Grossman, E.D. and Blake, R. (2002) Brain areas active during
Calder, A.J., Lawrence, A.D., Keane, J., Scott, S.K., Owen, visual perception of biological motion. Neuron, 35:
A.M., Christoffels, I. and Young, A.W. (2002) Reading the 1167–1175.
mind from eye gaze. Neuropsychologia, 40: 1129–1138. Hart, A.J., Whalen, P.J., Shin, L.M., McInerney, S.C., Fischer,
Calder, A.J., Young, A.W., Rowland, D., Perrett, D.I., Hod- H. and Rauch, S.L. (2000) Differential response in the human
ges, J.R. and Etcoff, N.L. (1996) Facial emotion recognition amygdala to racial outgroup vs ingroup face stimuli. Neuro-
after bilateral amygdala damage: differentially severe impair- report, 11: 2351–2355.
ment of fear. Cogn. Neuropsychol., 13: 699–745. Haxby, J.V., Gobbini, M.I., Furey, M.L., Ishai, A., Schouten,
Campbell, R., Heywood, C.A., Cowey, A., Regard, M. and J.L. and Pietrini, P. (2001) Distributed and overlapping rep-
Landis, T. (1990) Sensitivity to eye gaze in prosopagnosic resentations of faces and objects in ventral temporal cortex.
patients and monkeys with superior temporal sulcus ablation. Science, 293: 2425–2430.
Neuropsychologia, 28: 1123–1142. Haxby, J.V., Hoffman, E.A. and Gobbini, M.I. (2000) The
Canli, T., Sivers, H., Whitfield, S.L., Gotlib, I.H. and Gabrieli, distributed human neural system for face perception. Trends
J.D. (2002) Amygdala response to happy faces as a function Cogn. Sci., 4: 223–233.
of extraversion. Science, 296: 2191. Heberlein, A.S., Adolphs, R., Tranel, D. and Damasio, H. (2004)
Damasio, A.R. (1996) The somatic marker hypothesis and the Cortical regions for judgments of emotions and personality
possible functions of the prefrontal cortex. Philos. Trans. R. traits from point-light walkers. J. Cogn. Neurosci., 16:
Soc. Lond. B Biol. Sci., 351: 1413–1420. 1143–1158.
de Gelder, B., Vroomen, J., Pourtois, G. and Weiskrantz, L. Hennenlotter, A., Schroeder, U., Erhard, P., Castrop, F., Has-
(1999) Non-conscious recognition of affect in the absence of linger, B., Stoecker, D., Lange, K.W. and Ceballos-Bau-
striate cortex. Neuroreport., 10: 3759–3763. mann, A.O. (2005) A common neural basis for receptive and
Dicks, D., Myers, R.E. and Kling, A. (1968) Uncus and am- expressive communication of pleasant facial affect. Neuro-
ygdala lesions: effects on social behavior in the free-ranging image, 26: 581–591.
rhesus monkey. Science, 165: 69–71. Hooker, C.I., Paller, K.A., Gitelman, D.R., Parrish, T.B., Me-
Downing, P.E., Jiang, Y., Shuman, M. and Kanwisher, N. sulam, M.M. and Reber, P.J. (2003) Brain networks for anal-
(2001) A cortical area selective for visual processing of the yzing eye gaze. Brain Res. Cogn. Brain Res., 17: 406–418.
human body. Science, 293: 2470–2473. Izquierdo, A., Suda, R.K. and Murray, E.A. (2005) Compar-
Duhamel, J.R., Colby, C.L. and Goldberg, M.E. (1998) Ventral ison of the effects of bilateral orbital prefrontal cortex lesions
intraparietal area of the macaque: congruent visual and so- and amygdala lesions on emotional responses in rhesus mon-
matic response properties. J. Neurophysiol., 79: 126–136. keys. J. Neurosci., 25: 8534–8542.
Emery, N.J. (2000) The eyes have it: the neuroethology, func- Jones, E.G. and Burton, H. (1976) A projection from the
tion and evolution of social gaze. Neurosci. Biobehav. Rev., medial pulvinar to the amygdala in primates. Brain Res., 104:
24: 581–604. 142–147.
377

Kalin, N.H., Shelton, S.E. and Davidson, R.J. (2004) The role Mesulam, M.M., Van Hoesen, G.W., Pandya, D.N. and Gesch-
of the central nucleus of the amygdala in mediating fear and wind, N. (1977) Limbic and sensory connections of the in-
anxiety in the primate. J. Neurosci., 24: 5506–5515. ferior parietal lobule (area PG) in the rhesus monkey: a study
Kalin, N.H., Shelton, S.E., Davidson, R.J. and Kelley, A.E. with a new method for horseradish peroxidase histochemis-
(2001) The primate amygdala mediates acute fear but not the try. Brain Res., 136: 393–414.
behavioral and physiological components of anxious tem- Meunier, M. and Bachevalier, J. (2002) Comparison of emo-
perament. J. Neurosci., 21: 2067–2074. tional responses in monkeys with rhinal cortex or amygdala
Kanwisher, N., McDermott, J. and Chun, M.M. (1997) The lesions. Emotion, 2: 147–161.
fusiform face area: a module in human extrastriate cortex Meunier, M., Bachevalier, J., Murray, E.A., Malkova, L. and
specialized for face perception. J. Neurosci., 17: 4302–4311. Mishkin, M. (1999) Effects of aspiration versus neurotoxic
Kawashima, R., Sugiura, M., Kato, T., Nakamura, A., Ha- lesions of the amygdala on emotional responses in monkeys.
tano, K., Ito, K., Fukuda, H., Kojima, S. and Nakamura, K. Eur. J. Neurosci., 11: 4403–4418.
(1999) The human amygdala plays an important role in gaze Morris, J.S., DeGelder, B., Weiskrantz, L. and Dolan, R.J.
monitoring. A PET study. Brain, 122(Pt 4): 779–783. (2001) Differential extrageniculostriate and amygdala re-
Kim, H., Somerville, L.H., Johnstone, T., Polis, S., Alexander, sponses to presentation of emotional faces in a cortically
A.L., Shin, L.M. and Whalen, P.J. (2004) Contextual mod- blind field. Brain, 124: 1241–1252.
ulation of amygdala responsivity to surprised faces. J. Cogn. Morris, J.S., Friston, K.J., Buchel, C., Frith, C.D., Young,
Neurosci., 16: 1730–1745. A.W., Calder, A.J. and Dolan, R.J. (1998a) A neuromodu-
Kling, A. (1968) Effects of amygdalectomy and testosterone on latory role for the human amygdala in processing emotional
sexual behavior of male juvenile macaques. J. Comp. Physiol. facial expressions. Brain, 121(Pt 1): 47–57.
Psychol., 65: 466–471. Morris, J.S., Frith, C.D., Perrett, D.I., Rowland, D., Young,
Kling, A. (1974) Differential effects of amygdalectomy in A.W., Calder, A.J. and Dolan, R.J. (1996) A differential
male and female nonhuman primates. Arch. Sex Behav., 3: neural response in the human amygdala to fearful and happy
129–134. facial expressions. Nature, 383: 812–815.
Kling, A.S. and Brothers, L.A. (1992) The amygdala and social Morris, J.S., Ohman, A. and Dolan, R.J. (1998b) Conscious
behavior. In: Aggleton, A.P. (Ed.), The Amygdala: Neuro- and unconscious emotional learning in the human amygdala.
biological Aspects of Emotion, Memory, and Mental Dys- Nature, 393: 467–470.
function. Wiley-Liss, New York. Morris, J.S., Ohman, A. and Dolan, R.J. (1999) A subcortical
Kling, A., Lancaster, J. and Benitone, J. (1970) Amygdalec- pathway to the right amygdala mediating ‘‘unseen’’ fear.
tomy in the free-ranging vervet (Cercopithecus aethiops). J. Proc. Natl. Acad. Sci. USA, 96: 1680–1685.
Psychiatr. Res., 7: 191–199. Ohman, A. (2005) The role of the amygdala in human fear:
Kling, A., Steklis, H.D. and Deutsch, S. (1979) Radioteleme- automatic detection of threat. Psychoneuroendocrinology,
tered activity from the amygdala during social interactions in 30: 953–958.
the monkey. Exp. Neurol., 66: 88–96. Patterson II, J.C., Ungerleider, L.G. and Bandettini, P.A.
Kluver, H. and Bucy, P.C. (1937) Psychic blindness and other (2002) Task-independent functional brain activity correlation
symptoms following bilateral temporal amygdalectomy in with skin conductance changes: an fMRI study. Neuroimage,
Rhesus monkeys. Am. J. Physiol., 119: 352–353. 17: 1797–1806.
Kluver, H. and Bucy, P.C. (1939) Preliminary analysis of func- Pegna, A.J., Khateb, A., Lazeyras, F. and Seghier, M.L. (2005)
tions of the temporal lobes in monkeys. Arch. Neurol. Discriminating emotional faces without primary visual cor-
Psychiat., 2: 979–1000. tices involves the right amygdala. Nat. Neurosci., 8: 24–25.
Kluver, H. and Bucy, P.C. (1997) Preliminary analysis of func- Phelps, E.A., O’Connor, K.J., Cunningham, W.A., Funayama,
tions of the temporal lobes in monkeys. 1939. J. Neuropsych. E.S., Gatenby, J.C., Gore, J.C. and Banaji, M.R. (2000)
Clin. Neurosci., 9: 606–620. Performance on indirect measures of race evaluation predicts
Knight, D.C., Nguyen, H.T. and Bandettini, P.A. (2005) The amygdala activation. J. Cogn. Neurosci., 12: 729–738.
role of the human amygdala in the production of conditioned Pinsk, M.A., DeSimone, K., Moore, T., Gross, C.G. and
fear responses. Neuroimage, 26: 1193–1200. Kastner, S. (2005) Representations of faces and body parts in
Liddell, B.J., Brown, K.J., Kemp, A.H., Barton, M.J., Das, P., macaque temporal cortex: a functional MRI study. Proc.
Peduto, A., Gordon, E. and Williams, L.M. (2005) A direct Natl. Acad. Sci. USA, 102: 6996–7001.
brainstem-amygdala-cortical ‘alarm’ system for subliminal Pourtois, G., Sander, D., Andres, M., Grandjean, D., Reveret,
signals of fear. Neuroimage, 24: 235–243. L., Olivier, E. and Vuilleumier, P. (2004) Dissociable roles of
McCarthy, G. (1999) Physiological studies of face process- the human somatosensory and superior temporal cortices for
ing in humans. In: Gazzaniga, M.S. (Ed.), The New Cogni- processing social face signals. Eur. J. Neurosci., 20:
tive Neurosciences. MIT Press, Cambridge, MA, pp. 3507–3515.
393–410. Prather, M.D., Lavenex, P., Mauldin-Jourdain, M.L., Mason,
McGaugh, J.L. (2004) The amygdala modulates the consolida- W.A., Capitanio, J.P., Mendoza, S.P. and Amaral, D.G. (2001)
tion of memories of emotionally arousing experiences. Annu. Increased social fear and decreased fear of objects in monkeys
Rev. Neurosci., 27: 1–28. with neonatal amygdala lesions. Neuroscience, 106: 653–658.
378

Pribram, K.H., Reitz, S., McNeil, M. and Spevack, A.A. (1979) Weiskrantz, L. (1956) Behavioral changes associated with ab-
The effect of amygdalectomy on orienting and classical con- lation of the amygdaloid complex in monkeys. J. Comp.
ditioning in monkeys. Pavlov J. Biol. Sci., 14: 203–217. Physiol. Psychol., 49: 381–391.
Price, J.L. (2003) Comparative aspects of amygdala connectiv- Whalen, P.J. (1999) Fear, vigilance, and ambiguity: initial ne-
ity. Ann. NY Acad. Sci., 985: 50–58. uroimaging studies of the human amygdala. Curr. Dir. Psy-
Price, J.L. and Amaral, D.G. (1981) An autoradiographic study chol. Sci., 7: 177–187.
of the projections of the central nucleus of the monkey am- Whalen, P.J., Shin, L.M., McInerney, S.C., Fischer, H., Wright,
ygdala. J. Neurosci., 1: 1242–1259. C.I. and Rauch, S.L. (2001) A functional MRI study of hu-
Puce, A., Allison, T., Bentin, S., Gore, J.C. and McCarthy, G. man amygdala responses to facial expressions of fear versus
(1998) Temporal cortex activation in humans viewing eye and anger. Emotion, 1: 70–83.
mouth movements. J. Neurosci., 18: 2188–2199. Wicker, B., Keysers, C., Plailly, J., Royet, J.P., Gallese, V. and
Rizzolatti, G. and Craighero, L. (2004) The mirror–neuron Rizzolatti, G. (2003a) Both of us disgusted in My insula: the
system. Annu. Rev. Neurosci., 27: 169–192. common neural basis of seeing and feeling disgust. Neuron,
Romanski, L.M., Giguere, M., Bates, J.F. and Goldman-Rakic, 40: 655–664.
P.S. (1997) Topographic organization of medial pulvinar Wicker, B., Michel, F., Henaff, M.A. and Decety, J. (1998)
connections with the prefrontal cortex in the rhesus monkey. Brain regions involved in the perception of gaze: a PET
J. Comp. Neurol., 379: 313–332. study. Neuroimage, 8: 221–227.
Rosvold, H.E., Mirsky, A.F. and Pribram, K.H. (1954) Influ- Wicker, B., Perrett, D.I., Baron-Cohen, S. and Decety, J.
ence of amygdalectomy on social behavior in monkeys. J. (2003b) Being the target of another’s emotion: a PET study.
Comp. Physiol. Psychol., 47: 173–178. Neuropsychologia, 41: 139–146.
Spiridon, M. and Kanwisher, N. (2002) How distributed is vis- Williams, LM., Brammer, M.J., Skerrett, D., Lagopolous, J.,
ual category information in human occipito-temporal cortex? Rennie, C., Kozek, K., Olivieri, G., Peduto, T. and Gordon,
An fMRI study. Neuron, 35: 1157–1165. E. (2000) The neural correlates of orienting: an integration of
Sugase, Y., Yamane, S., Ueno, S. and Kawano, K. (1999) Glo- fMRI and skin conductance orienting. Neuroreport, 11:
bal and fine information coded by single neurons in the tem- 3011–3015.
poral visual cortex. Nature, 400: 869–873. Williams, M.A., McGlone, F., Abbott, D.F. and Mattingley,
Tranel, D. and Damasio, H. (1989) Intact electrodermal skin J.B. (2005) Differential amygdala responses to happy and
conductance responses after bilateral amygdala damage. Ne- fearful facial expressions depend on selective attention. Ne-
uropsychologia, 27: 381–390. uroimage, 24: 417–425.
Vuilleumier, P., Richardson, M.P., Armony, J.L., Driver, J. and Williams, L.M., Phillips, M.L., Brammer, M.J., Skerrett, D.,
Dolan, R.J. (2004) Distant influences of amygdala lesion on Lagopoulos, J., Rennie, C., Bahramali, H., Olivieri, G.,
visual cortical activa during emotional face processing. Nat. David, A.S., Peduto, A. and Gordon, E. (2001) Arousal dis-
Neurosci., 7: 1271–1278. sociates amygdala and hippocampal fear responses: evidence
Vuilleumier, P. and Schwartz, S. (2001a) Beware and be aware: from simultaneous fMRI and skin conductance recording.
capture of spatial attention by fear-related stimuli in neglect. Neuroimage, 14: 1070–1079.
Neuroreport, 12: 1119–1122. Winston, J.S., O’Doherty, J. and Dolan, R.J. (2003) Common
Vuilleumier, P. and Schwartz, S. (2001b) Emotional facial ex- and distinct neural responses during direct and incidental
pressions capture attention. Neurology, 56: 153–158. processing of multiple facial emotions. Neuroimage, 20: 84–97.
Wang, L., McCarthy, G., Song, A.W. and Labar, K.S. (2005) Winston, J.S., Strange, B.A., O’Doherty, J. and Dolan, R.J.
Amygdala activation to sad pictures during high-field (4 tesla) (2002) Automatic and intentional brain responses during eval-
functional magnetic resonance imaging. Emotion, 5: 12–22. uation of trustworthiness of faces. Nat. Neurosci., 5: 277–283.
Ward, R., Danziger, S. and Bamford, S. (2005) Response to Young, A.W., Aggleton, J.P., Hellawell, D.J., Johnson, M.,
visual threat following damage to the pulvinar. Curr. Biol., Broks, P. and Hanley, J.R. (1995) processing impairments
15: 571–573. after amygdalotomy. Brain, 118(Pt 1): 15–24.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 21

Towards a unifying neural theory of social cognition

Christian Keysers and Valeria Gazzola

BCN Neuro-Imaging-Centre, University Medical Center Groningen, University of Groningen, A. Deusinglaan 2,


9713AW Groningen, The Netherlands

Abstract: Humans can effortlessly understand a lot of what is going on in other peoples’ minds. Under-
standing the neural basis of this capacity has proven quite difficult. Since the discovery of mirror neurons, a
number of successful experiments have approached the question of how we understand the actions of others
from the perspective of sharing their actions. Recently we have demonstrated that a similar logic may apply
to understanding the emotions and sensations of others. Here, we therefore review evidence that a single
mechanism (shared circuits) applies to actions, sensations and emotions: witnessing the actions, sensations
and emotions of other individuals activates brain areas normally involved in performing the same actions
and feeling the same sensations and emotions. We propose that these circuits, shared between the first (I do,
I feel) and third person perspective (seeing her do, seeing her feel) translate the vision and sound of what
other people do and feel into the language of the observers own actions and feelings. This translation could
help understand the actions and feelings of others by providing intuitive insights into their inner life. We
propose a mechanism for the development of shared circuits on the basis of Hebbian learning, and un-
derline that shared circuits could integrate with more cognitive functions during social cognitions.

Keywords: mirror system; social cognition; emotions; actions; sensations; empathy; theory of mind

Humans are exquisitely social animals. The pro- actors slip off the roof, we shiver if we see an actor
gress of our species and technology is based on our cut himself, we grimace in disgust as the character
capacity for social learning. Social learning and has to eat disgusting food. This sharing experience
skilled social interactions rest upon our capacity to begs two related questions: How do we manage to
gain insights into the mind of others. Not surpris- slip into the skin of other people so effortlessly?
ingly, humans are indeed excellent at understand- Why do we share the experiences we observe in-
ing the inner life of others. This is exemplified in stead of simply understanding them?
our inner experience of watching a Hollywood The goal of this chapter will be to propose that a
feature film: we relax while effortlessly attributing single principle — shared circuits — could provide
a vast range of emotions and motivations to the a unifying perspective on both of these questions.
main character simply by witnessing the actions of To foreshadow the main message of our proposal,
the character, and the events that occur to him. we claim that a circuit composed of the temporal
Not only do we feel that we need very little explicit lobe (area STS (superior temporal sulcus) in mon-
thoughts to understand the actors, we actually keys or MTG (middle temporal gyrus) in humans),
share their emotions and motivations: our hands the rostral inferior parietal lobule (PF/IPL) and
sweat and our heart beats faster while we see the ventral premotor cortex (F5/BA44+6) is
involved both in our own actions and those of
Corresponding author. Fax: +31-503638875; others, thereby forming a shared circuit for per-
E-mail: c.keysers@rug.nl, c.keysers@med.umcg.nl forming and observing actions. We will show that

DOI: 10.1016/S0079-6123(06)56021-2 379


380

the somatosensory cortices are involved both in Primates


experiencing touch on our own body and in view-
ing other human beings or objects being touched; Three brain areas have been shown to contain neu-
that the anterior cingulate and insular cortices are rons that are selectively activated by the sight of the
involved in the experience of pain, and the per- actions of other individuals: the STS (Bruce et al.,
ception of other people’s pain; and finally that the 1981; Perrett et al., 1985, 1989; Oram and Perrett,
anterior insula is also involved both in the expe- 1994, 1996), the anterior inferior parietal lobule (an
rience of disgust and in the observation of disgust area sometimes called 7b and sometimes PF, but
in others (for the case of actions, this model the two names refer to the same area, and we will
is similar to those put forward by other authors: use PF in this manuscript; Gallese et al., 2002) and
Gallese et al. (2004), Rizzolatti and Craighero the ventral premotor cortex (area F5; Dipellegrino
(2004) and Hurley S.L. [http://www.warwick- et al., 1992; Gallese et al., 1996; Rizzolatti et al.,
.ac.uk/staff/S.L.Hurley]). 1996; Keysers et al., 2003) (Fig. 1). These three
Common to all these cases is that some of the brain areas are anatomically interconnected: STS
brain areas involved in the first person perspective has reciprocal connections with PF (Seltzer and
(I do or I feel) are also involved in the third person Pandya, 1978; Selemon and Goldmanrakic, 1988;
perspective (she does or she feels). We will argue Harries and Perrett, 1991; Seltzer and Pandya,
that this sharing transforms what we see other 1994; Rizzolatti and Matelli 2003) and PF is recip-
people do or feel into something very well known rocally connected with F5 (Matelli et al., 1986;
to us: what we do and feel ourselves. By doing so Luppino et al., 1999; Rizzolatti and Luppino, 2001;
it provides an intuitive grasp of the inner life of Tanne-Gariepy et al., 2002), while there are no di-
others. rect connections between F5 and the STS (see
We will review separately key evidence for Keysers and Perrett, in press, for a recent review).
shared circuits for actions, sensations and emo- All three areas contain neurons that appear to se-
tions. We will then show that these systems appear lectively respond to the sight of hand–object inter-
to generalize beyond the animate world. We will actions, with particular neurons responding to the
conclude by suggesting how Hebbian learning sight of particular actions, such as grasping, tearing
could account for the emergence of these shared or manipulating (Perrett et al., 1989; Dipellegrino et
circuits. al., 1992; Gallese et al., 1996, 2002; Keysers et al.,
2003). There is however a fundamental difference
among the three areas. Virtually all neurons in F5
Shared circuits for actions that respond when the monkey observes another
individual perform a particular action also respond
The first evidence that certain brain areas might be when the monkey performs the same action
involved both in the processing of first and third whether he is able to see his own actions or not
person perspectives comes from the study of ac- (Gallese et al., 1996). These neurons called mirror
tions in monkeys. Understanding the actions of neurons therefore constitute a link between what
others is a pragmatic need of social life. Surpris- the monkey sees other people do and what the
ingly, some areas involved in the monkey’s own monkey does himself. A substantial number of
actions are activated by the sight of someone else’s neurons in PF shows a similar behaviour (Gallese
actions (Dipellegrino et al., 1992; Gallese et al., et al., 2002). While in F5 and PF, motor informa-
1996). Today, we start to understand more about tion has an excitatory effect on activity, the situa-
the circuitry that might be responsible for the tion in the STS is quite different. None of the
emergence of this phenomenon (Keysers et al., neurons in the STS responding to the sight of a
2004a; Keysers and Perrett, 2004). Imaging studies particular action have been shown to robustly re-
suggest that a similar system exists in humans (see spond when the monkey performs the same action
Rizzolatti and Craighero, 2004 and Rizzolatti with his eyes closed (Keysers et al., 2004a; Keysers
et al., 2001 for a review). and Perrett, 2004). While some neurons in the STS
381

c ip
a IPL
PF BA6/
F5
F5 BA44
s TG
M
S
T
S

Fig. 1. (a) Lateral view of the macaque brain with the location of F5, PF and STS together with their anatomical connections (arrows).
The following sulci are shown: a ¼ arcuate, c ¼ central, ip ¼ intraparietal, s ¼ sylvian sulcus. (b) Corresponding view of the human brain.

respond similarly when the monkey sees himself kinds of grasps, but not to other actions not re-
perform an action and when it sees someone else lated to grasping). This combination of precisely
perform the same action (Perrett et al., 1989, 1990), and broadly tuned neurons is very important: the
many actually cease to respond to the sight of their precisely tuned neurons can give very detailed in-
preferred movement if the monkey himself is caus- sights into the actions of others, but require that
ing this movement (Hietanen and Perrett, 1993, these actions are within the motor vocabulary of
1996). For these latter neurons, the motor/prop- the observing monkey. The more broadly tuned
rioceptive signal therefore assumes an inhibitory neurons on the other hand will also respond to the
function, in contrast to the excitatory function ob- sight of novel actions that are not within the motor
served in F5 and PF. As a result, half of the cells in vocabulary of the monkey, but resemble actions
the STS appear to treat self and other in similar that are within the monkey’s vocabulary. Exam-
ways, the other half of the STS sharply distin- ples of the latter are the neurons responding
guishes other- from self-caused actions. to tool use, which have now been found in F5
Considering the STS-PF-F5 circuit as a whole, (Ferrari et al., 2005): the monkeys used in this ex-
we therefore have a system that responds to the periment have never used tools (e.g. a pincer) and
actions of others. Two of its components (PF and yet the sight of someone using a tool activated
F5) link the actions of others to our own motor some F5 neurons that responded when the monkey
programs, and may therefore give us an intuitive performed similar but different actions (grasping
insight into the actions of others because they with its hands).
transform the sight of these actions into something The STS-PF-F5 circuit also responds in cases
very well known to ourselves: our own actions where we recognize the actions of others but are
(Gallese et al., 2004; Keysers and Perrett, in press). unable to fully see these actions. In the STS, some
An essential property of mirror neurons is their neurons respond strongly to the invisible presence of
congruent selectivity, namely, the fact that if they a human hiding behind an occluding screen in a
respond more to a particular action (e.g. precision particular location. The same human hiding in a
grip) during execution, they also respond more to different location often caused no response (Baker
that same action during observation, compared to et al., 2001; Fig. 2a). Although this capacity has
other actions (Gallese et al., 1996). Importantly, been demonstrated for hidden humans, similar re-
not all mirror neurons show the same selectivity: sponses may exist for hidden objects. In F5, about
some are very precisely tuned for a particular ac- half of the mirror neurons responding when the
tion (e.g. they respond strongly to a precision grip, monkey himself grasps an object also respond to the
but not to a whole-hand prehension), while others sight of a human reaching behind an occluder but
are much more broadly tuned (responding to all only when the monkey previously saw an object
382

(a) STS (b) F5

5spks/s 5spk/s
1s 1s

Fig. 2. (a) Response of a neuron in STS while the monkey observes a human walk towards, hide behind and then reappear from an
occluding screen. The top and bottom histograms show its activity when hiding behind the left and centre occluder, respectively (see
cartoon on the left). The different experimental phases are shown on top, and coded as a white background when the subject is fully
visible, light grey when partially and dark grey when fully occluded by the screen. The discharge is stronger in the top compared to the
bottom occluded phase although in both cases, there were only three occluders to be seen without any visible individual (Baker et al.,
2001). (b) An F5 neuron while a human demonstrator grasps behind an occluding screen. In the top but not the bottom case, the
monkey previously saw an object being placed on a tray before the occluder was sled in front of the tray. The discharge starting as the
hand begins to be occluded (light and dark grey background) is much stronger in the top case, yet at that moment both visual stimuli
(top and bottom) are equal (Umilta et al., 2001). The scales are different in (a) and (b).

being placed behind the occluder (Umilta et al., behind trees. In such cases, understanding the leop-
2001; Fig. 2b). This observation begs the question of ards action, although it is not fully visible, will make
where the information necessary for this type of F5 the difference between life and death for the ob-
responses originates. As shown above, the STS serving monkey.
could provide a representation of the reaching and a Both STS and F5 also contain neurons that re-
representation of the hidden object. The STS-PF-F5 spond to the sound of actions. Neurons were found
circuit may then extrapolate from these two pieces in the STS that respond to the sound and/or the
of information towards a complete representation of vision of walking, with much smaller responses to
the action, causing F5 grasping neurons to fire. The other actions such as clapping hands (Barraclough
circuit is particularly well suited for such extrapo- et al. 2005; Fig. 3a). Similar neurons have been
lations because it is an inherent function of the pre- found in F5, but responding to seeing and/or hear-
motor cortex to code movement sequences ing a peanut being broken (Fig. 3b; Kohler et al.
unfolding in time. The same hardware could be 2002; Keysers et al., 2003). The latter neurons in F5
used to extrapolate the visible beginning of a grasp- also respond when the monkey breaks a rubber
ing into the full action. Many important actions peanut out of sight (i.e. without sound or vision of
around us are not fully visible: a leopard may be his own action). It therefore appears as though the
approaching a monkey, intermittently disappearing entire STS-PF-F5 circuit is multimodal: some of its
383

a b an actor with 490% accuracy based either on the


STS F5 sound or the vision of the action alone (Keysers
1st footstep crack of peanut et al., 2003).

Humans
V+S
A mirror system similar to that found in the mon-
key has now been described in humans. Regarding
the observation of actions, a number of imaging
studies, including fMRI, PET and MEG experi-
ments, have reported the three following areas be-
ing particularly involved in the observation of
actions: the caudal inferior frontal gyrus and ad-
jacent premotor cortex (Broadman areas [BAs] 44
and 6) corresponding to the monkey’s area F5, the
rostral inferior parietal lobule (IPL) corresponding
S to the monkey’s area PF, and caudal sectors of the
temporal lobe, in particular the posterior superior
temporal sulcus (pSTS) and the adjacent MTG
corresponding to the monkey’s STS (see Fig. 1;
Grafton et al., 1996; Rizzolatti et al., 1996; Decety
1s et al., 1997; Grezes et al., 1998; Iacoboni et al.,
1999; Nishitani and Hari, 2000; Buccino et al.,
Fig. 3. (a) response of an STS neuron while the monkey heard 2001; Grezes et al., 2001; Iacoboni et al., 2001;
(S ¼ sound), saw (V ¼ vision) or saw and heard (V+S) an ex-
Perani et al., 2001; Decety et al., 2002; Nishitani
perimenter walk. Note the strong response in all three cases. (b)
response of an F5 neuron in the same three conditions but for and Hari, 2002; Grezes et al., 2003; Manthey et al.,
the action of breaking a peanut. This neuron also responded 2003; Buccino et al., 2004b; Wheaton et al. 2004).
while the monkey broke a rubber peanut out of sight. The Two of these three areas, the IPL and BA44/6 are
curves at the bottom are sonographs (figure adapted from known to play an important role in motor control.
Keysers and Perrett, in press).
A smaller number of studies have also measured
brain activity during the execution of actions in the
same individuals in order to check if certain parts
neurons respond in similar ways to an action inde- of the brain are involved both during motor ex-
pendently of whether it is seen or heard. Given its ecution and the observation of similar actions
connections to both the auditory and visual corti- (Grafton et al., 1996; Iacoboni et al., 1999;
ces, STS appears to be a likely site for this audi- Buccino et al., 2004b). These studies found sec-
ovisual integration (see Ethofer and Wildgruber, tors of the IPL and BA44/6 to be involved both in
this volume). In the F5-PF-STS circuit, this audi- the observation and execution of actions, repre-
ovisual action representation then appears to be senting a human equivalent of the monkey’s mir-
integrated with the motor program of the matching ror neurons found in PF and F5.
action. With such a multimodal system, the mere The situation in the pSTS/MTG is less clear:
sound of someone knocking on the door would Iacoboni et al. (2001) find the STS to be active
activate a multimodal, audio-visuo motor represen- both during motor execution and observation,
tation of the action, leading to a deep understand- while Grafton et al. (1996) and Buccino et al.
ing and sharing of the heard action. Indeed, mirror (2004b) fail to find robust STS activation during
neurons with audiovisual properties are able to dis- motor execution. Two explanations have been
criminate which of two actions was performed by offered for this STS/MTG activation during the
384

execution of actions. The first holds that an effer- execution, and therefore congruent somatotopy
ence copy of executed actions is sent to congruent cannot be proven by their study either. It is note-
visual neurons in the STS/MTG to create a for- worthy, that during motor execution in other stud-
ward model of what the action should look like ies (e.g. Rijntjes et al., 1999; Hauk et al., 2004), a
(Iacoboni et al., 2001). The second, based on the somatotopy for action execution was observed,
fact that in monkeys the execution of actions re- which apparently resembled the visual one found in
duces the spiking activity of STS neurons, holds the above-cited studies.
that an efference copy is sent in order to cancel the Corroborating evidence for the existence of se-
visual consequences of our own actions (Keysers lective mirror neurons in humans stems from a
and Perrett, in press). Why, though, should a re- number of transcranial magnetic stimulation
duction in spiking show up as an increase in blood (TMS) studies (Fadiga et al., 1995; Gangitano et
oxygen level dependent (BOLD) signal? Logo- al., 2001; see Fadiga et al., 2005 for a review),
thetis (2003) has suggested that the BOLD effect is which suggests that observing particular hand/arm
dominated by synaptic activity, not spiking activ- movements selectively facilitates the motor execu-
ity; the metabolic demands of inhibitory synaptic tion of the specific muscles involved in the obser-
input could thus outweigh a reduction of spiking vation.
activity and thus be measured as an overall in- Evidence that BA44 is essential for recognizing
crease in BOLD signal (but see Waldvogel et al., the actions of others comes from studies that show
2000). Either way, the STS/MTG is an important that patients with premotor lesions show deficits in
element of the ‘mirror circuitry’ involved both in pantomime recognition that cannot be accounted
the observation and execution of actions (Keysers for by verbal problems alone (Bell, 1994; Halsband
and Perrett, in press). et al., 2001). Also, repetitive TMS induced virtual
A key property of the mirror system in monkeys lesions of BA44 impair the capacity to imitate ac-
is its congruent selectivity: mirror neurons respond- tions, even though they do not impair the capacity
ing for instance to a precision grip more than to a to perform the same actions when cued through
whole-hand prehension during motor execution spatial stimuli instead of a demonstrator’s actions
also respond more to the observation of a preci- (Heiser et al., 2003).
sion grip compared to a whole-hand prehension The mirror system in monkeys was shown to
(Gallese et al., 1996). Can the same be demon- also respond to the sound of actions (Kohler et al.,
strated for the human mirror system? A promising 2002; Keysers et al., 2003). In a recent study, we
alley for providing proof of such selectivity stems could demonstrate that a similar system also exists
from studies looking at somatotopy in the premotor in humans (Gazzola et al., 2005). In this study, the
activations. Buccino et al. (2001) and Wheaton et same participants were scanned during execution
al. (2004) showed participants’ foot, hand and of hand and mouth actions and when they listened
mouth actions, and observed that these actions ac- to the sound of similar actions. The entire circuit
tivated partially distinct cortical sites. They inter- composed of MTG-IPL-BA44/6 responded both
pret these activations as reflecting the mapping of during the execution and the sound of hand and
the observation of hand actions onto the execution mouth actions. Most importantly, the voxels in the
of hand actions, and so on for foot and mouth. premotor cortex that responded more during the
Unfortunately, neither of these studies contained sound of hand actions compared to mouth actions
motor execution tasks, and both therefore fail to also responded more during the execution of hand
establish the congruence of the somatotopical or- actions compared to mouth actions, and vice versa
ganization during observation and execution. Leslie for the mouth actions, demonstrating for the first
et al. (2004) asked participants to imitate facial and time a somatotopical organization of the mirror
manual actions, and observed the existence of system in humans, albeit for sounds.
patches of premotor cortex involved in either man- If the observation of other individuals’ actions
ual or facial imitation. Unfortunately, they did not are mapped onto our own motor programs, one
separate the vision of faces/hands from the motor may wonder how the perception of actions change
385

when we acquire new skills. Seung et al. (2005) and the observation and execution of actions, and will
Bangert et al. (2006) show that pianists demon- be termed ‘shared-circuits for actions’. The impli-
strate stronger activations of BA6/44, IPL and cations of having shared circuits for actions are
MTG while listening to piano pieces compared widespread. By transforming the sight of some-
with nonpianists, suggesting that the acquisition of one’s actions into our motor representation of
the novel motor skill of piano playing enhanced these actions, we achieve a very simple and yet
also the auditory mirror representation of these very powerful understanding of the actions of oth-
actions while listening — an observation that ers (Gallese et al., 1996; Keysers, 2003; Gallese et
might relate to the fact that pianists often find it al., 2004). In addition to providing insights into
harder to keep their fingers still while listening to the actions of others, activating motor programs
piano pieces. Calvo-Merino et al. (2005) showed similar to the ones we have observed/heard is of
male and female dancers’ dance movements that obvious utility for imitating the actions of others,
were specific for one of the genders. They found and shared circuits for actions have indeed been
that female dancers activated their premotor cor- reported to be particularly active during the imi-
tex more to female dance moves, and male dancers tation of actions (Iacoboni et al., 1999; Buccino et
more to male dance moves. This finding is partic- al., 2004b). Finally, as will be discussed below in
ularly important, as both male and female dancers more detail, by associating the execution and the
rehearse together and have therefore similar de- sound of actions, mirror neurons might be essen-
grees of visual expertise with both types of move- tial for the acquisition of spoken language (Kohler
ments, but have motor expertise only of their own et al., 2002; Keysers et al., 2003).
gender-specific movements. The premotor differ-
ences observed therefore truly relate to motor ex-
pertise. It is interesting, that although in both Sensations
examples, responses were stronger for experts
compared to nonexperts, mirror activity was not Observation and experience of touch
absent in people devoid of firsthand motor exper-
tise of the precise actions they were witnessing. If shared circuits may be essential to our under-
These weaker activations thus probably reflect the standing of the actions of others, how about the
activity of more broadly tuned mirror neurons sensations of others? If we see a spider crawling on
(Gallese et al., 1996) that may discharge maximally James Bond’s chest in the movie Dr. No, we lit-
to other, similar actions (e.g. walking, jumping), erally shiver, as if the spider crawled on our own
but also respond slightly to these different actions skin. What brain mechanisms might be responsible
(e.g. a specific dance move involving steps and for this automatic sharing of the sensations of
jumps). With these more widely tuned neurons, we others? May shared circuits exist for the sensation
can gain insights into actions that are novel to us, of touch?
by drawing on analogies with similar actions al- To investigate this possibility, we showed sub-
ready within our motor vocabulary. jects movies of other subjects being touched on
their legs. In control movies, the same legs were
approached by an object, but never touched. In
Conclusions separate runs finally, we touched the legs of the
participant. We found that touching the subjects’
Both monkeys and humans appear to activate a own legs activated the primary and secondary so-
circuit composed of temporal, parietal and frontal matosensory cortex of the subjects. Most interest-
neurons while observing the actions of others. The ingly, we found that large extents of the secondary
frontal and parietal nodes of this circuit are active somatosensory cortex also respond to the sight of
both when the subjects perform an action and someone else’s legs being touched. The control
when they perceive someone else perform a similar movies produced much smaller activations (Fig. 4;
action. These nodes are therefore shared between Keysers et al., 2004b).
386

From touch to pain

Painful stimulation of the skin and the observation


of a similar stimulation applied to others also ap-
pear to share a common circuitry including the
anterior cingulate cortex (ACC) and the anterior
insula. First, a neuron was recorded in the ACC
responding both to pinpricking off the patients
hand and to the sight of the surgeon pinpricking
himself (Hutchison et al., 1999). Later, this anec-
dotic finding was corroborated by an elegant
fMRI investigation, where on some trials the par-
ticipant received a small electroshock on her hand;
on other trials she saw a signal on a screen signi-
fying that her partner was receiving a similar elec-
troshock. Some voxels in the ACC and the
anterior insula were activated in both cases (Singer
et al., 2004), and the amount of that activation
correlated with how empathic the subjects were
according to two paper-and-pencil empathy scales
that measure specifically how much an observer
Fig. 4. Brain activity when a human is touched on his leg in the
scanner (red), and when he sees another individual being
shares the emotional distress of others. The pres-
touched on his leg (blue). The white voxels represent voxels ence of activations in the anterior cingulate and
active in both cases. (Adapted from Keysers et al., 2004b). The anterior insula during the observation of pain oc-
right hemisphere is shown on the right of the figure (neurolog- curring to others was corroborated by Jackson et
ical conventions) al. (2005). In a TMS study, Avenanti et al. (2005)
observed that observing someone else being pin-
Intrigued by the observation of a patient C, who pricked on his hand selectively facilitated TMS
reported that when she sees someone else being induced movements of the hand, suggesting that
touched on the face she literally feels the touch on the sharing of pain influences the motor behaviour
her own skin (Blakemore et al., 2005), she scanned of the observer. This observation supports the
both C and a group of normal controls while existence of cross-talks between different shared
touching them on their faces and necks. In a fol- circuits.
lowing session they showed video clips of someone
else being touched on the same locations. As in our
study, the experience of touch activated primary Emotions
and secondary somatosensory cortices. During ob-
servation, they found SI and SII activation. In C, The insula and disgust
these activations were significantly stronger, poten-
tially explaining why she literally felt the touch that Do shared circuits exist also for emotions? A series
happened to others. of elegant imaging studies by Phillips and collab-
It therefore appears as seeing someone else being orators (Phillips et al., 1997, 1998) suggested that
touched activated a somatosensory representation the anterior insula is implicated in the perception
of touch in the observers, as if they had been of the disgusted facial expressions of others. The
touched themselves. This finding is particularly im- same area has been implicated in the experience
portant as it demonstrates that the concept of of disgust (Small et al., 2003). In addition, both
shared circuits put forward for actions appears to be Calder et al. (2000) and Adolphs et al. (2003) re-
applicable to a very different system: that of touch. ported patients with insular lesions that lost both
387

the capacity to experience disgust and to recognize


disgust in the faces of others. It therefore appears
as though the insula may provide a shared circuit
for the experience and the perception of disgust.
Using fMRI we measured brain activity while
subjects viewed short movie clips of actors sniffing
the content of a glass and reacting with a pleased,
neutral or disgusted facial expression. Thereafter,
we exposed the subjects to pleasant or disgusting
odorants through an anaesthesia mask. The latter
manipulation induced the experience of disgust in
the subjects. We found that the anterior insula was
activated both by the experience of disgust and the
observation of the disgusted facial expressions of
others (Wicker et al., 2003) (Fig. 5, yellow circles).
These voxels were not significantly activated by the
pleasant odorants or the vision of the pleased fa- Fig. 5. Sagittal T1-weighted anatomical MRI of patient NK
(Calder et al., 2000) normalized to MNI space. The blue outline
cial expressions of others. We then superimposed
marks the zone of the left insular infarction. The red outline
the location of the voxels involved in the experi- shows the zone we found to be activated during the experience
ence of disgust and in the observation of disgust of disgust; the yellow outline indicates those zones found to be
onto an MRI image of a patient with insular dam- common to this experience and the observation of someone
age reporting a reduced experience of disgust and a else’s facial expression of disgust (Wicker et al., 2003). Adapted
from Gallese et al. (2004).
deficient capacity to recognize disgust in others
(Fig. 5, blue zone; Calder et al., 2000). The lesion
encompassed our activations. The amygdala and fear
Penfield and Faulk (1955) demonstrated that
electrical stimulation of the anterior insula can A similar logic has been put forward for the re-
cause sensations of nausea supporting the idea that lationship between fear and the amygdala, sug-
the observation of the disgusted facial expressions gesting that the amygdala responds to the sight of
of others actually triggered an internal represen- fearful facial expressions and during the experience
tation of nausea in the participant. of fear. According to this logic, without amygdala,
It therefore appears that the anterior insula in- both the capacity to perceive fear in the face of
deed forms a shared circuit for the first and third others and that to experience fear would be greatly
person perspective of disgust, a conclusion cor- affected. The state of that literature is undergoing
roborated by electrophysiological studies (Krolak- a recent re-evaluation (Spezio and Adolphs, this
Salmon et al., 2003). The lesion data support the volume). Below we will describe the arguments
idea that this circuit is indeed necessary for our first in favour, then against the role of the am-
understanding of disgust in others. Interestingly, ygdala as a central site both for the experience and
just as we showed for the shared circuits for ac- recognition of fear.
tions, the insula also appears to receive auditory For: Anatomically, the amygdala is linked both
information about the disgusted emotion state of to face processing and to bodily states. The am-
others. Adolphs et al. (2003) showed that their ygdala is a complex anatomical structure that re-
patient B with extensive insular lesions was unable ceives highly processed sensory information from
to recognize disgust, even if it was acted out with higher sensory cortices (Amaral and Price, 1984),
distinctive sounds of disgust, such as retching and including the temporal lobe where single neurons
vocal prosody. Imaging studies still fail to find in- respond to the sight of faces and facial expressions
sular activation to vocal expressions of disgust (Perrett et al., 1984; Hasselmo et al., 1989). These
(Phillips et al., 1998). connections would enable the amygdala to process
388

Table 1. The amygdala and the emotion of fear

Damage Ethiology Perceptual deficits References

Subject Left Right Fear Other

SM +++ +++ UW Yes Surprised a,b,c,g


JM +++ +++ E Yes Sad, disgusted, angry c,g
RH +++ +++ E No Angry c,g
SE +++ +++ E Yes Surprised d,g
DR ++ + S Yes Sad, disgusted, angry, surprised e,g
GT +++ +++ E No f,g
EP +++ +++ E No Angry f,g
SP ++ +++ S Yes Sad, disgusted g
DBB +++ ++ S No Sad, disgusted, angry g
NM ++ +++ ? Yes Sad h
SZ +++ ++ No Angry k
JC ++ +++ E Yes Angry i
YW ++ +++ E Yes i
RB +++ – E Yes i
JK ++ ++ UW No j
MA +++ +++ UW No j
FC +++ +++ UW No j
AF +++ +++ UW Yes j
AW +++ +++ UW No j
EW +++ +++ UW No j
WS +++ +++ UW No j
AvdW ++ ++ UW Yes j
RL ++ ++ UW No j
BR +++ +++ UW Yes j

Note: A number of neuropsychological studies have asked subjects with bilateral amygdala damage to rate how afraid six photographs of the Ekman
series of emotional facial expression photographs looked. Here we show a table reviewing all these studies, reporting for each patient whether he rated
these facial expressions as looking less afraid than do healthy control subjects. This information is taken from the referenced publications except for
patients JK to BR. For these patients, the original publication (Ref. j) reported only group data. M. Siebert and H. Markowitsch gave us the single subject
ratings of their patients and healthy subjects, and we considered deficient those patients that fell below 1.64 standard deviations of the healthy controls. In
total, 12 of 24 subjects with bilateral amygdala damage rated scared facial expressions as less afraid than normal subjects do. Abbreviations: ‘–’: no
damage, or no deficit; ‘+’: minimal damage; ‘++’: partial damage; ‘+++’: extensive or complete damage; UW: Urbach-Wiethe disease, a congenital
disease that causes bilateral calcifications in the amygdala; E: encephalitis, usually affecting extensive regions of the brain; S: surgical removal, usually for
treatment of epilepsy. References: a: Adolphs et al. (1994); b: Adolphs et al. (1995); c: Adolphs et al. (1998); d: Calder et al. (1996); e: Young et al. (1995);
f: Hamann et al. (1996); g: Adolphs et al. (1999); h: Sprengelmeyer et al. (1999); i: Broks et al. (1998); j: Siebert et al. (2003). k: Adolphs and Tranel (2003).

facial expressions. It sends fibres back to subcor- 1976) appeared, 12 of 24 subjects rated facial ex-
tical structures such as the hypothalamus, enabling pressions of fear as less afraid than did control
it to induce the kind of changes in the state of the subjects without bilateral amygdala lesions (see
body that are so typical of fear. It also sends fibres Table 1). This ‘fear-blindness’ was not due to gen-
back to the cortex, including the STS, which could eral facial recognition deficits (the patients never
enable it to influence the way faces are processed. had problems recognizing happy faces as happy),
In humans, bilateral amygdala damage does nor was it due to the patients not understanding
affect the capacity of subjects to recognize fear in the concept of fear (all patients specifically tested
the face of other individuals, but only in about half could provide plausible scenarios of situations in
the subjects. A review of the literature reveals re- which people are scared). Other negative emotions
ports of 24 subjects with bilateral amygdala dam- such as anger were often also affected.
age (see Table 1). When asked to rate how afraid, Imaging studies using static facial expressions
angry, sad, happy, disgusted or surprised the emo- corroborate the idea that the amygdala is impor-
tional facial photographs of (Ekman and Friesen, tant for the perception of fear in others: in the
389

majority of the cases, the amygdala was activated require that parts of the neural representations of
preferentially when subjects viewed fearful or an- the experience of fear should be triggered by the
gry facial expressions as compared to neutral facial observation of other peoples fear. This prediction
expressions (Zald, 2003). Studies using movies receives support from a study by Williams et al.
provide a different message (see below). (2001). They showed subjects Ekman faces of fear,
Lesions of the amygdala also corroborate to some and simultaneously recorded brain activity and
extent the idea of its involvement in generating fear. skin conductance. They found that the trials in
Monkeys with lesions in the amygdala appear to be which the fear-faces produced increases of skin
disinhibited: unlike their unlesioned counterparts, conductance were accompanied by increased
they immediately engage in social contacts with to- BOLD responses in the amygdala. It therefore ap-
tal strangers and in play with normally scary objects pears as though the vision of a fearful facial ex-
such as rubber snakes — as if, without amygdala, pression activates the amygdala and induces a
the monkeys fail to be scared of other individuals body state of fear/arousal in the observer, as in-
and objects (Amaral et al., 2003). In addition, three dicated by augmented skin conductance. This link
of the amygdala patients of Table 1 (SM, NM and between amygdala and body state is also corrob-
YW) were tested with regards to their own emotions orated by Anders et al. (2004).
of fear. SM appears to have reduced psychophys- Against: While there is evidence both from le-
iological reactions to fear (Adolphs et al., 1996); sion studies and imaging supporting the dual role
NM only remembers having been scared once in his of the amygdala in experiencing and recognizing
life and enjoyed activities that would be terrifying to fear, there is a number of recent studies that shed
most of us (e.g. bear hunting in Siberia, hanging doubts on this interpretation.
from a helicopter, Sprengelmeyer et al., 1999, p. First, half of the patients with bilateral am-
2455); YW did not even experience fear while being ygdalar lesions show no impairments in rating fear
mugged at night. This suggests that without am- in fearful faces. Authors have failed to find etio-
ygdala, there is something different and reduced in logical or anatomical differences between the pa-
the subjective experience of fear. tients with and without fear-blindness (Adolphs
Electrical stimulations of the amygdala in hu- et al., 1998).
mans lead to a variety of experiences, but when- Second, a recent study on SM, one of the sub-
ever it evoked an emotion, it was that of fear jects with bilateral amygdala damage, indicate that
(Halgren et al., 1978). Taken together with the the patient’s problem in identifying the expression
neuroimaging data in humans and the lesion data of fear in others is not due to an inability to rec-
in monkeys, the amygdala thus appears to be im- ognize fear per se, but an inappropriate explora-
portant for the normal associations of stimuli with tion of the stimuli (Adolphs et al., 2005): unlike
our personal, first person perspective of fear. control individuals, she failed to look at the eye
The role of the amygdala in experiencing fear is region of photographs. If she was encouraged to
corroborated by a number of imaging studies. do so, her fear recognition became entirely normal.
Arachnophobic individuals, when viewing spiders, In the context of the connections of the amygdala
experience more fear and show stronger BOLD with the STS, the function of the amygdala may
signals in their amygdala compared with control not be to recognize the facial expression of fear,
subjects (Dilger et al., 2003). Cholecystokinin-tet- but to render the eye region of facial expressions a
rapeptide (cck-4) injections induce panic attacks salient stimulus, selectively biasing the stimulus
that are accompanied by intense feeling of fear and processing in the STS towards the eye region (see
cause augmentation of regional cerebral blood also Spezio and Adolphs, this volume).
flow (rCBF) in the amygdala (Benkelfat et al., If the amygdala is indeed not responsible for the
1995). recognition of fear but only in orienting visual in-
The above evidence therefore suggests a role for spection towards the eye region, one would predict
the amygdala both in the recognition and the ex- equal activation of the amygdala to all facial ex-
perience of fear. The idea of shared circuits would pressions. While this is often not the case when
390

static images of facial expressions were used (Phan but its role in doing so may be indirect, by biasing
et al., 2004 for a review), using short movies of attention towards informative sections of facial
facial expressions we found that the amygdala was stimuli and by relaying information towards brain
indeed activated similarly by all facial expressions, areas responsible for the experience of fear. The
be they emotional or not (van der Gaag et al., other nodes of this circuitry remain to be investi-
2005). We used movies of happiness, disgust, fear gated.
and a neutral expression that contained as much
movement as the other facial expressions (blowing
up the cheeks). This finding sheds doubt on the Shared circuits for actions, sensations and emotions
idea of the amygdala as showing direct fear selec-
tivity, and supports the idea of the amygdala par- Subsuming the above evidence, it appears that in
ticipating in the processing of all facial expressions three systems — actions, sensations and emotions
(for instance by biasing visual processing to the — certain brain areas are involved both in first
eyes). The reason why we found neutral faces to person experience (I do, I feel) and third person
cause as much activation as emotional and fearful perspective (knowing what he does or he feels).
expressions using movies while studies using static These areas or circuits, that we call shared circuits,
stimuli have often reported differences remains to are the premotor cortex and inferior parietal lob-
be fully understood. Ecologically, facial expres- ule interconnected with the STS/MTG for actions,
sions are dynamic stimuli, not still photographs: the insula for the emotion of disgust, the ACC and
the task of detecting emotions from photos is evo- the anterior insula for pain, and somatosensory
lutionary rather new. We thus suggest that the lack cortices for touch. Possibly, the amygdala may be
of amygdalar selectivity found using movies, al- part of a shared circuit for fear. In all these cases,
though needing replication, may be a more valid observing what other people do or feel is therefore
picture of amygdalar function than the selectivity transformed into an inner representation of what
often observed using photographs. we would do or feel in a similar situation — as if
Doubt must also be shed on the importance of we would be in the skin of the person we observe.
the amygdala in feeling fear. Monkeys with very The idea of shared circuits, initially put forward
early lesions in the amygdala still demonstrate for actions (Gallese and Goldman, 1998) therefore
signs of panic, although they occur in contexts that appears much broader.
are not normally inducing fear (Amaral et al., In the light of this evidence, it appears as though
2003). In addition, there is no good evidence that social situations are processed by the STS to a high
patient SM completely lacks fear as an emotion, degree of sophistication, including multimodal au-
although it may well be that she does not exhibit dio-visual representations of complex actions.
fear appropriately in context — this is a difficult These representations privilege the third person
issue to measure in humans, and still remains un- perspective, with lesser responses if the origin of
resolved (Adolphs, personal communication). the stimulus is endogenous. Through the recruit-
However, Anderson and Phelps (2002) have as- ment of shared circuits, the brain then adds spe-
sessed this question in patients with amygdala cific first person elements to this description. If an
damage, and also found no evidence that they lack action is seen, the inferior parietal and premotor
fear as an emotion. Together, it might thus be areas add an inner representation of actions to the
speculated that the amygdala has a role both in the sensory third person description. If touch is wit-
normal experience of fear and in the recognition of nessed, the somatosensory cortices add an inner
fear in others, but that this role may be indirect, representation of touch. If pain is witnessed, the
through focusing gaze on the eye region and by ACC and the anterior insula add a sense of pain. If
linking normally fear-producing stimuli with other disgust is witnessed, the insula adds a sense of
brain areas that, in turn, are responsible for fear. disgust. What emerges from the resulting neural
The amygdala may thus be part of a circuit that activity is a very rich neural description of what
enables us to share the fear of other individuals, has been perceived, adding the richness of our
391

subjective experience of actions, emotions and someone else being touched is clearly nonmotor.
sensations to the objective visual and auditory de- Instead we think that each modality (actions, sen-
scription of what has been seen. sations and emotions) is understood and shared in
We are not normally confused about where the our brain using its own specific circuitry. The neu-
third person ends and the first starts, because ral representation of actions, emotions and sensa-
although the shared areas react in similar ways to tions that results from the recruitment of shared
our own experience and the perception of others, representations are then the intuitive key to un-
many other areas clearly discriminate between derstanding the other person, without requiring
these two cases. Our own actions include strong that they have to pass necessarily a somatosensory
M1 activation and week STS activations, while or motor common code to be interpreted.
those of others fail to normally activate M1 but Of course many social situations are complex,
strongly activate the STS. When we are touched, and involve multiple modalities: witnessing some-
our SI is strongly active, while it is much less active one hitting his finger with a hammer contains an
while we witness touch occurring to others. In- action, an emotion and a sensation. In most social
deed, patient C who is literally confused about situations, the different shared circuits mentioned
who is being touched shows reliable SI activity above thus work in concert.
during the sight of touch (Blakemore et al., 2005). Once shared circuits have transformed the ac-
In this context, the distinction between self and tions, sensations and emotions of others into our
other is quite simple, but remains essential for a own representations of actions, sensations and
social cognition based on shared circuits to work emotions, understanding other people’s boils
(Gallese and Goldman, 1998; Decety and Som- down to understanding ourselves — our own ac-
merville, 2003). Some authors now search for tions, sensations and emotions, an aspect that we
brain areas that explicitly differentiate self from will return to later in relation to theory of mind.
other. Both the right inferior parietal lobule and
the posterior cingulate gyrus have been implicated
in this function (Decety and Sommerville, 2003 Demystifying shared circuits through a Hebbian
and Vogt, 2005 for reviews). perspective
The account based on shared representation we
propose differs from those of other authors in that Neuroscientific evidence for the existence of shared
it does not assume that a particular modality is circuits is rapidly accumulating. The importance of
critical. Damasio and coworkers (Damasio, 2003) these circuits for social cognitions is evident. Yet,
emphasize the importance of somatosensory rep- for many readers, the existence of single neurons
resentation, stating that it is only once our brain responding to the sight, sound and execution of an
reaches a somatosensory representation of the action — to take a single example — remains a
body state of the person we observe that we un- very odd observation. How can single neurons
derstand the emotion he/she are undergoing. We, with such marvellous capacities emerge? The plau-
on the other hand, believe that somatosensory sibility of a neuroscientific account of social cog-
representations are important for understanding nitions based on shared circuits stands and falls
the somatosensory sensations of others, but may with our capacity to give a plausible explanation of
not be central to our understanding of other in- how such neurons can emerge. As outlined in de-
dividuals’ emotions and actions. The current pro- tail elsewhere (Keysers and Perrett, in press), we
posal represents an extension from our own propose that shared circuits are a simple conse-
previous proposals (e.g. Gallese et al., 2004), quence of observing ourselves and others (please
where we emphasized the motor aspect of under- refer to Keysers and Perrett, in press, for citations
standing other people. We believe that motor supporting the claims put forward below).
representations are essential for understanding the When they are young, monkeys and humans
actions of others, yet the activity in somatosensory spend a lot of time watching themselves. Each
cortices observed during the observation of time, the child’s hand wraps around an object, and
392

brings it towards him, a particular set of neural sounds with our mouth, the sound and the motor
activities overlaps in time. Neurons in the premo- program are correlated in time. As the sound will
tor cortex responsible for the execution of this recruit multimodal neurons in the STS, the estab-
action will be active at the same time as the audio- lished link also ties the sight of other people pro-
visual neurons in the STS responding to the sight ducing similar sounds to our motor program. The
and sound of grasping. Given that STS and F5 are visual information thereby rides on the wave of the
connected through PF, ideal Hebbian learning auditory associations (Keysers and Perrett, in
conditions are met (Hebb, 1949): what fires to- press).
gether wires together. As a result, the synapses The case of emotions might be similar, yet
going from STS grasping neurons to PF and then slightly more difficult. How can the sight of a dis-
F5 will be strengthened as the grasping neurons at gusted facial expression trigger our own emotion
all three levels will be repeatedly coactive. After of disgust, despite the fact that we do not usually
repeated self-observation, neurons in F5 receiving see our own disgusted facial expression? First, dis-
the enhanced input from STS will fire at the mere gust can often have a cause that will trigger si-
sight of grasping. Given that many neurons in the multaneous disgust in many individuals (e.g., a
STS show reasonably viewpoint-invariant re- disgusting smell). In this case, one’s own disgust
sponses (Perrett et al. 1989, 1990, 1991; Logo- then correlates directly with the disgusted facial
thetis et al., 1995), responding in similar ways to expression of others. Second, in parent–child rela-
views of a hand taken from different perspective, tionships, facial imitation is a prominent observa-
the sight of someone else grasping in similar ways tion (e.g. Stern, 2000). In our Hebbian perspective,
then suffices to activate F5 mirror neurons. All this imitation means that the parent acts as a mir-
that is required for the emergence of such mirror ror for the facial expression of the child, leading
responses is the availability of connections bet- again to the required correlation between the
ween STS-PF-F5 that can show Hebbian learning, child’s own emotion and the facial expression of
and there is evidence that Hebbian learning can that emotion in others. As described above, the
occur in many places in the neocortex (Bi and Poo, insula indeed receives the required visual input
2001; Markram et al., 1997). from the STS, where neurons have been shown to
The same Hebbian argument can be applied to respond to facial expressions (Mufson and Mesu-
the case of sensations and emotions. While seeing lam, 1982; Perrett et al., 1987, 1992; Puce and
ourselves being touched, somatosensory activations Perrett, 2003; Keysers et al., 2001). The insula also
overlap in time with visual descriptions of an object receives highly processed somatosensory informa-
moving towards and touching our body. After tion about our own facial expressions. These so-
Hebbian association the sight of someone else being matosensory data will be Hebbianly associated
touched can trigger somatosensory activations with both the sight of other individual’s facial ex-
(Keysers et al., 2004b; Blakemore et al., 2005). pressions and our own state of disgust. After that
Multimodal responses are particularly impor- Hebbian training, seeing someone else’s facial ex-
tant for cases where we do not usually see our own pressions may trigger a neuronal representation of
actions. How, for example, associating the sight of the somatosensory components of our own match-
someone’s lip movements with our own lip move- ing facial expressions. The debilitating effect of
ments is an important step in language acquisition. somatosensory lesions in understanding the emo-
How can we link the sight of another individual’s tions of others (Adolphs et al., 2000) may indicate
mouth producing a particular sound with our own that this triggering is indeed important in under-
motor programs given that we cannot usually see standing the emotions of others.
our own mouth movements? While seeing other To summarize, Hebbian association (a simple
individuals producing certain sounds with their and molecularly well-understood process) can
mouth, the sound and sight of the action are cor- therefore predict the emergence of associations
related in time, and can lead to STS multimodal between the first and third person perspective of
neurons. During our own attempts to produce actions, sensations and emotions.
393

Shared circuits and the inanimate world performing the same actions. Fig. 6a illustrates a
frame from a typical stimulus, as well as the
The world around us is not inhabited only by BOLD signal measured in BA 44 as defined by the
other human beings: we often witness events that probabilistic maps (Amunts et al., 1999). As seen
occur to inanimate objects. Do the shared circuits in 1.2, this area has been shown to be activated
we described above react to the sight of an inan- both during the execution of actions and the ob-
imate object performing actions or being touched? servation of another human being performing a
To investigate the first question, we showed similar action. Here, we see that the same area was
subjects movies of an industrial robot interacting also activated during the sight of a robot perform-
with everyday life objects (Gazzola et al., 2004, ing an action involving everyday objects. This re-
Society for Cognitive Neuroscience Annual Meet- sult is in contrast with previous reports in the
ing). The robot for instance was grasping a wine literature that failed to find premotor activation to
glass or closing a salt box. These actions the sight of a robot performing actions (Tai et al.,
were contrasted against the sight of a human 2004). Our experiment differs in some important

Fig. 6. (a) Top: location of the right BA44 according to Amunts et al. (1999), defined as the voxels where at least 1 of her 10 subjects
satisfied the cytoarchitectonic criterions for BA44. Below: the brain activity in this right BA44 for 14 subjects, expressed in terms of
parameter estimates in the GLM while subjects looked at a fixation cross or at a human or a robot opening a bottle. A star indicates
significant differences from the fixation. (b) Location of the region of interest (white) defined in Keysers et al. (2004b), and below, the
mean BOLD signal of eight independent subjects while being touched on their legs, seeing another human being touched, and seeing
objects being touched. All three cases differ significantly from fixation, but not from one another (adapted from Keysers et al., 2004b).
All error bars refer to standard error of the means.
394

aspects from these studies: first, we used more the actions, sensations and emotions of others into
complex actions instead of grasping of a ball; sec- neural representations normally associated with our
ond our blocks contained a variety of actions and own execution of similar actions, and our own ex-
not the same action repeated over and over again. perience of similar sensations and emotions. He-
Both of these factors could account for the ob- bbian learning could explain how these automatic
served difference. In the light of our results, it thus associations arise. Once these associations have
appears as though the shared circuit for actions been learned, they transform what other people do
responds to complex meaningful actions regardless and feel into our own experience of these actions,
of whether they are performed by humans and ro- sensations and emotions. This transformation rep-
bots. Half way along this human–robot contin- resents a intuitive and powerful form of communi-
uum, the premotor cortex also responds to the cation: it transmits the experience of doing and
sight of animals from another species performing feeling from one brain to another. This simple form
actions that resemble ours, such as biting (Buccino of communication has obvious adaptive value: be-
et al., 2004a). ing able to peek into someone else’s mind, and to
To test whether the secondary somatosensory share his experiences renders constructive social in-
cortex responds to the sight of objects being teractions faster and more effective. For instance,
touched, we showed subjects movies of inanimate sharing the disgust of a conspecific probing a po-
objects such as ring binders and rolls of paper tow- tential source of food will prevent the observer from
els being touched by a variety of rods. These con- tasting potentially damaging items.
ditions were compared against the activity when the Most forms of communication have a funda-
subject himself was touched, and when he saw an- mental problem: the sender transforms a content
other human leg being touched in similar ways. into a certain transmittable form according to a
Results, shown in Fig. 6b indicate that the SII/PV certain encoding procedure. The receiver then re-
complex was at least as activated by the sight of ceives the encoded message, and has to transform
objects being touched as by the sight of humans it back into the original content. How does the
being touched. Blakemore et al. (2005) also showed receiver learn how to decode the message? When
movies of objects being touched, and found that we learn a spoken language we spend years of our
compared to seeing a human face being touched, life guessing this encoding/decoding procedure.
seeing a fan being touched induced smaller activat- For the case of actions, the shared circuits we
ions. Unfortunately, the authors did not indicate if propose use the correlation in time in the STS-PF-
the activation to seeing an object being touched was F5 circuit during self-observation to determine the
significant. Why Blakemore et al. found stronger reciprocal relationship between the motor repre-
activity to the sight of a human face being touched sentation of actions and their audio-visual conse-
compared to an object, while we found similar ac- quences. Similar procedures may apply to
tivity to a human leg being touched compared to sensations and emotions. The acquired reciprocal
toilet paper rolls remains to be investigated. relationships can then be used to decode the mo-
Together, data appear to emerge suggesting that tor, somatosensory and emotional contents con-
the sight of the actions and tactile ‘experiences’ of tained in the behaviour of other individuals and in
the inanimate world may be transformed into our the situation they are exposed to (see also Leiberg
own experience of these actions and sensations, and Anders, this volume).
but further investigations of this aspect are im- When dealing with other human beings this de-
portant considering the somewhat contradictory coding procedure is generally very successful. Our
findings from different studies. brain appears to use the same procedure to un-
derstand members of other species and even inan-
Shared circuits and communication imate objects and robots. In the case of members
of other animal species, the decoded motivations,
We show that the brain appears to automatically emotions and feelings are anthropocentric, and
transform the visual and auditory descriptions of imperfect: when monkeys for instance open their
395

lips to show their teeth, interpreting this as a smile beings through shared circuits. C. Keysers, E.
is a misinterpretation — it is actually a display of Kohler and M. Goodale (unpublished observation)
threat. Often such interpretations will enable us to have for instance examined brain activity while
predict the forthcoming behaviour of the animal watching the eye movements of other individuals in
better than if we make no interpretation at all. In the hope to find evidence that brain areas such as
the case of inanimate objects, the interpretations the frontal eye field (FEF), normally responsible for
are very often wrong (e.g. the ring binders prob- our own eye movements, are critical for our un-
ably ‘felt’ nothing when being touched, and the derstanding of the eye movements of others. We
robot was not thirsty when it grasped for the glass found very little evidence for such a system: the
of whine). This overgeneralization may simply be a sight of eye movements activated the FEF no more
bug in our brain. Alternatively, overall, it might be than the sight of random dots moving by the same
better to apply the rule of the thumb: everything is amount. Despite the difficulty of interpreting neg-
probably a bit like myself, than to make no as- ative results, this finding is not too surprising: if two
sumption at all. A clear implication of this ten- people face each other, and one suddenly stares at
dency is that to make the human–machine the wall behind the other person, the other person
communication as smooth as possible, robots will tend to look behind him. The motor programs
should be made to behave as similarly to humans involved are very different: a very small saccade for
as possible. the first individual, and a turning of the head and
torso for the second. There being so little in com-
mon in motor terms, it makes no sense to analyse
The limits of simulation — a word of caution the gaze direction of others through one’s own mo-
tor programs. An external frame of reference, and
The shared circuits we describe have received con- an analysis of gaze in this external frame is needed
siderable interest. Often they now tend to be seen to understand what the other person is looking at
as a panacea to explain any issues of social cog- — a task that our motor system, working in ego-
nition. It is important to note that while we believe centric coordinates, is very poorly equipped for.
shared circuits to be very important for our intu- Shared circuits and mirror neurons therefore have
ition of the inner life of others, they cannot explain little to contribute to this task. It will remain for
everything. future research to outline the limits of what shared
We can for instance try to imagine what it feels circuits can explain.
like to fly like a bird, although we do not have the
motor programs to do so. Such abstract imagina-
tions are detached from our own bodily experi- Simulation and theory of mind — a hypothesis
ence, and should thus not be attributed to shared
circuits. We can of course imagine what it feels like Social cognitions are not restricted to the simula-
to flap our hands, as kids do to pretend to fly, but tions that shared circuits provide. Explicit
that would still leave us with doubts about what thoughts exist in humans and clearly supplement
real flying feels like. these automatic simulations. It is hard for instance
These limitations are often cruelly clear to us to imagine how a gossip of the type: ‘Did you
during imitation: we have often seen our mothers know that Marry still believes that her husband is
knit, feeling that we can truly understand the faithful while everyone else knows that he is hav-
movement, yet when we tried for the first time to ing an affair with another women every week?’ can
knit something ourselves, we realise that our un- be the result of simulation, yet thinking about the
derstanding had been quite superficial indeed, as (false) beliefs of others is clearly an important part
we were lacking the true motor programs on which of our social intelligence.
to mirror the actions we saw. The words theory of mind (ToM) and ‘mental-
But even with the required motor skills, we do izing’ have often been used to describe the set of
not understand all the inner life of other human cognitive skills involved in thinking about the mind
396

of others, in particular their beliefs (Frith and Frith, people’s minds might be a process related to think-
2003). People are considered to have a ToM if they ing about one’s self. If seen in the context of shared
are able to deal with the fact that other people can circuits, this leads to a simple framework for asso-
have beliefs that differ from reality, a capacity that ciating shared circuits and ToM. The mPFC may
is tested with so called false belief tasks such as the originally interpret states of our own mind and
famous ‘Sally and Anne’ test (Baron-Cohen et al., body, as evidenced by experiments such as that of
1985). In that test, an observer sees Sally hide an Gusnard et al. (2001). In this experiment, subjects
object in a basket. Sally then goes away for a while, saw emotional pictures, and were asked to judge if
and unbeknown to her, Anne moves the object the image evoked pleasant or unpleasant emotions
from the basket into a box. Sally then returns, and in themselves, or whether the images were taken
the observer is asked: ‘where will Sally [first] look indoors or outdoors. The mPFC was more acti-
for her object?’ If the observer answers ‘in the bas- vated in the former task, where subjects had to in-
ket, because she doesn’t know that Anne moved it’, trospect their own emotions.
the observer is thought to have a ToM. If the an- The mPFC receives indirect inputs about all as-
swer is ‘in the box’, the observer failed. Children pects of our own body, including motor, somato-
from the age of 4 years pass this test, while autistic sensory and visceral representations, which could
individuals often fail the test even in their teens allow it to create secondary representation of our
(Baron-Cohen et al., 1985). bodily state (Frith and Frith, 2003). Considering
Comparing this ToM task with the tasks in- our first person perspective, one could thus differ-
volved in the neuroimaging of shared circuits, it is entiate a first level representation, being our ac-
quite clear that these two fields of research tap into tions, emotions and sensations as they occur, and a
phenomena that differ dramatically in the amount second level representation in the mPFC of these
of explicit thoughts that are involved (see also states more related to our conscious understanding
Leiberg and Anders this volume). In research on and interpretation of ourselves. To illustrate that
shared circuits, subjects simply watch short video difference, if you vomit, you will feel disgust, and
clips of actions, emotions and sensations, without activate your insula (primary representation). If
being asked to reflect upon the meaning of these asked what you feel, you may start reflecting upon
stimuli or the beliefs and thoughts of the actors. In what you are feeling in a more conscious way, one
ToM tasks, subjects are directly invited to reflect that you can formulate in words (‘I felt like having
about states of minds of others. Strangely enough, a a stone in my stomach, I guess those mussels
number of authors have introduced a dichotomy weren’t that freshy’) and you are likely to acti-
between simulation processes and more theory vate your mPFC in addition to the primary rep-
driven processes involved in understanding others resentations.
(e.g. Gallese and Goldman, 1998; Saxe, 2005), sug- This is where simulation ties into the concepts of
gesting that either ToM or simulation should ex- theory of mind. Through the shared circuits we
plain all of social cognitions. Here we will attempt have described, the actions, emotions and sensa-
to provide a model that proposes that ToM might tions of others are ‘translated’ into the neural lan-
utilize simulations to reflect on the mind of others. guage of our own actions, emotions and
ToM-type tests have now been investigated a sensations. By doing so, they have been trans-
number of times using fMRI and PET (see Frith formed into what we called primary representa-
and Frith, 2003 for a comprehensive review), and tions of these states. This could generate an
all tasks have activated medial prefrontal cortex implicit sharing and hence understanding of the
(mPFC) compared to conditions requiring less states of others. If asked explicitly what went on in
mentalizing. What is intriguing, is that similar sites the mind of that other person, you would need to
of the mPFC are also involved in reflecting about generate a secondary, more conscious and cogni-
ourselves and our own emotions (Gusnard et al., tive representation of his state. Given that his state
2001; Frith and Frith, 2003), which lead Uta and has already been translated into the language of
Chris Frith to speculate that thinking about other our own states, one may hypothesize, that this task
397

would now be no different from reflecting about Adolphs, R., Damasio, H., Tranel, D. and Damasio, A.R.
your own states, and therefore activate the same (1996) Cortical systems for the recognition of emotion in fa-
cial expressions. J. Neurosci., 16: 7678–7687.
mPFC sites. Testing this hypothesis directly will be
Adolphs, R., Gosselin, F., Buchanan, T.W., Tranel, D., Schyns,
an exciting issue in future neuroimaging work. P. and Damasio, A.R. (2005) A mechanism for impaired fear
In this concept, shared circuits act like a trans- recognition after amygdala damage. Nature, 433: 68–72.
lator, converting the states of others into our own Adolphs, R. and Tranel, D. (2003) Amygdala damage impairs
primary state representations. Often social process- emotion recognition from scenes only when they contain fa-
ing can stop at that: we share some of the states of cial expressions. Neuropsychologia, 41: 1281–1289.
Adolphs, R., Tranel, D. and Damasio, A.R. (1998) The human
our partner, her/his sadness or happiness for in- amygdala in social judgment. Nature, 393: 470–474.
stance, without thinking any further. These are the Adolphs, R., Tranel, D. and Damasio, A.R. (2003) Dissociable
cases that simulation proponents have concentrated neural systems for recognizing emotions. Brain Cogn., 52:
on. In some cases, we might reflect further upon 61–69.
Adolphs, R., Tranel, D., Damasio, H. and Damasio, A. (1994)
her/his mind, just like we often reflect about our
Impaired recognition of emotion in facial expressions fol-
own states. Such reflections then provide much lowing bilateral damage to the human amygdala. Nature,
more elaborate, cognitive and differentiated under- 372: 669–672.
standings of other individuals. These latter are the Adolphs, R., Tranel, D., Damasio, H. and Damasio, A.R.
processes that ToM investigators are excited about. (1995) Fear and the human amygdala. J. Neurosci., 15:
With those mentalizing processes on top of sim- 5879–5891.
Adolphs, R., Tranel, D., Hamann, S., Young, A.W., Calder,
ulation, thinking about others can reach levels of A.J., Phelps, E.A., Anderson, A., Lee, G.P. and Damasio,
sophistications that go far beyond using simula- A.R. (1999) Recognition of facial emotion in nine individuals
tion alone. Using simulation, we inherently assume with bilateral amygdala damage. Neuropsychologia, 37:
that we are all equal. This is not the case: actions 1111–1117.
that may make us happy may make other people Amaral, D.G., Capitanio, J.P., Jourdain, M., Mason, W.A.,
Mendoza, S.P. and Prather, M. (2003). The amygdala: is it an
sad, reflecting biological and cultural differences, essential component of the neural network for social cogni-
and keeping in mind those differences may be a tion? Neuropsychologia, 41: 235–240.
critical role for higher processes (see also Leiberg Amaral, D.G. and Price, J.L. (1984) Amygdalo-cortical pro-
and Anders, this volume). jections in the monkey (Macaca fascicularis). J. Comp. Ne-
urol., 230: 465–496.
Amunts, K., Schleicher, A., Burgel, U., Mohlberg, H., Uylings,
Acknowledgements H.B. and Zilles, K. (1999) Broca’s region revisited: cytoar-
chitecture and intersubject variability. J. Comp. Neurol., 412:
319–341.
C.K. was financed by a VIDI grant by NWO and a Anders, S., Lotze, M., Erb, M., Grodd, W. and Birbaumer, N.
Marie Curie Excellence grant. The authors are (2004) Brain activity underlying emotional valence and
particularly thankful to Michaela Siebert for shar- arousal: a response-related fMRI study. Hum. Brain Mapp.,
ing her raw data with us to calculate Table 1, to 23: 200–209.
Anderson, A.K. and Phelps, E.A. (2002) Is the human am-
Andy Calder for providing the anatomical scan of ygdala critical for the subjective experience of emotion? Ev-
patient NK, to Ralph Adolphs for helpful discus- idence of intact dispositional affect in patients with amygdala
sions regarding the function of the amygdala and lesions. J. Cogn. Neurosci., 14: 709–720.
to Vittorio Gallese, Bruno Wicker and Giacomo Avenanti, A., Bueti, D., Galati, G. and Aglioti, S.M. (2005)
Rizzolatti for many inspiring discussions that led Transcranial magnetic stimulation highlights the sensorimo-
tor side of empathy for pain. Nat. Neurosci., 8: 955–960.
to the development of these models. Baker, C.I., Keysers, C., Jellema, T., Wicker, B. and Perrett,
D.I. (2001) Neuronal representation of disappearing and
hidden objects in temporal cortex of the macaque. Exp. Brain
References Res., 140: 375–381.
Bangert, M., Peschel, T., Schlaug, G., Rotte, M., Drescher, D.,
Adolphs, R., Damasio, H., Tranel, D., Cooper, G. and Dam- Hinrichs, H., Heinze, H.J. and Altenmuller, E. (2006) Shared
asio, A.R. (2000) A role for somatosensory cortices in the networks for auditory and motor processing in professional
visual recognition of emotion as revealed by three-dimen- pianists: evidence from fMRI conjunction. Neuroimage, 30:
sional lesion mapping. J. Neurosci., 20: 2683–2690. 917–926.
398

Baron-Cohen, S., Leslie, A.M. and Frith, U. (1985) Does the Decety, J., Grezes, J., Costes, N., Perani, D., Jeannerod, M.,
autistic child have a ‘‘theory of mind’’. Cognition, 21: 37–46. Procyk, E., Grassi, F. and Fazio, F. (1997) Brain activity
Barraclough, N.E., Xiao, D.K., Baker, C.I., Oram, M.W. and during observation of actions Influence of action content and
Perrett, D.I. (2005) Integration of visual and auditory infor- subject’s strategy. Brain, 120(Pt 10): 1763–1777.
mation by STS neurons responsive to the sight of actions. Decety, J. and Sommerville, J.A. (2003) Shared representations
J. Cogn. Neurosci., 17(3): 377–391. between self and other: a social cognitive neuroscience view.
Bell, B.D. (1994) Pantomime recognition impairment in apha- Trends Cogn. Sci., 7: 527–533.
sia: an analysis of error types. Brain Lang., 47: 269–278. Dilger, S., Straube, T., Mentzel, H.J., Fitzek, C., Reichenbach,
Benkelfat, C., Bradwejn, J., Meyer, E., Ellenbogen, M., Milot, J.R., Hecht, H., Krieschel, S., Gutberlet, I. and Miltner,
S., Gjedde, A. and Evans, A. (1995) Functional neuroanat- W.H. (2003) Brain activation to phobia-related pictures in
omy of CCK4-induced anxiety in normal healthy volunteers. spider phobic humans: an event-related functional magnetic
Am. J. Psychiatry, 152: 1180–1184. resonance imaging study. Neurosci. Lett., 348: 29–32.
Bi, G. and Poo, M. (2001) Synaptic modification by correlated Dipellegrino, G., Fadiga, L., Fogassi, L., Gallese, V. and Ri-
activity: Hebb’s postulate revisited. Ann. Rev. Neurosci., 24: zzolatti, G. (1992) Understanding motor events — a neuro-
139–166. physiological study. Exp. Brain Res., 91: 176–180.
Blakemore, S.J., Bristow, D., Bird, G., Frith, C. and Ward, J. Ekman, P. and Friesen, M.V. (1976) Pictures of Facial Affect.
(2005) Somatosensory activations during the observation of Consulting Psychologists Press, Palo Alto CA.
touch and a case of vision-touch synaesthesia. Brain, 128: Fadiga, L., Craighero, L. and Olivier, E. (2005) Human motor
1571–1583. cortex excitability during the perception of others’ action.
Broks, P., Young, A.W., Maratos, E.J., Coffey, P.J., Calder, Curr. Opin. Neurobiol., 15: 213–218.
A.J., Isaac, C.L., Mayes, A.R., Hodges, J.R., Montaldi, D., Fadiga, L., Fogassi, L., Pavesi, G. and Rizzolatti, G. (1995)
Cezayirli, E., Roberts, N. and Hadley, D. (1998) Face Motor facilitation during action observation: a magnetic
processing impairments after encephalitis: amygdala damage stimulation study. J. Neurophysiol., 73: 2608–2611.
and recognition of fear. Neuropsychologia, 36: 59–70. Ferrari, P.F., Rozzi, S. and Fogassi, L. (2005) Mirror neurons
Bruce, C., Desimone, R. and Gross, C.G. (1981) Visual prop- responding to observation of actions made with tools in
erties of neurons in a polysensory area in superior temporal monkey ventral premotor cortex. J. Cogn. Neurosci., 17:
sulcus of the macaque. J. Neurophysiol., 46: 369–384. 212–226.
Buccino, G., Binkofski, F., Fink, G.R., Fadiga, L., Fogassi, L., Frith, U. and Frith, C.D. (2003) Development and neurophys-
Gallese, V., Seitz, R.J., Zilles, K., Rizzolatti, G. and Freund, iology of mentalizing. Philos. Trans. R. Soc. Lond. B Biol.
H.J. (2001) Action observation activates premotor and pa- Sci., 358: 459–473.
rietal areas in a somatotopic manner: an fMRI study. Eur. J. Gallese, V., Fadiga, L., Fogassi, L. and Rizzolatti, G. (1996) Ac-
Neurosci., 13: 400–404. tion recognition in the premotor cortex. Brain, 119: 593–609.
Buccino, G., Lui, F., Canessa, N., Patteri, I., Lagravinese, G., Gallese, V., Fadiga, L., Fogassi, L. and Rizzolatti, G. (2002)
Benuzzi, F., Porro, C.A. and Rizzolatti, G. (2004a) Neural Action representation and the inferior parietal lobule. Com-
circuits involved in the recognition of actions performed by mon Mech. Percept. Act., 19: 334–355.
nonconspecifics: an FMRI study. J. Cogn. Neurosci., 16: Gallese, V. and Goldman, A. (1998) Mirror neurons and the sim-
114–126. ulation theory of mind-reading. Trends Cogn. Sci., 2: 493–501.
Buccino, G., Vogt, S., Ritzl, A., Fink, G.R., Zilles, K., Freund, Gallese, V., Keysers, C. and Rizzolatti, G. (2004) A unifying
H.J. and Rizzolatti, G. (2004b) Neural circuits underlying view of the basis of social cognition. Trends Cogn. Sci., 8:
imitation learning of hand actions: an event-related fMRI 396–403.
study. Neuron, 42: 323–334. Gangitano, M., Mottaghy, F.M. and Pascual-Leone, A. (2001)
Calder, A.J., Keane, J., Manes, F., Antoun, N. and Young, Phase-specific modulation of cortical motor output during
A.W. (2000) Impaired recognition and experience of disgust movement observation. Neuroreport, 12: 1489–1492.
following brain injury. Nat. Neurosci., 3: 1077–1078. Gazzola, V., Aziz-Zadeh, L., Formisano, E., and Keysers, C.,
Calder, A.J., Young, A.W., Rowland, D., Perrett, D.I., Hod- 2005. Hearing what you are doing — an FMRI study of
ges, J.R. and Etcoff, N.L. (1996) Facial emotion recognition auditory empathy. J. Cogn. Neurosci. Suppl. S, 82.
after bilateral amygdala damage: differentially severe impair- Grafton, S.T., Arbib, M.A., Fadiga, L. and Rizzolatti, G.
ment of fear. Cogn. Neuropsychol., 13: 699–745. (1996) Localization of grasp representations in humans by
Calvo-Merino, B., Glaser, D.E., Grezes, J., Passingham, R.E. positron emission tomography. 2. Observation compared
and Haggard, P. (2005) Action observation and acquired with imagination. Exp. Brain Res., 112: 103–111.
motor skills: an fMRI study with expert dancers. Cerebral Grezes, J., Armony, J.L., Rowe, J. and Passingham, R.E.
Cortex, 15: 1243–1249. (2003) Activations related to ‘‘mirror’’ and ‘‘canonical’’ neu-
Damasio, A.R. (2003) Looking for Spinoza. Harcourt, New rones in the human brain: an fMRI study. Neuroimage, 18:
York. 928–937.
Decety, J., Chaminade, T., Grezes, J. and Meltzoff, A.N. (2002) Grezes, J., Costes, N. and Decety, J. (1998) Top-down effect of
A PET exploration of the neural mechanisms involved in strategy on the perception of human biological motion: a
reciprocal imitation. Neuroimage, 15: 265–272. PET investigation. Cogn. Neuropsychol., 15: 553–582.
399

Grezes, J., Fonlupt, P., Bertenthal, B., Delon-Martin, C., Se- Keysers, C., 2003. Mirror neurons. In: Encyclopedia of Neuro-
gebarth, C. and Decety, J. (2001). Does perception of bio- science, 3rd edn. Elsevier, Amsterdam.
logical motion rely on specific brain regions? Neuroimage, Keysers, C., Gallese, V., Tereshenko, L., Nasoyan, A.,
13: 775–785. Sterbizzi, I., and Rizzolatti, G., 2004a. Investigation of au-
Gusnard, D.A., Akbudak, E., Shulman, G.L. and Raichle, ditory and motor properties in the STS (Unpublished).
M.E. (2001) Medial prefrontal cortex and self-referential Keysers, C., Kohler, E., Umilta, M.A., Nanetti, L., Fogassi, L.
mental activity: relation to a default mode of brain function. and Gallese, V. (2003) Audiovisual mirror neurons and ac-
Proc. Natl. Acad. Sci. USA, 98: 4259–4264. tion recognition. Exp. Brain Res., 153: 628–636.
Halgren, E., Walter, R.D., Cherlow, D.G. and Crandall, P.H. Keysers, C. and Perrett, D.I. (2004) Demystifying social cog-
(1978) Mental phenomena evoked by electrical stimulation of nitions: a Hebbian perspective. Trends Cogn. Sci., 8:
the human hippocampal formation and amygdala. Brain, 501–507.
101: 83–117. Keysers, C., Wicker, B., Gazzola, V., Anton, J.L., Fogassi, L.
Halsband, U., Schmitt, J., Weyers, M., Binkofski, F., Grutzner, and Gallese, V. (2004b) A touching sight: SII/PV activation
G. and Freund, H.J. (2001) Recognition and imitation of during the observation and experience of touch. Neuron, 42:
pantomimed motor acts after unilateral parietal and premo- 335–346.
tor lesions: a perspective on apraxia. Neuropsychologia, 39: Keysers, C., Xiao, D.K., Foldiak, P. and Perrett, D.I. (2001)
200–216. The speed of sight. J. Cogn. Neurosci., 13: 90–101.
Hamann, S.B., Stefanacci, L., Squire, L.R., Adolphs, R., Kohler, E., Keysers, C., Umilta, M.A., Fogassi, L., Gallese, V.
Tranel, D., Damasio, H. and Damasio, A. (1996) Recogniz- and Rizzolatti, G. (2002) Hearing sounds, understanding ac-
ing facial emotion. Nature, 379: 497. tions: action representation in mirror neurons. Science, 297:
Harries, M.H. and Perrett, D.I. (1991) Visual processing of 846–848.
faces in temporal cortex — physiological evidence for a Krolak-Salmon, P., Henaff, M.A., Isnard, J., Tallon-Baudry,
modular organization and possible anatomical correlates. C., Guenot, M., Vighetto, A., Bertrand, O. and Ma-
J. Cogn. Neurosci., 3: 9–24. uguiere, F. (2003) An attention modulated response to dis-
Hasselmo, M.E., Rolls, E.T. and Baylis, G.C. (1989) The role of gust in human ventral anterior insula. Ann. Neurol., 53:
expression and identity in the face-selective responses of neu- 446–453.
rons in the temporal visual-cortex of the monkey. Behav. Leslie, K.R., Johnson-Frey, S.H. and Grafton, S.T. (2004)
Brain Res., 32: 203–218. Functional imaging of face and hand imitation: towards a
Hauk, O., Johnsrude, I. and Pulvermuller, F. (2004) Somato- motor theory of empathy. Neuroimage, 21: 601–607.
topic representation of action words in human motor and Logothetis, N.K. (2003) The underpinnings of the BOLD func-
premotor cortex. Neuron, 41: 301–307. tional magnetic resonance imaging signal. J. Neurosci., 23:
Hebb, D. (1949) The Organisation of Behavior. Wiley, New 3963–3971.
York. Logothetis, N.K., Pauls, J. and Poggio, T. (1995) Shape rep-
Heiser, M., Iacoboni, M., Maeda, F., Marcus, J. and Mazzi- resentation in the inferior temporal cortex of monkeys. Curr.
otta, J.C. (2003) The essential role of Broca’s area in imita- Biol., 5: 552–563.
tion. Eur. J. Neurosci., 17: 1123–1128. Luppino, G., Murata, A., Govoni, P. and Matelli, M. (1999)
Hietanen, J.K. and Perrett, D.I. (1993) Motion sensitive cells in Largely segregated parietofrontal connections linking rostral
the macaque superior temporal polysensory area1. Lack of intraparietal cortex (areas AIP and VIP) and the ventral
response to the sight of the animals own limb movement. premotor cortex (areas F5 and F4). Exp. Brain Res., 128:
Exp. Brain Res., 93: 117–128. 181–187.
Hietanen, J.K. and Perrett, D.I. (1996) Motion sensitive cells in Manthey, S., Schubotz, R.I. and von Cramon, D.Y. (2003)
the macaque superior temporal polysensory area: response Premotor cortex in observing erroneous action: an fMRI
discrimination between self-generated and externally gener- study. Brain Res. Cogn. Brain Res., 15: 296–307.
ated pattern motion. Behav. Brain Res., 76: 155–167. Markram, H., Lubke, J., Frotscher, M. and Sakmann, B. (1997)
Hutchison, W.D., Davis, K.D., Lozano, A.M., Tasker, R.R. Regulation of synaptic efficacy by coincidence of postsynap-
and Dostrovsky, J.O. (1999) Pain-related neurons in the hu- tic APs and EPSPs. Science, 275: 213–215.
man cingulate cortex. Nat. Neurosci., 2: 403–405. Matelli, M., Camarda, R., Glickstein, M. and Rizzolatti, G.
Iacoboni, M., Koski, L.M., Brass, M., Bekkering, H., Woods, (1986) Afferent and efferent projections of the inferior area-6
R.P., Dubeau, M.C., Mazziotta, J.C. and Rizzolatti, G. (2001) in the macaque monkey. J.Comp. Neurol., 251: 281–298.
Reafferent copies of imitated actions in the right superior tem- Mufson, E.J. and Mesulam, M.M. (1982) Insula of the old-
poral cortex. Proc. Natl. Acad. Sci. USA, 98: 13995–13999. world monkey 2 Afferent cortical input and comments on the
Iacoboni, M., Woods, R.P., Brass, M., Bekkering, H., Mazzi- claustrum. J. Comp. Neurol., 212: 23–37.
otta, J.C. and Rizzolatti, G. (1999) Cortical mechanisms of Nishitani, N. and Hari, R. (2000) Temporal dynamics of cor-
human imitation. Science, 286: 2526–2528. tical representation for action. Proc. Natl. Acad. Sci. USA,
Jackson, P.L., Meltzoff, A.N. and Decety, J. (2005) How do we 97: 913–918.
perceive the pain of others? A window into the neural proc- Nishitani, N. and Hari, R. (2002) Viewing lip forms: cortical
esses involved in empathy. Neuroimage, 24: 771–779. dynamics. Neuron, 36: 1211–1220.
400

Oram, M.W. and Perrett, D.I. (1994) Responses of anterior Puce, A. and Perrett, D. (2003) Electrophysiology and brain
superior temporal polysensory (Stpa) neurons to biological imaging of biological motion. Philos. Trans. R. Soc. Lond. B
motion stimuli. J. Cogn. Neurosci., 6: 99–116. Biol. Sci., 358: 435–445.
Oram, M.W. and Perrett, D.I. (1996) Integration of form and Rijntjes, M., Dettmers, C., Buchel, C., Kiebel, S., Frackowiak,
motion in the anterior superior temporal polysensory area R.S. and Weiller, C. (1999) A blueprint for movement: func-
(STPa) of the macaque monkey. J. Neurophysiol., 76: tional and anatomical representations in the human motor
109–129. system. J. Neurosci., 19: 8043–8048.
Penfield, W. and Faulk, M.E. (1955) The insula: further obser- Rizzolatti, G. and Craighero, L. (2004) The mirror-neuron sys-
vations on its function. Brain, 78: 445–470. tem. Ann. Rev. Neurosci., 27: 169–192.
Perani, D., Fazio, F., Borghese, N.A., Tettamanti, M., Ferrari, Rizzolatti, G., Fadiga, L., Gallese, V. and Fogassi, L. (1996)
S., Decety, J. and Gilardi, M.C. (2001) Different brain cor- Premotor cortex and the recognition of motor actions. Cogn.
relates for watching real and virtual hand actions. Neuroim- Brain Res., 3: 131–141.
age, 14: 749–758. Rizzolatti, G., Fogassi, L. and Gallese, V. (2001) Neurophys-
Perrett, D.I., Harries, M.H., Bevan, R., Thomas, S., Benson, iological mechanisms underlying the understanding and im-
P.J., Mistlin, A.J., Chitty, A.J., Hietanen, J.K. and Ortega, itation of action. Nat. Rev. Neurosci., 2: 661–670.
J.E. (1989) Frameworks of analysis for the neural represen- Rizzolatti, G. and Luppino, G. (2001) The cortical motor sys-
tation of animate objects and actions. J. Exp. Biol., 146: tem. Neuron, 31: 889–901.
87–113. Rizzolatti, G. and Matelli, M. (2003) Two different streams
Perrett, D.I., Hietanen, J.K., Oram, M.W. and Benson, P.J. form the dorsal visual system: anatomy and functions. Exp.
(1992) Organization and functions of cells responsive to faces Brain Res., 153: 146–157.
in the temporal cortex. Philos.Trans. R. Soc. London Series Saxe, R. (2005) Against simulation: the argument from error.
B-Biol. Sci., 335: 23–30. Trends Cogn. Sci., 9: 174–179.
Perrett, D.I., Mistlin, A.J. and Chitty, A.J. (1987) Visual neu- Selemon, L.D. and Goldmanrakic, P.S. (1988) Common cor-
rons responsive to faces. Trends Neurosci., 10: 358–364. tical and subcortical targets of the dorsolateral prefrontal
Perrett, D.I., Mistlin, A.J., Harries, M.H. and Chitty, A.J. and posterior parietal cortices in the rhesus-monkey — ev-
(1990) Understanding the visual appearance and conse- idence for a distributed neural network subserving spatially
quences of hand actions. In: Goodale, M.A. (Ed.), Vision guided behavior. J. Neurosci., 8: 4049–4068.
and Action: The Control of Grasping. Ablex Publishing, Seltzer, B. and Pandya, D.N. (1978) Afferent cortical connec-
Norwood, NJ, pp. 163–180. tions and architectonics of superior temporal sulcus and sur-
Perrett, D.I., Oram, M.W., Harries, M.H., Bevan, R., Hie- rounding cortex in rhesus-monkey. Brain Res., 149: 1–24.
tanen, J.K., Benson, P.J. and Thomas, S. (1991) Viewer-cen- Seltzer, B. and Pandya, D.N. (1994) Parietal, temporal, and
tered and object-centered coding of heads in the macaque occipital projections to cortex of the superior temporal sulcus
temporal cortex. Exp. Brain Res., 86: 159–173. in the rhesus-monkey — a retrograde tracer study. J. Comp.
Perrett, D.I., Smith, P.A.J., Potter, D.D., Mistlin, A.J., Head, Neurol., 343: 445–463.
A.S., Milner, A.D. and Jeeves, M.A. (1984) Neurones re- Seung, Y., Kyong, J.S., Woo, S.H., Lee, B.T. and Lee, K.M.
sponsive to faces in the temporal cortex: studies of functional (2005) Brain activation during music listening in individuals
organization, sensitivity to identity and relation to percep- with or without prior music training. Neurosci. Res., 52:
tion. J. Comp. Physiol. Psychol., 3: 197–208. 323–329.
Perrett, D.I., Smith, P.A.J., Potter, D.D., Mistlin, A.J., Head, Siebert, M., Markowitsch, H.J. and Bartel, P. (2003) Amy-
A.S., Milner, A.D. and Jeeves, M.A. (1985) Visual cells in the gdala, affect and cognition: evidence from 10 patients with
temporal cortex sensitive to face view and gaze direction. Urbach-Wiethe disease. Brain, 126: 2627–2637.
Proc. R. Soc. Lond. B Biol. Sci., 223: 293–317. Singer, T., Seymour, B., O’Doherty, J., Kaube, H., Dolan, R.J.
Phan, K.L., Wager, T.D., Taylor, S.F. and Liberzon, I. (2004) and Frith, C.D. (2004) Empathy for pain involves the affec-
Functional neuroimaging studies of human emotions. CNS tive but not sensory components of pain. Science, 303:
Spectrom., 9: 258–266. 1157–1162.
Phillips, M.L., Young, A.W., Scott, S.K., Calder, A.J., Andrew, Small, D.M., Gregory, M.D., Mak, Y.E., Gitelman, D., Me-
C., Giampietro, V., Williams, S.C., Bullmore, E.T., Bra- sulam, M.M. and Parrish, T. (2003) Dissociation of neural
mmer, M. and Gray, J.A. (1998) Neural responses to facial representation of intensity and affective valuation in human
and vocal expressions of fear and disgust. Proc. R. Soc. gustation. Neuron, 39: 701–711.
London B Biol. Sci., 265: 1809–1817. Sprengelmeyer, R., Young, A.W., Schroeder, U., Grossenbac-
Phillips, M.L., Young, A.W., Senior, C., Brammer, M., An- her, P.G., Federlein, J., Buttner, T. and Przuntek, H. (1999)
drew, C., Calder, A.J., Bullmore, E.T., Perrett, D.I., Row- Knowing no fear. Proc. R. Soc. Lond. B Biol. Sci., 266:
land, D., Williams, S.C., Gray, J.A. and David, A.S. (1997) A 2451–2456.
specific neural substrate for perceiving facial expressions of Stern, D.N. (2000) The Interpersonal World of the Infant. Basic
disgust. Nature, 389: 495–498. Books, New York.
401

Tai, Y.F., Scherfler, C., Brooks, D.J., Sawamoto, N. and Wheaton, K.J., Thompson, J.C., Syngeniotis, A., Abbott, D.F.
Castiello, U. (2004) The human premotor cortex is ‘mirror’ and Puce, A. (2004) Viewing the motion of human body parts
only for biological actions. Curr. Biol., 14: 117–120. activates different regions of premotor, temporal, and pari-
Tanne-Gariepy, J., Rouiller, E.M. and Boussaoud, D. (2002) etal cortex. Neuroimage, 22: 277–288.
Parietal inputs to dorsal versus ventral premotor areas in the Wicker, B., Keysers, C., Plailly, J., Royet, J.P., Gallese, V. and
macaque monkey: evidence for largely segregated visuomotor Rizzolatti, G. (2003) Both of us disgusted in my insula: the
pathways. Exp. Brain Res., 145: 91–103. common neural basis of seeing and feeling disgust. Neuron,
Umilta, M.A., Kohler, E., Gallese, V., Fogassi, L., Fadiga, L., 40: 655–664.
Keysers, C. and Rizzolatti, G. (2001) I know what you are Williams, L.M., Phillips, M.L., Brammer, M.J., Skerrett, D.,
doing: a neurophysiological study. Neuron, 31: 155–165. Lagopoulos, J., Rennie, C., Bahramali, H., Olivieri, G.,
van der Gaag, C., Minderaa, R., and Keysers, C., 2005. Emo- David, A.S., Peduto, A. and Gordon, E. (2001) Arousal dis-
tion observation, recognition and imitation: towards an un- sociates amygdala and hippocampal fear responses: evidence
derstanding of empathy of individual emotions. J. Cogn. from simultaneous fMRI and skin conductance recording.
Neurosci., Suppl S, 166. Neuroimage, 14: 1070–1079.
Vogt, B.A. (2005) Pain and emotion interactions in subregions Young, A.W., Aggleton, J.P., Hellawell, D.J., Johnson, M.,
of the cingulate gyrus. Nat. Rev. Neurosci., 6: 533–544. Broks, P. and Hanley, J.R. (1995) Face Processing impair-
Waldvogel, D., van Gelderen, P., Muellbacher, W., Ziemann, ments after amygdalotomy. Brain, 118: 15–24.
U., Immisch, I. and Hallett, M. (2000) The relative meta- Zald, D.H. (2003) The human amygdala and the emotional
bolic demand of inhibition and excitation. Nature, 406: evaluation of sensory stimuli. Brain Res. Brain Res. Rev., 41:
995–998. 88–123.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 22

Empathizing: neurocognitive developmental


mechanisms and individual differences

Bhismadev Chakrabarti and Simon Baron-Cohen

Autism Research Centre, University of Cambridge, Psychiatry Department, Douglas House, 18B Trumpington Rd,
Cambridge CB2 2AH, UK

Abstract: This chapter reviews the Mindreading System model encompassing four neurocognitive mech-
anisms (ID, EDD, SAM, and ToMM) before reviewing the revised empathizing model encompassing two
new neurocognitive mechanisms (TED and TESS). It is argued that the empathizing model is more com-
prehensive because it entails perception, interpretation, and affective responses to other agents. Sex differ-
ences in empathy (female advantage) are then reviewed, as a clear example of individual differences in
empathy. This leads into an illustration of individual differences using the Empathy Quotient (EQ). Finally,
the neuroimaging literature in relation to each of the neurocognitive mechanisms is briefly summarized and
a new study is described that tests if different brain regions respond to the perception of different facial
expressions of emotion, as a function of the observer’s EQ.

Keywords: empathy; theory of mind; mindreading; neuroimaging; sex differences; psychopathology;


individual differences; basic emotions

Introduction to empathy by investigating if an individual’s level


of empathy affects how their brain processes dis-
In this chapter, we take the concept of empathy, crete emotions.
and consider it in terms of neurocognitive devel-
opmental mechanisms and in terms of individual What is empathizing?
differences. The first part of the chapter deals with
two conceptual approaches to the development of Empathizing is the drive to identify another per-
the empathizing ability. The second part of the son’s emotions and thoughts, and to respond to
chapter presents some empirical evidence on a these with an appropriate emotion (Davis, 1994).
quantitative trait measure of empathy. A simple We use the term ‘drive’ but recognize that it also
definition of ‘empathizing’ is that it is the lens overlaps with the concept of a skill or an ability.
through which we perceive and process emotions. We also focus on the definition of empathy given
We therefore review the literature from neuroim- by Davis while recognizing that other authors may
aging studies, which suggests that perception of have a slightly different definition. Empathizing
discrete basic emotions is processed in different does not just entail the cold calculation of what
neural regions and networks. Finally, we describe someone else thinks and feels (or what is some-
a recent study that reconciles these two approaches times called mindreading). Psychopaths can do
that much. Empathizing is also about having an
Corresponding author. Fax: +44-1223-746033; E-mail: appropriate emotional reaction inside you, an
bc249@cam.ac.uk emotion triggered by the other person’s emotion.

DOI: 10.1016/S0079-6123(06)56022-4 403


404

Empathizing is done in order to understand an- Cognitive empathy is involved in explicit under-
other person, predict their behaviour, and to con- standing of another’s feelings and switching to
nect or resonate with them emotionally. Imagine take their perspective. Piaget referred to empathy
you could recognize that ‘‘Jane is in pain,’’ but this as ‘decentering’, or responding nonegocentrically
left you cold, or detached, or happy, or preoccu- (Piaget and Inhelder, 1956). More recent develop-
pied. This would not be empathizing. Now imag- mental psychologists refer to this aspect of empa-
ine you do not just see Jane’s pain, but you also thy in terms of using a ‘theory of mind’, or
automatically feel concerned, wincing yourself, ‘mindreading’ (Astington et al., 1988; Whiten,
and feeling a desire to run across and help alleviate 1991). Essentially, the cognitive component of em-
her pain. This is empathizing. And empathizing pathizing entails setting aside your own current
extends to recognizing and responding to any perspective, attributing a mental state (sometimes
emotion or state of mind, not just the more ob- called an ‘attitude’) to the other person, and then
vious ones, like pain. Empathy is a skill (or a set of inferring the likely content of their mental state,
skills). As with any other skill, such as athleticism given their experience. The cognitive element also
or mathematical or musical ability, we all vary in allows you to predict the other person’s mental
it. In the same way that we can think about why state or behaviour.
someone is talented or average or even disabled in The second aspect to empathy is the ‘affective’
these other areas, so we can think about individual component (Hobson, 1993). A similar component in
differences in empathy. other accounts has been called ‘emotional conta-
Empathy is a defining feature of human rela- gion’, defined as the tendency to mimic and syn-
tionships. Empathy stops you doing things that chronise facial expressions automatically,
would hurt another person’s feelings. Empathy vocalizations, postures, and movements with those
also stops you inflicting physical pain on a person of another person, to converge emotionally (Hat-
or animal. Empathy allows you to tune into some- field et al., 1992). This may be the most primitive
one else’s world, setting aside your own world — component of empathy. For example, when wit-
your perception, knowledge, assumptions, or feel- nessing someone else in a state of fear, if the observer
ings. It allows you to see another side of an ar- ‘catches’ a similar state of fear, this acts as a ‘quick-
gument easily. Empathy drives you to care for, or and-easy’ route to alerting oneself to environmental
offer comfort to, another person, even if they are dangers without having to face the dangers oneself.
unrelated to you and you stand to gain nothing in A third component involves a ‘concern mecha-
return. Empathy also makes real communication nism’ (Nichols, 2001) often associated with a pro-
possible. Talking ‘‘at’’ a person is not real com- social/altruistic component, also termed
munication. It is a monologue. Real conversation ‘sympathy’. This is distinct from emotional conta-
is sensitive to this listener at this time. Empathy gion in not necessarily involving matched states
also provides a framework for the development of between the observer and the person experiencing
a moral code. Moral codes are built out of fellow- the emotion, and being possibly specific to a cer-
feeling and compassion. tain class of emotions (sadness and pain, but not
disgust or happiness) in the other person. It rep-
resents a case where the observer feels both an
Fractionating empathy emotional response to someone else’s distress and
a desire to alleviate their suffering.
Philosophical (Stein, 1989) and evolutionary
(Brothers, 1990; Levenson, 1996; Preston and de
Waal, 2002) accounts have suggested that empa- How does empathizing develop? The Mindreading
thizing is not a unitary construct. Possible constit- System
uent ‘fractions’ of empathy include (1) ‘emotional
contagion/affective empathy’, (2) ‘cognitive empa- In 1994, Baron-Cohen proposed a model to specify
thy’, and (3) sympathy. the neurocognitive mechanisms that comprise the
405

‘Mindreading System’ (Baron-Cohen, 1994, 1995). developmentally prior to the other two mecha-
Mindreading is defined as the ability to interpret nisms, and are active early in infancy, if not from
one’s own or another agent’s actions as driven by birth.
mental states. The model was proposed in order to SAM is developmentally more advanced. SAM
explain (1) ontogenesis of a theory of mind and (2) automatically represents or interprets if the self
neurocognitive dissociations that are seen in chil- and another agent are (or are not) perceiving the
dren with or without autism. The model is shown same event. SAM does this by building ‘triadic’
in Fig. 1 and contains four components: the in- representations. For example, where ID can build
tentionality detector (ID), the eye direction detec- the dyadic representation ‘‘Mother wants the cup’’
tor (EDD), the shared attention mechanism and where EDD can build the dyadic representa-
(SAM), and finally the theory-of-mind mechanism tion ‘‘Mother sees the cup’’, SAM can build the
(ToMM). triadic representation ‘‘Mother sees that I see the
ID and EDD build ‘Dyadic’ representations of cup’’. As is apparent, triadic representations in-
simple mental states. ID automatically represents volve embedding or recursion. (A dyadic repre-
or interprets an agent’s self-propelled movement as sentation ‘‘I see a cup’’ is embedded within
a desire or goal-directed movement, a sign of its another dyadic representation ‘‘Mum sees the
agency, or an entity with volition (Premack, 1990). cup’’ to produce this triadic representation.)
For example, ID interprets an animate-like mov- SAM takes its input from ID and EDD, and tri-
ing shape as ‘‘it wants x’’ or ‘‘it has goal y.’’ EDD adic representations are made out of dyadic rep-
automatically interprets or represents eye-like resentations. SAM typically functions from 9 to 14
stimuli as ‘‘looking at me’’ or ‘‘looking at some- months of age, and allows ‘joint attention’ behav-
thing else.’’ That is, EDD picks out that an entity iours such as protodeclarative pointing and gaze
with eyes can perceive. Both ID and EDD are monitoring (Scaife and Bruner, 1975).
ToMM is the jewel in the crown of the 1994
model of the Mindreading System. It allows
ID EDD 0-9m epistemic mental states to be represented (e.g.,
‘‘Mother thinks this cup contains water’’ or
‘‘Mother pretends this cup contains water’’), and
it integrates the full set of mental-state concepts
(including emotions) into a theory. ToMM devel-
ops between two and four years of age, and allows
pretend play (Leslie, 1987), understanding of false
belief (Wimmer and Perner, 1983), and under-
SAM 9-14m standing of the relationships between mental states
(Wellman, 1990). An example of the latter is the
seeing-leads-to-knowing principle (Pratt and Bry-
ant, 1990), where a typical 3-year-old can infer
that if someone has seen an event, then they will
ToMM 2-4yrs know about it.
The model shows the ontogenesis of a theory of
Key: IDD = Intentionality Detector
mind in the first 4 years of life, and justifies
EDD = Eye Direction Detector
SAM = Shared Attention Mechanism the existence of four components on the basis of
ToMM = Theory of Mind Mechanism developmental competence and neuropsychologi-
m-months cal dissociation. In terms of developmental
competence, joint attention does not appear
Fig. 1. Baron-Cohen’s (1994) model of mindreading. Key:
IDD ¼ Intentionality Detector; EDD ¼ Eye Direction Detec- possible until 9–14 months of age, and joint at-
tor; SAM ¼ Shared Attention Mechanism; ToMM ¼ Theory- tention appears to be a necessary but not sufficient
of-Mind Mechanism. condition for understanding epistemic mental
406

states (Baron-Cohen, 1991; Baron-Cohen and TED ID EDD 0 -9m


Swettenham, 1996). There appears to be a devel-
opmental lag between acquiring SAM and
ToMM, suggesting that these two mechanisms
are dissociable. In terms of neuropsychological
dissociation, congenitally blind children can ulti-
mately develop joint (auditory or tactile) attention
(i.e., SAM), using the amodal ID rather than the
visual EDD route. They can therefore go on to SAM 9 -14m
develop ToMM. Children with autism appear to
be able to represent the dyadic mental states of
seeing and wanting, but show delays in shared at-
tention (Baron-Cohen, 1989b) and in understand-
ing false belief (Baron-Cohen et al., 1985; Baron-
Cohen, 1989a) — that is, in acquiring SAM and
ultimately ToMM. It is this specific developmental TESS (14m) ToMM (48m)
delay that suggests that SAM and ToMM are dis-
Key: As in Figure 1, but:
sociable from EDD. TED = The Emotion Detector; and
TESS = The Empathising SyStem

Fig. 2. Baron-Cohen’s (2005) model of empathizing. Key: As in


Shortcomings of the Mindreading System model: Fig. 1, with TED ¼ The Emotion Detector and TESS ¼ The
the Empathizing SyStem Empathizing SyStem.

The 1994 model of the Mindreading System was


revised in 2005 because of certain omissions and typically motivates the observer to respond to the
too narrow a focus. The key omission is that in- other person).
formation about affective states, available to the Like the other infancy perceptual input mecha-
infant perceptual system, has no dedicated neuro- nisms of ID and EDD, the new component of
cognitive mechanism. In Fig. 2, the revised model TED can build dyadic representations of a special
(Baron-Cohen, 2005) is shown and now includes a kind, namely, it can represent affective states. An
new fifth component: the emotion detector (TED). example would be ‘‘Mother — is unhappy’’ or
But the concept of mindreading (or theory of even ‘‘Mother — is angry with me.’’ Formally, we
mind) makes no reference to the affective state in can describe this as agent-affective state proposi-
the observer triggered by recognition of another’s tion. We know that infants can represent affective
mental state. This is a particular problem for any states from as early as three months of age
account of the distinction between autism and (Walker, 1982). As with ID, TED is amodal, in
psychopathy. For this reason, the model is no that affective information can be picked up from
longer of ‘mindreading’ but is of ‘empathizing’, facial expression, or vocal intonation, ‘motherese’
and the revised model also includes a new sixth being a particularly rich source of the latter (Field,
component: The Empathizing SyStem (TESS). 1979). Another person’s affective state is presum-
(TESS is spelt as it is to playfully populate the ably also detectable from their touch (e.g., tense
Mindreading Model with apparently anthropo- vs. relaxed), which implies that congenitally blind
morphic components.) Where the 1994 Mindread- infants should find affective information accessible
ing System was a model of a passive observer through both auditory and tactile modalities. TED
(because all the components had simple decoding allows the detection of the basic emotions (Ekman
functions), the 2005 Empathizing SyStem is a and Friesen, 1969). The development of TED is
model of an observer impelled towards action (be- probably aided by simple imitation that is typical
cause an emotion is triggered in the observer which of infants (e.g., imitating caregiver’s expressions),
407

which in itself would facilitate emotional conta- that you are in pain’’. Here, TESS is needed, since
gion (Meltzoff and Decety, 2003). an appropriate affective state has been triggered in
When SAM becomes available, at 9–14 months the observer by the emotional state identified in the
of age, it can receive inputs from any of the three other person. And where ToMM employs M-rep-
infancy mechanisms, ID, EDD, or TED. Here, we resentations1 (Leslie, 1995) of the form agent-atti-
focus on how a dyadic representation of an affec- tude-proposition (e.g., Mother — believes —
tive state can be converted into a triadic represen- Johnny took the cookie), TESS employs a new
tation by SAM. An example would be that the class of representations, which we can all E-repre-
dyadic representation ‘‘Mother is unhappy’’ can be sentations2 of the form self-affective state [agent-
converted into a triadic representation ‘‘I am un- affective state-proposition] (e.g., ‘‘I feel sorry that
happy that Mother is unhappy’’, ‘‘Mother is un- — Mom feels sad about — the news in the letter’’)
happy that I am unhappy’’, etc. Again, as with (Baron-Cohen, 2003). The critical feature of this E-
perceptual or volitional states, SAM’s triadic rep- representation is that the self’s affective state is ap-
resentations of affective states have this special propriate to and triggered by the other person’s
embedded or recursive property. affective state. Thus, TESS can represent [I am
TESS in the 2005 model is the real jewel in the horrified — that you are in pain], or [I am con-
crown. This is not to minimize the importance of cerned — that you are in pain], or [I want to al-
ToMM, which has been celebrated for the last 20 leviate — that you are in pain], but it cannot
years in research in developmental psychology represent [I am happy — that you are in pain]. At
(Leslie, 1987; Wimmer et al., 1988; Whiten, 1991). least, it cannot do so if TESS is functioning nor-
ToMM is of major importance in allowing the mally. One could imagine an abnormality in TESS
child to represent the full range of mental states, leading to such inappropriate emotional states be-
including epistemic ones (such as false belief), and ing triggered, or one could imagine them arising
is important in allowing the child to pull mental- from other systems (such as a competition system
istic knowledge into a useful theory with which to or a sibling-rivalry system), but these would not be
predict behaviour (Wellman, 1990; Baron-Cohen, evidence of TESS per se.
1995). But TESS allows more than behavioural
explanation and prediction (itself a powerful
achievement). TESS allows an empathic reaction Dissociations between TED, ToMM, and TESS
to another’s emotional state. This is, however, not from neuropsychiatry
to say that these two modules do not interact.
Knowledge of mental states of others made pos- Before leaving this revision of the model, it is
sible by ToMM could certainly influence the way worth discussing why the need for this has arisen.
in which an emotion is processed and/or expressed First, emotional states are an important class of
by TESS. TESS also allows for sympathy. It is this mental states to detect in others, and yet the earlier
element of TESS that gives it the adaptive benefit model focused only on volitional, perceptual, in-
of ensuring that organisms feel a drive to help each formational, and epistemic states. Second, when it
other. comes to pathology, it would appear that in autism
TED might function, although this may be delayed
(Hobson, 1986; Baron-Cohen et al., 1993, 1997c),
M-representations versus E-representations at least in terms of detecting basic emotions. Even
high-functioning people with autism or Asperger
To see the difference between TESS and ToMM, Syndrome have difficulties both in ToMM (when
consider this example: ‘‘I see you are in pain’’. Here, measured with mental-age appropriate tests)
ToMM is needed to interpret your facial ex- (Happé, 1994; Baron-Cohen et al., 1997b, 2001)
pressions and writhing body movements in terms
of your underlying mental state (pain). But now 1
M stands for mental.
2
consider this further example: ‘‘I am devastated — E stands for empathy.
408

and TESS (Attwood, 1997; Baron-Cohen et al., Sex differences in empathizing


1999a, b, 2003, 2004). This suggests that TED and
TESS may be fractionated. In the introduction to this chapter we promised to
In contrast, the psychiatric condition of psy- consider sex differences in empathizing. Some of the
chopathy may entail an intact TED and ToMM, best evidence for individual differences in empa-
alongside an impaired TESS. The psychopath (or thizing comes from the study of sex differences,
sociopath) can represent that you are in pain, or where many studies converge on the conclusion that
that you believe — that he is the gas-man, thereby there is a female superiority in empathizing. Sex
gaining access to your house or your credit card. differences are best viewed as summated individual
The psychopath can go on to hurt you or cheat differences, on multiple dimensions that include ge-
you without having the appropriate affective re- netic and epigenetic factors. Some of the observed
action to your affective state. In other words, he or behavioural differences are reviewed here.
she does not care about your affective state (Me-
aley, 1995; Blair et al., 1997). Lack of guilt or (1) Sharing and turn-taking. On average, girls
shame or compassion in the presence of another’s show more concern for fairness, while boys
distress are diagnostic of psychopathy (Cleckley, share less. In one study, boys showed 50
1977; Hare et al., 1990). Separating TESS and times more competition, while girls showed
ToMM thus allows a functional distinction to be 20 times more turn-taking (Charlesworth
drawn between the neurocognitive causes of au- and Dzur, 1987).
tism and psychopathy. (2) Rough and tumble play or ‘‘rough housing’’
(wrestling, mock fighting, etc). Boys show
more of this than girls do. Although there
Developmental dissociations is a playful component, it can hurt or be
intrusive, so it needs lower empathizing to
Developmentally, one can also distinguish TED carry it out (Maccoby, 1999).
from TESS. We know that at three months of age, (3) Responding empathically to the distress of
infants can discriminate facial and vocal expres- other people. Girls aged 1 year or more
sions of emotion (Walker, 1982; Trevarthen, show greater concern through more sad
1989), but that it is not until about 14 months looks, sympathetic vocalizations, and com-
that they can respond with appropriate affect forting. More women than men also report
(e.g., a facial expression of concern) to another’s frequently sharing the emotional distress of
apparent pain (Yirmiya et al., 1990) or show ‘‘so- their friends. Women also show more com-
cial referencing.’’ Clearly, this account is skeletal forting, even of strangers, than men do
in not specifying how many emotions TED is ca- (Hoffman, 1977).
pable of recognizing. Our recent survey of emo- (4) Using a ‘theory of mind’. By three years of
tions identifies that there are 412 discrete emotion age, little girls are already ahead of boys in
concepts that the adult English language user their ability to infer what people might be
recognizes (Baron-Cohen et al., submitted). How thinking or intending (Happe, 1995). This
many of these are recognized in the first year of sex difference appears in some but not all
life is not clear. It is also not clear exactly how studies (Charman et al., 2002).
empathizing changes during the second year of (5) Sensitivity to facial expressions. Women are
life. We have assumed the same mechanism that better at decoding nonverbal communica-
enables social referencing at 14-months-old also tion, picking up subtle nuances from tone
allows sympathy and the growth of empathy of voice or facial expression, or judging a
across development. This is the most parsimoni- person’s character (Hall, 1978).
ous model, though it may be that future research (6) Questionnaires measuring empathy. Many
will justify further mechanisms that affect the de- of these find that women score higher than
velopment of empathy. men (Davis, 1994).
409

(7) Values in relationships. More women value exchange with a partner going for longer.
the development of altruistic, reciprocal re- When girls disagree, they are more likely to
lationships, which by definition require em- express their different opinion sensitively,
pathizing. In contrast, more men value in the form of a question rather than an
power, politics, and competition (Ahlgren assertion. Boys’ talk is more ‘single-voiced
et al., 1979). Girls are more likely to en- discourse’ (the speaker presents their own
dorse co-operative items on a questionnaire perspective alone). The female speech style
and to rate the establishment of intimacy as is more ‘double-voiced discourse’ (girls
more important than the establishment of spend more time negotiating with the other
dominance. Boys are more likely than girls person, trying to take the other person’s
to endorse competitive items and to rate wishes into account) (Smith, 1985).
social status as more important than inti- (13) Talk about emotions. Women’s conversa-
macy (Knight et al., 1989). tion involves much more talk about feel-
(8) Disorders of empathy (such as psycho- ings, while men’s conversation with each
pathic personality disorder or conduct dis- other tends to be more object or activity
order) are far more common among males focused (Tannen, 1991).
(Dodge, 1980; Blair, 1995). (14) Parenting style. Fathers are less likely than
(9) Aggression, even in normal quantities, can mothers to hold their infant in a face-to-
only occur with reduced empathizing. Here face position. Mothers are more likely to
again, there is a clear sex difference. Males follow through the child’s choice of topic in
tend to show far more ‘direct’ aggression play, while fathers are more likely to im-
(pushing, hitting, punching, etc.) while fe- pose their own topic. And mothers fine-
males tend to show more ‘indirect’ (or re- tune their speech more often to match what
lational, covert) aggression (gossip, the child can understand (Power, 1985).
exclusion, bitchy remarks, etc.). Direct ag- (15) Face preference and eye contact. From
gression may require an even lower level of birth, females look longer at faces, and
empathy than indirect aggression. Indirect particularly at people’s eyes, and males are
aggression needs better mindreading skills more likely to look at inanimate objects
than does direct aggression, because its im- (Connellan et al., 2001).
pact is strategic (Crick and Grotpeter, (16) Finally, females have also been shown to
1995). have better language ability than males. It
(10) Murder is the ultimate example of a lack of seems likely that good empathizing would
empathy. Daly and Wilson (1988) analysed promote language development (Baron-
homicide records dating back over 700 Cohen et al., 1997a) and vice versa, so
years, from a range of different societies. these may not be independent.
They found that ‘male-on-male’ homicide
was 30–40 times more frequent than ‘fe- Leaving aside sex differences as one source of
male-on-female’ homicide. evidence for individual differences, one can see
(11) Establishing a ‘dominance hierarchy’. Males that empathy is normally distributed within the
are quicker to establish these. This in part population. Figure 3 shows the data from the Em-
may reflect their lower empathizing skills, pathy Quotient (EQ), a validated 60-item self-re-
because often a hierarchy is established by port questionnaire (Baron-Cohen and
one person pushing others around, to be- Wheelwright, 2004). It has been factor analysed
come the leader (Strayer, 1980). to suggest the existence of three distinct compo-
(12) Language style. Girls’ speech is more co- nents, which roughly correspond to the three-com-
operative, reciprocal, and collaborative. In ponent model of empathy (Lawrence et al., 2004).
concrete terms, this is also reflected in girls Scores on the EQ show a quasi-normal distribu-
being able to keep a conversational tion in several populations, with scores from
410

25 proposed above, presented in order of their devel-


AS/HFA group
Controls
opment.
20
Number of subjects

15 1. Studies of emotional contagion have demon-


strated involuntary facial mimicry (Dimberg
10 et al., 2000) as well as activity in regions of
the brain where the existence of ‘mirror’ neu-
5
rons has been suggested (Carr et al., 2003;
0 Wicker et al., 2003; Jackson et al., 2005).
0 to 5
6 to 10
11 to 15
16 to 20
21 to 25
26 to 30
31 to 35
36 to 40
41 to 45
46 to 50
51 to 55
56 to 60
61 to 65
66 to 70
71 to 75
76 to 80
2. ID has been tested (Brunet et al., 2000) in a
PET study in a task involving attribution of
intentions to cartoon characters. Reported
EQ score activation clusters included the right medial
prefrontal (BA 9), inferior frontal (BA 47)
Fig. 3. The distribution of EQ in the general population (dot-
ted line). Also shown is the distribution of empathy scores from cortices, superior temporal gyrus (BA 42),
people with Asperger Syndrome (AS) or High Functioning and bilateral anterior cingulate cortex. In an
Autism (HFA). (From Baron-Cohen and Wheelwright, 2004.) elegant set of experiments that required par-
ticipants to attribute intentions to animations
people with Autism Spectrum Conditions (ASCs) of simple geometric shapes (Castelli et al.,
clustering toward the lower end (see Fig. 3). The 2000), it was found that the ‘intentionality’
EQ shows significant sex differences (Goldenfeld et score attributed by the participants to indi-
al., 2006). vidual animations was positively correlated
The search for the neural correlates of empathy to the activity in superior temporal sulcus
has had two traditions of research, one focusing on (STS), the temporo-parietal junction, and the
theory-of-mind studies (involving largely intention medial prefrontal cortex. A subsequent study
attribution or emotion attribution) and another (Castelli et al., 2002) demonstrated a group
focusing on action understanding. The latter has difference in activity in the same set of struc-
gained considerable importance in recent years tures between people with Autism/Asperger
since the discovery of mirror neurons (Gallese et Syndrome and neurotypical controls.
al., 2004). 3. EDD has been studied in several neuroimaging
On finding increasing evidence of sex differences studies on gaze direction perception (Calder et
in the EQ in the general population, we sought to al., 2002; Pelphrey et al., 2003; see Grosbras et
investigate the neural correlates of this trait meas- al., 2005 for a review) and have implicated the
ure of empathizing across the population. Since posterior STS bilaterally. This evidence, taken
empathizing can be viewed as a lens through which together with similar findings from primate lit-
we perceive and process emotions, we attempted to erature (Perrett and Emery, 1994), suggests this
marry the two fields of emotion perception and area to be a strong candidate for the anatom-
empathizing. The following section briefly intro- ical equivalent of the EDD.
duces the current state of the literature on the 4. A recent imaging study (Williams et al., 2005)
neural bases of basic emotions and the results of a investigated the neural correlates of SAM and
recent study from our lab. reported bilateral activation in anterior cingu-
late (BA 32,24), and medial prefrontal cortex
(BA 9,10) and the body of caudate nucleus in
Neuroimaging studies of empathizing and emotion a joint attention task, when compared to a
control task involving nonjoint attention (see
Neuroimaging studies have implicated the follow- Frith and Frith, 2003 for a review).
ing different brain areas for performing tasks that 5. Traditional ‘theory-of-mind’ (cognitive em-
tap components of the model of empathy pathy) tasks have consistently shown activity
411

in medial prefrontal cortex, superior tempo- of sad expressions are comparatively less consist-
ral gyrus, and the temporo-parietal junctions ent. Perception of sad face and induction of sad
(Frith and Frith, 2003; Saxe et al., 2004). This mood are both known to be associated with an
could be equated to the brain basis of increased response in the subgenual cingulate cor-
ToMM. tex (Mayberg et al., 1999; Liotti et al., 2000), the
6. Sympathy has been relatively less investigated, hypothalamus in humans (Malhi et al., 2004) and
with one study implicating the left inferior in rats (Shumake, Edwards et al., 2001) as well as
frontal gyrus, among a network of other in the middle temporal gyrus (Eugene, Levesque et
structures (Decety and Chaminade, 2003). al., 2003). There have been very few studies on the
Work on ‘moral’ emotions has suggested the passive viewing of surprise. One study by (Schroe-
involvement of a network comprising the me- der et al., 2004) has reported bilateral activation in
dial frontal gyrus, the medial orbitofrontal the parahippocampal region, which is known for
cortex, and the STS (Moll et al., 2002). its role in novelty detection from animal literature.
While the discrete emotions model holds well for
these relatively ‘simple’ emotions, the dimensional
Neuroimaging of discrete emotions models (e.g. see (Rolls, 2002)) become increasingly
relevant as we consider the more ‘socially complex’
An increasing body of evidence from lesion, ne- emotions, e.g. pride, shame and guilt – since it
uroimaging and electrophysiological studies sug- would not be very economical to have discrete
gest that these affect programs might have discrete neural substrates for the whole gamut of emotions.
neural bases (Calder et al., 2001). Fear is possibly These two models, however, need not be in con-
the single most well investigated emotion. Passive flict, since the more complex emotions can be con-
viewing of fear expressions as well as experiencing ceptualized as being formed of a combination of
fear (as induced through recalling a fear memory, the basic ones (i.e. with each of the ‘basic’ emo-
or seeing fearful stimuli) activates the amygdala, tions representing a dimension in emotion space).
orbitofrontal cortex and the anterior cingulate Two major meta-analytic studies of neuroimag-
cortex (Morris et al., 1999; Damasio et al., 2000). ing literature on emotions highlight the role of
There is considerable evidence from non-human discrete regions in primarily visual processing of
primates (Kalin et al., 2001; Prather et al., 2001) different basic emotions (Phan et al., 2002; Mu-
and rats (LeDoux, 2000) to suggest a crucial role rphy et al., 2003). Some studies using stimuli in
for these regions in processing fear. Visual or au- other sensory modalities (olfactory (Anderson et
ditory perception of disgust expressions as well as al., 2003) gustatory (Small et al., 2003), auditory
experiencing disgust is known to activate the an- (Lewis et al., in press)) have shown the possibly
terior insula and the pallidum (Phillips et al., 1997, dissociable role for the amygdala and the orbito-
1998; Wicker et al., 2003). An increasing consensus frontal cortex in processing emotions along the
on the role of the ventral striatum in processing two dimensions of valence and arousal.
reward from different sensory domains [receiving The relative absence of neuroimaging studies on
food rewards (O’Doherty et al., 2002), viewing ‘complex’ emotions could be possibly due to the
funny cartoons (Mobbs et al., 2003), remembering increased cultural variability of the elicitors as well
happy events (Damasio et al., 2000)] concurs well as the display rules that these expressions entail.
with studies that report activation of this region in Among the few exceptions, guilt and embarrass-
response to viewing happy faces (Phillips et al., ment have been investigated by Takahashi et al.,
1998; Lawrence et al., 2004). 2004 who reported activation in ventromedial pre-
Perception of angry expressions have been frontal cortex, left Superior Temporal Sulcus
shown to evoke a response in the premotor cor- (STS) and higher visual cortices when participants
tex and the striatum (Grosbras and Paus, in press) read sentences designed to evoke guilt or embar-
as well as the lateral orbitofrontal cortex (Blair et rassment. This, taken together with the areas un-
al., 1999). The results of studies on the processing derlying the ‘ToMM’ system could possibly
412

suggest an increased role of ‘theory-of-mind’ to with EQ scores for different basic emotions
make sense of these emotions. (Chakrabarti et al., in press).
Different regional responses were found to cor-
relate with the EQ for different emotions, suggest-
ing that there is no unitary correlate of the
‘Empathizing’ with discrete emotions? personality trait of empathizing across these emo-
tions. Specifically, for perception of happy faces, a
Returning to the concept of individual differences parahippocampal-ventral striatal cluster response
in empathizing, this poses an interesting question was positively correlated with the EQ. The role of
for the brain basis of perception of discrete emo- this region in reward processing is well known
tions. Do we use a centralized ‘empathy circuit’ to (O’Doherty et al., 2004). This suggests that the
make sense of all emotions? If so, can one detect more ‘‘empathic’’ a person is, the higher is his/her
differences in how discrete emotions are processed reward response to a happy face. Interestingly, the
in individuals who are at different points on the response from the same region correlated nega-
EQ continuum? tively with the EQ during perception of sad faces.
A direct approach to investigate individual This fits perfectly with the earlier results, i.e., the
differences in empathizing has been to test for more empathic a person is, the lower is his/her
sex differences in perception of emotions. Using reward response to a sad face.
facial Electromyography (EMG), one study (Hel- For happy and sad faces therefore, empathizing
land, 2005) observed that females tend to show seems to involve mirroring. The higher a person’s
increased facial mimicry to facial expressions of EQ, the stronger the reward response to happy
happiness and anger when compared to males. In a faces and vice versa for sad faces. This is in con-
metareview of neuroimaging results on sex differ- cordance with suggestions from earlier studies on
ences on emotion perception, Wager et al. (2003) pain perception (Singer et al., 2004) and disgust
reported that females show increased bilaterality in perception (Wicker et al., 2003), where observation
emotion-relevant activation compared to males. and experience have been shown to be mediated by
This is not always found (Lee et al., 2002; Schienle the same set of structures. One of the issues with
et al., 2005). One of the reasons for this might have the previous studies is a possible confound be-
been the fact that sex differences are summated tween ‘personal distress’ and empathizing. The
individual differences. Instead of such a broad novel element in our study is that we explicitly
category-based approach (as in sex-difference tested for the personality trait of empathizing in
studies), an approach based on individual differ- relation to perception of specific emotions.
ences in self-report personality scores (Canli et al., However, empathizing does not appear to be
2002) or genetic differences (e.g., Hariri et al., purely an index of mirroring. For perception of
2002) may be more finely tuned. angry faces, EQ correlated positively to clusters
To test this model of individual variability, we centered on the precuneus/inferior parietal lobule,
asked if an individual’s score on the EQ predicted the superior temporal gyrus, and the dorsolateral
his/her response to four basic emotions (happy, prefrontal cortex. The posterior cingulate region is
sad, angry, and disgust). If empathizing were known to be involved in self/other distinction
modulated by a unitary circuit, then individual (Vogt, 2005), and the superior temporal gyrus is
differences in empathizing would correlate with known for its role in ToMM tasks (Saxe et al.,
activity in the same structures for all basic emo- 2004). This suggests that higher EQ corresponds to
tions. Twenty-five volunteers (13 female and 12 higher activation in areas related to the distinction
male) selected across the EQ space were scanned in of self versus other, as well as those that are re-
a 3 T functional magnetic resonance imaging cruited to determine another person’s intentions.
(fMRI) scanner on a passive viewing task using The dorsolateral prefrontal cortex is known for its
dynamic facial expressions as stimuli. It was found role in decision making and context evaluation
that activity in different brain regions correlated (Rahm et al., 2006). Higher EQ would therefore
413

predict better evaluation of the threat from an an- the other’s emotional state (e.g., the Superior
gry expression. Since expressions of anger are usu- Temporal Gyrus (STG) and Inferior Perietal Lob-
ally more socially urgent for attention than those ule (IPL)/precuneus response to angry faces cor-
of either sadness or happiness, it is essential that relating with EQ).
highly empathic persons do not merely ‘mirror’ the This study provides support for the discrete
expression. A high empathizer’s perception of an emotions model discussed above, but reveals how
angry face would therefore need to be accompa- empathy at the neural level is subtle and complex:
nied by an accurate determination of the inten- Neural networks activated by perception of dis-
tions of the person as well as an evaluation of the crete emotions depend on the observer’s EQ. Em-
posed threat. pathy is likely to be determined by other individual
In response to disgust faces, a cluster containing differences, such as fetal testosterone (Knickmeyer
the dorsal anterior cingulate cortex and medial et al., 2005; Knickmeyer and Baron-Cohen, 2006),
prefrontal cortices is negatively correlated with genetic variation (Skuse et al., 1997; Chakrabarti
EQ, suggesting that the areas involved in attribu- et al., 2006), as well as early care or neglect
tion of mental states (primarily required for deci- (Fonagy et al., 1997). We conclude that more basic
phering the ‘complex’ emotions) are selectively neuroscience into empathy will enrich our under-
recruited less by people of high EQ. This is what standing of this most fundamental human quality.
might be expected, since disgust as an emotion is
less interpersonal than anger or sadness, so re-
Abbreviations
sources for decoding complex emotional signals
need not be utilized. Another cluster that includes
EDD Eye Direction Detector
the right insula and Inferior Frontal Gyrus (IFG)
EQ Empathy Quotient
is negatively correlated with EQ. Given the well-
ID Intentionality Detector
established role of this region in processing dis-
SAM Shared Attention Mechanism
gust, this was a surprising result. We expected that
TED The Emotion Detector
an increased ability to empathize would result in
TESS The Empathy SyStem
an increased disgust response to facial expressions
ToMM Theory-of-Mind Mechanism
of disgust. The negative correlation suggests that
people with high EQ had a lower insula-Inferior
Frontal Gyrus (IFG) response to disgust expres-
sions. A re-examination of the behavioural litera-
ture on disgust sensitivity reveals a similar result Acknowledgements
since Haidt et al. (1994) suggested that increased
socialization leads to lower disgust sensitivity. In- S.B-C. was supported by the MRC and the Lurie
dividuals with high EQ may socialize more than Marks Family Foundation during the period of
those with low EQ. this work. B.C. was supported by Trinity College,
Together, these results demonstrate variability Cambridge. Parts of this chapter are reprinted
among different basic emotions in how empathy from elsewhere (Baron-Cohen, 2005; Goldenfeld et
interacts with them. This fits with the core idea al., 2006; Chakrabarti et al., in press).
that the different basic emotions have relatively
independent evolutionary antecedents (Panksepp, References
1998) and social communicatory functions (Izard
and Ackerman, 2000). While some of the emotions Anderson, A., Christoff, K., Stappen, I., Panitz, D., Ghahre-
involve more ‘mirroring’ (the same areas show ac- mani, D., Glover, G., Gabrieli, J. and Sobel, N. (2003) Dis-
sociated neural representations of intensity and valence in
tivation during recognition and experience, e.g.,
human olfaction. Nature Neuroscience, 6: 196–202.
the striatal response to happy faces correlating Ahlgren, A. and Johnson, D.W. (1979) Sex differences in co-
positively with EQ), others require an increased operative and competitive attitudes from the 2nd to the 12th
distinction between one’s own emotional state and grades. Dev. Psychol., 15: 45–49.
414

Astington, J., Harris, P. and Olson, D. (1988) Developing The- Asperger Syndrome or high-functioning autism. J. Child
ories of Mind. Cambridge University Press, New York. Psychol. Psychiat., 42: 241–252.
Attwood, T. (1997) Asperger’s Syndrome. Jessica Kingsley, Baron-Cohen, S., Wheelwright, S. and Jolliffe, T. (1997c) Is
London, UK. there a ‘‘language of the eyes’’? Evidence from normal adults
Baron-Cohen, S. (1989a) The autistic child’s theory of mind: a and adults with autism or Asperger Syndrome. Vis. Cogn., 4:
case of specific developmental delay. J. Child Psychol. Psych- 311–331.
iat., 30: 285–298. Baron-Cohen, S., Wheelwright, S., Stone, V. and Rutherford,
Baron-Cohen, S. (1989b) Perceptual role taking and protode- M. (1999b) A mathematician, a physicist, and a computer
clarative pointing in autism. Br. J. Devl. Psychol., 7: 113–127. scientist with Asperger Syndrome: performance on folk psy-
Baron-Cohen, S. (1991) Precursors to a theory of mind: un- chology and folk physics test. Neurocase, 5: 475–483.
derstanding attention in others. In: Whiten, A. (Ed.), Natural Blair, R.J. (1995) A cognitive developmental approach to mo-
Theories of Mind. Basil Blackwell, Oxford. rality: investigating the psychopath. Cognition, 57: 1–29.
Baron-Cohen, S. (1994) The mindreading system: new direc- Blair, R.J.R., Morris, J., Frith, C., Perrett, D.I. and Dolan, R.J.
tions for research. Curr. Psychol. Cogn., 13: 724–750. (1999) Dissociable neural responses to facial expressions of
Baron-Cohen, S. (1995) Mindblindness: An Essay on Autism sadness and anger. Brain, 122: 883–893.
and Theory of Mind. MIT Press/Bradford Books, Boston. Blair, R.J., Jones, L., Clark, F. and Smith, M. (1997). The
Baron-Cohen, S. (2003) The Essential Difference: Men, Women psychopathic individual: a lack of responsiveness to distress
and the Extreme Male Brain. Penguin, London. cues? Psychophysiology, 34: 192–198.
Baron-Cohen, S. (2005) The empathizing system: a revision of Brothers, L. (1990) The neural basis of primate social commu-
the 1994 model of the mindreading system. In: Ellis, B. and nication. Motiv. Emotion, 14: 81–91.
Bjorklund, D. (Eds.), Origins of the Social Mind. Guilford, Brunet, E., Sarfati, Y., Hardy-Bayle, M.-C. and Decety, J.
New York, USA. (2000) A PET investigation of the attribution of intentions
Baron-Cohen, S., Baldwin, D. and Crowson, M. (1997a) Do with a non-verbal task. NeuroImage, 11: 157–166.
children with autism use the Speaker’s Direction of Gaze Calder, A.J., Lawrence, A.D., Keane, J., Scott, S.K., Owen,
(SDG) strategy to crack the code of language? Child Dev., A.M., Christoffels, I. and Young, A.W. (2002) Reading the
68: 48–57. mind from eye gaze. Neuropsychologia, 40: 1129–1138.
Baron-Cohen, S., Jolliffe, T., Mortimore, C. and Robertson, M. Calder, A.J., Lawrence, A.D. and Young, A.W. (2001) Neuro-
(1997b) Another advanced test of theory of mind: evidence psychology of fear and loathing. Nature Rev. Neurosci., 2:
from very high functioning adults with autism or Asperger 352–363.
Syndrome. J. Child Psychol. Psychiat., 38: 813–822. Canli, T., Sivers, H., Whitfield, S.L., Gotlib, I. and Gabrieli, J.
Baron-Cohen, S., Leslie, A.M. and Frith, U. (1985). Does (2002) Amygdala response to happy faces as a function of
the autistic child have a ‘‘theory of mind’’? Cognition, 21: extraversion. Science, 296: 2191.
37–46. Carr, L.M., Iacoboni, M., Dubeau, M.-C., Mazziotta, J.C. and
Baron-Cohen, S., O’Riordan, M., Jones, R., Stone, V. and Lenzi, G.L. (2003) Neural mechanisms of empathy in hu-
Plaisted, K. (1999a) A new test of social sensitivity: detection mans: a relay from neural systems for imitation to limbic
of faux pas in normal children and children with Asperger areas. Proc. Natl. Acad. Sci. USA, 100: 5497–5502.
Syndrome. J. Autism Dev. Disord., 29: 407–418. Castelli, F., Frith, C., Happe, F. and Frith, U. (2002) Autism,
Baron-Cohen, S., Richler, J., Bisarya, D., Gurunathan, N. and Asperger Syndrome and brain mechanisms for the attribu-
Wheelwright, S. (2003) The systemising quotient (SQ): an tion of mental states to animated shapes. Brain, 125:
investigation of adults with Asperger Syndrome or high 1839–1849.
functioning autism and normal sex differences. Philos. Trans. Castelli, F., Happe, F., Frith, U. and Frith, C. (2000) Move-
R. Soc., 358: 361–374. ment and mind: a functional imaging study of perception and
Baron-Cohen, S., Spitz, A. and Cross, P. (1993) Can children interpretation of complex intentional movement patterns.
with autism recognize surprise? Cogn. Emotion,, 7: 507–516. NeuroImage, 12: 314–325.
Baron-Cohen, S. and Swettenham, J. (1996) The relationship Chakrabarthi, B., Bullmore, E.T. and Baron-Cohen, S. (in
between SAM and ToMM: the lock and key hypothesis. In: press) Empathizing with basic emotions: common and dis-
Carruthers, P. and Smith, P. (Eds.), Theories of Theories of crete neural substrates.
Mind. Cambridge University Press, Cambridge. Chakrabarti, B., Kent, L., Suckling, J., Bullmore, E.T. and
Baron-Cohen, S. and Wheelwright, S. (2004) The empathy Baron-Cohen, S. (2006) Variations in human cannabinoid
quotient (EQ): an investigation of adults with Asperger Syn- receptor (CNR1) gene modulate striatal response to happy
drome or high functioning autism, and normal sex differ- faces. Eur. J. Neurosci., 23: 1944–1948.
ences. J. Autism Dev. Disord., 34: 163–175. Charlesworth, W.R. and Dzur, C. (1987) Gender comparisons
Baron-Cohen, S., Wheelwright, S., Hill, J. and Golan, O. (sub- of preschoolers’ behavior and resource utilization in group
mitted) Development of the emotion lexicon. problem-solving. Child Dev., 58: 191–200.
Baron-Cohen, S., Wheelwright, S., Hill, J., Raste, Y. and Charman, T., Ruffman, T. and Clements, W. (2002) Is there a
Plumb, I. (2001) The ‘‘reading the mind in the eyes’’ test gender difference in false belief development. Socl. Dev., 11:
revised version: a study with normal adults, and adults with 1–10.
415

Cleckley, H.M. (1977) The Mask of Sanity: An Attempt to Happe, F. (1994) An advanced test of theory of mind: under-
Clarify Some Issues About The So-called Psychopathic Per- standing of story characters’ thoughts and feelings by able
sonality. Mosby, St Louis. autistic, mentally handicapped, and normal children and
Connellan, J., Baron-Cohen, S., Wheelwright, S., Ba’tki, A. and adults. J. Autism Dev. Disord., 24: 129–154.
Ahluwalia, J. (2001) Sex differences in human neonatal social Happe, F. (1995) The role of age and verbal ability in the theory
perception. Infant Behav. Dev., 23: 113–118. of mind task performance of subjects with autism. Child
Crick, N.R. and Grotpeter, J.K. (1995) Relational aggression, Dev., 66: 843–855.
gender, and social-psychological adjustment. Child Dev., 66: Hare, R.D., Hakstian, T.J., Ralph, A., Forth-Adelle, E., et al.
710–722. (1990) The revised psychopathy checklist: reliability and fac-
Daly, M. and Wilson, M. (1988) Homicide. Aldine de Gruyter, tor structure. Psychol. Assess., 2: 338–341.
New York. Hariri, A.R., Mattay, V.S., Tessitore, A., Kolachana, B., Fera,
Damasio, A.R., Grabowski, T.J., Bechara, A., Damasio, H., F., Goldman, D., Egan, M. and Weinberger, D.R. (2002)
Ponto, L.L.B., Parvizi, J. and Hichwa, R.D. (2000) Subcor- Serotonin transporter genetic variation and the response of
tical and cortical brain activity during the feeling of self- the human amygdala. Science, 297: 400–403.
generated emotions. Nat. Neurosci., 3: 1049–1056. Hatfield, E., Cacioppo, J.T. and Rapson, R.L. (1992) Emo-
Davis, M.H. (1994) Empathy: A Social Psychological Ap- tional contagion. In: Clark, M.S. (Ed.), Review of Person-
proach. Westview Press, CO. ality and Social Psychology: Emotion and Behaviour. Sage
Decety, J. and Chaminade, T. (2003) Neural correlates of feel- Publications, Newbury Park, p. CA.
ing sympathy. Neuropsychologia, 41: 127–138. Helland, S. (2005) Gender differences in facial imitation, Un-
Dimberg, U., Thunberg, M. and Elmehed, K. (2000) Uncon- published thesis, University of Lund, available online at http://
scious facial reactions to emotional facial expressions. Psy- theses.lub.lu.se/archive/sob/psy/psy03022/PSY03022.pdf
chol. Sci., 11: 86–89. Hobson, R.P. (1986) The autistic child’s appraisal of expres-
Dodge, K. (1980) Social cognition and children’s aggressive sions of emotion. J. Child Psychol. Psychiat., 27: 321–342.
behaviour. Child Dev., 51: 162–170. Hobson, R.P. (1993) Autism and The Development of Mind.
Ekman, P. and Friesen, W. (1969) The repertoire of non-verbal Lawrence Erlbaum Associates, NJ.
behavior: categories, origins, usage, and coding. Semiotica, 1: Hoffman, M.L. (1977) Sex differences in empathy and related
49–98. behaviors. Psychol. Bull., 84: 712–722.
Eugene, F., Levesque, J., Mensour, B., Leroux, J.M., Beaudoin, Izard, C. and Ackerman, B. (2000) Motivational, organiza-
G., Bourgouin, P. and Beauregard, M. (2003) The impact of tional, and regulatory functions of discrete emotions. In:
individual differences on the neural circuitry underlying sad- Haviland-Jones, J. and Lewis, M. (Eds.), Handbook of emo-
ness. NeuroImage, 19: 354–364. tions. Guilford Press, New York, pp. 253–264.
Field, T. (1979) Visual and cardiac responses to animate and Jackson, P., Meltzoff, A.N. and Decety, J. (2005) How do we
inanimate faces by term and preterm infants. Child Dev., 50: perceive the pain of others? A window into the neural proc-
188–194. esses involved in empathy. NeuroImage, 24: 771–779.
Fonagy, P., Steele, H., Steele, M. and Holder, J. (1997). At- Kalin, N.H., Shelton, S.E. and Davidson, R.J. (2001) The pri-
tachment and theory of mind: overlapping constructs? ACPP mate amygdala mediates acute fear but not the behavioral
Occasional Papers, 14: 31–40. and physiological components of anxious temperament.
Frith, U. and Frith, C. (2003) Development and neurophysi- Journal of Neuroscience, 21: 2067–2074.
ology of mentalizing. Philos. Trans. R. Soc., 358: 459–473. Knickmeyer, R., Baron-Cohen, S., Raggatt, P. and Taylor, K.
Gallese, V., Keysers, C. and Rizzolatti, G. (2004) A unifying (2006) Foetal testosterone and empathy. Hormones and Be-
view of the basis of social cognition. Trends Cogn. Sci., 8: haviour, 49: 282–292.
396–403. Knickmeyer, R., Baron-Cohen, S., Raggatt, P. and Taylor, K.
Goldenfeld, N., Baron-Cohen, S., Wheelwright, S., Ashwin, C. (2005) Foetal testosterone, social cognition, and restricted
and Chakrabarti, B. (2005) Empathizing and systemizing in interests in children. J. Child Psychol. Psychiat., 45: 1–13.
males and females, and autism: a test of neural competition Knight, G.P., Fabes, R.A. and Higgins, D.A. (1989)
theory. In: Farrow T. and Woodruff P. (Eds.), Empathy and Gender differences in the cooperative, competitive, and in-
Mental Illness. Cambridge University Press, Cambridge. dividualistic social values of children. Motiv. Emotion, 13:
Grosbras, M.-H. and T. Paus. (in press) Brain networks in- 125–141.
volved in viewing angry hands or faces. Cereb. Cortex. Lawrence, A.D., Chakrabati, B. et al. (2004) Looking at happy
Grosbras, M.-H., Laird, A.R. and Paus, T. (2005) Cortical re- and sad faces: an fMRI study. Annual meeting of the Cog-
gions involved in eye movements, shifts of attention and gaze nitive Neuroscience Society, San Diego, USA, Cognitive Ne-
perception. Hum. Brain Mapp., 25: 140–154. uroscience Society.
Haidt, J., McCauley, C. and Rozin, P. (1994) Individual differ- Lawrence, E.J., Shaw, P., Baker, D., Baron-Cohen, S. and
ences in sensitivity to disgust: a scale sampling seven domains David, A.S. (2004) Measuring Empathy — reliability and
of disgust elicitors. Person. Individ. Diff., 16: 701–713. validity of the empathy quotient. Psychol. Med., 34: 911–919.
Hall, J.A. (1978) Gender effects in decoding nonverbal cues. LeDoux, J. (2000) Emotion circuits in the brain. Annual Re-
Psychol. Bull., 85: 845–857. view of Neuroscience, 23: 155–184.
416

Lee, T., Liu, H.-L., Hoosain, R., Liao, W.-T., Wu, C.-T., Yuan, O’Doherty, J., Deichmann, R., Critchley, H.D. and Dolan, R.J.
K., Chan, C., Fox, P. and Gao, J. (2002) Gender differences (2002) Neural responses during anticipation of a primary
in neural correlates of recognition of happy and sad faces in taste reward. Neuron, 33: 815–826.
humans assessed by functional magnetic resonance imaging. Panksepp, J. (1998) Affective neuroscience: the foundations of
Neurosci. Lett., 333: 13–16. human and animal emotions. Oxford University Press, New
Leslie, A.M. (1987) Pretence and representation: The origins of York.
‘‘theory of mind’’. Psychol. Rev., 94: 412–426. Pelphrey, K.A., Singerman, J.D., Allison, T. and McCarthy, G.
Leslie, A. (1995) ToMM, ToBy, and agency: core architecture (2003) Brain activation evoked by perception of gaze shifts:
and domain specificity. In: Hirschfeld, L. and Gelman, S. the influence of context. Neuropsychologia, 41: 156–170.
(Eds.), Domain Specificity in Cognition and Culture. Cam- Perrett, D.I. and Emery, N. (1994) Understanding the inten-
bridge University Press, New York. tions of others from visual signals: neurophysiological evi-
Levenson, R.W. (1996) Biological substrates of empathy and dence. Curr. Psychol. Cogn., 13: 683–694.
facial modulation of emotion: two facets of the scientific leg- Phan, K.L., Wager, T., Taylor, S.F. and Liberzon, I. (2002)
acy of John Lazetta. Motiv. Emotion, 20: 185–204. Functional neuroanatomy of emotion: a meta-analysis of
Lewis, P., Critchley, H., Rotshtein, P. and Dolan, R. (in press) emotion activation studies in PET and fMRI. Neuroimage,
Neural correlates of processing valence and arousal in affec- 16: 331–348.
tive words. Cerebral Cortex. Phillips, W., Baron-Cohen, S. and Rutter, M. (1998) Under-
Liotti, M., Mayberg, H.S., Brannan, S.K., McGinnis, S., standing intention in normal development and in autism. Br.
Jerabek, P. and Fox, P.T. (2000) Differential limbic–cortical J. Dev. Psychol., 16: 337–348.
correlates of sadness and anxiety in healthy subjects: impli- Phillips, M.L., Young, A.W., Scott, S., Calder, A.J., Andrew,
cations for affective disorders. Biol. Psychiat., 48: 30–42. C., Giampietro, V., Williams, S., Bullmore, E.T., Brammer,
Maccoby, E. (1999) The Two Sexes: Growing Up Apart, Com- M.J. and Gray, J. (1998) Neural responses to facial and vocal
ing Together. Harvard University Press, Cambridge, USA. expressions of fear and disgust. Proceedings of the Royal
Malhi, G., Lagopoulos, J., Ward, P., Kumari, V., Mitchell, D., Society London B, 265: 1809–1817.
Parker, G., Ivanovski, B. and Sachdev, P. (2004) Cognitive Phillips, M., Young, A., Senior, C., Brammer, M., Andrew, C.,
generation of affect in bipolar depression: an fMRI study. Calder, A., Bullmore, E., Perrett, D., Rowland, D., Williams,
Eur. J. Neurosci., 19: 741–754. S., Gray, J. and David, A. (1997) A specific neural substrate for
Mayberg, H.S., Liotti, M., Brannan, S.K., McGinnis, S., Mahurin, perceiving facial expressions of disgust. Nature, 389: 495–498.
R.K., Jerabek, P.A., Silva, J.A., Tekell, J.L., Martin, C.C., Lan- Piaget, J. and Inhelder, B. (1956) The Child’s Conception of
caster, J.L. and Fox, P.T. (1999) Reciprocal limbic-cortical Space. Routledge and Kegan Paul, London.
function and negative mood: converging PET findings in Power, T.G. (1985) Mother- and father-infant play: a develop-
depression and normal sadness. Am. J. Psychiat., 156: 675–682. mental analysis. Child Dev., 56: 1514–1524.
Mealey, L. (1995) The sociobiology of sociopathy: an inte- Prather, M., Lavenex, P., Mauldin-Jourdain, M., Mason, W.,
grated evolutionary model. Behav. Brain Sci., 18: 523–599. Capitanio, J., Mendoza, S. and Amaral, D. (2001) Increased
Meltzoff, A.N. and Decety, J. (2003) What imitation tells us social fear and decreased fear of objects in monkeys with
about social cognition: a rapproachement between develop- neonatal amygdala lesions. Neuroscience, 106: 653–658.
mental psychology and cognitive neuroscience. Philos. Trans. Pratt, C. and Bryant, P. (1990) Young children understand that
R. Soc., 358: 491–500. looking leads to knowing (so long as they are looking into a
Mobbs, D., Greicius, M.D., Abdel-Azim, E., Menon, V. and single barrel). Child Dev., 61: 973–983.
Reiss, A.L. (2003) Humor modulates the mesolimbic reward Premack, D. (1990) The infant’s theory of self-propelled ob-
centres. Neuron, 40: 1041–1048. jects. Cognition, 36: 1–16.
Moll, J., de Oliveira-Souza, R., Eslinger, P., Bramati, I., Mo- Preston, S.D. and de Waal, F.B.M. (2002) Empathy: its ulti-
urao-Miranda, J., Andreiuolo, P. and Pessoa, L. (2002) The mate and proximate bases. Behav. Brain Sci., 25: 1–72.
neural correlates of moral sensitivity: A functional magnetic Rahm, B., Opwis, K., Kaller, C., Spreer, J., Schwarzvald, R., et
resonance imaging investigation of basic and moral emo- al. (2006) Tracking the subprocesses of decision-based action
tions. J. Neurosci., 22: 2730–2736. in the human frontal lobes. Neuroimage, 30: 656–667.
Morris, J., Ohman, A. and Dolan, R.J. (1999) A subcortical Rolls, E.T. (2002). Neural basis of emotions. In: Smelsner N.
pathway to the right amygdala mediating ‘unseen’ fear. and Baltes, P. (Eds.), International Encyclopedia of the So-
PNAS, 96: 1680–1685. cial and Behavioral Sciences. Elsevier, Amsterdam.
Murphy, F.C., Nimmo-Smith, I. and Lawrence, A.D. (2003) Saxe, R., Carey, S. and Kanwisher, N. (2004) Understanding
Functional neuroanatomy of emotions: a meta-analysis. other minds: linking developmental psychology and func-
Cogn. Affect. Behav. Neurosci., 3: 207–233. tional neuroimaging. Annu. Rev. Psychol., 55: 87–124.
Nichols, S. (2001) Mindreading and the cognitive architecture Scaife, M. and Bruner, J. (1975) The capacity for joint visual
underlying altruistic motivation. Mind Lang., 16: 425–455. attention in the infant. Nature, 253: 265–266.
O’Doherty, J. (2004) Reward representations and reward-re- Schienle, A., Schafer, A., Stark, R., Walter, B. and Vaitl, D. (2005)
lated learning in the human brain: insights from neuroimag- Gender differences in the processing of disgust- and fear-in-
ing. Current Opinion in Neurobiology, 14: 776–796. ducing pictures: an fMRI study. Neuroreport, 16: 277–280.
417

Schroeder, U., Hennenlotter, A., Erhard, P., Haslinger, B., Trevarthen, C., 1989. The relation of autism to normal socio-
Stahl, R., Lange, K. and Ceballos-Baumann, A. (2004) cultural development: the case for a primary disorder in reg-
Functional neuroanatomy of perceiving suprised faces. Hum. ulation of cognitive growth by emotions. In: Lelord, G., Muk,
Brain Mapp., 23: 181–187. J., Petit, M. (Eds.), Autisme er troubles du developpment glo-
Shumake, J., Edwards, E. and Gonzalez-Lima, F. (2001) bal de l’enfant. Expansion Scientifique Francaise, Paris.
Hypermetabolism of paraventricular hypothalamus in the Vogt, B.A. (2005) Pain and emotion interactions in subregions
congenitally helpless rat. Neurosci. Lett., 311: 45–48. of the cingulate gyrus. Nat. Rev. Neurosci., 6: 533–544.
Singer, T., Seymour, B., O’Doherty, J., Kaube, H., Dolan, R.J. Wager, T., Phan, K.L., Liberzon, I. and Taylor, S. (2003) Va-
and Frith, C.D. (2004) Empathy for pain involves the affec- lence gender and lateralization of functional brain anatomy:
tive but not sensory components of pain. Science, 303: A meta-analysis of findings from neuroimaging. NeuroIm-
1157–1167. age, 19: 513–531.
Skuse, D.H., James, R.S., Bishop, D.V.M., Coppins, B., Dal- Walker, A.S. (1982) Intermodal perception of exptessive be-
ton, P., Aamodt-Leeper, G., Bacarese-Hamilton, M., Cre- haviours by human infants. J. Exp. Child Psychol., 33:
swell, C., McGurk, R. and Jacobs, P.A. (1997) Evidence from 514–535.
Turner’s syndrome of the imprinted x-linked locus affecting Wellman, H. (1990) Children’s Theories of Mind. Bradford/
cognitive function. Nature, 287: 705–708. MIT Press, Cambridge MA.
Small, D., Gregory, M., Mak, Y., Gitelman, D., Mesulam, M. Whiten, A. (1991) Natural Theories of Mind. Basil Blackwell,
and Parrish, T. (2003) Dissociation of neural representation Oxford.
of intensity and affective valuation in human gustation. Wicker, B., Perrett, D.I., Baron-Cohen, S. and Decety, J. (2003)
Neuron, 39: 701–711. Being the target of another’s emotion: a PET study. Ne-
Smith, P.M. (1985) Language, The Sexes and Society. Basil uropsychologia, 41: 139–146.
Blackwell, Oxford. Williams, J.H.G., Waiter, G.D., Perra, O., Perrett, D.I. and
Stein, E. (1989) On The Problem of Empathy. ICS Publications, Whiten, A. (2005) An fMRI study of joint attention expe-
Washington DC. rience. NeuroImage, 25: 133–140.
Strayer, F.F. (1980) Child ethology and the study of preschool Wimmer, H., Hogrefe, J. and Perner, J. (1988) Children’s un-
soical relations. In: Foot, H.C., Chapman, A.J. and Smith, derstanding of informational access as a source of knowl-
J.R. (Eds.), Friendship and Social Relations in Children. edge. Child Dev., 59: 386–396.
Wiley, New York. Wimmer, H. and Perner, J. (1983) Beliefs about beliefs: repre-
Takahashi, H., Yahata, N., Koeda, M., Matsuda, T., Asai, K. sentation and constraining function of wrong beliefs in young
and Okubo, Y. (2004) Brain activation associated with eval- children’s understanding of deception. Cognition, 13:
uative process of guilt and embarrassment: An fMRI study. 103–128.
NeuroImage, 23: 967–974. Yirmiya, N., Kasari, C., Sigman, M. and Mundy, P. (1990)
Tannen, D. (1991) You Just Don’t Understand: Women and Facial expressions of affect in autistic, mentally retarded, and
Men in Conversation. Virago, London. normal children. J. Child Psychol. Psychiat.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 23

The multiple facets of empathy: a survey of theory


and evidence

Susanne Leiberg1,2, and Silke Anders1,3

1
Institute of Medical Psychology and Behavioral Neurobiology, University of Tübingen, Tübingen, Germany
2
Department of Psychiatry and Psychotherapy, RWTH Aachen University, Aachen, Germany
3
Section for Experimental MR of the CNS, University of Tübingen, Tübingen, Germany

Abstract: Empathy is the ability to perceive and understand other people’s emotions and to react appro-
priately. This ability is a necessary prerequisite for successful interpersonal interaction. Empathy is a
multifaceted construct including low-level mechanisms like emotional contagion as well as high-level
processes like perspective-taking. The ability to empathize varies between individuals and is considered a
stable personality trait: some people are generally more successful in empathizing than others. In this
chapter we will first present different conceptualizations of the construct of empathy, and refer to empathy-
regulating processes as well as to the relationship between empathy and social behavior. Then, we will
review peripheral physiological and brain imaging studies pertaining to low- and high-level empathic
processes, empathy-modulating processes, and the link between empathy and social behavior. Further, we
will present evidence regarding interindividual differences in these processes as an important source of
information for solving the conundrum of how the comprehension of others’ emotions is achieved by our
brains.

Keywords: empathy; emotional contagion; simulation theory; perspective-taking; theory of mind; emotion
regulation; neuroimaging; psychophysiology

Introduction interpersonal interactions. In most general terms,


empathy refers to the ability to accurately perceive
As humans, we live in a highly complex social and understand another person’s emotions and to
environment. Not only do we interact with our react appropriately. How this capability of un-
closest kin, but maintain a complicated social derstanding others’ emotions can be conceptual-
network with friends, colleagues, acquaintances, ized and how it is influenced by psychological and
and the like. To achieve our goals in daily life, we social factors has been debated since the begin-
have to deal with complete strangers or, worse, ning of the last century. With the advent of mod-
with people we would rather choose to avoid. ern neuroimaging methods, interest in empathy
Empathy — the capability to understand other has been spurred, which has produced a consid-
people’s emotions — is crucial for comprehending erable number of studies shedding light on the
intentions and behaviors as well as for adapting neural bases of empathy.
our own behavior in order to achieve smooth This review is divided into two sections: The first
part pertains to the theoretical background and
Corresponding author. Tel.: + 49-241-8088550; reviews different conceptualizations of empathy. It
Fax: +49-241-8082401; E-mail: sleiberg@ukaachen.de will be shown that most researchers view empathy

DOI: 10.1016/S0079-6123(06)56023-6 419


420

as a multifaceted construct involving low-level ence (Decety and Jackson, 2004; Blair, 2005). The-
affective processes, like emotional contagion, odor Lipps (1903) used the German word
sometimes regarded as automatic processes, and ‘‘einfühlen,’’ meaning ‘‘to feel into,’’ to explain
high-level cognitive processes, like perspective- how a person comes to know about the inner state
taking. The second part illustrates how peripheral of others. He believed that we achieve an under-
physiological and neuroimaging studies contrib- standing of another person by internally imitating
uted to the understanding of empathy. We will the other’s gestures and actions (an idea that later
review studies related to contagion-like processes came to be called ‘‘motor mimicry’’). Later on,
and perspective-taking. Additionally, we will ded- ‘‘einfühlen’’ was translated into ‘‘empathy’’ by
icate one paragraph to individual differences in Titchener (1909). Since the time of Lipps and
these processes, since we believe that the analysis Titchener, many definitions of empathy have been
of individual differences is useful for elucidating put forward (see Table 1), but still today a concise,
the neurobiology of empathy. Finally, studies will agreed-upon definition of empathy is missing. The
be reviewed that have investigated processes that common denominator of the definitions listed in
may modulate empathic responses and studies that Table 1 is the compassionate responding to another
have addressed the relationship between empathy person’s emotions (perhaps excluding Ickes’ (1997)
and social behavior. definition). Disagreement subsists regarding the
implementation of empathy and whether it occurs
in a contagion-like fashion or depends on higher
Theoretical background level cognitive processes. ‘‘Contagion-like’’ refers to
‘‘the tendency to automatically mimic and synchro-
Empathy has been the subject of investigation in nize facial expressions, vocalizations, postures, and
many different research areas, including personality movements with those of another person and, con-
psychology (Allport, 1961), psychotherapy research sequently, to converge emotionally’’ (Hatfield et al.,
(Rogers, 1975), social psychology (Batson and 1994). In contrast, higher level cognitive processes
Coke, 1981), developmental psychology (Eisenberg in empathy refer to actively taking the perspective
and Strayer, 1987) and, recently, social neurosci- of another person.

Table 1. Definitions of the term ‘‘empathy’’

Group Definition

Ax (1964) An autonomic nervous system state, which tends to simulate that of another person
Batson et al. (1987) Other-oriented feelings of concern, compassion, and tenderness experienced as a result of witnessing
another person’s suffering
Eisenberg (2000) Affective response that stems from the apprehension or comprehension of another’s emotional state
or condition and is similar to what the other person is feeling or would be expected to feel
Hoffman (1984) Affective response more appropriate to someone else’s situation than to ones own
Ickes (1997) Complex form of psychological inference in which observation, memory, knowledge, and reasoning
are combined to yield insights into the thoughts and feelings of others
Preston and de Waal (2002) Any process where the attended perception of the object’s state generates a state in the subject that
is more applicable to the object’s state or situation than to the subject’s own prior state or situation
Rogers (1959) To perceive the internal frame of reference of another with accuracy and with the emotional
components and meanings, which pertain thereto as if one were the person, but without ever losing
the ‘‘as if’’ condition
Webster’s Dictionary (1971) The capacity for participating in, or a vicarious experiencing of another’s feelings, volitions, or ideas
and sometimes another’s movements to the point of executing bodily movements resembling his
Wispé (1986) Attempt by one self-aware self to comprehend unjudgmentally the positive and negative experiences
of another self
421

Simulation theory and Theory– theory in this debate still differ in the importance they
concede to either approach.
Researchers in line with Lipps stress the impor-
tance of contagion-like imitative processes, en-
dorsing a simulationist account of empathy, Contagion-like processes
whereas others emphasize the contribution of
higher level cognitive processes favoring a the- On the basis of the ideomotor theory (Prinz, 1987;
ory–theory (TT) account of empathy. Simulation Hommel et al., 2001), which assumes that an ac-
theory (ST) posits that we understand other tion is stored as a sensory feedback representation
people’s mental states, including their emotions, in our brains, Preston and de Waal (2002) pro-
by reproducing or simulating the other’s mental posed a perception-action model of empathy, rem-
state in our own minds. While traditional STs iniscent of Lipps’ internal imitation. This
assume different simulation mechanisms varying representation is activated when observing some-
in their degree of automaticity (Gordon, 1986; body perform the action, and will in turn prime the
Goldman, 2005, 2006), others appear to reserve activation of the corresponding motor representa-
the term for an automatic process (Gallese et al., tion in the observer because of their overlap. Pres-
2004). TT, on the other hand, holds that we use a ton and de Waal transferred this idea to the
lay theory of psychology to attribute mental states concept of empathy. They surmise that the obser-
(such as beliefs, intentions, desires, or emotions) vation of a person in an emotional state automat-
to others, thereby understanding their behavior. ically activates a representation of that emotional
TT is closely related, if not equivalent, to the state, including its associated autonomic, somatic,
concept of theory of mind (ToM). It is argued that and motor responses in the observer. The greater
an organism possesses a ToM if it (a) imputes the overlap between the emotional state we ob-
mental states to itself and others, (b) recognizes serve and our representation of this emotional
that others have mental states that can differ from state, the greater the activation of this emotional
its own, and (c) is aware that the behavior of other state in ourselves. This model is in accordance with
persons can be explained by their mental states other recent theories, which emphasize the impor-
(Premack and Woodruff, 1978; Frith and Frith, tance of the activation of representations for re-
2003). In the earlier literature, the term ToM enactment of associated feelings like the influential
appears to refer to the knowledge of the fact that somatic marker hypothesis (Damasio, 1994) and
others have a mind with mental states possibly conceptualizations of ST in terms of neural mirror
differing from one’s own (Premack and Woodruff, networks (Gallese and Goldman, 1998; Gallese et
1978), whereas in recent years it has come to al., 2004).
denote the ability or process of inferring others’ STs as outlined so far can account for low-level
mental states, also called ‘‘mentalizing’’ or ‘‘min- contagion-like empathic responses to clear dis-
dreading’’ (Gallagher and Frith, 2003). plays of emotions of similar beings in unequivocal
A dichotomy has been (artificially) established situations. However, we are not only able to em-
between those researchers favoring ST (Gallese et pathize with someone displaying a familiar emo-
al., 2004) and those who support TT (Saxe, 2005a). tional expression or with someone in a well-known
Following this discussion, one gets the impression emotional situation, but are capable to understand
that both sides talk at cross-purposes although the emotions of others in diverse situations. And
neither of them holds that all aspects of under- our empathic reactions to the very same emotional
standing others can be sufficiently explained by cues vary markedly depending on the situational
only one of the theories (Gordon, 2005; Goldman context and our relationship with the other person.
and Sebanz, 2005; Mitchell, 2005; Saxe, 2005b, For instance, our empathic reaction to a person
2005c). Instead, a hybrid model of ST and TT is expressing fear will be much stronger when this
advocated for explaining our ability to understand person is actually being attacked than when this
others’ minds. However, the researchers involved person’s fear is due to experiencing the thrills of a
422

haunted house. Furthermore, we can understand trol. These functions enable us to infer emotional
others’ emotions even if they do not explicitly states from more symbolic information, to inte-
show them. Knowing your friend, you will prob- grate expressive and other cues, to differentiate
ably know that she is sad after she had failed her between our own emotions and those of another
exam, although she does not have a sad facial ex- person, and to inhibit prepotent responses.
pression, voice, or posture. Empathic reactions can
be triggered by propositional information like ver-
bal and written accounts of emotions, or com- Perspective-taking
pletely internally, by thoughts or imagination.
Finally, and probably most importantly, we can A higher level cognitive process by which under-
derive the inner state of a person when our own standing others’ emotions can be achieved is
situation is completely different and even when we perspective-taking. Perspective-taking incorporates
have never faced a similar situation — situations integrating information from different sources
likely compromising successful simulation. Imag- when inferring the other’s mental state and inhib-
ine, for example, how reading the letter of your iting one’s own perspective if necessary. A role of
sister in which she describes how she had to resign perspective-taking in empathy (Feshbach, 1978;
in her last horse jumping competition: you will be Davis, 1996; Batson et al., 2003) and understand-
more likely to share her feelings when you are an ing others’ minds in general (Gordon, 1986;
ambitious sportsman who eagerly participates in Goldman, 2006) has been acknowledged. Two
competitions, but even if you hate sports and fear different forms of perspective-taking have been
horses, you might have an idea what your sister distinguished (Stotland, 1969; Batson et al., 1997a;
felt. Davis et al., 2004): a self-focused form, in which
How can low-level simulation models accom- one imagines how oneself would feel in the target’s
modate these instances of empathizing? We argue situation (first-person perspective-taking), and an
that a model describing our ability to empathize other-focused form, in which one imagines what
cannot stop at simple contagion-like simulation the target is thinking and feeling (third-person
mechanisms but has to add elaborate, cognitive perspective-taking). While both types of perspec-
processes, which can account for empathic re- tive-taking have been shown to generate empathic
sponding in the mentioned examples. Baron-Co- responses, only the former induces distress in the
hen and colleagues (Baron-Cohen, 2005; see observer (Stotland, 1969; Batson et al., 1997a).
Chakrabarti and Baron-Cohen, this volume) have Interestingly, the term perspective-taking is used
proposed a model of empathy comprising several by both, traditional simulation theorists and the-
lower and higher level mechanisms. The compo- ory theorists. In traditional ST, perspective-taking
nents of this model are thought to develop at is a second simulation process, which differs from
different time points in a human’s life. Lower level contagion-like simulation in automaticity and level
mechanisms develop early and constitute the affec- of awareness. In comparison to the contagion-like
tive component of empathy in that they are emo- simulation processes where the observation of an
tional contagion-like processes. Higher level emotional state in the target leads instantaneously
mechanisms, like ToM and the so-called Empa- to the activation of the representation of that
thizing System, develop later on. Preston and de emotional state in the observer, perspective-taking
Waal (2002) recognize the involvement of higher requires the observer to deliberately activate a
level cognitive processes but do not specify exactly representation of the target’s state and to integrate
when these processes are employed and how they context information and knowledge about the
interact with the more basic simulation processes. other’s beliefs, desires, and preferences. Theory
These authors state that their development is theorists have used the term perspective-taking to
closely related to the size of prefrontal cortex describe how one makes inferences about another
(PFC), which accommodates functions like work- person’s inner state using theoretical knowledge
ing memory, goal assessment, and inhibitory con- about the other’s situation.
423

As perspective-taking incorporates the integra- match situational demands (Gross, 1999, 2002). In
tion of information about the person and the con- empathy, two types of regulatory processes can be
text it can improve the accuracy of simulation of or distinguished: one that controls one’s own emo-
the predictions about the other’s emotional state. tions (self-focused) and one that controls the level
By knowing that your friend’s goal is to graduate of engagement with the other person (other-fo-
from university you can understand that she is sad cused) (Reik, 1949; Eisenberg, 2000). The former
failing the exam, although her demeanor may not might be similar to regulatory processes employed
indicate this. In the same line, when you know that to modulate our emotional reactions to emotional
your sister loves the feeling of taking one seemingly stimuli in a nonsocial context.
insurmountable hurdle after the other, you can
understand that she was disappointed having to
resign from the race. Nevertheless, perspective- Empathy and social behavior
taking does not necessarily lead to a correct
representation of the other’s mind and may thus, Empathy is a prerequisite for prosocial behavior
like contagion-like processes, entail mistakes in (Eisenberg and Fabes, 1991; Trobst et al., 1994;
inferring the other’s emotion. Batson et al., 1997b; Batson, 1998). Only if we
understand that our friend is sad about failing the
exam we will make an effort to console her. More-
Modulatory processes over, it is believed that the ability to regulate one’s
emotions or to disengage oneself from the other
To meet situational demands but also to insure our person’s emotional state also figures prominently in
personal well-being, it is sometimes indispensable prosocial behavior (Batson, 1991; Eisenberg, 2000).
to modulate our empathic responding. In some According to Eisenberg (Eisenberg and Strayer,
instances, our intentions or goals require enhance- 1987; Eisenberg, 2000), empathy can lead either to
ment of empathic responses. As a psychotherapist, an other-oriented response or to personal distress,
for example, one is often confronted with people depending on how well someone can regulate
that appraise and react to situations differently one’s own emotion and the engagement with the
than oneself. In this and other settings, where con- observed person. If we become too distressed by
tagion-like processes are likely to be compromised, empathizing with another person and are not ca-
perspective-taking can substitute or amplify pable of regulating our empathic response, we will
contagion-like processes and thus help to better rather try to alleviate our own distress than attend
understand the other’s emotions and to intensify to the other person. Interestingly, the psychoana-
our empathic response. In other instances, it is vital lyst (Reik, 1949) subsumed both processes —
for our personal well-being to control contagion- making the other’s experience one’s own (incorpo-
like processes (Bandura, 1997). In our daily life we ration) and moving back from the merged inner
are often confronted with negative emotions, dis- relationship to a position of separate identity
tress, and pain of other people. When we watch TV (detachment) — under the concept of empathy. It
or go on the Internet, we are flooded with reports should be stressed, though, that the ability to
of ongoing wars and other inhumanities. If conta- empathize with, and to distance oneself from a
gion-like simulation processes were automatic in a person in distress is not a sufficient condition for
sense that they could not be controlled, all these prosocial behavior to arise. First, other factors, like
instances would lead to an internal reproduction of the perceived urgency and the observer’s perceived
stressful emotional states. One simple way to efficacy of helping, influence the likelihood of help-
prevent empathic responding is to divert attention ing behavior. As Preston and de Waal (2002) have
away from the aversive situation. Other ways to suggested, the occurrence of prosocial behavior will
inhibit unwanted vicarious emotional reactions are be the outcome of ‘‘a complex cost/benefit analysis
emotion-regulation processes, i.e., processes that on the perceived effectiveness of helping and the
aim to modify emotional responses in order to effect of helping on short and long-term goals.’’
424

Second, in many real-life social situations, other are far from being understood, consensual opinions
people will be present, which may either foster or on some aspects of empathy are emerging.
inhibit prosocial behavior (see, e.g., the phenome- Empathy is a multifaceted construct incorporating
non of responsibility diffusion, Darley and Latane, contagion-like processes subserved by a perception-
1968). Finally, understanding another person’s representation coupling and higher level proc-
emotions can as well lead to betrayal or deception esses like perspective-taking. Perspective-taking
of that person. can modulate or substitute contagion-like proc-
Summarizing these theoretical accounts of esses. Furthermore, empathic responses can be
empathy, we conclude that although an agreed- subject to modulation. Empathy and the ability to
upon definition of empathy is still missing and the modulate empathic responses are essential for social
neural underpinnings of empathy-related processes behavior. Fig. 1 depicts a schematic overview of

Fig. 1. Schematic overview of how understanding of others’ emotional states and an appropriate response toward the other person
may evolve. Numbers on the arrows signify studies, which have addressed the process marked by the arrow. The numbers correspond
to the following studies: 1Lundquist and Dimberg (1995), 2Dimberg et al. (2000), 3Vaughan and Lanzetta (1980), 4Dimberg (1982),
5
Wiesenfeld et al. (1984), 6McHugo et al. (1985), 7Levenson and Ruef (1992), 8Sonnby-Borgström (2002), 9Carr et al. (2003), 10Wicker
et al. (2003), 11Leslie et al. (2004), 12Lanzetta and Englis (1989), 13Singer et al. (2006), 14Dimberg (1988), 15Stotland (1969), 16Hynes et
al. (2006), 17Völlm et al. (2006), 18Shamay-Tsoory et al. (2005a), 19Shamay-Tsoory et al. (2005b), 20Beauregard et al. (2001), 21Critchley
et al. (2002), 22Ochsner et al. (2002), 23Levesque et al. (2003), 24Ochsner et al. (2004b), 25Phan et al. (2005), 26Eippert et al. (in press),
27
Eisenberg and Fabes (1991), 28Trobst et al. (1994), 29Batson et al. (1997b) and 30Batson (1998).
425

how these processes could work together in under- imitation of emotional facial expressions results in
standing other’s emotion and the generation of an corresponding subjective emotional experiences
other-oriented response. (Dimberg, 1988; Lundquist and Dimberg, 1995),
others have failed to find a consistent relationship
between motor imitation and induced emotions
Evidence from peripheral physiological and
(Gump and Kulik, 1997; Hess and Blairy, 2001).
neuroimaging studies
Evidence favoring a facilitative effect of motor
imitation on emotion understanding includes the
Contagion-like processes
finding that subjects prevented from imitating by
putting a pen between their lips detect emotional
As noted in the introduction, emotional contagion
changes in facial expressions significantly later than
refers to the unintentional imitation of an observed
subjects free to imitate (Niedenthal et al., 2001).
person’s facial expression, tone of vocalization,
Furthermore, the anecdotic finding that facial re-
posture, and movements, which consequently leads
semblance between long-term spouses, ascribed to
to an emotional state in the observer similar to the
repeated facial imitation through the course of the
one of the target (Hatfield et al., 1994). Emotional
relationship, is related to marital happiness (Zajonc
contagion, in turn, will facilitate the understanding
et al., 1987), supports the notion that facial imita-
of the other’s emotion (Lipps, 1903; Hoffman,
tion plays a role in understanding others’ emotions.
1984). In this section we will present evidence
On the contrary, in an electromyography (EMG)
for emotional contagion on different levels of
study by Hess and Blairy (2001) where subjects
emotional reactions, including motor as well as
viewed video clips and were asked to judge their
autonomic nervous system (ANS) responses and
own and the displayed emotions, no relation was
brain activity.
found between facial imitation and induced emo-
tions or judgment accuracy. In another study
Contagion-like motor responses (McHugo et al., 1985), subjects viewed videotapes
More than a hundred years ago, Lipps (1903) of the former president of the U.S.A. Ronald
assumed that we involuntarily imitate others’ facial Reagan showing emotional facial expressions.
expressions to understand their feelings. Less than Subjects exhibited congruent EMG responses to
30 years ago, first proof for this hypothesis arose Reagan’s positive and negative facial expressions
from electromyographic studies showing that look- regardless of whether they approved of him or not.
ing at different emotional facial expressions evokes Interestingly, though, self-reported feelings varied
corresponding differential changes in facial muscle with liking. Subjects in favor of Reagan reported
activity in the observer (Vaughan and Lanzetta, emotions similar to the displayed facial expressions
1980; Dimberg, 1982; McHugo et al., 1985). When while subjects opposed to him reported negative
subjects are exposed to pictures of angry and happy feelings to all expressions. These latter results ques-
facial expressions, they react with increased activity tion the assumed connection between facial imita-
in the corrugator supercilii muscle and the zygo- tion, emotional experience, and recognition of
maticus major muscle, respectively (Dimberg, 1982, emotional facial expressions (Blairy et al., 1999;
1988). Notably, these facial reactions occur as early Hess and Blairy, 2001). Thus, it has been suggested
as 300 ms after stimulus onset and are also observed that facial imitation has a more communicative
after subliminal stimulus presentation, ruling out function and serves to signal others that we com-
that subjects voluntarily imitate the presented prehend their emotional state (Bavelas et al., 1986;
facial expressions (Lundquist and Dimberg, 1995; Hess et al., 1999). This hypothesis has been
Dimberg et al., 2000). corroborated by the finding that observers’ motor
The existence of motor imitation is well proven, imitation increased with the availability of eye
but its relation to emotional experience and emo- contact with the target. When eye contact with the
tion understanding is still under debate. While target was not achieved, subjects did not imitate
some studies have demonstrated that involuntary the target’s facial expression (Bavelas et al., 1986).
426

Furthermore it was shown that subjects whose videotape (Gottman and Levenson, 1985). Subse-
behavior was imitated perceive an interaction as quently, Levenson and Ruef (1992) demonstrated
smoother and the partner as more likable than that this ‘‘physiological linkage’’ is important for
subjects whose behavior was not imitated the understanding of others’ emotions. In this study,
(Chartrand and Bargh, 1999). subjects watching video segments of marital inter-
Further dispute concerns the question whether actions were instructed to rate the feelings of one of
facial imitation occurs automatically in the sense the spouses. The greater the synchrony between the
that it does not depend on attention or intention. subject’s and the target’s changes in heart rate, the
While Dimberg et al. (2000) demonstrated that greater was the accuracy with which the subject
facial imitation occurs rapidly and is elicited by rated the target’s feelings (Levenson and Ruef,
stimuli the subject is not aware of, Lanzetta and 1992). Levenson states that ‘‘synchrony of the
Englis (1989) found a striking influence of the autonomic responses of two people is associated
context on facial imitation in social interactions. with heightened emotional closeness and greater
Depending on whether subjects were engaged in a capacity for empathic accuracy’’ (Levenson, 1996).
cooperative or competitive interaction, the ob- Lanzetta and Englis (1989) showed that ANS
server’s EMG responses were either congruent or activity in response to a target’s facial expression
incongruent with the target’s facial expressions; is, like facial activity, modulated by context.
Lanzetta and colleagues termed this behavior Subjects expecting cooperation with the target
empathy and counterempathy, respectively. Since showed increased autonomic arousal, indicated by
the EMG was sampled at a very low rate in this elevated skin conductance responses (SCRs), when
study (1 Hz), it is impossible to decide at what the target’s facial expression signaled distress,
point in time these context-dependent responses while they evinced decreased arousal in response
occurred. Studies investigating whether different to displays of pleasure. Subjects expecting compe-
EMG response components exist, namely a rapid tition with the target, in contrast, showed in-
contagion-like response component and a later creased arousal to pleasure displays and relaxation
component that is influenced by contextual to distress displays.
factors, are lacking.

Contagion-like brain responses


Contagion-like autonomic responses The discovery of mirror neurons (Gallese et al.,
A considerable body of research indicates that emo- 1996; Rizzolatti et al., 1996) revealed a possible
tional contagion is also found in ANS activity. The neural implementation of contagion-like empathic
linkage of the peripheral physiological responses processes. Originally, the term mirror neurons was
between two people during communication was first introduced by Rizzolatti and colleagues (Gallese
demonstrated in studies of psychotherapist–client et al., 1996; Rizzolatti et al., 1996) to describe neu-
interaction. It was observed that heart rates of rons in area F5 of the monkey premotor cortex that
psychotherapists and clients changed synchronously are activated during the execution as well as the
during therapy sessions (DiMascio et al., 1955, observation of goal-directed actions (Iacoboni
1957; Coleman et al., 1956; Stanek et al., 1973). On et al., 1999; Rizzolatti et al., 2001; Keysers and
the basis of these findings, Kaplan and Bloom Perrett, 2004). Mirror neurons with similar proper-
(1960) hypothesized a physiological component of ties are thought to exist in the inferior gyrus of the
empathy. The significance of shared physiology in human PFC (Brodmann area [BA] 44). This area
empathy was later stressed by Levenson (Levenson has been seen activated in neuroimaging studies
and Ruef, 1992; Levenson, 2003). His first evidence during the execution and observation of goal-
for contagion-like autonomic responses was the directed behavior (Iacoboni et al., 1999;
finding that subjects experienced the same physio- Molnar-Szakacs et al., 2005). Notably, this mirror
logical change when participating in a conversation neuron network is also activated during first-person
and when later watching this conversation on and third-person motor imagery (Fourkas et al.,
427

2005). However, neural networks with mirror prop- representations of corresponding emotions and
erties seem to be neither restricted to goal-directed thereby generate empathy. Corroborating evidence
behavior nor confined to the PFC. Evidence is for a mirror neuron network for facial expressions
accumulating for the existence of mirror networks comes from a study by Leslie et al. (2004) showing
for facial expressions, sensations, and emotions. an overlap of activation for deliberate imitation and
Recordings of single-neuron activity from the viewing emotional facial expressions in the right
human anterior cingulate cortex (ACC) revealed premotor cortex, the fusiform gyrus, and the right
increased activity during both reception and obser- superior temporal cortex. Both studies found over-
vation of painful stimulation (Hutchison et al., lapping activity in the right premotor and the
1999). This evidence for a mirror neuron network superior temporal cortex, pointing to a likely in-
for pain was supported by functional magnetic volvement of these areas in a mirror network of
resonance imaging (fMRI) studies, which found in- emotional facial expressions. Since these studies
creased activity in the ACC when subjects received did not directly induce emotions in the subjects
a painful stimulus, observed a signal indicating that (although emotions might have been induced indi-
a loved one will receive a painful stimulus (Singer rectly by producing and viewing facial expressions),
et al., 2004), directly observed a person getting a they do not conclude that the areas found are part
painful stimulus (Morrison et al., 2004; Jackson of a shared network for affect. Wicker et al. (2003)
et al., 2005), or observed facial expressions of pain were the first to investigate overlapping neural ac-
(Botvinick et al., 2005). Some of these studies iden- tivity during the actual experience and observation
tified the anterior insula as another component of of an emotion. They elicited disgust by having their
this circuit for observing and receiving pain (Singer subjects inhale repulsive odorants. In the viewing
et al., 2004; Jackson et al., 2005; Botvinick et al., condition, subjects observed someone inhaling
2005). The insula as well as the ACC has been im- something from a jar and subsequently showing a
plicated in mediating the affective component of facial expression of disgust. Overlapping activity
pain (Peyron et al., 2000). Contagion-like responses was found in the anterior insula and the ACC.
to another person’s pain do not seem to be re- While this is the first evidence of common neural
stricted to affective networks. A recent transcranial activity during experience and observation of an
magnetic stimulation (TMS) study (Avenanti et al., emotion, it has to be shown whether the involve-
2005) has shown that motor excitability is altered ment of the insula is restricted to the experience and
during observation of painful stimulation. Observ- recognition of disgust, as suggested by lesion studies
ing pinpricks to another person’s hand led to a (Calder et al., 2000; Adolphs et al., 2003).
reduction of motor evoked potentials to TMS over Although these studies suggest that the obser-
the motor cortex. vation of emotionally aroused people does lead to
The existence of mirror systems in emotion has an activation of emotional mirror networks in a
been postulated by many researchers (Adolphs, contagion-like manner, it has to be noted that in
1999; Gallese, 2003a, b). They suggested that mirror all neuroimaging studies discussed so far subjects
networks constitute the neural base of contagion- were simply instructed to watch the other person
like processes in empathy. First evidence for the in the viewing condition. None of the studies con-
existence of mirror networks for facial expressions trolled for higher level cognitive processes. Thus, it
emerged from a study by Carr et al. (2003). Viewing remains to be shown how far the activations found
and deliberately imitating different facial expression in these studies depend on higher level processing.
resulted in increased activation in the pars opercul-
aris of the inferior frontal gyrus, the premotor
cortex, the insula, the superior temporal cortex, and Perspective-taking
the amygdala. In line with other research groups
(Buck, 1980; Dimberg, 1988; Preston and de Waal, While contagion-like simulation processes may be
2002), Carr et al. (2003) posit that represen- at the core of our capacity to empathize, it is gen-
tations of observed facial expressions activate erally believed that higher level processes including
428

perspective-taking also play a role in empathy (conceptual perspective-taking). Both conditions re-
(Feshbach, 1978; Batson et al., 1997a). Owing to its sulted in increased activity in medial PFC and right
complexity, perspective-taking is much more diffi- temporo-parietal regions. Additionally, emotional
cult to investigate with peripheral physiological perspective-taking elicited a stronger hemodynamic
measures than contagion-like processes. Neuroim- response in the vmPFC and the vlPFC than con-
aging methods on the other hand are especially ceptual perspective-taking. The authors interpreted
suitable for studying processes that underlie these findings as evidence for the existence of dis-
perspective-taking in that they allow dissociation tinct neural systems: one underlying perspective-
of different processes and are sensitive to subtle taking in general and others that are engaged
experimental manipulations. depending on the type of information that has to
Stotland (1969) conducted the first and, to our be inferred. Specifically, the vmPFC is believed to be
knowledge, the only study that tried to elucidate engaged in inferring another person’s emotional
the relationship between perspective-taking, state. The finding of selective vmPFC activation
subjective reports of empathy and distress, and during emotional perspective-taking is in accord-
a peripheral physiological component of empathic ance with studies demonstrating that patients with
responding: the SCR. He distinguished between vmPFC lesions have striking difficulties with men-
two forms of perspective-taking: one in which the talizing tasks involving emotional processing but
subjects imagined how they would feel if they not with those devoid of emotional processing
were exposed to the painful heat stimulus that was (Shamay-Tsoory et al., 2005b). However, the
applied to a person they were observing (first- vmPFC has also been implicated in passively view-
person perspective-taking) and another one in ing social as opposed to nonsocial emotional scenes
which the subjects had to imagine what the other (Moll et al., 2002) as well as in linking a specific
person was feeling in that situation (third-person situation with its emotional value (Damasio, 1996)
perspective-taking). Both types of deliberate and emotional self-reflection (Lane et al., 1997;
perspective-taking led to more empathic feelings Mitchell et al., 2005). Thus, it is not clear whether
and an increase in SCR compared to passively the vmPFC is particularly involved in adopting the
viewing the painful treatment. Additionally, the emotional perspective of another person or other
self-focused perspective-taking resulted in distress aspects of empathy.
feelings. Of note, in both conditions the SCR did Other neuroimaging studies contrasting mental-
not begin to increase until 30 s after the experi- izing tasks with and without emotional processing
menter had announced that the painful heat was demands do not support a role of the vmPFC in
applied to the victim, suggesting the dominance of emotional perspective-taking (Shamay-Tsoory
more complex cognitive processes in both kinds of et al., 2005a; Völlm et al., 2006). In a positron
perspective-taking. emission tomography study, healthy subjects
Neuroimaging studies point to a role of the PFC listened to interviews in which characters described
in higher level empathic processes. An early fMRI a distressful conflict or a neutral situation, and
study by Farrow et al. (2001) found ventromedial, were asked questions about the characters’ feelings
dorsomedial, and ventrolateral PFC (vmPFC, and thoughts. The dorsomedial, but not the vent-
dmPFC, and vlPFC, respectively) as well as tempo- romedial PFC, was more strongly activated
ral regions activated when subjects were asked to during listening to and answering questions about
judge the emotional state of a person in a written the distressful interview than during listening to
scenario. Hynes et al. (2006) investigated whether and answering questions about the neutral one
inferring others’ emotions (emotional perspective- (Shamay-Tsoory et al., 2005a). Using cartoon-
taking) and inferring others’ thoughts (conceptual mentalizing tasks, Völlm et al. (2006) found in-
perspective-taking) rely on distinct neural networks: creased activation common to both emotional and
subjects were presented with written scenarios conceptual perspective-taking in bilateral dmPFC,
and required to indicate what the character was bilateral temporoparietal junction, and left
feeling (emotional perspective-taking) or thinking temporal pole. Emotional perspective-taking was
429

associated with stronger activity in the lower part 32) as a key structure in inferring other people’s
of the right dmPFC, the right ACC, the right mental states (Fletcher et al., 1995; Brunet et al.,
posterior cingulate cortex, and the left amygdala. 2000; Castelli et al., 2000; McCabe et al., 2001;
All neuroimaging studies reviewed above Vogeley et al., 2001; Berthoz et al., 2002; Gallagher
contrasted mentalizing tasks with and without and Frith, 2003; Walter et al., 2004; but see Bird et
emotional processing and found ventral and/or al., 2004). However, the aPCC might not be crucial
dorsal medial prefrontal areas more strongly acti- for all types of mentalizing. Walter et al. (2004)
vated during emotional mentalizing (see Fig. 2 for have shown that the aPCC is more strongly
a summary of these studies’ results). Nevertheless, activated during mentalizing tasks requiring the
discrepancies in the topography of activation understanding of a person who is engaged in a so-
between these studies are evident, particularly cial interaction than the understanding of a person
concerning the involvement of the vmPFC or who is not. The authors propose that the aPCC is
dmPFC. While some of these differences might be primarily involved in understanding other people’s
due to the different stimuli used, interpretation of intentions in social interactions. While it is possible
these findings is further complicated by the fact that the aPCC is exclusively involved in reasoning
that one study (Hynes et al., 2006) restricted the about social interactions, other areas of the medial
analysis to the ventral PFC (zo0 and y43, MNI PFC could play a role in mentalizing tasks in
coordinates). general, irrespective of the nature of the stimuli.
In addition to neuroimaging studies investigating But what exact role do these structures play in
possible differences between emotional and con- mentalizing processes? One idea is that medial PFC
ceptual perspective-taking, also studies focusing on regions support the decoupling of one’s own and
mentalizing in general have identified the medial the other’s perspective (Leslie, 1994; Gallagher and
PFC, especially the anterior paracingulate cortex Frith, 2003). This hypothesis is bolstered by the
(aPCC; the region at the border of BAs 9, 10, and finding that the medial PFC is activated in

Fig. 2. Activation maxima in the medial and lateral prefrontal cortex found during emotional as opposed to conceptual perspective-
taking (filled icons) or in both, emotional and conceptual perspective-taking (open icons).
430

competitive games only when subjects believe they had to say in which of two boxes a person would
are competing with another participant and not look for an object, knowing themselves where the
when they think they are playing against a compu- object was and knowing that the person did not
ter (McCabe et al., 2001; Gallagher et al., 2002). have this knowledge. This condition would thus
Other researchers (Vorauer and Ross, 1999; Ruby require the inhibition of one’s own belief and the
and Decety, 2001, 2003, 2004) have proposed a inference of the other’s belief. The task was mod-
different contribution of the medial PFC to men- ified for the low-inhibition condition in that it no
talizing. They advocate that while we adopt the longer entailed discrepant knowledge of the subject.
perspective of another person, the medial PFC In this condition the subject did not know where
inhibits the self-perspective we predominantly view the object was placed. The low-inhibition condition
situations from (for a review concerning the thus only necessitates the inference of the other’s
dominance of the self-perspective see Birch and belief. A patient with a lesion to the right PFC
Bloom, 2004). Several studies have investigated extending into the right superior temporal gyrus
differences in brain activation when subjects had to performed normally on the low self-perspective
adopt a first-person or third-person perspective inhibition task but was markedly impaired in the
during visual, motor, conceptual, and emotional high self-perspective inhibition task. This supports
perspective-taking (Ruby and Decety, 2001, 2003, the notion that prefrontal areas are involved in the
2004; Vogeley et al., 2001; Grezes et al., 2004; suppression of the self-perspective when we try to
Vogeley et al., 2004). Taking the perspective of take the perspective of another person but not in
another person resulted in stronger activation of the inference of another person’s perspective per se.
medial prefrontal areas (dorsal and ventral alike), The fact that impairments in tasks with high self-
left temporal regions, and right inferior parietal perspective inhibition demands were found for both
cortex, regardless of the task, suggesting a role of conceptual and emotional perspective-taking sug-
these brain regions in the inhibition of the self- gests that the inhibition of the self-perspective
perspective. Anticipating the emotional reactions of might rely on similar networks in both conceptual
one’s mother in an imagined emotional situation and emotional tasks.
compared to anticipating one’s own emotional re- Interestingly, the impairments of three patients
actions in the same situation additionally activated with intact PFC and STG but lesions to the left
the left temporal pole (Ruby and Decety, 2004), a temporoparietal junction were not restricted to
region that has been associated with autobiograph- perspective-taking that required inhibition of the
ical recall (Fink et al., 1996). This finding nicely fits self-perspective (Apperly et al., 2004; Samson
with the assumption that putting oneself into an et al., 2004). The authors argue that the deficits
emotional situation (first-person perspective- of these three patients on false-belief tasks are not
taking, see theoretical background) relies less on due to impaired inhibition of the self-perspective
activation of past knowledge than anticipating the but to a different dysfunctional process, namely
emotional reactions of another person in a specific the inference of another person’s mental states.
situation (third-person perspective-taking). Thus, inhibition of the self-perspective and infer-
While these studies provide valuable information ence of the other’s perspective are putatively
concerning the neural networks mediating third- distinct neural processes subserved by prefrontal
person perspective-taking, they do not strictly and temporoparietal regions, respectively.
differentiate between the two mechanisms likely in-
volved in this process: inhibition of one’s own per-
spective and inference of the other’s perspective. A Interindividual differences in empathic processes
recent study has tried to unravel these two mech-
anisms by modulating the self-perspective People differ in their ability to understand and share
inhibition demands in false-belief tasks (Samson other’s feelings. Considering these individual differ-
et al., 2005). A standard false-belief task was used ences might further elucidate some of the mecha-
for the high-inhibition condition in which subjects nisms that underlie empathy (see Chakrabarti and
431

Baron-Cohen, this volume). Studies investigating only more mimicking behavior but also a higher
individual differences in dispositional empathy have correspondence between facial EMG activity and
used a wealth of different self-report measures to self-reported feelings than low-empathic subjects.
assess a person’s tendency to understand and share In high dispositional empathic subjects, activity in
another person’s feelings. These self-report measures the zygomaticus major muscle correlated positively
usually assess either ‘‘emotional empathy,’’ i.e., the with the valence of the self-reported feeling,
tendency to vicariously experience another person’s whereas in low dispositional empathic subjects this
emotion (similar to emotional contagion), or ‘‘cog- correlation was negative. Low empathic subjects
nitive empathy,’’ i.e., the ability to take another per- smiled more when they reported to feel angry than
son’s perspective (Mehrabian, 1997; Baron-Cohen when they reported to feel happy. Subjects scoring
and Wheelwright, 2004). A typical item in question- high on dispositional empathy report greater
naires assessing emotional empathy would be ‘‘I am emotional responding and show higher SCR than
often quite touched by things that I see happen,’’ low empathic subjects when viewing an empathy-
whereas ‘‘When I’m upset at someone, I usually try evoking film (Eisenberg et al., 1991). This confirms
to ‘put myself in his shoes’ for a while’’ would be earlier findings showing that high empathic women
representative for cognitive empathy (items taken exhibited larger SCR and more congruent facial
from the interpersonal reactivity index; Davis, 1983). expressions than low empathic women when
Some items in these questionnaires, though, are not watching videotapes of smiling, calm, and crying
easy to assign to either of these two classes, since babies (Wiesenfeld et al., 1984).
they contain aspects of both. In Table 2 we provide Hitherto there have not been many studies ex-
an overview of the measures used to assess dispos- plicitly investigating individual differences in higher
itional empathy, including their subscales, and level empathic processes. First evidence for a rela-
indicate when a relation with either contagion-like tion between interindividual differences in cognitive
or perspective-taking processes has been found. empathic processes and emotional responding
EMG studies have shown more pronounced comes from a study demonstrating a positive
mimicking behavior in high than in low disposit- correlation between the subjects’ dispositional abil-
ional empathic subjects when they passively viewed ity to take the perspective of another person and
emotional facial expressions (Sonnby-Borgström, their self-reported emotional responding to a film
2002; Dimberg et al., 2005). Interestingly, subjects when they were instructed to imagine the emotions
scoring high on dispositional empathy showed not of the film characters (Eisenberg et al., 1991).

Table 2. Empathy questionnaire and their relation to contagion-like processes and perspective-taking

Measure Subscales Studies finding relation to Studies finding relation to


contagion-like processes perspective-taking

Truax rating scale (Truax, 1961) Shamay-Tsoory et al. (2005a)


Questionnaire measure of Sonnby-Borgström (2002)
emotional empathy
(Mehrabian and Epstein, 1972)
Wiesenfeld et al. (1984)
Interpersonal reactivity index Perspective-taking
(Davis, 1983)
Fantasy
Empathic concern Singer et al. (2004)
Personal distress
Balanced emotional Empathy Singer et al. (2004)
scale (Mehrabian, 1997)
Empathy scale (Leibetseder et al., Empathic preparedness; Leiberg et al.
2001) concernment (unpublished data)
432

These results suggests that there might be marked Correlation analysis demonstrated a stronger ac-
differences in both contagion-like and higher level tivity in the vmPFC and the aPCC during engage-
empathic processes between subjects, which are ment for subjects who rated their attempts to
reflected in questionnaire scores and which should engage with the target as more successful than
be taken into account in order to resolve some of subjects who rated their attempts as less successful
the discrepancies between studies investigating (Leiberg et al., 2005). As discussed above, both
the relation between emotional responding and areas have been implicated in active empathic
emotional judgments reviewed above. processes. The aPCC has been linked to decoupling
To investigate whether interindividual differ- one’s own from another person’s perspective and
ences are also reflected in brain activity when to self-perspective inhibition. The vmPFC appears
different subjects are confronted with people in to serve a more general role in emotional process-
distress, we conducted an fMRI study in which ing and might be specific for emotional as opposed
subjects passively viewed photographs of victims of to conceptual perspective-taking. Interestingly, ac-
violence before they were asked to regulate their tivation in the vmPFC correlated not only with the
empathic responding (Leiberg et al., unpublished average success ratings of the subjects but also with
data). Using intersubject correlation analyses, we the average increase in startle eye blink responses
found that activity in the vmPFC during passively of each subject during empathizing trials, indicat-
viewing disturbing photographs increased with ing that deliberate perspective-taking might have
dispositional emotional empathy. Together with activated some emotional simulation networks.
the study by Singer et al. (2004), which found that Both aspects of empathic processing — taking the
overlapping activity in the ACC and the anterior perspective of another person and activating
insula during reception of a painful stimulus and appropriate emotional responses — might be more
observation of a signal indicating that a loved one developed in some persons than in others.
will receive a painful stimulus was stronger in However, in this study, self-reported success in
subjects who scored high on self-report measures of empathizing and differences in vmPFC and aPCC
empathy, our study provides evidence for a relation activity did not correlate with dispositional empa-
between dispositional empathy and activation thy, as assessed with self-report questionnaires,
differences in neural circuits underlying simulation suggesting that current psychological instruments
processes in empathy. might not capture all aspects of interindividual
In the study by Shamay-Tsoory et al. (2005a), differences in emotional perspective-taking or that
where subjects had to listen to and answer ques- situational empathic reactions are determined by
tions about empathy-inducing or neutral-theme more than the perception of one’s own ability to
interviews, the relationship between individual empathize.
differences in empathic accuracy and brain activ- The relation between individual differences in
ity was also investigated. The answers of the sub- dispositional empathy and empathic processes,
jects to the questions about the empathy-inducing i.e., contagion-like as well as more cognitive ones,
interview were rated for their level of empathic is far from being understood. This is partly due to
accuracy. Empathic accuracy correlated positively the item heterogeneity of self-report measures of
with activity in the dmPFC, one of the areas that dispositional empathy and the inconclusive assign-
were more activated during listening to empathy- ment of those items to the different aspects of
eliciting interviews than to neutral ones. Thus, empathy, i.e., emotional and cognitive empathy.
subjects who were more accurate in their under-
standing of the characters’ emotions showed
stronger dmPFC activity. In our study, after Modulatory processes
passive viewing, subjects were instructed to ac-
tively engage with or disengage from the displayed Two types of modulatory processes can be distin-
person in distress. After each trial, subjects rated guished that serve to attenuate vicarious emotional
their success in engagement or disengagement. responding in empathy: one that is focused on
433

one’s own emotions (self-focused emotion regula- None of these studies distinguished between po-
tion) and another one that is focused on one’s en- tentially empathy-eliciting social-emotional stimuli
gagement with the other person (other-focused and stimuli without any social content. In another
emotion regulation). The first one is also employed study, Harenski and Hamann (2006) explored
to modulate emotional reactions to nonsocial whether downregulating emotional responses to
emotional stimuli. social stimuli with moral or nonmoral content re-
Psychophysiological studies have shown that cruits differential neural networks. Notably, both
peripheral physiological measures like corrugator conditions entailed increased activity in ventrolat-
supercilii activity, startle eye blink amplitude, and eral and dorsal PFC, whereby decreasing emo-
SCR are sensitive to emotion regulation (Jackson tional response to stimuli with moral content
et al., 2000; Eippert et al., in press). In recent years, recruited additional regions. However, as presum-
neuroimaging studies (Beauregard et al., 2001; ably not all of these stimuli induced empathic
Ochsner et al., 2002, 2004b; Levesque et al., 2003; emotional responses, it is not clear how these re-
Phan et al., 2005; Eippert et al., in press) have sults relate to emotion regulation in empathy.
investigated the neural bases of emotion regulation It is still unclear whether the structures involved
(for a review see Ochsner and Gross, 2005). It has in regulating one’s own emotions also mediate the
been demonstrated that voluntarily downregulat- modulation of the engagement with another per-
ing electrodermal activity in a biofeedback task son. Recent brain imaging studies support the view
(Critchley et al., 2002) or downregulating sexual that networks subserving self-focused and other-
arousal (Beauregard et al., 2001) as well as reap- focused responses are partly distinct. Evaluation
praising negative visual stimuli in neutral terms of one’s own and another person’s emotions elicits
(Ochsner et al., 2002, 2004b; Levesque et al., 2003; common activation in the medial PFC and differ-
Phan et al., 2005; Eippert et al., in press) is asso- ential activation in other parts of the medial and
ciated with increased activity in the vmPFC, the the lateral PFC (Ochsner et al., 2004a). Internally
dorsolateral PFC (dlPFC), and the ACC and with and externally focused processing of stimuli en-
decreased activity in the amygdala. Deliberate up- gages different prefrontal regions (Christoff et al.,
regulation of negative emotions also produced in- 2003; Ochsner et al., 2004b). In our study (Leiberg
creased activity in the vmPFC, the dlPFC, and the et al., unpublished data), subjects were asked not
ACC, as well as increased activity in the amygdala. only to actively empathize with a depicted person
While the amygdala and possibly also the vmPFC in distress, but also, in a different condition, to
are core structures in emotional responding, the disengage from that person by thinking that the
dlPFC and the ACC have been implicated in the photograph was not real but from a movie, or that
cognitive control of behavior (Cabeza and Nyberg, the person was not a person but an actor or a doll.
2000; Miller and Cohen, 2001), conflict monitoring Disengaging from, vs. engaging with the victim,
(Botvinick et al., 2004), and the inhibition of pre- yielded stronger activation in vlPFC, dmPFC, and
dominant responses (Garavan et al., 1999; Braver dlPFC. Thus, it appears that other-focused disen-
et al., 2001). The studies on emotion regulation gagement engages additional brain regions com-
suggest that the dlPFC and the ACC play a role in pared to self-focused emotion regulation (Ochsner
modulating emotional responses by exerting top- et al., 2004b; Eippert et al., in press), supporting
down control on areas involved in the processing the notion of two distinct modulatory processes.
of emotional stimuli.
However, the stimulus material employed in
these studies was rather homogeneous. All but two Empathy and social behavior
studies (Beauregard et al., 2001; Levesque et al.,
2003) used pictures from the International Affec- The capability to understand and share other peo-
tive Picture System (IAPS; Lang et al., 2005) var- ple’s emotions has an impact on our behavior to-
ying largely in content and emotional meaning. ward these people (see also Chakrabarti and
The other two studies employed film excerpts. Baron-Cohen, this volume). On the one hand, a
434

positive relation has been found between empathy activity in the vmPFC and significantly smaller
and helping behavior (Eisenberg and Fabes, 1991; SCR during adjustment of the stimulus intensity
Trobst et al., 1994; Batson et al., 1997b; Batson, than subjects who scored low on this measure.
1998), which is inverted if observers of a person in Diminished autonomic and vmPFC responding in
need of help are overwhelmed by their vicarious highly callous subjects was associated with stron-
emotional experience and are not able to distance ger self-reported feelings of aggression and more
themselves from the observed person’s emotional aggressive behavior. Given that the vmPFC has
state (Batson, 1991; Eisenberg and Fabes, 1998). been implicated in emotion processing in general
On the other hand, an inverse relation has been and emotional perspective-taking in particular
observed between empathy and aggressive be- (Hynes et al., 2006), these findings support a con-
havior (Miller and Eisenberg, 1988; Davis, 1996; nection between vicarious emotional responses
Mehrabian, 1997). and the inhibition of aggressive behavior.
Supporting the notion that deviant social
behavior might be a result of diminished empathic
responding, boys with disruptive behavior disorder Summary and conclusions
have been found to show lower scores on an empa-
thy questionnaire and significantly less corrugator The investigation of the nature of empathy, i.e.,
supercilii activity to angry expressions than age- the ability to perceive, share, and understand
matched healthy controls (de Wied et al., 2006). others’ emotions, has attracted many researchers
Neuroimaging studies are only beginning to identify from different fields. Two main approaches to
the neural substrates underlying the link between the construct of empathy can be identified (see
empathy and behavior. Mathiak and Weber (in Table 1): one that is focusing on the contagion-like
press) investigated neural activity underlying aggres- manner in which the perception of a person in an
sive behavior in virtual environments. During emotional state can result in a similar affective re-
violent acts, subjects exhibited reduced activity in sponse in the observer, and one that concentrates
the medial PFC and the amygdala — areas that have on the understanding of another person’s emotion
been found active when engaging with, or taking the by means of perspective-taking. Two processes
perspective of, a person in distress (Hynes et al., have been proposed to underlie empathy. One that
2006; Völlm et al., 2006). Reduced activity in these assumes that the understanding of other people’s
areas suggests that subjects distanced themselves emotions depends on an internal simulation proc-
from and suppressed empathic responding to the ess, which relies on shared representations between
victims-to-be. It would be of great interest to see the observer and the target (simulation theory),
whether activity in these areas is increased when the and another that proposes that humans possess a
subjects’ task is not to kill opponents but to help theory of mind, which they use to attribute mental
associates. Virtual reality appears to be well suited to states like emotions to other people and thereby
study the relationship between empathy and social understand their emotions (theory–theory). Most
behavior. neuroscientists favor the view that empathy can be
In another study from our laboratory (Lotze achieved by both types of processes that work
et al., submitted), subjects played a competitive closely intertwined. In most of the situations when
reaction time task against an opponent and which- we observe someone in an emotional state, some
ever person was faster in responding to a cue was contagion-like processes will be more or less
allowed to administer an aversive stimulus of his automatically initiated. To what extent conta-
choice to the other. While in fact the second gion-like processes are employed and result in a
participant was a collaborator of the experimenter, ‘‘correct’’ representation depends on the current
subjects believed that their opponent was a fellow emotional state of the observer and the experience
subject and saw him receiving the aversive stimu- the observer has with the target’s situation. When
lus. Subjects who scored high on a self-report a situation is ambiguous or too complex, other
psychopathy scale exhibited significantly smaller processes, like perspective-taking, which do not
435

necessarily produce the same emotional state in the another person’s position is required to under-
observer as in the target, will be used to under- stand her mental state. On the other hand, more
stand the other’s emotions. ventromedial prefrontal regions seem to be in-
Peripheral physiological and neuroimaging volved in the processing of emotional information
studies provide converging evidence for the sig- and possibly represent emotional responses elicited
nificance of both contagion-like processes and when adopting another person’s perspective. If
perspective-taking in empathy. They demonstrate perspective-taking is a secondary process eliciting
automatic imitation of observed motor behavior, simulation in a top-down manner, one would as-
synchrony in ANS activity during empathic judg- sume that in addition to the medial prefrontal ar-
ments, and shared neural circuits for executing and eas involved in inhibition of the first-person
viewing facial expressions, and experiencing and perspective, similar brain regions as in studies on
observing emotions. It remains to be seen whether contagion-like processes would be activated. The
the activation of a representation of the target’s reviewed emotional perspective-taking studies do
emotional state in the observer is necessary and point this way (Völlm et al., 2006), but bearing in
sufficient for understanding other people’s emo- mind that the employed paradigms vary, it is likely
tions. The results so far are equivocal concerning too early to come to a definite conclusion.
the relationship between activated representations, Questionnaire and behavioral studies have
subjective feelings, and the understanding of the shown that the ability to empathize with another
other person’s emotions. While some studies on person does not only vary depending on the
contagion-like effects on motor and ANS activity situation but also interindividually. Contagion
find a connection between these three parameters, effects on motor, ANS, and brain activity are pre-
others do not. dicted by dispositional empathy as assessed with
There is tentative evidence suggesting that con- questionnaire measures. Subjects scoring high on
tagion-like processes do not proceed in a strictly these questionnaires exhibit stronger mimicking
automatic manner but may be influenced by con- behavior, increases in SCR, and activity in empa-
textual factors (Lanzetta and Englis, 1989). A very thy-related brain areas. A relation between inter-
recent fMRI study (Singer et al., 2006) employing individual differences in the ability to take the
the Prisoner’s Dilemma paradigm has shown that perspective of another person and brain activity
the brain activity of male subjects during observ- during higher level empathic processes has so far
ing an opponent in pain was modulated by the only been shown for measures of situational
perceived fairness of the opponent. Activity in empathy, like empathic accuracy or success of em-
pain-related brain regions was significantly re- pathizing, rather than dispositional empathy. This
duced when observing an unfair compared to a fair could be due to the heterogeneity of the question-
opponent in pain. Because of the rather coarse naire items, some assessing affective, some cogni-
time resolution of fMRI, it is not possible to tive aspects of empathy.
decide whether this modulating influence of con- Intentional processes regulate empathic proc-
text is already present at the initial neural response esses and the ensuing responses. Both the ability to
or this response is completely stimulus-driven and empathize with other people and the ability to
later on altered by contextual factors. regulate vicarious emotional responses are prereq-
Neuroimaging studies suggest that prefrontal uisites for prosocial behavior. Diminished em-
areas, specifically the medial prefrontal cortex, pathic responding has been shown to be related to
play a role in higher level empathic processes that deviant social behavior like enhanced aggression.
enable us to infer another person’s emotional state Medial prefrontal regions are implicated in the
in the presence of insufficient or conflicting infor- suppression of empathic responding likely neces-
mation, or when the other person’s emotional sit- sary to perform acts of aggression. Presently,
uation is completely alien to us. The aPCC in neuroimaging studies are beginning to elucidate
particular has been implicated in the inhibition the neural substrates underlying the link between
of the first-person perspective when adoption of empathy and social behavior.
436

Abbreviations Bandura, A. (1997) Self-Efficacy: The Exercise of Control.


Freeman, New York.
Baron-Cohen, S. (2005) The empathizing system: A revision of
ACC anterior cingulate cortex
the 1994 model of the Mindreading System. In: Ellis, B. and
aPCC anterior paracingulate cortex Bjorklund, D. (Eds.), Origins of the Social Mind. Guilford,
BA Brodmann area New York, pp. 468–492.
dlPFC dorsolateral prefrontal cortex Baron-Cohen, S. and Wheelwright, S. (2004) The empathy
dmPFC dorsomedial prefrontal cortex quotient: an investigation of adults with Asperger syndrome
EMG electromyography or high functioning autism, and normal sex differences.
J. Autism. Dev. Disord., 34: 163–175.
fMRI functional magnetic resonance Batson, C.D. (1991) The Altruism Question: Towards a Social-
imaging Psychological Answer. Erlbaum, Mahwah, NJ.
PFC prefrontal cortex Batson, C.D. (1998) Altruism and prosocial behavior. In:
SCR skin conductance response Gilbert, D.T., Fiske, S.T. and Lindzey, G. (Eds.), The Hand-
book of Social Psychology. McGraw-Hill, Boston, pp. 282–316.
ST simulation theory
Batson, C.D. and Coke, J. (1981) Empathy: a source of altru-
ToM theory of mind istic motivation for helping. In: Rushton, J. and Sorrentino,
TMS transcranial magnetic stimulation R. (Eds.), Altruism and Helping Behavior. Erlbaum,
TT theory theory Hillsdale, NJ, pp. 167–187.
vlPFC ventrolateral prefrontal cortex Batson, C.D., Early, S. and Salvarani, G. (1997a) Perspective
vmPFC ventromedial prefrontal cortex taking: imagining how another feels versus imagining how
you would feel. Pers. Soc. Psychol. Bull., 23: 751–758.
Batson, C.D., Fultz, J. and Schoenrade, P.A. (1987) Adult’s
emotional reactions to the distress of other’s. In: Eisenberg,
N. and Strayer, J. (Eds.), Empathy and Its Development.
Cambridge University Press, Cambridge, pp. 163–184.
Acknowledgments Batson, C.D., Lishner, D.A., Carpenter, A., Dulin, L.,
Harjusola-Webb, S., Stocks, E.L., Gale, S., Hassan, O. and
Sampat, B. (2003) ‘‘yAs you would have them do unto
This work was supported by Deutsche Forsch- you’’: Does imagining yourself in the other’s place stimulate
ungsgemeinschaft and the Junior Science moral action? Pers. Soc. Psychol. Bull., 29: 1190–1201.
Program of the Heidelberger Academy of Sciences Batson, C.D., Sager, K., Garst, E., Kang, M., Rubchinsky, K.
and Humanities. We thank Robert Langner for and Dawson, K. (1997b) Is empathy-induced helping due to
valuable discussions and Krystyna Swirszcz for self-other merging? J. Pers. Soc. Psychol., 73: 495–509.
Bavelas, J.B., Black, A., Chovil, N. and Lemery, C.R. (1986)
helpful comments. ‘I show how you feel’: motor mimicry as a communicative
act. J. Pers. Soc. Psychol., 50: 322–329.
Beauregard, M., Levesque, J. and Bourgouin, P. (2001) Neural
References correlates of conscious self-regulation of emotion. J. Neuro-
sci., 21: RC165.
Adolphs, R. (1999) Social cognition and the human brain. Berthoz, S., Armony, J.L., Blair, R.J. and Dolan, R.J. (2002)
Trends Cogn. Sci., 3: 469–479. An fMRI study of intentional and unintentional (embarrass-
Adolphs, R., Tranel, D. and Damasio, A.R. (2003) Dissociable ing) violations of social norms. Brain, 125: 1696–1708.
Birch, S.A. and Bloom, P. (2004) Understanding children’s and
neural systems for recognizing emotions. Brain Cogn.,
52: 61–69. adults’ limitations in mental state reasoning. Trends Cogn.
Allport, G. (1961) Pattern and Growth in Personality. Holt, Sci., 8: 255–260.
Rinehart & Winston, New York. Bird, C.M., Castelli, F., Malik, O., Frith, U. and Husain, M.
(2004) The impact of extensive medial frontal lobe damage
Apperly, I.A., Samson, D., Chiavarino, C. and Humphreys,
G.W. (2004) Frontal and temporo-parietal lobe contributions on ‘Theory of Mind’ and cognition. Brain, 127: 914–928.
to theory of mind: neuropsychological evidence from a false- Blair, R.J. (2005) Responding to the emotions of others:
belief task with reduced language and executive demands. dissociating forms of empathy through the study of typical
and psychiatric populations. Conscious Cogn., 14: 698–718.
J. Cogn. Neurosci., 16: 1773–1784.
Avenanti, A., Bueti, D., Galati, G. and Aglioti, S.M. (2005) Blairy, S., Herrera, P. and Hess, U. (1999) Mimicry and the
Transcranial magnetic stimulation highlights the sensorimo- judgment of emotional facial expressions. JNB, 23: 5–41.
Botvinick, M.M., Cohen, J.D. and Carter, C.S. (2004) Conflict
tor side of empathy for pain. Nat. Neurosci., 8: 955–960.
Ax, A.A. (1964) Goals and methods of psychophysiology. monitoring and anterior cingulate cortex: an update. Trends
Psychophysiology, 1: 8–25. Cogn. Sci., 8: 539–546.
437

Botvinick, M., Jha, A.P., Bylsma, L.M., Fabian, S.A., attempts to empathize: how do we imagine the perspective of
Solomon, P.E. and Prkachin, K.M. (2005) Viewing facial another? Pers. Soc. Psychol. Bull., 30: 1625–1635.
expressions of pain engages cortical areas involved in the Decety, J. and Jackson, P.L. (2004) The functional architecture
direct experience of pain. Neuroimage, 25: 312–319. of human empathy. Behav. Cogn. Neurosci. Rev., 3: 71–100.
Braver, T.S., Barch, D.M., Gray, J., Molfese, D.L. and Snyder, de Wied, M., van Boxtel, A., Zaalberg, R., Goudena, P.P. and
A. (2001) Anterior cingulate cortex and response conflict: Matthys, W. (2006) Facial EMG responses to dynamic
effects of frequency, inhibition and errors. Cereb. Cortex, emotional facial expressions in boys with disruptive behavior
11: 825–836. disorders. J. Psychiatr. Res., 40: 112–121.
Brunet, E., Sarfati, Y., Hardy-Baylé, M.-C. and Decety, J. DiMascio, A., Boyd, R.W. and Greenblatt, M. (1957) Physio-
(2000) A PET investigation of the attribution of intentions logical correlates of tension and antagonism during psycho-
with a nonverbal task. Neuroimage, 11: 157–166. therapy: a study of ‘‘interpersonal physiology’’. Psychosom.
Buck, R. (1980) Nonverbal behavior and the theory of emotion: Med., 19: 99–104.
the facial feedback hypothesis. J. Pers. Soc. Psychol., DiMascio, A., Boyd, R.W., Greenblatt, M. and Solomon, H.C.
38: 811–824. (1955) The psychiatric interview: a sociophysiologic study.
Cabeza, R. and Nyberg, L. (2000) Imaging cognition II: an Dis. Nervous Systems, 16: 4–9.
empirical review of 275 PET and fMRI studies. J. Cogn. Dimberg, U. (1982) Facial reactions to facial expressions.
Neurosci., 12: 1–47. Psychophysiology, 19: 643–647.
Calder, A.J., Keane, J., Manes, F., Antoun, N. and Young, Dimberg, U. (1988) Facial electromyography and the experi-
A.W. (2000) Impaired recognition and experience of disgust ence of emotion. J. Psychophysiol., 3: 277–282.
following brain injury. Nat. Rev. Neurosci., 3: 1077–1078. Dimberg, U., Andreasson, P. and Thunberg, M. (2005)
Carr, L., Iacoboni, M., Dubeau, M.C., Mazziotta, J.C. and Empathy and facial reactions to facial expressions. Psycho-
Lenzi, G.L. (2003) Neural mechanisms of empathy in physiology, 42(Suppl 1): 50.
humans: a relay from neural systems for imitation to limbic Dimberg, U., Thunberg, M. and Elmehed, K. (2000) Uncon-
areas. Proc. Natl. Acad. Sci. USA, 100: 5497–5502. scious facial reactions to emotional facial expressions.
Castelli, F., Happe, F., Frith, U. and Frith, C.D. (2000) Move- Psychol. Sci., 11: 86–89.
ment and mind: a functional imaging study of perception and Eippert, F., Veit, R., Weiskopf, N., Erb, M. and Birbaumer, N.
interpretation of complex intentional movement patterns. Coping with fear: brain activity during emotion regulation.
Neuroimage, 12: 314–325. Hum. Brain Mapp., In press.
Chartrand, T.L. and Bargh, J.A. (1999) The chameleon effect: Eisenberg, N. (2000) Emotion, regulation, and moral develop-
the perception-behavior link and social interaction. J. Pers. ment. Annu. Rev. Psychol., 51: 665–697.
Soc. Psychol., 76: 893–910. Eisenberg, N. and Fabes, R.A. (1991) Prosocial behavior and
Christoff, K., Ream, J.M., Geddes, L.P. and Gabrieli, J.D. empathy: a multimethod, developmental perspective. In:
(2003) Evaluating self-generated information: anterior pre- Clark, M. (Ed.), Review of Personality and Social Psychol-
frontal contributions to human cognition. Behav. Neurosci., ogy. Sage, Newbury Park, CA, pp. 34–61.
117: 1161–1168. Eisenberg, N. and Fabes, R.A. (1998) Prosocial development.
Coleman, R.M., Greenblatt, M. and Solomon, H.C. (1956) In: Damon, W. (Ed.) Handbook of Child Psychology: Vol. 3.
Physiological evidence of rapport during psychotherapeutic Social, Emotional, and Personality Development. Wiley,
interviews. Dis. Nervous. Systems, 17: 71–77. New York, pp. 701–778.
Critchley, H.D., Melmed, R.N., Featherstone, E., Mathias, C.J. Eisenberg, N., Fabes, R.A., Schaller, M., Miller, P., Carlo, G.,
and Dolan, R.J. (2002) Volitional control of autonomic Poulin, R., Shea, C. and Shell, R. (1991) Personality and
arousal: a functional magnetic resonance study. Neuroimage, socialization correlates of vicarious emotional responding.
16: 909–919. J. Pers. Soc. Psychol., 61: 459–470.
Damasio, A.R. (1994) Descartes’ Error: Emotion, Reason, and Eisenberg, N. and Strayer, J. (1987) Empathy and Its Devel-
the Human Brain. Putnam, New York. opment. Cambridge University Press, Cambridge.
Damasio, A.R. (1996) The somatic marker hypothesis and the Farrow, T.F., Zheng, Y., Wilkinson, I.D., Spence, S.A.,
possible functions of the prefrontal cortex. Philos. Trans. R. Deakin, J.F., Tarrier, N., Griffiths, P.D. and Woodruff,
Soc. Lond. B Biol. Sci., 351: 1413–1420. P.W. (2001) Investigating the functional anatomy of empathy
Darley, J.M. and Latane, B. (1968) Bystander intervention and forgiveness. Neuroreport, 12: 2433–2438.
in emergencies: diffusion of responsibility. J. Pers. Soc. Feshbach, N.D. (1978) Studies of empathic behavior in
Psychol., 8: 377–383. children. In: Maher, B.A. (Ed.), Progress in Experimental
Davis, M.H. (1983) Measuring individual differences in empa- Personality Research. Academic Press, New York, pp. 1–47.
thy: evidence for a multidimensional approach. J. Pers. Soc. Fink, G.R., Markowitsch, H.J., Reinkemeier, M., Bruckbauer,
Psychol., 44: 113–126. T., Kessler, J. and Heiss, W.D. (1996) Cerebral representa-
Davis, M.H. (1996) Empathy a Social Psychological Approach. tions of one’s own past: neural networks involved in auto-
Westview, Boulder, CO. biographical memory. J. Neurosci., 16: 4275–4282.
Davis, M.H., Soderlund, T., Cole, J., Gadol, E., Kute, M., Fletcher, P.C., Happe, F., Frith, U., Baker, S.C., Dolan, R.J.,
Myers, M. and Weihing, J. (2004) Cognitions associated with Frackowiak, R.S. and Frith, C.D. (1995) Other minds in the
438

brain: a functional imaging study of ‘theory of mind’ in story Hatfield, E., Cacioppo, J.T. and Rapson, R. (1994) Emotional
comprehension. Cognition, 57: 109–128. Contagion. Cambridge University Press, New York.
Fourkas, A.D., Avenanti, A., Urgesi, C. and Aglioti, S.M. Hess, U. and Blairy, S. (2001) Facial mimicry and emotional
(2006) Corticospinal facilitation during first and third person contagion to dynamic emotional facial expressions and their
imagery. Exp. Brain Res., 168: 143–151. influence on decoding accuracy. Int. J. Psychophysiol.,
Frith, U. and Frith, C.D. (2003) Development and neurophys- 40: 129–141.
iology of mentalizing. Philos. Trans. R. Soc. Lond. B Biol. Hess, U., Blairy, S. and Philippot, P. (1999) Facial mimicry. In:
Sci., 358: 459–473. Philippot, P., Feldman, R. and Coats, E. (Eds.), The Social
Gallagher, H.L. and Frith, C.D. (2003) Functional imaging of Context of Nonverbal Behavior. Cambridge University
‘theory of mind’. Trends Cogn. Sci., 7: 77–83. Press, New York, pp. 213–241.
Gallagher, H.L., Jack, A.I., Roepstorff, A. and Frith, C.D. Hoffman, M.L. (1984) Interaction of affect and cognition in
(2002) Imagining the intentional stance in a competitive empathy. In: Izard, C., Kagan, J. and Zajonc, R. (Eds.),
game. Neuroimage, 16: 814–821. Emotions, Cognition, and Behavior. Cambridge University
Gallese, V. (2003a) The manifold nature of interpersonal rela- Press, New York, pp. 103–131.
tions: the quest for a common mechanism. Philos. Trans. R. Hommel, B., Müsseler, J., Aschersleben, G. and Prinz, W.
Soc. Lond. B. Biol. Sci., 358: 517–528. (2001) The theory of event coding (TEC): a framework
Gallese, V. (2003b) The roots of empathy: the shared manifold for perception and action planning. Behav. Brain Sci.,
hypothesis and the neural basis of intersubjectivity. Psycho- 24: 849–878.
pathology, 36: 171–180. Hutchison, W.D., Davis, K.D., Lozano, A.M., Tasker, R.R.
Gallese, V., Fadiga, L., Fogassi, L. and Rizzolatti, G. (1996) and Dostrovsky, J.O. (1999) Pain-related neurons in the
Action recognition in the premotor cortex. Brain, 119(Pt 2): human cingulate cortex. Nat. Neurosci., 2: 403–405.
593–609. Hynes, C.A., Baird, A.A. and Grafton, S.T. (2006) Differential
Gallese, V. and Goldman, A. (1998) Mirror neurons and the role of the orbital frontal lobe in emotional versus cognitive
simulation theory of mind-reading. Trends Cogn. Sci., perspective-taking. Neuropsychologia, 44: 374–383.
2: 493–501. Iacoboni, M., Woods, R.P., Brass, M., Bekkering, H.,
Gallese, V., Keysers, C. and Rizzolatti, G. (2004) A unifying view Mazziotta, J.C. and Rizzolatti, G. (1999) Cortical mecha-
of the basis of social cognition. Trends Cogn. Sci., 8: 396–403. nisms of human imitation. Science, 286: 2526–2528.
Garavan, H., Ross, T.J. and Stein, E.A. (1999) Right hemispheric Ickes, W. (1997) Empathic Accuracy. Guilford, New York.
dominance of inhibitory control: an event-related functional Jackson, D.C., Malmstadt, J.R., Larson, C.L. and Davidson, R.J.
MRI study. Proc. Natl. Acad. Sci. USA, 96: 8301–8306. (2000) Suppression and enhancement of emotional responses
Goldman, A. (2005) Mirror systems, social understanding to unpleasant pictures. Psychophysiology, 37: 515–522.
and social cognition. http://www.interdisciplines.org/mirror/ Jackson, P.L., Meltzoff, A.N. and Decety, J. (2005) How do we
papers/3/6/1/. perceive the pain of others? A window into the neural proc-
Goldman, A. (2006) Simulating Minds: The Philosophy, esses involved in empathy. Neuroimage, 24: 771–779.
Psychology, and Neuroscience of Mindreading. Oxford Kaplan, H.B. and Bloom, S.W. (1960) The use of sociological
University Press, New York. and social-psychological concepts in physiological research: a
Goldman, A.I. and Sebanz, N. (2005) Simulation, mirroring, and review of selected experimental studies. J. Nervous Mental
a different argument from error. Trends Cogn. Sci., 9: 320. Disord., 131: 128–134.
Gordon, R.M. (1986) Folk psychology as simulation. Mind Keysers, C. and Perrett, D.I. (2004) Demystifying social
Lang., 1: 158–170. cognition: a Hebbian perspective. Trends Cogn. Sci., 8:
Gordon, R.M. (2005) Simulation and systematic errors in 501–507.
prediction. Trends Cogn. Sci., 9: 361–362. Lane, R.D., Fink, G.R., Chau, P.M. and Dolan, R.J. (1997)
Gottman, J.M. and Levenson, R.W. (1985) A valid measure for Neural activation during selective attention to subjective
obtaining self-report of affect. J. Consult. Clin. Psychol., emotional responses. Neuroreport, 8: 3969–3972.
53: 151–160. Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (2005) Interna-
Grezes, J., Frith, C.D. and Passingham, R.E. (2004) Inferring tional affective picture system (IAPS): digitized photographs,
false beliefs from the actions of oneself and others: an fMRI instruction manual and affective ratings. Technical Report
study. Neuroimage, 24: 744–750. A-6.
Gross, J.J. (1999) Emotion regulation: past, present, future. Lanzetta, J.T. and Englis, B.G. (1989) Expectations of coop-
Cogn. Emotion, 13: 551–573. eration and competition and their effects on observers’
Gross, J.J. (2002) Emotion regulation: affective, cognitive, and vicarious emotional responses. J. Pers. Soc. Psychol.,
social consequences. Psychophysiology, 39: 281–291. 56: 543–554.
Gump, B.B. and Kulik, J.A. (1997) Stress, affiliation, and Leiberg, S., Eippert, F., Veit, R., Birbaumer, N. and Anders, S.
emotional contagion. J. Pers. Soc. Psychol., 72: 305–319. (2005) Prefrontal networks in empathy [abstract]. Presented
Harenski, C.L. and Hamann, S. (2006) Neural correlates of at the 11th Conference on Functional Mapping of the
regulating negative emotions related to moral violations. Human Brain, June 12–16, 2005, Toronto, Canada. Availa-
Neuroimage, 30: 313–324. ble on CD-Rom in Neuroimage, Vol. 26, Supplement 1.
439

Leibetseder, M., Laireiter, A.-R., Riepler, A. and Köller, T. Molnar-Szakacs, I., Iacoboni, M., Koski, L. and Mazziotta, J.C.
(2001) E-SKala: Fragebogen zur Erfassung von Empathie- (2005) Functional segregation within pars opercularis of the
Beschreibung und psychometrische Eigenschaften. Z. Differen. inferior frontal gyrus: evidence from fMRI studies of imita-
Diagn. Psychol., 22: 70–85. tion and action observation. Cereb. Cortex, 15: 986–994.
Leslie, A.M. (1994) Pretending and believing: issues in the the- Moll, J., Oliveira-Souza, R., Eslinger, P.J., Bramati, I.E.,
ory of ToMM. Cognition, 50: 211–238. Mourao-Miranda, J., Andreiuolo, P.A. and Pessoa, L. (2002)
Leslie, K.R., Johnson-Frey, S.H. and Grafton, S.T. (2004) The neural correlates of moral sensitivity: a functional mag-
Functional imaging of face and hand imitation: towards a netic resonance imaging investigation of basic and moral
motor theory of empathy. Neuroimage, 21: 601–607. emotions. J. Neurosci., 22: 2730–2736.
Levenson, R.W. (1996) Biological substrates of empathy and Morrison, I., Lloyd, D., di Pellegrino, G. and Roberts, N.
facial modulation of emotion: two facets of the scientific (2004) Vicarious responses to pain in anterior cingulate
legacy of John Lanzetta. Motiv. Emotion, 20: 185–204. cortex: is empathy a multisensory issue? Cogn. Affect. Behav.
Levenson, R.W. (2003) Blood, sweat, and fears. Ann. N Y Neurosci., 4: 270–278.
Acad. Sci., 1000: 348–366. Niedenthal, P.M., Brauer, M., Halberstadt, J.B. and Innes-
Levenson, R.W. and Ruef, A.M. (1992) Empathy: a physio- Ker, A.H. (2001) When did her smile drop? Facial mimicry
logical substrate. J. Pers. Soc. Psychol., 63: 234–246. and the influences of emotional state on the detection of
Levesque, J., Eugene, F., Joanette, Y., Paquette, V., Mensour, B., change in emotional expression. Cogn. Emotion, 15:
Beaudoin, G., Leroux, J.M., Bourgouin, P. and Beauregard, M. 853–864.
(2003) Neural circuitry underlying voluntary suppression of Ochsner, K.N. and Gross, J.J. (2005) The cognitive control of
sadness. Biol. Psychiatry, 53: 502–510. emotion. Trends Cogn. Sci., 9: 242–249.
Lipps, T. (1903) Einfühlung, innere Nachahmung, und Organ- Ochsner, K.N., Bunge, S.A., Gross, J.J. and Gabrieli, J.D.
empfindungen. Arch. gesamte Psychol., 1: 185–204. (2002) Rethinking feelings: an FMRI study of the cognitive
Lotze, M., Veit, R., Anders, S. and Birbaumer, N. The role of regulation of emotion. J. Cogn. Neurosci., 14: 1215–1229.
medial prefrontal cortex in the control of social-interactive Ochsner, K.N., Knierim, K., Ludlow, D.H., Hanelin, J.,
aggression, Submitted. Ramachandran, T., Glover, G. and Mackey, S.C. (2004a)
Lundquist, L.-O. and Dimberg, U. (1995) Facial expressions Reflecting upon feelings: an fMRI study of neural systems
are contagious. J. Psychophysiol., 9: 203–211. supporting the attribution of emotion to self and other.
Mathiak, K. and Weber, D.L. Towards brain correlates of J. Cogn. Neurosci., 16: 1746–1772.
natural behavior: fMRI during violent video games. Hum. Ochsner, K.N., Ray, R.D., Cooper, J.C., Robertson, E.R.,
Brain Mapp., electronically published. Chopra, S., Gabrieli, J.D. and Gross, J.J. (2004b) For better
McCabe, K., Houser, D., Ryan, L., Smith, V. and Trouard, T. or for worse: neural systems supporting the cognitive
(2001) A functional imaging study of cooperation in down- and up-regulation of negative emotion. Neuroimage,
two-person reciprocal exchange. Proc. Natl. Acad. Sci. 23: 483–499.
USA, 98: 11832–11835. Peyron, R., Laurent, B. and Garcia-Larrea, L. (2000) Func-
McHugo, G.J., Lanzetta, J.T., Sullivan, D.G., Masters, R.D. tional imaging of brain responses to pain. a review and
and Englis, B.G. (1985) Emotional reactions to a political meta-analysis (2000). Neurophysiol. Clin., 30: 263–288.
leader’s expressive displays. J. Pers. Soc. Psychol., Phan, K.L., Fitzgerald, D.A., Nathan, P.J., Moore, G.J., Uhde,
49: 1513–1529. T.W. and Tancer, M.E. (2005) Neural substrates for volun-
Mehrabian, A. (1997) Relations among personality scales of tary suppression of negative affect: a functional magnetic
aggression, violence, and empathy: validational evidence resonance imaging study. Biol. Psychiatry, 57: 210–219.
bearing on the risk of eruptive violence scale. Aggress. Premack, D. and Woodruff, G. (1978) Does the chimpanzee
Behav., 23: 433–445. have a theory of mind? Behav. Brain Sci., 1: 515–526.
Mehrabian, A. and Epstein, N. (1972) A measure of emotional Preston, S.D. and de Waal, F.B. (2002) Empathy: its ultimate
empathy. J. Pers., 40: 525–543. and proximate bases. Behav. Brain Sci., 25: 1–20.
Miller, E.K. and Cohen, J.D. (2001) An integrative theory Prinz, W. (1987) Ideomotor action. In: Heuer, H. and Sanders,
of prefrontal cortex function. Annu. Rev. Neurosci., A.F. (Eds.), Perspectives on Perception and Action.
24: 167–202. Erlbaum, Hillsdale, NJ, pp. 47–76.
Miller, P.A. and Eisenberg, N. (1988) The relation of empathy Reik, T. (1949) Character Analysis. Farrar, Strauss, Giroux,
to aggressive and externalizing/antisocial behavior. Psychol. New York.
Bull., 103: 324–344. Rizzolatti, G., Fadiga, L., Gallese, V. and Fogassi, L. (1996)
Mitchell, J.P. (2005) The false dichotomy between simulation Premotor cortex and the recognition of motor actions. Brain
and theory-theory: the argument’s error. Trends Cogn. Sci., Res. Cogn. Brain Res., 3: 131–141.
9: 363–364. Rizzolatti, G., Fogassi, L. and Gallese, V. (2001) Neurophys-
Mitchell, J.P., Banaji, M.R. and Macrae, C.N. (2005) The iological mechanisms underlying the understanding and
link between social cognition and self-referential thought imitation of action. Nat. Rev. Neurosci., 2: 661–670.
in the medial prefrontal cortex. J. Cogn. Neurosci., 17: Rogers, C. (1959) A theory of therapy, personality, and inter-
1306–1315. personal relationships as developed in the client-centered
440

framework. In: Koch, J.S. (Ed.) Psychology: a Study of a cardiophobic patients and their doctor during psychoana-
Science: Vol. 3. Formulations of the Person in the Social lytical inital interviews. Psychother. Psychosom., 22:
Context. McGraw-Hill, New York, pp. 184–256. 289–299.
Rogers, C. (1975) Empathic: an unappreciated way of being. Stotland, E. (1969) Exploratory investigations of empathy.
The Couns. Psychol., 2: 2–10. In: Berkowitz, L. (Ed.), Advances in Experimental Social
Ruby, P. and Decety, J. (2001) Effect of subjective perspective Psychology. Academic Press, New York, pp. 271–314.
taking during simulation of action: a PET investigation of Titchener, E. (1909) Experimental Psychology of the Thought
agency. Nat. Neurosci., 4: 546–550. Processes. Macmillan, New York.
Ruby, P. and Decety, J. (2003) What you believe versus what Trobst, K.K., Collins, R.L. and Embree, J.M. (1994) The role
you think they believe: a neuroimaging study of conceptual of emotion in social support provision: gender, empathy and
perspective-taking. Eur. J. Neurosci., 17: 2475–2480. expressions of distress. J. Soc. Pers. Relat., 11: 45–62.
Ruby, P. and Decety, J. (2004) How would you feel versus how Truax, C.B. (1961) A scale for measurement of empathy.
do you think she would feel? a neuroimaging study of Psychiatr. Inst., 1: 12.
perspective-taking with social emotions. J. Cogn. Neurosci., Vaughan, K.B. and Lanzetta, J.T. (1980) Vicarious instigation
16: 988–999. and conditioning of facial expressive and autonomic re-
Samson, D., Apperly, I.A., Chiavarino, C. and Humphreys, sponses to a model’s expressive display of pain. J. Pers. Soc.
G.W. (2004) Left temporoparietal junction is necessary Psychol., 38: 909–923.
for representing someone else’s belief. Nat. Neurosci., Vogeley, K., Bussfeld, P., Newen, A., Herrmann, S., Happe, F.,
7: 499–500. Falkai, P., Maier, W., Shah, N.J., Fink, G.R. and Zilles, K.
Samson, D., Apperly, I.A., Kathirgamanathan, U. and (2001) Mind reading: neural mechanisms of theory of mind
Humphreys, G.W. (2005) Seeing it my way: a case of a and self-perspective. Neuroimage, 14: 170–181.
selective deficit in inhibiting self-perspective. Brain, 128: Vogeley, K., May, M., Ritzl, A., Falkai, P., Zilles, K. and Fink,
1102–1111. G.R. (2004) Neural correlates of first-person perspective as
Saxe, R. (2005a) Against simulation: the argument from error. one constituent of human self-consciousness. J. Cogn. Ne-
Trends Cogn. Sci., 9: 174–179. urosci., 16: 817–827.
Saxe, R. (2005b) Hybrid vigour: reply to Mitchell. Trends Völlm, B.A., Taylor, A.N., Richardson, P., Corcoran, R., Stir-
Cogn. Sci., 9: 364. ling, J., McKie, S., Deakin, J.F. and Elliott, R. (2006) Neu-
Saxe, R. (2005c) On ignorance and being wrong: reply to ronal correlates of theory of mind and empathy: a functional
Gordon. Trends Cogn. Sci., 9: 362–363. magnetic resonance imaging study in a nonverbal task. Ne-
Shamay-Tsoory, S.G., Lester, H., Chisin, R., Israel, O., uroimage, 29: 90–98.
Bar-Shalom, R., Peretz, A., Tomer, R., Tsitrinbaum, Z. Vorauer, J.D. and Ross, M. (1999) Self-awareness and feeling
and Aharon-Peretz, J. (2005a) The neural correlates of transparent: failing to suppress one’s self. J. Exp. Soc.
understanding the other’s distress: a positron emission tomo- Psychol., 35: 415–440.
graphy investigation of accurate empathy. Neuroimage, Walter, H., Adenzato, M., Ciaramidaro, A., Enrici, I., Pia, L.
27: 468–472. and Bara, B.G. (2004) Understanding intentions in social
Shamay-Tsoory, S.G., Tomer, R., Berger, B.D., Goldsher, D. interaction: the role of the anterior paracingulate cortex.
and Aharon-Peretz, J. (2005b) Impaired ‘‘affective theory of J. Cogn. Neurosci., 16: 1854–1863.
mind’’ is associated with right ventromedial prefrontal Wicker, B., Keysers, C., Plailly, J., Royet, J.P., Gallese, V. and
damage. Cogn. Behav. Neurol., 18: 55–67. Rizzolatti, G. (2003) Both of us disgusted in My insula: the
Singer, T., Seymour, B., O’Doherty, J., Kaube, H., Dolan, R.J. common neural basis of seeing and feeling disgust. Neuron,
and Frith, C.D. (2004) Empathy for pain involves the affective 40: 655–664.
but not sensory components of pain. Science, 303: 1157–1162. Wiesenfeld, A.R., Whitman, P.B. and Malatesta, C.Z. (1984)
Singer, T., Seymour, B., O’Doherty, J.P., Stephan, K.E., Individual differences among adult women in sensitivity to
Dolan, R.J. and Frith, C.D. (2006) Empathic neural re- infants: evidence in support of an empathy concept. J. Pers.
sponses are modulated by the perceived fairness of others. Soc. Psychol., 46: 118–124.
Nature, 439: 466–469. Wispé, L. (1986) The distinction between sympathy and empa-
Sonnby-Borgström, M. (2002) Automatic mimicry reactions as thy: to call forth a concept, a word is needed. J. Pers. Soc.
related to differences in emotional empathy. Scand. J. Psychol., 50: 314–321.
Psychol., 43: 433–443. Zajonc, R.B., Adelmann, P.K., Murphy, S.T. and Niedenthal,
Stanek, B., Hahn, R. and Mayer, H. (1973) Biometric findings P.M. (1987) Convergence in the physical appearance of
on cardiac neurosis. III. Changes in ECG and heart rate in spouses. Motiv. Emotion (Hist. Arch.), 11: 335–346.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 24

Partly dissociable neural substrates for recognizing


basic emotions: a critical review

Andreas Hennenlotter1, and Ulrike Schroeder2

1
Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain Sciences, StephanstraX e 1A,
D-04103 Leipzig, Germany
2
Klinik Holthausen, Am Hagen 20, D-45527 Hattingen, Germany

Abstract: Facial expressions are powerful non-verbal displays of emotion which signal valence information
to others and constitute an important communicative element in social interaction. Six basic emotional
expressions (fear, disgust, anger, surprise, happiness, and sadness) have been shown to be universal in their
performance and perception. Recently, a growing number of clinical and functional imaging studies have
aimed at identifying partly dissociable neural subsystems for recognizing basic emotions. Convincing re-
sults have been obtained for fearful and disgusted facial expressions only. Empirical evidence for a spe-
cialized neural representation of anger, surprise, sadness, or happiness is more limited, primarily due to lack
of clinical cases with selective impairments in recognizing these emotions. In functional imaging research,
the detection of dissociable neural responses requires direct comparisons of signal changes associated with
the perception of different emotions, which are often not provided. Only recently has evidence been
obtained that the recruitment of emotion-specific neural subsystems may be closely linked to characteristic
facial features of single expressions such as the eye region for fearful faces. Investigations into the neural
systems underlying the processing of such diagnostic cues for each of the six basic emotions may be helpful
to further elucidate their neural representation.

Introduction While the argument for phylogenetic continuity


plays an important role in contemporary explana-
The formal evolutionary treatment of human fa- tions of emotions, Darwin’s vestigialism has largely
cial expressions began with Charles Darwin’s ‘‘The been replaced by the view that expressions of
Expression of the Emotions in Man and Animals’’ emotion are adaptive and had been selected
(Darwin, 1872). Darwin found evidence for con- for social communication (Schmidt and Cohn,
tinuity in bodily movements and facial gestures 2001). Emotion that is manifested by facial ex-
that humans shared with animals. He used these pression signals occurrences of value, and
resemblances across species to argue for common being able to transfer and receive such information
descent. Darwin’s view of facial expressions, how- undoubtedly confers a survival advantage. It is
ever, was not ‘‘evolutionary’’ at all, because he did generally accepted today that six basic emotional
not consider them as adaptations but accidents or expressions (happiness, surprise, fear, sadness, an-
vestiges of earlier evolutionary stages in which the ger, and disgust) are universal in their performance
intellect was of less importance (Fridlund, 1994). and in their perception (Ekman et al., 1969, 1987).
Motivated by recent advances in cognitive
Corresponding author. Tel.: +49-341-99-40-266; Fax: +49- neurosciences, a growing number of clinical and
341-99-40-260; E-mail: hennen@cbs.mpg.de functional imaging studies have aimed at identifying

DOI: 10.1016/S0079-6123(06)56024-8 443


444

common as well as dissociable neural substrates as- information to conspecifics that a novel stimulus is
sociated with the recognition of basic emotions. aversive and should be avoided (Mineka and
Posterior occipito-temporal regions have been asso- Cook, 1993). Thereby, fearful faces enable a quick
ciated with the perceptual analysis of facial expres- and defensive response that is critical for survival.
sive features (Haxby et al., 2000). The extraction of Accordingly, fear expressions are recognized sig-
emotional meaning from faces has been linked to nificantly faster out of visual noise than neutral
the ventral prefrontal cortex and somatosensory- faces (Reinders et al., 2005) and their processing
related regions of the right hemisphere where lesions occurs automatically, largely independently of at-
result in a general impairment in facial emotion tention and awareness (Esteves et al., 1994; Oh-
recognition (Rapcsak et al., 1993; Adolphs et al., man et al., 2001).
1996, 2000; Hornak et al., 1996). Consistent with Human lesion studies have found impaired rec-
these findings, functional imaging studies have re- ognition of fearful faces following amygdala dam-
vealed ventral prefrontal and somatosensory-related age (Adolphs et al., 1994; Young et al., 1995;
activations primarily in association with explicit Calder et al., 1996; Sprengelmeyer et al., 1999;
recognition tasks such as labeling expressed emotion Anderson and Phelps, 2000; Sato et al., 2002). The
and facial expression matching (Nakamura et al., first case (SM) was described by Adolphs et al.
1999; Narumoto et al., 2000; Gorno-Tempini et al., (1994, 1995). SM suffers from Urbach–Wiethe
2001; Winston et al., 2003). disease that has caused a nearly complete bilateral
Findings from clinical studies in patients destruction of the amygdala as well as a small
with focal brain lesions or specific neurological dis- portion of the adjacent entorhinal cortex, while
orders have also provided evidence that recognition sparing other subcortical and cortical regions. In
can be impaired for specific emotions such as fear comparison to control subjects (unilateral am-
(e.g., Adolphs et al., 1999) and disgust (e.g., Spreng- ygdala damage, other brain damage, and healthy
elmeyer et al., 1996; Gray et al., 1997). Accordingly, subjects), SM showed abnormal ratings of facial
functional imaging studies revealed that certain expressions of fear and, to a lesser extent, anger
brain regions are primarily implicated in the process- and surprise that could not be explained by deficits
ing of specific emotions (e.g., Morris et al., 1996; in basal cognitive or visual functions. Impairment
Phillips et al., 1997), suggesting that processing of of fear recognition was also observed for patients
individual emotions might be subserved by partly with right unilateral anteromedial temporal lobe-
separable neural subsystems (Calder et al., 2001). ctomies that included the amygdala (Anderson
In this review, we will critically examine findings et al., 2000). In a larger group involving nine pa-
from lesion, neurophysiologic, and functional im- tients with bilateral amygdala damage, individual
aging studies with respect to the question of whether performances of fear recognition varied consider-
recognition of basic emotions is subserved by partly ably, ranging from severely impaired to completely
dissociable neural subsystems. By including each of normal (Adolphs et al., 1999). Preserved fear rec-
the six basic emotions, we provide an important ex- ognition ability for two of those patients with bi-
tension to previous reviews that either focused on lateral amygdala lesions (GT and EP) is consistent
the two most extensively studied emotions, namely with an earlier study that found no significant im-
fear and disgust (Calder et al., 2001), or examined pairments in GT’s and EP’s ratings of any emotion
the neural basis of face (Posamentier and Abdi, (Hamann et al., 1996). However, a re-examination
2003) and facial expression processing (Adolphs, of these two patients and a third patient, GP
2002a, 2002b; Blair, 2003) more generally. (Schmolck and Squire, 2001), using a different
method of analyzing these patients’ ratings and an
additional forced-choice labeling task, revealed
Fear impairments of all three patients, particularly for
recognizing fear and sadness. Besides deficits
Facial expressions of fear have been described as in fear recognition, patients with amygdala dam-
unconditioned stimuli that allow to rapidly convey age are often impaired on recognition of other
445

emotions (e.g., Young et al., 1995; Calder et al., of Morris et al. (1996, 1998) additionally allowed
1996; Young et al., 1996; Broks et al., 1998; Ado- comparing signal changes associated with the
lphs et al., 1999; Rapcsak et al., 2000; Schmolck perception of different expression intensities of
and Squire, 2001; Sato et al., 2002). However, the same emotion. Both studies found a differen-
given the heterogeneity of amygdala lesions (bi- tial response to fearful faces relative to happy
lateral/unilateral, complete/partial) and various faces within the left amygdala, which increased
etiologies (congenital, encephalitis, and surgical), with increasing intensity of fear expressions. The
the finding of impaired fear recognition is sur- failure to detect activation of the right amygdala
prisingly consistent. Furthermore, varying de- may be explained by the finding of more rapid
grees of additional damage to other brain regions habituation of the right than the left amygdala
observed in these patients may also contribute to (Phillips et al., 2001; Wright et al., 2001). It is
recognition deficits in more than just one emotion worth mentioning that dissociable responses
(Rapcsak et al., 2000). within the left amygdala (revealed by direct com-
Using various techniques (e.g., ‘‘Bubbles,’’ parisons of signal changes associated with the
Gosselin and Schyns, 2001; Schyns et al., 2002) perception of different emotions) have not always
to estimate which aspects of a facial expression are been found for perception of fearful faces (Wins-
most important when recognizing specific emo- ton et al., 2003), and occasionally have been ob-
tions, a recent study elucidated the mechanism by served for emotions other than fear, such as
which amygdala damage may compromise fear sadness (Blair et al., 1999) and disgust (Gorno-
recognition (Adolphs et al., 2005). The authors Tempini et al., 2001).
could show that the fear recognition deficit of a In line with evidence from a recent lesion study
patient with complete bilateral amygdala lesions (Adolphs et al., 2005), findings derived from func-
(SM) is caused by an inability to use information tional imaging research have also provided evi-
from the eye region of faces, which is the most dence that the amygdala may be specifically
important feature for identifying this emotion. Her involved in the processing of information from
deficit in recognizing fearful expressions com- the eye region of faces. These studies showed that
pletely disappeared when she was instructed to amygdala responses to fearful faces are modulated
explicitly look at the eyes. Interestingly, the eye by the degree of threat-related ambiguity as a
region is also important for recognizing anger and function of gaze direction (Adams et al., 2003),
sadness (Smith et al., 2005), which, apart from that fearful eyes alone are sufficient to produce
fear, has often been reported to be impaired after amygdala responses (Morris et al., 2002), and that
amygdala damage (Adolphs et al., 1999). these responses appear to be driven by the size of
Numerous imaging studies on facial emotion the white scleral field and not by the outline of the
recognition have found amygdala activation in eye (Whalen et al., 2004). Findings from lesion and
response to fearful faces (Breiter et al., 1996; functional imaging research therefore provide con-
Morris et al., 1996, 1998; Phillips et al., 1997, verging evidence concerning both, the important
1998b, 2004; Sprengelmeyer et al., 1998; Whalen role of the amygdala in fear recognition and the
et al., 1998, 2001; Thomas et al., 2001; Wright et specific facial cues that it may be attuned to.
al., 2001; Yang et al., 2002; Fischer et al., 2003).
Some of these studies reported a dissociable re-
sponse to fearful faces, particularly within the left Disgust
amygdala. In these studies, signal changes asso-
ciated with the perception of fear were directly The facial expression of disgust signals important
compared to signal changes elicited by the per- information regarding the quality of food, poten-
ception of a different emotion such as happiness tial physical contamination and disease, and
(Morris et al., 1996, 1998; Whalen et al., 1998; induces avoidance behavior in conspecifics. The
Wright et al., 2001), disgust (Phillips et al., 1998b, adaptive significance of disgust has been related to
2004), or anger (Whalen et al., 2001). The studies a specific form of threat response associated with
446

an internal defense system, as opposed to an 1998; Wicker et al., 2003; Schroeder et al., 2004a).
external defense system related to fear (Calder The insula was found to be activated in response
et al., 2001). to disgusted faces in two imaging studies that in-
The first evidence that perception of disgusted volved direct comparisons with fearful (Phillips et
faces might be associated with a particular brain al., 1998b) and surprised facial expressions
region came from an investigation on emotion (Schroeder et al., 2004a), while the putamen was
recognition in people with Huntington’s disease, a found to be activated in one of these studies only
dominantly inherited neurodegenerative disorder. (Phillips et al., 1998b). Recently, findings from a
These patients showed a disproportionately severe combined clinical and functional imaging study
impairment in recognizing disgusted faces (Spreng- elucidated the neural mechanisms that may
elmeyer et al., 1996), which was later replicated in underlie impaired disgust recognition in pre-symp-
a sample of Chinese patients suffering from Hunt- tomatic Huntington’s disease (Hennenlotter et al.,
ington’s disease (Wang et al., 2003). Subsequent 2004). In Huntington’s disease gene carriers,
studies in pre-symptomatic Huntington’s disease perception of disgusted relative to neutral faces
gene carriers revealed a selective deficit in recog- was associated with significantly decreased activa-
nizing disgust, i.e., recognition of other emotions tion of the left anterior insulo-opercular region,
was intact or only mildly impaired (Gray et al., closely corresponding to the location of the pri-
1997; Hennenlotter et al., 2004; Sprengelmeyer mary taste cortex in humans (Frey and Petrides,
et al., 2005). Only one study reported a more gen- 1999; Small et al., 1999, 2003). This finding is
eralized deficit of emotion recognition in people consistent with the notion that perception of oth-
with manifest Huntington’s disease and no emo- ers’ disgust and that of taste are closely linked
tion recognition deficit in people at risk of carrying (Rozin and Fallon, 1987) and probably share a
the Huntington gene (Milders et al., 2003). Since similar neural substrate (Phillips et al., 1997).
Huntington’s disease has been conceptualized pri- In conclusion, both clinical and functional
marily as a basal ganglia disorder, Sprengelmeyer imaging studies suggest a specific involvement of
et al. (1996) proposed that disgust recognition may the insula and basal ganglia in disgust recognition,
be closely associated with the basal ganglia. How- although the differential functions of these regions
ever, pathological changes in Huntington’s disease in recognizing disgust are still open to question.
are not confined to the striatum, but also affect
cortical regions (de la Monte et al., 1988; Jernigan
et al., 1991), including the insular cortex (Thieben Anger
et al., 2002). The latter region has been implicated
in impaired disgust recognition by a single case Recognition of angry faces has been proposed to
study of a patient who suffered from focal damage serve several, partly overlapping adaptive func-
to the left insula and putamen (Calder et al., 2000). tions each associated with specific neuroanatomi-
Moreover, Krolak-Salmon et al. (2003) demon- cal circuitries. In the context of a defense system
strated intracerebral event-related potentials to fa- relating to threats to the acquisition of valuable
cial expressions of disgust from insular contacts in resources, the ventral striatum dopamine system
patients suffering from drug-refractory temporal has been suggested to be implicated in the recog-
lobe epilepsy. Apart from neural degeneration of nition of angry faces as signals of conspecific chal-
the basal ganglia, early neural loss in the insula lenge (Lawrence et al., 2002). Angry expressions
may be considered as a possible explanation have also been proposed to signal discontent in
for impaired disgust recognition in Huntington’s order to discourage socially inappropriate and un-
disease. expected behavior (Averill, 1982). They may thus
Converging evidence for a role of both the ins- serve as a cue for behavioral extinction and re-
ula and basal ganglia in disgust recognition comes versal learning which has been closely associated
from functional imaging studies in healthy subjects with the orbitofrontal cortex (Dias et al., 1996;
(Phillips et al., 1997, 1998b; Sprengelmeyer et al., Rolls, 1996). Finally, as threatening stimuli, angry
447

expressions might engage the amygdala similar to impairments. Given that functional imaging studies
expressions of fear (Blair et al., 1999). have implicated ventrolateral prefrontal regions in
In support of the notion that angry expressions processing of angry faces (Sprengelmeyer et al.,
are processed as threatening stimuli, similar to ex- 1998; Blair et al., 1999; Phillips et al., 1999), the
pressions of fear, recognition of anger has been authors propose that human signals of aggression
most consistently reported to be impaired after are processed by a frontostriatal system.
amygdala damage in addition to fear (Adolphs A role for frontostriatal circuitries in anger rec-
et al., 1994, 1994; Young et al., 1995; Calder et al., ognition is in line with findings of a recent study
1996; Broks et al., 1998; Sato et al., 2002). A role (Schroeder et al., 2004b) investigating facial ex-
for the amygdala in anger recognition is further pression processing in Parkinson’s disease patients
consistent with the finding that bilateral amygdala with subthalamic nucleus (STN) deep brain stim-
damage impairs the use of information from the ulation. STN stimulation is an accepted form of
eye region (Adolphs et al., 2005) which is, apart treatment for patients with Parkinson’s disease
from fear, also the diagnostic region for recogniz- who have medically intractable motor symptoms.
ing anger expressions (Smith et al., 2005). How- Schroeder et al. (2004b) found that STN stimula-
ever, there is also evidence for preserved anger tion selectively reduces recognition of angry faces,
recognition in some patients with amygdala lesions but leaves intact recognition of other emotions.
(Adolphs et al., 1994; Calder et al., 1996) and post- Since in animals the STN is also targeted by limbic
encephalitic patients with amygdala lesions (Broks cortices (Parent and Hazrati, 1995) such as the
et al., 1998). orbitofrontal and anterior cingulate cortex (Cant-
On the basis of observations from comparative eras et al., 1990), the results of the study point at a
research demonstrating altered dopamine activity possible role of these regions in anger recognition.
during aggressive encounters between conspecifics In fact, lesions involving the orbitofrontal cortex
(Redolat et al., 1991; Miczek et al., 2002; Ferrari have been associated with changes in aggressive
et al., 2003) and evidence pointing at a contribution behavior in humans (Blair, 2001; Brower and
of dopamine to humans’ aggressive behavior (Tii- Price, 2001) and difficulties in recognizing angry
honen et al., 1995; Lawrence et al., 2003), a sep- and disgusted faces (Blair and Cipolotti, 2000).
arate line of investigations focused on the role of Furthermore, transcranial magnetic stimulation
the dopamine system and the ventral striatum in (TMS) over the medial-frontal cortex has been
anger recognition (Lawrence et al., 2002; Calder found to impair processing of angry, but not
et al., 2004). In a recent study, Lawrence et al. happy facial expressions (Harmer et al., 2001b).
(2002) showed that after acute administration of Findings from functional imaging studies have
the dopamine D2-class receptor antagonist sulpi- implicated the orbitofrontal cortex (Blair et al.,
ride, subjects are significantly worse at recognizing 1999), anterior cingulate (Sprengelmeyer et al.,
angry faces, though there are no such impairments 1998; Blair et al., 1999; Strauss et al., 2005), as well
in recognizing other facial expressions. In a fol- as the amygdala (Whalen et al., 2001; Yang et al.,
lowing study, Calder et al. (2004) investigated 2002; Adams et al., 2003) in processing of angry
whether recognition of facial and vocal signals of faces. The finding of orbitofrontal cortex activa-
anger and self-reported experience of anger would tion in response to angry faces by Blair et al.
be affected by damage to the ventral striatal do- (1999) might be of particular interest, since this
pamine system. To address this question, they region showed a differential response to angry
studied a case series of four human subjects with faces (when compared to sad faces). Additionally,
focal lesions affecting the ventral striatum. All four increasing intensity of angry expressions was
demonstrated a disproportionately severe impair- found to be associated with enhanced activity of
ment in recognizing human signals of aggression, the orbitofrontal cortex. Convergent evidence has
in particular facial expressions, whereas a control been provided by two meta-analyses that indicated
group of individuals with damage to more dorsal a particular role for the lateral orbitofrontal cortex
basal ganglia regions showed no signs of anger in processing of angry faces (Phan et al., 2002;
448

Murphy et al., 2003). When activated by angry negative interpretations of surprised faces were
expressions, the orbitofrontal cortex has been sug- associated with greater signal changes in the right
gested to suppress current behavior either through ventral amygdala, while more positive interpreta-
inhibition or by activation of an alternative be- tions were associated with greater signal changes
havioral response (Blair et al., 1999). Other studies in the ventral medial prefrontal cortex. Perception
reported amygdala activation in response to angry of surprised faces relative to neutral faces resulted
faces (Whalen et al., 2001; Yang et al., 2002), in in dorsal amygdala activation. Similar results were
particular when they signal ambiguous threats in obtained when the interpretation of surprised faces
the case of averted gaze (Adams et al., 2003). was determined by contextual experimental stim-
However, none of these imaging studies reported a uli, rather than subjective judgment (Kim et al.,
dissociable response of the amygdala to angry 2004). However, since none of these studies in-
expressions. Notably, a recent study found that cluded emotional expressions other than surprise,
angry faces evoke sensitization in several regions the question of whether a specific neural subsystem
including the anterior cingulate and basal ganglia is involved in surprise perception remains open.
but not within the amygdala (Strauss et al., 2005) The first study that aimed at identifying a spe-
where fearful faces have reproducibly evoked ha- cialized neural system for surprise perception
bituation (Breiter et al., 1996; Wright et al., 2001; (Schroeder et al., 2004a) was based on a psycho-
Fischer et al., 2003). Findings from functional im- logical model where surprise is conceived as an
aging research are therefore more consistent with evolutionary old mechanism to analyze unex-
the proposed role of frontal regions such as the pected events in order to update knowledge for
orbitofrontal (Blair et al., 1999) and anterior successful individual-environmental transaction
cingulate cortex (Sprengelmeyer et al., 1998; Blair (Schutzwohl, 1998). Regions implicated in the
et al., 1999; Strauss et al., 2005) in anger recog- detection of novel or unexpected events were
nition than with a specific involvement of the therefore expected to be specifically involved in
amygdala. the perception of surprised faces. Compared to
both neutral and disgusted faces, perception of
surprised faces consistently yielded activation of
Surprise the right posterior parahippocampal gyrus. In fact,
the right parahippocampal gyrus was the only
The facial expression of surprise has already been region that showed significant activation in the
described by Charles Darwin in 1872, who pro- direct comparison with disgusted faces. Since
posed novelty and unexpectedness as elicitors. the parahippocampal gyrus has been implicated
Since surprise can predict a positive as well as a in the processing of novel stimuli as compared to
negative outcome (Tomkins and McCarter, 1964), familiar stimuli (Stern et al., 1996; Gabrieli et al.,
it may be considered as the most controversial ex- 1997), perception of surprise in others may be
pression of all six basic emotions. To date only a closely related to the recognition or evaluation
single case study reported impaired surprise rec- of novel stimuli in the environment which is
ognition following bilateral amygdala damage thought of as an initial step in memory formation
(Adolphs et al., 1994). However, SM’s impairment (Fernandez et al., 1998).
in recognizing surprise was not selective since she Given the small number of functional imaging
also showed deficits in recognizing other emotions studies that focused on surprise perception and the
that were most pronounced for fear. considerable variations in experimental designs
Three imaging studies have investigated the and data analyses, the different roles of the am-
neural basis of surprise perception so far (Kim ygdala and the parahippocampal gyrus in recog-
et al., 2003, 2004; Schroeder et al., 2004a). The nizing this emotion remain unclear. With respect
study by Kim et al. (2003) aimed at investigating to a possible role of the amygdala in surprise per-
the neural correlates implicated in positive versus ception, imaging paradigms are needed that also
negative evaluation of surprised faces. More involve the presentation of fearful faces that have
449

been found to specifically engage the amygdala. is the diagnostic region for happiness (Smith et al.,
Studies on emotion recognition in patients with 2005), remains normal (Adolphs et al., 2005).
amygdala lesions, however, generally fail to find Findings from neuroimaging studies on percep-
impairments in surprise recognition. It is therefore tion of happy faces revealed no consistent pattern
interesting to note that the only case of a patient of activation. Various regions have been impli-
with concomitant deficits in surprise recognition cated in the perception of happiness, including the
(SM) suffered from lesions of the amygdala that basal ganglia (Morris et al., 1996, 1998), inferior/
extended into the enthorinal cortex (Adolphs et orbitofrontal cortex (Dolan et al., 1996; Gorno-
al., 1994), a part of the parahippocampal gyrus. Tempini et al., 2001), anterior cingulate cortex
Given the robust activation of the posterior para- (Dolan et al., 1996; Kesler-West et al., 2001), and
hippocampal gyrus during perception of surprised amygdala (Breiter et al., 1996; Pessoa et al., 2002;
faces (Schroeder et al., 2004a), one might speculate Yang et al., 2002; Winston et al., 2003; Hennenl-
that this region is specifically involved in the per- otter et al., 2005). In their meta-analysis, Phan et
ception of surprise. In order to substantiate this al. (2002) found that nearly 70% happiness induc-
finding in the future, studies are needed that tion studies reported activation in the basal gan-
investigate emotion recognition in patients with glia/ventral striatum, which is consistent with
focal lesions of the posterior medial temporal work implicating the dopaminergic system and
lobes. basal ganglia/ventral striatum in reward process-
ing and incentive reward motivation (Koepp et al.,
1998; Knutson et al., 2001). This finding, however,
could not be replicated in a second meta-analysis
Happiness by Murphy et al. (2003) for facial expressions of
happiness. In a sub-analysis of studies that used
Smile has been suggested to be an important signal facial expressions of emotion as stimuli, the rostral
of cooperative intention and affiliation during so- supracallosal anterior cingulate cortex and dorso-
cial interaction (Schmidt and Cohn, 2001). Human medial prefrontal cortex were most consistently
infants smile more when an adult’s eye gaze is fixed activated for happiness instead.
on them than when the gaze is averted (Haines and Whereas some studies reported signal increases
Muir, 1969). As a positive fitness consequence, in the amygdala to positively valenced facial ex-
these smiles subsequently elicit responsive and at- pressions (Breiter et al., 1996; Pessoa et al., 2002;
tentive parental behavior (Jones and Raag, 1989). Yang et al., 2002; Winston et al., 2003; Hennenl-
Smiling is also the most easily recognized expres- otter et al., 2005), others do not or have found
sion. Following the norms published by Ekman signal decreases in the amygdala (Morris et al.,
and Friesen (1976), mean accuracy for recognition 1996; Whalen et al., 1998). In the study of Whalen
of facial expressions of happiness reaches 100% et al. (1998), signal decreases in response to happy
(Young et al., 1996). faces were found in the ventral amygdala, whereas
To date, there is no evidence for impaired rec- both fearful and happy faces activated the sublen-
ognition of happy faces following damage of spe- ticular substantia innominata (SI), suggesting a
cific brain regions or in patients suffering from spatial dissociation of regions that respond to
neurological or psychiatric disorders. Only one emotional valence (ventral amygdala including the
patient with amygdala damage was reported to be basolateral amygdala) versus salience or arousal
impaired in her appraisal of happiness to a lesser value (dorsal amygdala/SI region including the
extent (Anderson and Phelps, 2000). The preserved central nucleus of the amygdala) (Whalen et al.,
ability to recognize happy faces found in most pa- 1994, 1998; Somerville et al., 2004). Given that
tients with amygdala lesion may be explained by the amygdala’s activity is enhanced by faces con-
the finding that damage of the amygdala impairs taining dynamic information (LaBar et al., 2003),
the use of information from the eyes, whereas the inconsistencies in findings concerning its involve-
use of information from the mouth region, which ment in pleasant facial affect may be related to the
450

lack of temporal cues in static facial displays that expressions (Anderson and Phelps, 2000; Anderson
have mostly been used in these studies. This notion et al., 2000). Most patients with bilateral amygdala
is supported by the finding of a recent study by lesions, however, have been found to be unim-
Hennenlotter et al. (2005) who reported robust paired in recognizing sad faces (e.g., Adolphs et al.,
activation of the bilateral amygdala during passive 1994, 1999; Calder et al., 1996). Rather than being
observation of dynamic smile expressions (short necessary for recognizing sad faces, the amygdala
video sequences) compared to observation of neu- may be involved in the activation of concomitant
tral expressions of the same individuals. Beyond its autonomic responses (Blair et al., 1999). In line
well-known involvement in processing fearful with this notion, administration of the beta-ad-
faces, these findings suggest that the amygdala renoceptor blocker propranolol was found to im-
(in particular its dorsal parts) may play a gener- pair the speed at which sad faces are recognized
alized role in facial emotion processing by mod- (Harmer et al., 2001a).
ulating the vigilance level in response to both In a functional magnetic resonance imaging
positive and negative facial expressions (Yang et (fMRI) study, Blair et al. (1999) reported that the
al., 2002; Winston et al., 2003). Since recognition neural response in the left amygdala, right inferior
of happy faces is usually not impaired in patients and middle temporal gyri was significantly greater
with amygdala lesion, it is, however, implausible toward sad expressions as opposed to angry ex-
that the amygdala is part of a neural network that pressions. Activation within these regions in-
is more specifically involved in the representation creased as a function of the degree of sadness
of pleasant facial affect. expressed in the morphed face stimuli. The ante-
rior cingulate cortex and right temporal pole cor-
related with increasing expression intensity of both
Sadness sad and angry faces. Unfortunately, the study did
not include fearful faces as control stimuli that
The facial expression of sadness has been linked to would have allowed testing whether there is a
the inhibition of aggression and the elicitation differential response in the amygdala to sad versus
of empathic reactions and prosocial behavior (Ei- fearful facial expressions (Posamentier and Abdi,
senberg et al., 1989; Harmer et al., 2001a). Ac- 2003). Activations of the right inferior and middle
cordingly, psychopaths, who are characterized by temporal gyri may reflect top-down modulatory
disregard for others and aggressive behaviors, effects of the amygdala onto the visual processing
fail to show normal autonomic responses to sad stream (Morris et al., 1998; Vuilleumier et al.,
facial expressions (House and Milligan, 1976; An- 2001) and have also been reported for perception
iskiewicz, 1979; Chaplin et al., 1995; Blair et al., of other emotions (e.g., Kesler-West et al., 2001;
1997). Psychopathy has been suggested to reflect Vuilleumier et al., 2001, 2003; Schroeder et al.,
early amygdala damage (Blair et al., 1999; Blair, 2004a). Involvement of the amygdala during per-
2001) since psychopathic individuals show impair- ception of sad faces has been confirmed recently by
ments in fear conditioning and augmentation of two fMRI studies (Yang et al., 2002; Winston
startle reflex (e.g., Patrick et al., 1993; Lucas et al., et al., 2003). In contrast to the findings of Blair
2000) similar to patients with amygdala lesions et al. (1999), however, Winston et al. (2003) failed
(Bechara et al., 1995; LaBar et al., 1995; Angrilli et to find evidence for a differential response of the
al., 1996). Given the hyporesponsiveness to sad amygdala to sad faces. Yang et al. (2002) only
faces and indications of amygdala dysfunction in compared patterns of activation for emotional
these patients, the amygdala has been suggested to faces (including sadness) relative to neutral faces
be involved in the processing of sad facial expres- leaving it open whether the amygdala showed a
sions (Blair et al., 1999). This suggestion is sup- dissociable response to sad faces. Two further
ported by the finding that some patients with right studies (Phillips et al., 1998a; Kesler-West et al.,
amygdaloid lesion show deficits in judging the in- 2001) found no activation of the amygdala for sad
tensity of several negative emotions including sad faces relative to neutral faces.
451

In conclusion, evidence for the existence of a emotion from regions more generally involved in
specific neural subsystem for recognizing sad faces facial expression processing. Such inferences are
is still lacking. Amygdala activations during per- not possible when only neutral facial expressions
ception of sad faces have been related to the ac- are used as baseline condition. Unfortunately,
tivation of concomitant autonomic responses studies often fail to provide direct comparisons,
(Blair et al., 1999) and may reflect modulation of making it difficult to evaluate the findings with
the vigilance level in response to emotionally respect to emotion-specific neural subsystems.
valenced stimuli in general (Yang et al., 2002; Some have argued that in functional imaging
Winston et al., 2003). studies on emotion recognition, differential habit-
uation effects in emotion processing regions might
account for the finding of emotion-specific re-
General conclusion sponses of some regions, particularly in the case of
repeated presentations of the same expression as in
Convincing results for specialized neural represen- block design paradigms (Winston et al., 2003;
tations of basic emotions have been obtained for Strauss et al., 2005). Converging evidence from
fearful and disgusted faces only. Clinical and func- clinical research on emotion recognition in pa-
tional imaging studies have provided converging tients with amygdala lesions (e.g., Adolphs et al.,
evidence for a double dissociation of fear and dis- 1999) and Huntington’s disease (e.g., Sprengelm-
gust recognition. Whereas the amygdala is prima- eyer et al., 1996), however, clearly backs up the
rily involved in recognizing facial expressions of findings derived from functional imaging studies
fear (Adolphs et al., 1994; Young et al., 1995; on fear and disgust recognition (e.g., Phillips et al.,
Breiter et al., 1996; Calder et al., 1996; Morris et 1997, 1998b; Sprengelmeyer et al., 1998). Moreo-
al., 1996, 1998; Phillips et al., 1997, 1998b, 2004; ver, imaging studies on fear recognition consist-
Sprengelmeyer et al., 1998, 1999; Whalen et al., ently yielded activation of the amygdala despite
1998, 2001; Anderson and Phelps, 2000; Thomas et the fact that considerable habituation effects have
al., 2001; Wright et al., 2001; Sato et al., 2002; been reported for this regions (Breiter et al., 1996;
Yang et al., 2002; Fischer et al., 2003), the insular Wright et al., 2001; Fischer et al., 2003).
cortex and basal ganglia are implicated in recog- It is further important to note that, in functional
nizing facial expressions of disgust but not fear imaging research mostly implicit tasks (e.g., gender
(Sprengelmeyer et al., 1996, 1998, 2005; Gray et classification) have been used, whereas in clinical
al., 1997; Phillips et al., 1997, 1998b; Calder et al., studies subjects are usually instructed to explicitly
2000; Wang et al., 2003; Wicker et al., 2003; Hen- rate or categorize emotional expressions. Im-
nenlotter et al., 2004; Schroeder et al., 2004a). aging studies on explicit emotion recognition (e.g.,
Empirical evidence for a specialized neural repre- Gorno-Tempini et al., 2001; Winston et al., 2003)
sentation of anger, surprise, sadness, or happiness frequently failed to find dissociable responses to
is more limited, primarily due to lack of clinical specific emotions within those regions where le-
data that allow drawing causal inferences from sions have been associated with impairments in
selective impairments in recognizing these emo- emotion recognition such as the amygdala (e.g.,
tions. In functional imaging research, the detection Adolphs et al., 1999) and insula/basal ganglia
of emotion-specific neural subsystems requires di- (Calder et al., 2000). One reason for this finding
rect comparisons of signal changes associated may be that frontal regions recruited by explicit
with the perception of different emotions, i.e., ex- recognition tasks may attenuate neural responses
pressions of one emotion are used as a baseline in limbic regions such as the amygdala (Critchley
condition for the target emotion. Since direct com- et al., 2000; Hariri et al., 2000).
parisons allow tracking the blood oxygen level Only recently has evidence been obtained that
dependent (BOLD) response across different emo- the activation of emotion-specific neural subsys-
tion conditions, it becomes possible to distinguish tems may be closely linked to characteristic facial
regions associated with the perception of a specific features of single expressions such as the eye
452

region for fearful faces (Whalen et al., 2001; Mor- Angrilli, A., Mauri, A., Palomba, D., Flor, H., Birbaumer, N.,
ris et al., 2002; Adolphs et al., 2005). These find- Sartori, G. and di Paola, F. (1996) Startle reflex and emotion
modulation impairment after a right amygdala lesion. Brain,
ings suggest that not all facial expressive features
119(Pt 6): 1991–2000.
are equally important in recognizing single Aniskiewicz, A.S. (1979) Autonomic components of vicarious
emotions. Smith et al. (2005) demonstrated that conditioning and psychopathy. J. Clin. Psychol., 35: 60–67.
diagnostic cues for recognizing specific emotions Averill, J.R. (1982) Anger and Aggression: An Essay on Emo-
could be identified by comparative analyses of the tion. Springer, New York.
diagnostic filtering functions of human observers. Bechara, A., Tranel, D., Damasio, H., Adolphs, R., Rockland,
C. and Damasio, A.R. (1995) Double dissociation of condi-
The use of such ‘‘effective faces’’ as stimuli in tioning and declarative knowledge relative to the amygdala
functional imaging research may therefore be and hippocampus in humans. Science, 269: 1115–1118.
a promising method to further differentiate the Blair, R.J. (2001) Neurocognitive models of aggression, the
neural subsystems implicated in the recognition of antisocial personality disorders, and psychopathy. J. Neurol.
Neurosurg. Psychiatry, 71: 727–731.
specific emotions.
Blair, R.J. (2003) Facial expressions, their communicatory
functions and neuro-cognitive substrates. Philos. Trans. R.
Soc. Lond. B Biol. Sci., 358: 561–572.
References Blair, R.J. and Cipolotti, L. (2000) Impaired social response
reversal. A case of ‘acquired sociopathy’. Brain, 123(Pt 6):
Adams Jr., R.B., Gordon, H.L., Baird, A.A., Ambady, N. and 1122–1141.
Kleck, R.E. (2003) Effects of gaze on amygdala sensitivity to Blair, R.J., Jones, L., Clark, F. and Smith, M. (1997) The psy-
anger and fear faces. Science, 300: 1536. chopathic individual: A lack of responsiveness to distress
Adolphs, R. (2002a) Neural systems for recognizing emotion. cues? Psychophysiology, 34: 192–198.
Curr. Opin. Neurobiol., 12: 169–177. Blair, R.J.R., Morris, J.S., Frith, C.D., Perrett, D.I. and Dolan,
Adolphs, R. (2002b) Recognizing emotion from facial expres- R.J. (1999) Dissociable neural responses to facial expressions
sions: psychological and neurological mechanisms. Behav. of sadness and anger. Brain, 122: 883–893.
Cognit. Neurosci. Rev., 1: 21–61. Breiter, H.C., Etcoff, N.L., Whalen, P.J., Kennedy, W.A., Ra-
Adolphs, R., Damasio, H., Tranel, D., Cooper, G. and Dam- uch, S.L., Buckner, R.L., Strauss, M.M., Hyman, S.E. and
asio, A.R. (2000) A role for somatosensory cortices in the Rosen, B.R. (1996) Response and habituation of the human
visual recognition of emotion as revealed by three-dimen- amygdala during visual processing of facial expression. Neu-
sional lesion mapping. J. Neurosci., 20: 2683–2690. ron, 17: 875–887.
Adolphs, R., Damasio, H., Tranel, D. and Damasio, A.R. Broks, P., Young, A.W., Maratos, E.J., Coffey, P.J.,
(1996) Cortical systems for the recognition of emotion in fa- Calder, A.J., Isaac, C.L., Mayes, A.R., Hodges, J.R., Mont-
cial expressions. J. Neurosci., 16: 7678–7687. aldi, D., Cezayirli, E., Roberts, N. and Hadley, D. (1998)
Adolphs, R., Gosselin, F., Buchanan, T.W., Tranel, D., Schyns, Face processing impairments after encephalitis: amygdala
P. and Damasio, A.R. (2005) A mechanism for impaired fear damage and recognition of fear. Neuropsychologia, 36:
recognition after amygdala damage. Nature, 433: 68–72. 59–70.
Adolphs, R., Tranel, D., Damasio, H. and Damasio, A. (1994) Brower, M.C. and Price, B.H. (2001) Neuropsychiatry of fron-
Impaired recognition of emotion in facial expressions fol- tal lobe dysfunction in violent and criminal behaviour: a
lowing bilateral damage to the human amygdala. Nature, critical review. J. Neurol. Neurosurg. Psychiatry, 71:
372: 669–672. 720–726.
Adolphs, R., Tranel, D., Damasio, H. and Damasio, A.R. Calder, A.J., Keane, J., Lawrence, A.D. and Manes, F. (2004)
(1995) Fear and the human amygdala. J. Neurosci., 15: Impaired recognition of anger following damage to the ven-
5879–5891. tral striatum. Brain, 127: 1958–1969.
Adolphs, R., Tranel, D., Hamann, S., Young, A.W., Calder, Calder, A.J., Keane, J., Manes, F., Antoun, N. and Young,
A.J., Phelps, E.A., Anderson, A., Lee, G.P. and Damasio, A.W. (2000) Impaired recognition and experience of disgust
A.R. (1999) Recognition of facial emotion in nine individuals following brain injury. Nat. Neurosci., 3: 1077–1078.
with bilateral amygdala damage. Neuropsychologia, 37: Calder, A.J., Lawrence, A.D. and Young, A.W. (2001) Neuro-
1111–1117. psychology of fear and loathing. Nat. Rev. Neurosci., 2:
Anderson, A.K. and Phelps, E.A. (2000) Expression without 352–363.
recognition: Contributions of the human amygdala to emo- Calder, A.J., Young, A.W., Rowland, D., Perrett, D.I., Hod-
tional communication. Psychol. Sci., 11: 106–111. ges, J.R. and Etcoff, N.L. (1996) Facial emotion recognition
Anderson, A.K., Spencer, D.D., Fulbright, R.K. and Phelps, after bilateral amygdala damage. Cogn. Neuropsychol., 13:
E.A. (2000) Contribution of the anteromedial temporal lobes 669–745.
to the evaluation of facial emotion. Neuropsychology, 14: Canteras, N.S., Shammah-Lagnado, S.J., Silva, B.A. and
526–536. Ricardo, J.A. (1990) Afferent connections of the subthalamic
453

nucleus: a combined retrograde and anterograde horseradish Frey, S. and Petrides, M. (1999) Re-examination of the human
peroxidase study in the rat. Brain Res., 513: 43–59. taste region: a positron emission tomography study. Eur.
Chaplin, T.C., Rice, M.E. and Harris, G.T. (1995) Salient vic- J. Neurosci., 11: 2985–2988.
tim suffering and the sexual responses of child molesters. J. Fridlund, A.J. (1994) Human Facial Expression: An Evolu-
Consult. Clin. Psychol., 63: 249–255. tionary View. Academic Press, San Diego.
Critchley, H., Daly, E., Phillips, M., Brammer, M., Bullmore, Gabrieli, J.D., Brewer, J.B., Desmond, J.E. and Glover, G.H.
E., Williams, S., Van Amelsvoort, T., Robertson, D., David, (1997) Separate neural bases of two fundamental memory
A. and Murphy, D. (2000) Explicit and implicit neural mech- processes in the human medial temporal lobe. Science, 276:
anisms for processing of social information from facial ex- 264–266.
pressions: a functional magnetic resonance imaging study. Gorno-Tempini, M.L., Pradelli, S., Serafini, M., Pagnoni,
Hum. Brain Mapp., 9: 93–105. G., Baraldi, P., Porro, C., Nicoletti, R., Umita, C. and Nic-
Darwin, C. (1872) The Expression of the Emotions in Man and helli, P. (2001) Explicit and incidental facial expression
Animals. John Murray, London. processing: an fMRI study. Neuroimage, 14: 465–473.
de la Monte, S.M., Vonsattel, J.P. and Richardson Jr., E.P. Gosselin, F. and Schyns, P.G. (2001) Bubbles: a technique to
(1988) Morphometric demonstration of atrophic changes in reveal the use of information in recognition tasks. Vision
the cerebral cortex, white matter, and neostriatum in Hunt- Res., 41: 2261–2271.
ington’s disease. J. Neuropathol. Exp. Neurol., 47: 516–525. Gray, J.M., Young, A.W., Barker, W.A., Curtis, A. and Gib-
Dias, R., Robbins, T.W. and Roberts, A.C. (1996) Dissociation son, D. (1997) Impaired recognition of disgust in Hunting-
in prefrontal cortex of affective and attentional shifts. ton’s disease gene carriers. Brain, 120: 2029–2038.
Nature, 380: 69–72. Haines, S.M. and Muir, D.W. (1969) Infant sensitivity to adult
Dolan, R.J., Fletcher, P., Morris, J., Kapur, N., Deakin, J.F. eye direction. Child Dev., 67: 1940–1951.
and Frith, C.D. (1996) Neural activation during covert Hamann, S.B., Stefanacci, L., Squire, L.R., Adolphs, R.,
processing of positive emotional facial expressions. Neuro- Tranel, D., Damasio, H. and Damasio, A. (1996) Recogniz-
image, 4: 194–200. ing facial emotion. Nature, 379: 497.
Eisenberg, N., Fabes, R.A., Miller, P.A., Fultz, J., Shell, R., Hariri, A.R., Bookheimer, S.Y. and Mazziotta, J.C. (2000)
Mathy, R.M. and Reno, R.R. (1989) Relation of sympathy Modulating emotional responses: effects of a neocortical
and personal distress to prosocial behavior: a multimethod network on the limbic system. Neuroreport, 11: 43–48.
study. J. Pers. Soc. Psychol., 57: 55–66. Harmer, C.J., Perrett, D.I., Cowen, P.J. and Goodwin, G.M.
Ekman, P. and Friesen, W.V. (1976) Pictures of Facial Affect. (2001a) Administration of the beta-adrenoceptor blocker
Consulting Psychologist Press, Palo Alto, CA. propranolol impairs the processing of facial expressions of
Ekman, P., Friesen, W.V., O’Sullivan, M., Chan, A., sadness. Psychopharmacology (Berl.), 154: 383–389.
Diacoyanni-Tarlatzis, I., Heider, K., Krause, R., LeCompte, Harmer, C.J., Thilo, K.V., Rothwell, J.C. and Goodwin, G.M.
W.A., Pitcairn, T., Ricci-Bitti, P.E., et al. (1987) Universals (2001b) Transcranial magnetic stimulation of medial-frontal
and cultural differences in the judgments of facial expressions cortex impairs the processing of angry facial expressions.
of emotion. J. Pers. Soc. Psychol., 53: 712–717. Nat. Neurosci., 4: 17–18.
Ekman, P., Sorenson, E.R. and Friesen, W.V. (1969) Pan-cul- Haxby, J.V., Hoffman, E.A. and Gobbini, M.I. (2000) The
tural elements in facial displays of emotion. Science, 164: distributed human neural system for face perception. Trends
86–88. Cogn. Sci., 4: 223–233.
Esteves, F., Parra, C., Dimberg, U. and Ohman, A. (1994) Hennenlotter, A., Schroeder, U., Erhard, P., Castrop, F., Has-
Nonconscious associative learning: Pavlovian conditioning of linger, B., Stoecker, D., Lange, K.W. and Ceballos-Bau-
skin conductance responses to masked fear-relevant facial mann, A.O. (2005) A common neural basis for receptive and
stimuli. Psychophysiology, 31: 375–385. expressive communication of pleasant facial affect. Neuro-
Fernandez, G., Weyerts, H., Schrader-Bolsche, M., Tendolkar, image, Jan: 26(2): 581–591.
I., Smid, H.G., Tempelmann, C., Hinrichs, H., Scheich, H., Hennenlotter, A., Schroeder, U., Erhard, P., Haslinger,
Elger, C.E., Mangun, G.R. and Heinze, H.J. (1998) Success- B., Stahl, R., Weindl, A., von Einsiedel, H.G., Lange,
ful verbal encoding into episodic memory engages the pos- K.W. and Ceballos-Baumann, A.O. (2004) Neural correlates
terior hippocampus: a parametrically analyzed functional associated with impaired disgust processing in pre-sympto-
magnetic resonance imaging study. J. Neurosci., 18: matic Huntington’s disease. Brain, 127: 1446–1453.
1841–1847. Hornak, J., Rolls, E.T. and Wade, D. (1996) Face and voice
Ferrari, P.F., Gallese, V., Rizzolatti, G. and Fogassi, L. (2003) expression identification in patients with emotional and be-
Mirror neurons responding to the observation of ingestive havioural changes following ventral frontal lobe damage.
and communicative mouth actions in the monkey ventral Neuropsychologia, 34: 247–261.
premotor cortex. Eur. J. Neurosci., 17: 1703–1714. House, T.H. and Milligan, W.L. (1976) Autonomic responses
Fischer, H., Wright, C.I., Whalen, P.J., McInerney, S.C., Shin, to modeled distress in prison psychopaths. J. Pers. Soc. Psy-
L.M. and Rauch, S.L. (2003) Brain habituation during chol., 34: 556–560.
repeated exposure to fearful and neutral faces: a functional Jernigan, T.L., Salmon, D.P., Butters, N. and Hesselink, J.R.
MRI study. Brain Res. Bull., 59: 387–392. (1991) Cerebral structure on MRI, Part II: specific changes in
454

Alzheimer’s and Huntington’s diseases. Biol. Psychiatry, 29: Morris, J.S., deBonis, M. and Dolan, R.J. (2002) Human am-
68–81. ygdala responses to fearful eyes. Neuroimage, 17: 214–222.
Jones, S.S. and Raag, T. (1989) Smile production in older in- Morris, J.S., Friston, K.J., Buchel, C., Frith, C.D., Young,
fants: the importance of a social recipient for the facial signal. A.W., Calder, A.J. and Dolan, R.J. (1998) A neuromodula-
Child Dev., 60: 811–818. tory role for the human amygdala in processing emotional
Kesler-West, M.L., Andersen, A.H., Smith, C.D., Avison, M.J., facial expressions. Brain, 121: 47–57.
Davis, C.E., Kryscio, R.J. and Blonder, L.X. (2001) Neural Morris, J.S., Frith, C.D., Perrett, D.I., Rowland, D., Young,
substrates of facial emotion processing using fMRI. Brain A.W., Calder, A.J. and Dolan, R.J. (1996) A differential
Res. Cogn. Brain Res., 11: 213–226. neural response in the human amygdala to fearful and happy
Kim, H., Somerville, L.H., Johnstone, T., Alexander, A.L. and facial expressions. Nature, 383: 812–815.
Whalen, P.J. (2003) Inverse amygdala and medial prefrontal Murphy, F.C., Nimmo-Smith, I. and Lawrence, A.D. (2003)
cortex responses to surprised faces. Neuroreport, 14: Functional neuroanatomy of emotions: a meta-analysis.
2317–2322. Cogn. Affect. Behav. Neurosci., 3: 207–233.
Kim, H., Somerville, L.H., Johnstone, T., Polis, S., Alexander, Nakamura, K., Kawashima, R., Ito, K., Sugiura, M., Kato, T.,
A.L., Shin, L.M. and Whalen, P.J. (2004) Contextual mod- Nakamura, A., Hatano, K., Nagumo, S., Kubota, K.,
ulation of amygdala responsivity to surprised faces. J. Cogn. Fukuda, H. and Kojima, S. (1999) Activation of the right
Neurosci., 16: 1730–1745. inferior frontal cortex during assessment of facial emotion. J.
Knutson, B., Adams, C.M., Fong, G.W. and Hommer, D. Neurophysiol., 82: 1610–1614.
(2001) Anticipation of increasing monetary reward selectively Narumoto, J., Yamada, H., Iidaka, T., Sadato, N., Fukui, K.,
recruits nucleus accumbens. J. Neurosci., 21: RC159. Itoh, H. and Yonekura, Y. (2000) Brain regions involved in
Koepp, M.J., Gunn, R.N., Lawrence, A.D., Cunningham, V.J., verbal or non-verbal aspects of facial emotion recognition.
Dagher, A., Jones, T., Brooks, D.J., Bench, C.J. and Grasby, Neuroreport, 11: 2571–2576.
P.M. (1998) Evidence for striatal dopamine release during a Ohman, A., Lundqvist, D. and Esteves, F. (2001) The face in
video game. Nature, 393: 266–268. the crowd revisited: a threat advantage with schematic stim-
Krolak-Salmon, P., Henaff, M.A., Isnard, J., Tallon-Baudry, uli. J. Pers. Soc. Psychol., 80: 381–396.
C., Guenot, M., Vighetto, A., Bertrand, O. and Mauguiere, Parent, A. and Hazrati, L.N. (1995) Functional anatomy of the
F. (2003) An attention modulated response to disgust in hu- basal ganglia. II. The place of subthalamic nucleus and ex-
man ventral anterior insula. Ann. Neurol., 53: 446–453. ternal pallidum in basal ganglia circuitry. Brain Res. Brain
LaBar, K.S., Crupain, M.J., Voyvodic, J.T. and McCarthy, G. Res. Rev., 20: 128–154.
(2003) Dynamic perception of facial affect and identity in the Patrick, C.J., Bradley, M.M. and Lang, P.J. (1993) Emotion in
human brain. Cereb. Cortex, 13: 1023–1033. the criminal psychopath: startle reflex modulation. J. Ab-
LaBar, K.S., LeDoux, J.E., Spencer, D.D. and Phelps, norm. Psychol., 102: 82–92.
E.A. (1995) Impaired fear conditioning following uni- Pessoa, L., McKenna, M., Gutierrez, E. and Ungerleider, L.G.
lateral temporal lobectomy in humans. J. Neurosci., 15: (2002) Neural processing of emotional faces requires atten-
6846–6855. tion. Proc. Natl. Acad. Sci. USA, 99: 11458–11463.
Lawrence, A.D., Calder, A.J., McGowan, S.W. and Grasby, Phan, K.L., Wager, T., Taylor, S.F. and Liberzon, I. (2002)
P.M. (2002) Selective disruption of the recognition of facial Functional neuroanatomy of emotion: a meta-analysis of
expressions of anger. Neuroreport, 13: 881–884. emotion activation studies in PET and fMRI. Neuroimage,
Lawrence, A.D., Evans, A.H. and Lees, A.J. (2003) Compulsive 16: 331–348.
use of dopamine replacement therapy in Parkinson’s disease: Phillips, M.L., Bullmore, E.T., Howard, R., Woodruff, P.W.,
reward systems gone awry? Lancet Neurol., 2: 595–604. Wright, I.C., Williams, S.C., Simmons, A., Andrew, C., Bra-
Lucas, J.A., Rippeth, J.D., Uitti, R.J., Shuster, E.A. and mmer, M. and David, A.S. (1998a) Investigation of facial
Wharen, R.E. (2000) Neuropsychological functioning in a recognition memory and happy and sad facial expression
patient with essential tremor with and without bilateral VIM perception: an fMRI study. Psychiatry Res., 83: 127–138.
stimulation. Brain Cogn., 42: 253–267. Phillips, M.L., Medford, N., Young, A.W., Williams, L., Will-
Miczek, K.A., Fish, E.W., De Bold, J.F. and De Almeida, R.M. iams, S.C., Bullmore, E.T., Gray, J.A. and Brammer, M.J.
(2002) Social and neural determinants of aggressive behavior: (2001) Time courses of left and right amygdalar responses to
pharmacotherapeutic targets at serotonin, dopamine and fearful facial expressions. Hum. Brain Mapp., 12: 193–202.
gamma-aminobutyric acid systems. Psychopharmacology Phillips, M.L., Williams, L.M., Heining, M., Herba, C.M.,
(Berl.), 163: 434–458. Russell, T., Andrew, C., Bullmore, E.T., Brammer, M.J.,
Milders, M., Crawford, J.R., Lamb, A. and Simpson, S.A. Williams, S.C., Morgan, M., Young, A.W. and Gray, J.A.
(2003) Differential deficits in expression recognition in gene- (2004) Differential neural responses to overt and covert pres-
carriers and patients with Huntington’s disease. entations of facial expressions of fear and disgust. Neuroim-
Neuropsychologia, 41: 1484–1492. age, 21: 1484–1496.
Mineka, S. and Cook, M. (1993) Mechanisms involved in the Phillips, M.L., Williams, L., Senior, C., Bullmore, E.T., Bra-
observational conditioning of fear. J. Exp. Psychol. Gen., mmer, M.J., Andrew, C., Williams, S.C. and David,
122: 23–38. A.S. (1999) A differential neural response to threatening
455

and non-threatening negative facial expressions in paranoid Small, D.M., Gregory, M.D., Mak, Y.E., Gitelman, D., Me-
and non-paranoid schizophrenics. Psychiatry Res., 92: 11–31. sulam, M.M. and Parrish, T. (2003) Dissociation of neural
Phillips, M.L., Young, A.W., Scott, S.K., Calder, A.J., Andrew, representation of intensity and affective valuation in human
C., Giampietro, V., Williams, S.C., Bullmore, E.T., Bra- gustation. Neuron, 39: 701–711.
mmer, M. and Gray, J.A. (1998b) Neural responses to facial Small, D.M., Zald, D.H., Jones-Gotman, M., Zatorre, R.J.,
and vocal expressions of fear and disgust. Proc. R. Soc. Pardo, J.V., Frey, S. and Petrides, M. (1999) Human cortical
Lond. B Biol. Sci., 265: 1809–1817. gustatory areas: a review of functional neuroimaging data.
Phillips, M.L., Young, A.W., Senior, C., Brammer, M., An- Neuroreport, 10: 7–14.
drew, C., Calder, A.J., Bullmore, E.T., Perrett, D.I., Row- Smith, M.L., Cottrell, G.W., Gosselin, F. and Schyns, P.G.
land, D., Williams, S.C., Gray, J.A. and David, A.S. (1997) A (2005) Transmitting and decoding facial expressions. Psy-
specific neural substrate for perceiving facial expressions of chol. Sci., 16: 184–189.
disgust. Nature, 389: 495–498. Somerville, L.H., Kim, H., Johnstone, T., Alexander, A.L. and
Posamentier, M.T. and Abdi, H. (2003) Processing faces and Whalen, P.J. (2004) Human amygdala responses during pres-
facial expressions. Neuropsychol. Rev., 13: 113–143. entation of happy and neutral faces: correlations with state
Rapcsak, S.Z., Comer, J.F. and Rubens, A.B. (1993) Anomia anxiety. Biol. Psychiatry, 55: 897–903.
for facial expressions: neuropsychological mechanisms and Sprengelmeyer, R., Rausch, M., Eysel, U.T. and Przuntek, H.
anatomical correlates. Brain Lang., 45: 233–252. (1998) Neural structures associated with recognition of facial
Rapcsak, S.Z., Galper, S.R., Comer, J.F., Reminger, S.L., expressions of basic emotions. Proc. R. Soc. Lond. B Biol.
Nielsen, L., Kaszniak, A.W., Verfaellie, M., Laguna, J.F., Sci., 265: 1927–1931.
Labiner, D.M. and Cohen, R.A. (2000) Fear recognition Sprengelmeyer, R., Schroeder, U., Young, A.W. and Epplen,
deficits after focal brain damage: a cautionary note. Neurol- J.T. (2006) Disgust in pre-clinical Huntington’s disease: a
ogy, 54: 575–581. longitudinal study. Neuropsychologia, 44(4): 518–533.
Redolat, R., Brain, P.F. and Simon, V.M. (1991) Sulpiride has Sprengelmeyer, R., Young, A.W., Calder, A.J., Karnat, A.,
an antiaggressive effect in mice without markedly depressing Lange, H., Homberg, V., Perrett, D.I. and Rowland, D.
motor activity. Neuropharmacology, 30: 41–46. (1996) Loss of disgust. Perception of faces and emotions in
Reinders, A.A., den Boer, J.A. and Buchel, C. (2005) The Huntington’s disease. Brain, 119: 1647–1665.
robustness of perception. Eur. J. Neurosci., 22: 524–530. Sprengelmeyer, R., Young, A.W., Schroeder, U., Grossenbac-
Rolls, E.T., (1996) The orbitofrontal cortex. Philos. Trans. R. her, P.G., Federlein, J., Buttner, T. and Przuntek, H. (1999)
Soc. Lond. B Biol. Sci., 351: 1433–1443; discussion Knowing no fear. Proc. R. Soc. Lond. B Biol. Sci., 266:
1443–1444. 2451–2456.
Rozin, P. and Fallon, A.E. (1987) A perspective on disgust. Stern, C.E., Corkin, S., Gonzalez, R.G., Guimaraes, A.R.,
Psychol. Rev., 94: 23–41. Baker, J.R., Jennings, P.J., Carr, C.A., Sugiura, R.M., Ved-
Sato, W., Kubota, Y., Okada, T., Murai, T., Yoshikawa, S. and antham, V. and Rosen, B.R. (1996) The hippocampal for-
Sengoku, A. (2002) Seeing happy emotion in fearful and an- mation participates in novel picture encoding: evidence from
gry faces: qualitative analysis of facial expression recognition functional magnetic resonance imaging. Proc. Natl. Acad.
in a bilateral amygdala-damaged patient. Cortex, 38: Sci. USA, 93: 8660–8665.
727–742. Strauss, M.M., Makris, N., Aharon, I., Vangel, M.G., Good-
Schmidt, K.L. and Cohn, J.F. (2001) Human facial expressions man, J., Kennedy, D.N., Gasic, G.P. and Breiter, H.C. (2005)
as adaptations: Evolutionary questions in facial expression fMRI of sensitization to angry faces. Neuroimage, 26:
research. Am. J. Phys. Anthropol., Suppl., 33: 3–24. 389–413.
Schmolck, H. and Squire, L.R. (2001) Impaired perception of Thieben, M.J., Duggins, A.J., Good, C.D., Gomes, L., Mahant,
facial emotions following bilateral damage to the anterior N., Richards, F., McCusker, E. and Frackowiak, R.S. (2002)
temporal lobe. Neuropsychology, 15: 30–38. The distribution of structural neuropathology in pre-clinical
Schroeder, U., Hennenlotter, A., Erhard, P., Haslinger, B., Huntington’s disease. Brain, 125: 1815–1828.
Stahl, R., Lange, K.W. and Ceballos-Baumann, A.O. (2004a) Thomas, K.M., Drevets, W.C., Whalen, P.J., Eccard, C.H.,
Functional neuroanatomy of perceiving surprised faces. Dahl, R.E., Ryan, N.D. and Casey, B.J. (2001) Amygdala
Hum. Brain Mapp., 23: 181–187. response to facial expressions in children and adults. Biol.
Schroeder, U., Kuehler, A., Hennenlotter, A., Haslinger, B., Psychiatry, 49: 309–316.
Tronnier, V.M., Krause, M., Pfister, R., Sprengelmeyer, R., Tiihonen, J., Kuikka, J., Bergstrom, K., Hakola, P., Karhu, J.,
Lange, K.W. and Ceballos-Baumann, A.O. (2004b) Facial Ryynanen, O.P. and Fohr, J. (1995) Altered striatal dopa-
expression recognition and subthalamic nucleus stimulation. mine re-uptake site densities in habitually violent and non-
J. Neurol. Neurosurg. Psychiatry, 75: 648–650. violent alcoholics. Nat. Med., 1: 654–657.
Schutzwohl, A. (1998) Surprise and schema strength. J. Exp. Tomkins, S.S. and McCarter, R. (1964) What and where are the
Psychol. Learn. Mem. Cogn., 24: 1182–1199. primary affects? Some evidence for a theory. Percept. Mot.
Schyns, P.G., Bonnar, L. and Gosselin, F. (2002) Show me the Skills, 18: 119–158.
features! Understanding recognition from the use of visual Vuilleumier, P., Armony, J.L., Driver, J. and Dolan, R.J. (2001)
information. Psychol. Sci., 13: 402–409. Effects of attention and emotion on face processing in the
456

human brain: an event-related fMRI study. Neuron, 30: man amygdala responses to facial expressions of fear versus
829–841. anger. Emotion, 1: 70–83.
Vuilleumier, P., Armony, J.L., Driver, J. and Dolan, R.J. (2003) Wicker, B., Keysers, C., Plailly, J., Royet, J.P., Gallese, V. and
Distinct spatial frequency sensitivities for processing faces Rizzolatti, G. (2003) Both of us disgusted in My insula: the
and emotional expressions. Nat. Neurosci., 6: 624–631. common neural basis of seeing and feeling disgust. Neuron,
Wang, K., Hoosain, R., Yang, R.M., Meng, Y. and Wang, 40: 655–664.
C.Q. (2003) Impairment of recognition of disgust in Chinese Winston, J.S., O’Doherty, J. and Dolan, R.J. (2003) Common
with Huntington’s or Wilson’s disease. Neuropsychologia, and distinct neural responses during direct and incidental
41: 527–537. processing of multiple facial emotions. Neuroimage, 20:
Whalen, P.J., Kagan, J., Cook, R.G., Davis, F.C., Kim, 84–97.
H., Polis, S., McLaren, D.G., Somerville, L.H., McLean, Wright, C.I., Fischer, H., Whalen, P.J., McInerney, S.C., Shin,
A.A., Maxwell, J.S. and Johnstone, T. (2004) Human am- L.M. and Rauch, S.L. (2001) Differential prefrontal cortex
ygdala responsivity to masked fearful eye whites. Science, and amygdala habituation to repeatedly presented emotional
306: 2061. stimuli. Neuroreport, 12: 379–383.
Whalen, P.J., Kapp, B.S. and Pascoe, J.P. (1994) Neuronal ac- Yang, T.T., Menon, V., Eliez, S., Blasey, C., White, C.D., Reid,
tivity within the nucleus basalis and conditioned neocortical A.J., Gotlib, I.H. and Reiss, A.L. (2002) Amygdalar activa-
electroencephalographic activation. J. Neurosci., 14: tion associated with positive and negative facial expressions.
1623–1633. Neuroreport, 13: 1737–1741.
Whalen, P.J., Rauch, S.L., Etcoff, N.L., McInerney, S.C., Lee, Young, A.W., Aggleton, J.P., Hellawell, D.J., Johnson, M.,
M.B. and Jenike, M.A. (1998) Masked presentations of emo- Broks, P. and Hanley, J.R. (1995) Face processing impair-
tional facial expressions modulate amygdala activity without ments after amygdalotomy. Brain, 118(Pt 1): 15–24.
explicit knowledge. J. Neurosci., 18: 411–418. Young, A.W., Hellawell, D.J., Van De Wal, C. and Johnson,
Whalen, P.J., Shin, L.M., McInerney, S.C., Fischer, H., Wright, M. (1996) Facial expression processing after amygdalotomy.
C.I. and Rauch, S.L. (2001) A functional MRI study of hu- Neuropsychologia, 34: 31–39.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 25

Integration of emotion and cognition in patients with


psychopathy

Monika Sommer1, Göran Hajak1, Katrin Döhnel1, Johannes Schwerdtner1,


Jörg Meinhardt2 and Jürgen L. Müller1

1
Department of Psychiatry, Psychotherapy and Psychosomatics, University of Regensburg, UniversitätsstraX e 84,
D-93053 Regensburg, Germany
2
Ludwig-Maximilian University, Munich, Germany

Abstract: Psychopathy is a personality disorder associated with emotional characteristics like impulsivity,
manipulativeness, affective shallowness, and absence of remorse or empathy. The impaired emotional
responsiveness is considered to be the hallmark of the disorder. There are two theories that attempt to
explain the emotional dysfunction and the poor socialization in psychopathy: (1) the low-fear model and (2)
the inhibition of violence model. Both approaches are supported by several studies. Studies using aversive
conditioning or the startle modulation underline the severe difficulties in processing negative stimuli in
psychopaths. Studies that explore the processing of emotional expressions show a deficit of psychopathic
individuals for processing sad or fearful facial expressions or vocal affect. In the cognitive domain, psy-
chopaths show performance deficits in the interpretation of the motivational significance of stimuli. Studies
investigating the impact of emotions on cognitive processes show that in psychopaths in contrast to healthy
controls negative emotions drain no resources from a cognitive task. It is suggested that dysfunctions in the
frontal cortex, especially the orbitofrontal cortex, the cingulate cortex and the amygdala are associated with
the emotional and cognitive impairments.

Keywords: psychopathy; emotional impairment; emotion–cognition interaction

Concepts of psychopathy unreliability, inability to accept blame or shame,


lack of emotions, egocentricity, failure to learn from
The concept of psychopathy was first introduced by experiences, and inability to follow goals. Cleckley’s
Phillippe Pinel approximately 200 years ago. Pinel definition of psychopathy was included in the
(1809) characterized the ‘‘Mania sans délir’’ by DSM-II. Subsequent revisions of the DSM changed
emotional instability and social drift (Sass and Her- primarily personality-based descriptions of the dis-
pertz, 1995). More recent conceptualizations are order to primarily behavioral-based descriptions
linked to the study by Cleckley (1941) and his book such as antisocial personality disorder, conduct dis-
‘‘The Mask of Sanity.’’ In addition to antisocial order, and oppositional defiant disorder. This con-
behavior, Cleckley provided extensive clinical de- ceptual change was associated with the expectation
scriptions of specific characterological features of that behavioral characteristics were more reliably
psychopathic patients, e.g., superficial charm, assessed than were personality traits (Cloninger,
1978). Lilienfeld (1994) criticized that behavioral
Corresponding author. Tel.: +49-941-9412050; classifications alone are too narrow and emphasized
Fax: +49-941-9412065; E-mail: monika.sommer@medbo.de the importance of a personality component to the

DOI: 10.1016/S0079-6123(06)56025-X 457


458

assessment of psychopathy. Moreover, these con- conditioning, most studies have revealed weaker
ceptualizations ignore the emotional characteristics aversive conditioning in psychopaths as compa-
of individuals with psychopathy. red to healthy controls (Hare and Quinn, 1971).
The current state-of-the-art diagnostic instru- Lykken (1957) describes reduced anxiety levels
ment in the identification of adult psychopathy is based on questionnaires, and less electrodermal re-
Hare’s Psychopathy Checklist — Revised (PCL-R; sponses to conditioned stimuli (CSs) that were pre-
Hare et al., 2000). Hare’s work on psychopathy viously associated with shocks. Most, but not all of
with adult offenders has shown that two corre- succeeding studies found a lack of skin conductance
lated factors are important for the description of responding to aversive stimuli (for a review see
psychopathy: Factor 1 includes emotional charac- Hare, 1998). Results regarding cardiovascular re-
teristics like affective shallowness, superficial sponses have varied. Some studies reported defi-
charm, and absence of remorse or empathy; Fac- cient phasic heart rate reactivity to aversive stimuli,
tor 2 describes juvenile and adult delinquency other studies found increased anticipatory re-
and aggression. These two factors contain both sponses — a result that might indicate a success-
personality components and concrete behavioral ful anticipatory coping with the stimulus (Hare
aspects. et al., 1978; Hare, 1982). Studies using event-related
potentials in response to or in anticipation of avers-
ive stimuli show very inconsistent results. Whereas
Psychopathy and emotion
some studies found no difference in components
like N100, P200, or P300 between psychopaths and
Impaired emotional responsiveness is considered
healthy controls (for a review see Raine, 1989),
to be the hallmark of psychopathy (Cleckley, 1976;
other studies found smaller P300 responses in psy-
Hare, 1991). A growing body of research has fo-
chopaths (Kiehl et al., 1999). Flor et al. (2002)
cused on the nature and bases of affective distur-
studied aversive Pavlovian conditioning in non-
bances in psychopaths. It is suggested that the
criminal psychopaths and healthy controls. An un-
emotional deficits associated with psychopathy in-
pleasant odor was used as unconditioned stimulus
terfere with the development of moral reasoning
(US) and neutral faces as conditioned stimuli (CSs).
and put the individual at risk for developing high
Event-related potentials to the CS were used as in-
levels of antisocial behavior.
dices. Whereas the healthy controls show significant
Currently, there are two models that attempt to
differentiation between CS+ and CS , the psy-
explain the emotional dysfunction and the poor
chopaths failed to exhibit a conditioned response
socialization in psychopaths: the low-fear model
while unconditioned responses were comparable
(e.g., Fowles, 1980; Patrick, 1994) and the empa-
between the two groups. N100, P200, and P300 to
thy dysfunction, violence inhibition mechanism
the CSs revealed that psychopaths were not defi-
(VIM) model (e.g. Blair, 1995, 2001).
cient in information processing and showed even
better anticipatory responding than the healthy
The low-fear model of psychopathy controls indicated by the terminal contingent nega-
tive variation (tCNV). In a fMRI paradigm this
The low-fear model suggests that the poor social- research group studied the brain circuits underlying
ization in psychopaths is the result of their reduced aversive conditioning. They used neutral faces as
ability to experience fear. As a result, they show a CSs and painful pressure as US and found that
lack of anticipation of aversive events and a re- psychopaths in contrast to healthy controls show
duced ability to adjust their behavior in response no activation in the limbic-prefrontal circuit includ-
to the negative consequences their behavior has led ing amygdala, orbitofrontal cortex, insula, and an-
to in the past (Birbaumer et al., 2005). terior cingulate. Additionally, they failed to show
Empirical support for the low-fear model is conditioned skin conductance and emotional va-
drawn from experiments studying fear condition- lence ratings, although contingency and arousal
ing and from startle experiments. Using Pavlovian ratings do not differ (Birbaumer et al., 2005). The
459

authors suggest that a hypoactive frontolimbic cir- (Blair et al., 2002, 2001a). The model postulates
cuit may underlie the deficient responses (Flor et al., that the functional integrity of this system is crucial
2002; Veit et al., 2002; Birbaumer et al., 2005). for moral socialization. Negative emotions of other
In contrast to the results by Birbaumer et al. people are aversive for healthy subjects; therefore,
(2005), Schneider et al. (2000) found in an aversive they learn to avoid initiating behavior that results in
conditioning paradigm using neutral faces as CS sadness or fear of others. The VIM model predicts
and an unpleasant odor as US that patients with that psychopaths should show particular difficulty
an antisocial personality disorder in contrast to when processing sad and fearful expressions (Blair
healthy controls show more activation in the amy- et al., 2002). Blair (2003) emphasizes that facial ex-
gdala and the dorsolateral prefrontal cortex pressions have a communicatory function. They
(PFC). They found no differences between the transmit rapidly information on the valence of ob-
two groups on the behavioral level. Schneider et al. jects or situations. The author suggests that expres-
(2000) suggest that this increase results from an sions of fearfulness, sadness and happiness are
additional effort to form negative emotional asso- reinforcers that modulate the probability of future
ciations. behavior. Blair (1995, 2003) suggests that for psy-
In summary, the results indicate that psycho- chopaths sad and fearful expressions are not avers-
paths are able to detect the association between CS ive unconditioned stimuli. Therefore, psychopaths
and US, but it seems that they do not process the do not learn to avoid committing behavior that
emotional significance of the association. causes harm to others.
Another way for studying the processing of Empirical support for this position is drawn
aversive affective states is the modulation of the from studies with children having psychopathic
startle response, especially the increase in the re- tendencies and with psychopathic adults showing
flexive startle response during presentation of an reduced autonomic arousal to sad but not angry
aversive stimulus (Lang et al., 1990, 1998). Patrick facial expressions (Blair, 1999, 2001, 2003; Stevens
et al. (1993) report that whereas the autonomic et al., 2001; Blair et al., 2001a; McMurran et al.,
or self-reported responses to affective pictures did 2002). In a study investigating the ability of adult
not differ between psychopaths and healthy con- psychopaths to process vocal affect, Blair et al.
trols, psychopaths show no startle potentiation to (2002) show that psychopathic inmates were par-
aversive pictures. Levenston et al. (2000) studied ticularly impaired in the recognition of fearful vo-
psychopathic and nonpsychopathic prisoners and cal affect. According to Blair’s (2003) expressions
showed that for psychopaths startle was inhibited that serve as positive or negative reinforcers pref-
during victim scenes and only weakly potenti- erentially activate the amygdala. He argues that
ated during threat. The authors interpreted the studies showing a reduced amygdaloid volume of
absence of a normal startle potentiation as a weak- psychopathic individuals relative to comparison
ness in the capacity of aversive cues to prime de- individuals (Tiihonen et al., 2000), reduced amy-
fensive actions. gdala activation during an emotional memory task
In sum, the aversive conditioning and the startle (Kiehl et al., 2001), and aversive conditioning
experiments underline the severe difficulties in tasks (Veit et al., 2002) support his assumptions.
processing negative stimuli in individuals with Some studies examined amygdala activation in
psychopathy. The neural correlates underlying psychopaths (Schneider et al., 2000; Veit et al.,
these impairments, especially the role of limbic 2002; Müller et al., 2003) and found inconsistent
areas and the PFC, is unclear. results. Whereas Veit et al. (2002) report reduced
activation of the amygdala, the orbitofrontal cor-
tex, the anterior cingulate, and the insula in psy-
The VIM of psychopathy chopaths, Schneider et al. (2000) studying patients
with antisocial personality disorder found en-
The VIM suggests that there is a system that hanced amygdala and PFC activation. In an emo-
preferentially responds to sad and fearful stimuli tion induction paradigm, Müller et al. (2003)
460

found increased activation of the amygdala and associations when they become inappropriate
prefrontal regions during processing of negative (Rolls, 2000), Newman et al. (1987) found that
affective pictures, whereas pleasant affective pic- psychopaths show impaired extinction of previ-
tures induce more activation in the orbitofrontal ously rewarding responses in a single-pack card
cortex. playing task. Blair et al. (2001b) found that the
In sum, psychopaths processing sad or fearful performance of boys with psychopathic tendencies
social stimuli, like facial expressions or vocal affect, on the four-pack gambling task developed by
make more errors and show different cortical ac- Bechara et al. (1999) was impaired relative to a
tivation patterns in contrast to nonpsychopaths. control group. Two other studies (Schmitt et al.,
Nevertheless, further investigation is needed to de- 1999; Blair and Cipolotti, 2000) found no differ-
termine the neural correlates of this dysfunction, ences between psychopaths and healthy controls on
particularly the role of the orbitofrontal cortex performance of the gambling task. Mitchell et al.
(OFC) and the amygdala. Though we know noth- (2002) investigated the performance of psycho-
ing about the processing of angry faces in psycho- paths on two neuropsychological tasks believed to
paths. Like sad and fearful expressions angry faces be sensitive to dorsolateral prefrontal and orbito-
have high communicatory functions. They signal frontal cortex functioning: the intradimensional/
discontent with a present behavior of an other extradimensional (ID/ED) shift task and the gam-
person or with a given situation (Lawrence et al., bling task. They found that psychopaths were less
2002). Therefore, they may support behavioral likely to avoid making risky selections over the
extinction and reversal learning (Rolls, 1996). Sim- course of the gambling task relative to comparison
ilarly to the processing of sad and fearful expres- individuals. On the ID/ED task the performance of
sions, the processing of angry expressions is closely psychopaths was not different from controls on
related to the amygdala (Adolphs et al., 1999). attentional set-shifting, but they show significant
Therefore, impairment of the amygdala should in- impairments on response reversal. The authors
fluence the processing of anger in psychopathic suggest that the performance impairment may be
patients. representative of a dysfunction within a neural cir-
cuit including the amygdala and the OFC. The
amygdala is involved in the formation of stimulus-
Cognition and psychopathy reinforcement associations (LeDoux, 1998). The
OFC is involved in encoding motivational signifi-
Although several studies report that there are no cance of cues and the incentive value of expected
differences in neuropsychological functions be- outcomes and is particularly important for appro-
tween psychopaths and nonpsychopaths (Sutker priate responding when reinforcement contingen-
and Allain, 1987; Hart et al., 1990), a possible im- cies change (Schoenbaum et al., 1998). Amygdala
pairment of the orbitofrontal cortex (OFC) in psy- and OFC together are considered to play a key role
chopathy is discussed. LaPierre et al. (1995) found in encoding and implementing associative infor-
in comparison to nonpsychopathic inmates that mation about the motivational significance of stim-
psychopathic inmates are impaired on all neuro- uli (Schoenbaum et al., 2000).
psychological tests considered sensitive to orbito- There are some limitations for considering the
frontal and ventromedial dysfunctions including a OFC as the key cortical structure in psychopathy.
visual go/no-go task, Porteus Maze Q-scores (i.e., Blair et al. (2001a) found differences in the OFC
rule braking errors), and an odor identification impairment between adults and boys with psycho-
task. LaPierre et al. (1995) observe no performance pathic tendencies. Considering the evidence for
deficits of psychopathic patients on measures sen- interdependence and functional connectivity of the
sitive to dorsolateral-prefrontal and poster- OFC and the amygdala, Mitchell et al. (2002) ar-
orolandic function (i.e., the Wisconsin card gue that a primary deficit within the amygdala
sorting test). Considering that the OFC is involved could give rise to deficits associated with OFC
in altering previously acquired stimulus-reward impairment and that the differences in OFC
461

impairment between adults and boys may be de- the cognitive consequences of emotions (e.g.
velopmental consequences of the disorder. affect-evoked intrusive thoughts), may increase
Thus it is unclear, whether the cognitive deficits the information processing load and drain attent-
of psychopathic patients are independent from the ional resources that otherwise might be devoted
emotional impairment or whether they are sec- to task performance. Many results of different
ondary consequences of the emotional impair- behavioral studies support this hypothesis (e.g.
ments. Forgas, 1995). Using a dual-task paradigm and
P300 as a direct index of resources, Meinhardt and
Pekrun (2003) examine the model assumptions
The influence of emotions on cognitive processes with event-related EEG potentials (ERPs) and
show that not only negative but also positive emo-
Cognitive processes like memory, attention, and tions compete with task-related processing re-
the inhibition of prepared responses and emotions sources. These results suggest that the RAM by
both function as control systems influencing and Ellis and Ashbrook (1988) can be expanded to
regulating behavior (Carver and Schreier, 1990; pleasant and appetitive experiences as well.
Kosslyn and Koenig, 1992; Braver and Cohen, In the past years, a growing number of studies
2000). But emotion and cognition are not two have tried to characterize the neural basis of the
separately working information-processing do- interrelationship between cognition and emotion.
mains. Both processes are closely intertwined. For healthy subjects, studies with positron emis-
Cognitive processes are able to regulate emotions sion tomography (PET) or functional magnetic res-
(Ochsner and Gross, 2005) and conversely, emo- onance imaging (fMRI) show a dynamic interplay
tional and motivational factors can significantly between cognition and emotion (Mayberg et al.,
affect cognitive performance (Simpson et al., 1999; Bush et al., 2000; Yamasaki et al., 2002).
2001b; Davidson, 2002; Davidson et al., 2003). Drevets and Raichle (1998) describe a reciprocal
In his model about the emotion–cognition in- association between emotional and cognitive brain
teraction, Gray (2001) postulates that approach areas. In their review of brain mapping studies, they
and withdrawal emotional states can enhance or mention that in areas involved in emotional
impair cognitive performance depending on the processing, such as the amygdala, the orbitofron-
particular emotion and cognitive process involved, tal cortex, and the ventral anterior cingulate cortex
and that different emotional states can have op- (ACC), blood flow increases during emotion-
posites effects. Heller and Nitschke (1998) found related tasks, but blood flow decreases during per-
that positively valenced emotion tends to facilitate formance of attention-demanding cognitive tasks.
performance on tasks that depend more on the left Conversely, blood flow increases during cognitive
PFC, whereas negatively valenced emotion tends tasks in the dorsal anterior cingulate and the dorso-
to facilitate performance on tasks dependent on lateral prefrontal cortices, and decreases in these
right PFC. According to Tomarken and Keener areas during experimentally induced or pathologi-
(1998) hypothesis approach- and withdrawal-re- cal emotional states. The authors assume that such
lated emotions bias the ability of PFC to organize reciprocal patterns of regional cerebral blood flow
behavior. may reflect an important cross-modal interaction
Cognitive resource models predict that all emo- during mental operations. Two studies demon-
tional states consume resources and that these re- strated that the degree of activity reduction in emo-
sources are no longer available for controlled tion processing brain areas depends on a combined
cognition. Ellis and Ashbrook (1988) introduced effect of attentional demands of the task and ac-
an approach for explaining the effects of primary companying performance anxiety (Simpson et al.,
negative emotions (e.g., depressive states) on task 2001a, b).
performance. On the basis of the concept of lim- Another way to evaluate the effects of emotion
ited processing capacity, their resource allocation on cognitive processes is to examine how affective
model (RAM) assumes that emotions, especially stimuli (e.g. words, pictures) modulate the activity
462

in brain regions known to participate in certain positive encoding context, recall was predicted by
cognitive functions (Whalen et al., 1998; Simpson activation of right anterior parahippocampal and
et al., 2000; Compton et al., 2003). Using fMRI extrastriate visual brain areas, whereas in a nega-
Whalen et al. (1998) found that the emotional tive encoding context recall was predicted by ac-
Stroop task activates a ventral subregion in the tivation of the amygdala.
ACC, whereas a nonemotional-counting Stroop In summary, studies investigating the impact of
task engages a more dorsal region of the ACC. emotions on cognitive processes, while using
Compton et al. (2003) found that dorsolateral different paradigms, show that there is a dynamic
frontal lobe activity was increased by negative and exchange between cognitive task performance and
incongruent color words, indicating that these re- emotional states. Especially areas of the PFC, the
gions are involved in sustaining selective attention cingulate cortex, and the limbic cortex play an
in the presence of salient distractors. Activity in the important role. It seems to be that the degree to
left lateral OFC increased for negative as compared which both domains recruit similar or different
with neutral emotional words. Ignoring negative brain regions depends on the degree to which they
emotional words increases bilateral occipitotempo- involve similar or different processing components
ral activity and decreases amygdala activity. The (Compton et al., 2003).
authors suggest that emotion and attention are re-
lated via a network of regions that monitor for
salient information, maintain attention on the task,
suppress irrelevant information, and select appro- The interaction of emotion and cognition in
priate responses. psychopathy
The majority of experiments interested in the
interrelationship between emotion and cognition In spite of empirical evidence that emotion
used emotional states (e.g. test anxiety) or patho- processing is critically imbalanced in psychopathy
logical traits (e.g. depression) to study the impact and that emotional and cognitive processes are
of emotions on cognitive processes, or they studied closely intertwined, it can be assumed that the
the cognitive processing of emotional stimuli (e.g. emotional impairment of psychopaths also influ-
emotional Stroop). ences their cognitive processing.
Only few studies evaluate the effects of induced Williamson et al. (1991) studied psychopaths
emotions on cognitive processes. Gray et al. (2002) during performance of an emotional lexical deci-
examine the conjoint effects of cognitive task and sion task. In contrast to healthy controls, psycho-
emotional state manipulations on lateral PFC. The paths show no difference in reaction time and no
authors suggest that the integration of emotion P200 differentiation between emotional and neu-
and cognition would allow goal-directed control of tral words. However, they show better accuracy
behavior dependent on the emotional context, and for emotional than neutral words. Lorenz and
this complex function is mediated by lateral PFC. Newman (2002) investigated the influence of emo-
They use fMRI to assess brain activity during the tional cues on a lexical-decision task and showed
performance of two different demanding cognitive that low-anxious psychopaths did not differ from
tasks (three-back memory task for verbal stimuli controls regarding appraisal of emotional cues and
or affective faces) after a preceding emotional lexical decisions. However, psychopaths were im-
video was seen. An emotion–cognition interaction paired in using the emotional informations. Intra-
was found in bilateral PFC and was interpreted as tor et al. (1997) used single photon emission
evidence that emotion and cognition conjointly computed tomography (SPECT) and found that
and equally contribute to the control of thought psychopaths show greater activation bilaterally in
and behavior. temporofrontal areas for affective words in con-
Erk et al. (2003) investigated the impact of an trast to neutral words. They suggest that the ac-
emotional context on subsequent recall of emo- tivation increase is a functional correlate of extra
tionally neutral material. They found that in a effort, which is required to solve the task.
463

Gordon et al. (2004) studied a nonpsychiatric negative emotions consume no resources. These re-
population using the psychopathy personality in- sults suggest that for psychopaths, negative emo-
ventory (PPI; Lilienfeld and Andrews, 1996) to tions do not drain attention and possibly have less
create subgroups with high and low trait measures informational content. Therefore, their regulatory
of psychopathy. Participants performed a recog- influence on behavior is lower than in nonpsycho-
nition task that required attention be given to ei- paths.
ther the affect or identity of face stimuli. No
significant behavioral differences were found. But
in the affect recognition task, low-scored partici- Conclusion
pants show greater activation in right inferior
frontal cortex, right amygdala, and medial PFC. Impaired emotional processing is a key factor in
Participants with high scores of trait psychopathy psychopathy. Psychopathic individuals show se-
show significantly greater activation in visual cor- vere difficulties in aversive conditioning (Flor
tex and right dorsolateral PFC. They found no et al., 2002; Schneider et al., 2000; Birbaumer et
differences between the two groups in response to al., 2005), the modulation of the startle response
the identity recognition task. Gordon et al. (2004) (Patrick et al., 1993; Levenston et al., 2000), and
suggest that individuals with high trait measures of the processing of sad and fearful social stimuli, like
psychopathy show a special neural signature as- affective facial expressions (Blair, 2001, 2003) or
sociated with different task processing strategies. vocal affect (Blair et al., 2002). In psychopaths,
All these studies used emotional stimuli to eval- negative emotions show no impact on cognitive
uate the affective influence on cognitive processes. processing (Müller et al., 2005). In summary,
But how does an affective context influence the whereas in healthy controls negative emotions
processing of a cognitive task in psychopathy? affect ongoing behavior, this regulatory influence
Müller et al. (submitted) studied the impact of is absent in psychopaths and this may be a key
pleasant and unpleasant emotions on interference factor for their socialization problems.
processing in psychopathic patients and healthy
controls. They use fMRI to assess brain activity
during the performance of a Simon task after emo- References
tion induction with positive, negative, and neutral
pictures from the International Affective Picture Adolphs, R., Tranel, D., Hamann, S., Young, A.W., Calder,
A.J., Phelps, E.A., Anderson, A., Lee, G.P. and Damasio,
System (IAPS; CSEA-NIMH, 1995) (Sommer et A.R. (1999) Recognition of facial emotion in nine individuals
al., 2004). Whereas control subjects made more er- with bilateral amygdala damage. Neuropsychologia, 37:
rors in the negative emotional context than in the 1111–1117.
positive or neutral context, psychopaths show no Bechara, A., Damasio, H., Damasio, A.R. and Lee, G.P. (1999)
Different contributions of the human amygdala and ventro-
influence of the induced emotion on error rates, al-
medial prefrontal cortex to decision-making. J. Neurosci., 19:
though they do not differ from healthy controls in 5473–5481.
their valence ratings. For healthy controls the ac- Birbaumer, N., Veit, R., Lotze, M., Erb, M., Hermann, C.,
tivation data show a network of areas, which was Grodd, W. and Flor, H. (2005) Deficient fear conditioning in
sensitive to the ‘‘Emotion  Task’’ interaction, in- psychopathy. Arch. Gen. Psychiat., 62: 799–805.
cluding superior and inferior frontal gyrus, anterior Blair, R.J.R. (1995) A cognitive developmental approach to
morality: investigating the psychopaths. Cognition, 57: 1–29.
cingulate, putamen, and thalamus. Especially nega- Blair, R.J.R. (1999) Psychophysiological responsiveness to the
tive emotions lead to less activation during incom- distress of others in children with autism. Pers. Individ. Diff.,
patible trials compared to compatible trials. For 26: 477–485.
psychopathic patients, no interaction effect was Blair, R.J.R. (2001) Neurocognitive models of aggression, the
antisocial personality disorders, and psychopathy. J. Neurol.
found. Whereas for healthy controls the negative
Neurosurg. Psychiatry, 71: 727–731.
emotions increase the information processing load Blair, R.J.R. (2003) Facial expressions, their communicatory
and drain resources that otherwise might be de- functions and neuro-cognitive substrates. Philos. Trans. R.
voted to the Simon task, in psychopathic patients Soc. Lond. B, 358: 561–572.
464

Blair, R.J.R. and Cipolotti, L. (2000) Impaired social response Flor, H., Birbaumer, N., Hermann, C., Ziegler, S. and Patrick,
reversal: A case of ‘acquired sociopathy’. Brain, 123: C.J. (2002) Aversive Pavlovian conditioning in psychopaths:
1122–1141. peripheral and central correlates. Psychophysiology, 39:
Blair, R.J., Colledge, E., Murray, L. and Mitchell, D.G. (2001a) 505–518.
A selective impairment in the processing of sad and fearful Forgas, J.P. (1995) Mood and judgement: the affect infusion
expressions in children with psychopathic tendencies. J. Ab- model (AIM). Psychol. Bull., 117: 39–66.
norm. Child Psychol., 29: 491–498. Fowles, D.C. (1980) The three arousal model: implications of
Blair, R.J.R., Colledge, E. and Mitchell, D.G.V. (2001b) So- Gray’s two-factor learning theory for heart rate, electrode-
matic markers and response reversal: is there orbitofrontal rmal activity, and psychopathy. Psychophysiology, 17:
cortex dysfunction in boys with psychopathic tendencies? 87–104.
J. Abnorm. Child Psychol., 29: 499–511. Gordon, H.L., Baird, A.A. and End, A. (2004) Functional
Blair, R.J.R., Mitchell, D.G.V., Richell, R.A., Kelly, S., Leo- differences among those high and low on a trait measure of
nard, A., Newman, C. and Scott, S.K. (2002) Turning a deaf psychopathy. Biol. Psychiatry, 56: 516–521.
ear to fear: impaired recognition of vocal affect in psycho- Gray, J.R. (2001) Emotional modulation of cognitive control:
pathic individuals. J. Abnorm. Psychol., 111: 682–686. approach-withdrawal states double-dissociate spatial from
Braver, T.S. and Cohen, J.D. (2000) On the control of control: verbal two-back task performance. J. Exp. Psychol. Gen.,
the role of dopamine in regulating prefrontal function and 130: 436–452.
working memory. In: Monsell, S. and Driver, J. (Eds.), Con- Gray, J.R., Braver, T.S. and Raichle, M.E. (2002) Integration
trol of Cognitive Processes: Attention and Performance of emotion and cognition in the lateral prefrontal cortex.
XVIII. MIT Press, Cambridge, MA, pp. 713–737. Proc. Natl. Acad. Sci. USA, 99: 4115–4120.
Bush, G., Luu, P. and Posner, M.I. (2000) Cognitive and emo- Hare, R.D. (1982) Psychopathy and physiological activity dur-
tional influence in anterior cingulate cortex. Trends Cogn. ing anticipation of an aversive stimulus in a distraction par-
Sci., 4: 215–222. adigm. Psychophysiology, 19: 266–271.
Carver, C.S. and Schreier, M.F. (1990) Origins and functions of Hare, R.D. (1991) Manual for the Hare Psychopathy Checklist-
positive and negative affect: a control process view. Psychol. Revised. Multi Health Systems, Toronto, Canada.
Rev., 97: 19–35. Hare, R.D. (1998) Psychopathy, affect, and behavior. In:
Center for the Study of Emotion and Attention (CSEA-NIMH) Cooke, D., Forth, A. and Hare, R. (Eds.), Psychopathy:
(1995) The International Affective Picture System: Digitized Theory, Research, and Implications for Society. Kluwer,
Photographs. The Center for Research in Psychophysiology, Dordrecht, pp. 105–137.
University of Florida, Gainesville, FL. Hare, R.D., Clark, D., Grann, M. and Thornton, D. (2000)
Cleckley, H. (1941) The Mask of Sanity. Mosby, St. Louis, Psychopathy and the predictive validity of the PCL-R: an
MO. international perspective. Behav. Sci. Law, 18: 623–645.
Cleckley, H. (1976) The Mask of Sanity (5th edition). Mosby, Hare, R.D., Frazelle, J. and Cox, D.N. (1978) Psychopathy and
St. Louis, MO. physiological responses to threat of an aversive stimulus.
Cloninger, C.R. (1978) The antisocial personality. Hosp. Pract., Psychophysiology, 15: 165–172.
13: 97–106. Hare, R.D. and Quinn, M.J. (1971) Psychopathy and auto-
Compton, R.J., Banich, M.T., Mohanty, A., Milham, M.P., nomic conditioning. J Abnorm. Psychol., 77: 223–235.
Herrington, J., Miller, G.A., Scalf, P.E., Webb, A. and Hell- Hart, S.D., Forth, A.E. and Hare, R.D. (1990) Performance of
er, W. (2003) Paying attention to emotion: a fMRI investi- criminal psychopaths on selected neuropsychological tests.
gation of cognitive and emotional Stroop tasks. Cogn. J. Abnorm. Psychol., 99: 374–379.
Affect. Behav. Neurosci., 3: 81–96. Heller, W. and Nitschke, J.B. (1998) The puzzle of regional
Davidson, R.J. (2002) Anxiety and affective style: role of pre- brain activity in depression and anxiety: the importance of
frontal cortex and amygdala. Biol. Psychiatry, 51: 68–80. subtypes and comorbidity. Cogn. Emotion, 12: 421–447.
Davidson, R.J., Scherer, K.R. and Goldsmith, H.H. (2003) Intrator, J., Hare, R., Stritzke, P., Brichtswein, K., Dorfman,
Handbook of Affective Sciences. Oxford University Press, D., Harpur, T., Bernstein, D., Handelsman, L., Schaefer, C.,
New York. Keilp, J., Rosen, J. and Machac, J. (1997) A brain imaging
Drevets, W.C. and Raichle, M.E. (1998) Reciprocal suppression (single photon emission computerized tomography) study of
of regional cerebral blood flow during emotional versus semantic and affective processing in psychopaths. Biol. Psy-
higher cognitive processes: implications for interactions be- chiatry, 42: 96–103.
tween emotion and cognition. Cogn. Emotion, 12: 353–385. Kiehl, K.A., Brink, J., Hare, R.D. and McDonald, J. (1999)
Ellis, H.C. and Ashbrook, P.W. (1988) Resource allocation Reduced P300 responses in criminal psychopaths during a
model of the effects of depressed mood states on memory. In: visual oddball task. Biol. Psychiatry, 45: 1498–1507.
Fiedler, K. and Forgas, J. (Eds.), Affect, Cognition and So- Kiehl, K.A., Smith, A.M., Hare, R.D., Mendrek, A., Forster,
cial Behavior. Hogrefe, Toronto, pp. 25–43. B.B., Brink, J. and Liddle, P.F. (2001) Limbic abnormalities
Erk, S., Kiefer, M., Grothe, J., Wunderlich, A.P., Spitzer, M. in affective processing by criminal psychopaths as revealed by
and Walter, H. (2003) Emotional context modulates subse- functional magnetic resonance imaging. Biol. Psychiatry, 50:
quent memory effect. Neuroimage, 18: 439–447. 677–684.
465

Kosslyn, S.M. and Koenig, O. (1992) Wet Mind. Free Press, blood: evidence from a fMRI study on the interaction of
New York. cognition and emotion. Am. J. Psychiatry, submitted.
Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1990) Newman, J.P., Patterson, C.M. and Kosson, D.S. (1987)
Emotion, attention, and the statle reflex. Psychol. Rev., 97: Response preservation in psychopaths. J. Abnorm. Psychol.,
377–395. 96: 145–148.
Lang, P.J., Bradley, M.M. and Cuthbert, B.N. (1998) Emotion Ochsner, K.N. and Gross, J.J. (2005) The cognitive control of
and motivation: measuring affective perception. J. Clin. Ne- emotion. Trends Cogn. Sci., 9: 242–249.
urophysiol., 15: 397–408. Patrick, C.J. (1994) Emotion and psychopathy: startling new
LaPierre, D., Braun, C.M.J. and Hodgins, S. (1995) Ventral insights. Psychophysiology, 31: 319–330.
frontal deficits in psychopathy: neuropsychological test find- Patrick, C.J., Bradley, M.M. and Lang, P.J. (1993) Emotion in
ings. Neuropsychologia, 33: 139–151. the criminal psychopath: startle reflex modulation. J.
Lawrence, A.D., Calder, A.J., McGowan, S.W. and Grasby, Abnorm. Psychol., 102: 82–92.
P.M. (2002) Selective disruption of the recognition of facial Pinel, P. (1809) Traité medico-philosophique sur l’aliénation
expressions of anger. Neuroreport, 13: 881–884. mentale (2nd edition). Brosson, Paris.
LeDoux, J. (1998) The Emotional Brain. Weidenfeld & Nicho- Raine, A. (1989) Evoked potentials and psychopathy. Int.
lson, New York. J. Psychophysiol., 8: 1–16.
Levenston, G.K., Patrick, C.J., Bradley, M.M. and Lang, P.J. Rolls, E.T. (1996) The orbitofrontal cortex. Philos. Trans. R.
(2000) The psychopath as observer: emotion and attention in Soc. Lond. B Biol. Sci., 351: 1433–1443.
picture processing. J. Abnorm. Psychol., 109: 373–385. Rolls, E.T. (2000) The orbitofrontal cortex and reward. Cereb.
Lilienfeld, S.O. (1994) Conceptual problems in the assessment Cortex, 10: 284–294.
of psychopathy. Clin. Psychol. Rev., 14: 17–38. Sass, H. and Herpertz, S. (1995) The history of personality
Lilienfeld, S.O. and Andrews, B.P. (1996) Development and disorders. In: Berrios, G. and Porter, R. (Eds.), A History of
preliminary validation of self-report measure of psychopathic Clinical Psychiatry. Athlone, London, pp. 633–644.
personality traits in noncirminal populations. J. Pers. Assess., Schmitt, W.A., Brinkley, C.A. and Newman, J.P. (1999) Test-
66: 488–524. ing Damasio’s somatic marker hypothesis with psychopathic
Lorenz, A.R. and Newman, J.P. (2002) Deficient response individuals: risk takers or risk averse? J. Abnorm. Psychol.,
modulation and emotion processing in low-anxious Cauca- 108: 538–543.
sian psychopathic offenders: results from a lexical decision Schneider, F., Habel, U., Kessler, C., Posse, S., Godd, W. and
task. Emotion, 2: 91–104. Müller-Gärtner, H.-W. (2000) Functional imaging of condi-
Lykken, D.T. (1957) A study of anxiety in the sociopathic per- tioned aversive emotional responses in antisocial personality
sonality. J. Abnorm. Soc. Psychol., 55: 6–10. disorder. Neuropsychobiology, 42: 192–201.
Mayberg, H.S., Liotti, M., Brannan, S.K., McGinnis, S., Mah- Schoenbaum, G., Chiba, A.A. and Gallagher, M. (1998) Or-
urin, R.K., Jerabek, P.A., Silva, J.A., Tekell, J.L., Martin, bitofrontal cortex and basolateral amygdala encode expected
C.C., Lancaster, J.L. and Fox, P.T. (1999) Reciprocal limbic- outcomes during learning. Nat. Neurosci., 1: 155–159.
cortical function and negative mood: converging PET find- Schoenbaum, G., Chiba, A.A. and Gallagher, M. (2000)
ings in depression and normal sadness. Am. J. Psychiatry, Changes in functional connectivity in orbitofrontal cortex
156: 675–682. and basolateral amygdala during learning and reversal train-
McMurran, M., Blair, M. and Egan, V. (2002) An investigation ing. J. Neurosci., 20: 5179–5189.
of the correlations between aggression, impulsiveness, social Simpson, J.R., Drevets, W.C., Snyder, A.Z., Gusnard, D.A.
problem-solving, and alcohol use. Aggress. Behav., 28: and Raichle, M.E. (2001a) Emotion-induced changes in hu-
439–445. man medial prefrontal cortex: II. during anticipatory anxiety.
Meinhardt, J. and Pekrun, R. (2003) Attentional resource al- Proc. Natl. Acad. Sci. USA, 98: 688–693.
location to emotional events: an ERP study. Cogn. Emotion, Simpson, J.R., Öngür, D., Akbudak, E., Conturo, T.E., Ollin-
17: 477–500. ger, J.M., Snyder, A.Z., Gusnard, D.A. and Raichle, M.E.
Mitchell, D.G.V., Colledge, E., Leonard, A. and Blair, R.J.R. (2000) The emotional modulation of cognitive processing: an
(2002) Risky decisions and response reversal: is there evi- fMRI study. J. Cogn. Neurosci., 12: 157–170.
dence of orbitofrontal cortex dysfunction in psychopathic Simpson, J.R., Snyder, A.Z., Gusnard, D.A. and Raichle, M.E.
individuals? Neuropsychologia, 40: 2013–2022. (2001b) Emotion-induced changes in human medial prefron-
Müller, J.L., Sommer, M., Wagner, V., Lange, K., Taschler, tal cortex: I. during cognitive task performance. Proc. Natl.
H., Röder, C.H., Schuierer, G., Klein, H. and Hajak, G. Acad. Sci. USA, 98: 683–687.
(2003) Abnormalities in emotion processing within cortical Sommer, M., Müller, J.L., Weber, T. and Hajak, G. (2004) Die
and subcortical regions in criminal psychoaths: evidence Bedeutung von Affekt und Emotion für psychiatrische Er-
from a functional magnetic resonance imaging study using krankungen. Psychiatr. Prax., 31: S64.
pictures with emotional content. Biol. Psychiatry, 54: Stevens, D., Charman, T. and Blair, R.J.R. (2001) Recognition
152–162. of emotion in facial expressions and vocal tones in children
Müller, J.L., Weber, T., Sommer, M., Döhnel, K., Meinhardt, with psychopathic tendencies. J. Genet. Psychol., 162:
J. and Hajak, G. Why criminal psychopaths act in cold 201–210.
466

Sutker, P.B. and Allain, A.N. (1987) Cognitive abstraction, emotional learning in antisocial behavior and social phobia
shifting and control: clinical sample comparisons on in humans. Neurosci. Lett., 328: 233–236.
psychopaths and non-psychopaths. J. Abnorm. Psychol., Whalen, P.J., Bush, G., McNally, R.J., Wilhelm, S., McIn-
96: 73–75. erney, S.C., Jenike, M.A. and Rauch, S.L. (1998) The emo-
Tiihonen, J., Hodgins, S., Vaurio, O., Laasko, M., Repo, E., tional counting stroop paradigm: a functional magnetic
Soinen, H., Aronen, H.J., Nieminen, P. and Savolainen, L. resonance imaging probe of the anterior cingulate affective
(2000) Amygdaloid vo lume loss in psychopathy. Soc. division. Biol. Psychiatry, 44: 1219–1228.
Neurosci. Abstr., 2017. Williamson, S., Harpur, T.J. and Hare, R.D. (1991) Abnormal
Tomarken, A.J. and Keener, A.D. (1998) Frontal brain asym- processing of affective words by psychopaths. Psychophyio-
metry and depression: a self-regulatory perspective. Cogn. logy, 28: 260–273.
Emotion, 12: 387–420. Yamasaki, H., LaBar, K.S. and McCarthy, G.M. (2002) Dis-
Veit, R., Flor, H., Erb, M., Hermann, C., Lotze, M., Grodd, W. sociable prefrontal brain systems for attention and emotion.
and Birbaumer, N. (2002) Brain circuits involved in Proc. Natl. Acad. Sci. USA, 99: 11447–11451.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 26

Disordered emotional processing in schizophrenia


and one-sided brain damage

Katarzyna Kucharska-Pietura1,

1
Whitchurch Hospital, Cardiff and Vale NHS Trust, Cardiff CF 14 7XB, UK

Abstract: The work concentrates on the problem of human emotions in healthy and pathologically
changed brains, mainly in persons afflicted with schizophrenia or with organic impairments localized in one
of the cerebral hemispheres. This chapter presents the state of current knowledge concerning the hemi-
spheric lateralization of emotions among healthy people, psychiatric patients, and patients with one-sided
brain lesion, on the basis of clinical observations, the results of experimental work, and the newest ne-
uroimaging techniques. The numerous experiments and scientific methods used to assess the hemispheric
lateralization of emotions and the discrepancies in their results point toward a lack of consistent theory in
the field of hemispheric specialization in the regulation of emotional processes. Particular scientific interest
was taken in the emotions of persons afflicted with schizophrenia, either in its early or late stages. This was
inspired by the emotional behavior of schizophrenic patients on a psychiatric ward and their ability to
perceive and express emotions during various stages of the schizophrenic process. In order to examine the
cerebral manifestations of emotional deficits and the specialization of cerebral hemispheres for emotional
processes, the author has described the emotional behavior of patients with unilateral cerebral stroke,
i.e., patients with damage to the right or left cerebral hemisphere. Overall, the inferior performance of
emotional tasks by right-hemisphere-damaged patients compared to other groups might support right-
hemisphere superiority for affect perception despite variations in the stimuli used.

Keywords: hemispheric specialization; chimeric faces; affects perception; emotion expression;


schizophrenia; one-sided brain damage

Hemispheric specialization for emotional processing two main hypotheses. The first one assumes that
structures analyzing both positive and negative
Increasingly more experimental data and clinical emotions are related exclusively to the right
studies demonstrate the brain asymmetry in the hemisphere (RH) (Borod et al., 1983, 1988;
processing of affect (Reuter-Lorenz and Davidson, David and Cutting, 1990; Kucharska-Pietura
1981; Sackeim et al., 1982; Dawson and Fisher, et al., 2002). According to the second standpoint
1994; Altenmüller et al., 2002; Noesselt et al., — valence hypothesis — the brain laterality re-
2005). The clinical observation of psychiatric and garding emotions is dependent on the emotional
neurological patients with right- or left-brain-hem- valence: The RH controls negatively valenced
isphere damaged, the findings of amytal trials, and emotions and the left hemisphere (LH) those that
neuropsychological experimental studies confirm are positively valenced (Reuter-Lorenz and Da-
vidson, 1981; Bryden et al., 1982; Mandal et al.,
Corresponding author. Tel.: +44-2920693941; 1998; Adolphs et al., 2001; Simon-Thomas et al.,
Fax: +44-2920627954; E-mail: katepietura@hotmail.com 2005). Heilman (1997) updated this hypothesis by

DOI: 10.1016/S0079-6123(06)56026-1 467


468

discussing a model of a modular cortical network activity was examined during a higher cognitive
influencing the limbic structures. According to this task (Simon-Thomas et al., 2005). The authors ob-
theory, the frontal lobes play a key role for valence served that the RH is more influenced by negative
with the left mediating positive emotions and the emotions and concurrent cognitive processing is
right negative affect. Furthermore, Heilman (1997) consistent with the valence hypothesis proposing,
highlighted the crucial role of the RH in activating that the RH is dominant for processing with-
arousal systems and of the LH in modulating in- drawal-related emotions, and that LH may play
hibition of these systems. Finally, he concluded crucial role in approach-related emotions.
that the orbito-frontal regions mediate avoidance
behaviors and the parietal lobes mediate approach
behaviors. The perception of emotional chimeric faces in
Emotional processing in the context of brain schizophrenia and brain damage: further evidence of
asymmetry has been studied utilizing the methods right-hemisphere dysfunction
of functional neuroimaging (Northoff et al., 2004;
Simon-Thomas et al., 2005). Blood et al. (1999) Neuropsychological research has established that
using positron emission tomography (PET) as- the right cerebral hemisphere plays a special role
sessed emotional responses to pleasant and un- in tasks involving the direction of spatial attention
pleasant music in healthy volunteers. The authors — especially to the contra-lateral hemispace —
showed a significant increase of cerebral blood recognition of faces, perception, and possibly gen-
flow in the right parahippocampal gyrus in re- eration of affect (Davidson, 1984; Borod and Koff,
sponse to unpleasant stimuli and a decrease of 1989; Pally, 1998; Lane et al., 1999; Kucharska-
blood flow in the orbitofrontal cortex and the Pietura et al., 2002). One of the simplest to ad-
subcallosal cingulum. Noesselt et al. (2005) used minister yet most robust neuropsychological test
event-related functional magnetic resonance imag- that elicits right cerebral hemisphere ‘‘dominance’’
ing (fMRI) to assess the asymmetrical brain acti- is the chimeric faces test. This was first developed
vation during processing of fearful versus neutral by Levy et al. (1983) using chimeras combining
faces in healthy subjects. They found right-later- photographs of open-mouthed smiles with neutral
alized brain activations for fearful minus neutral facial expressions, joined at the vertical meridian.
left-hemifield faces in right visual areas as well as Normal dextrals perceive those faces with a smile
more activity in the right than in left amygdale. to their left as happier. This ‘‘left hemiface bias’’ is
This might suggest right-lateralized emotional observed regardless of the means of presentation
processing during bilateral stimulation involving (in free vision or tachistoscopically) (Grega et al.,
enhanced coupling of the amygdale and RH ex- 1988; Phillips and David, 1997) and has been in-
trastriate cortex. terpreted as evidence of RH dominance in facial
Utilizing electroencephalography (EEG), Sch- processing (Kolb et al., 1983; Levy et al., 1983).
midt and Trainor (2001) assessed frontal brain Adaptations of this technique have been developed
electrical activity in healthy controls during their such as the use of happy–sad chimeric face draw-
rating on valence and intensity of four orchestral ings (David, 1989). The realism of the faces (Hopt-
pieces. The authors found a greater activity relative man and Levy, 1988), the valence of the expression
to the left frontal EEG in response to positive (Christman and Hackworth, 1993), the relevance
stimuli and greater activity relative to the right of affect to the task (Luh et al., 1991, 1994), and
frontal activity for negative stimuli. A similar age (Levine and Levy, 1986) have all been found to
lateralization pattern was shown by Altenmüller exert little effect on the perceptual bias which is
et al. (2002) who found positive emotional remarkably robust. Indeed a similar perceptual
attributions accompanied by an increase in the left asymmetry has been found with nonfacial stimuli
temporal activation and negative emotional attri- (Luh et al., 1991, 1994) but it is of smaller mag-
butions by an increase in the right frontotemporal nitude. So while stimuli utilizing facial expressions
cortex. Further, event-related potentials (ERPs) evoke the strongest bias, it is more accurate to
469

refer to the phenomenon as a left-hemispatial bias personality types (Jessimer and Markham, 1997).
(LHB). Further evidence for the RH ‘‘dominance’’ David and Cutting (1990) were first to demon-
on this task comes from the fact that non-right strate markedly reduced LHB in patients
handers show a reduction or even reversal of the with schizophrenia and this has been replicated
LHB (Levy et al., 1983; Hoptman and Levy, 1988; (Phillips and David, 1997; Lior and Nachson,
David, 1989; Luh et al., 1994). 1999). Furthermore reductions in LHB have
Versions of the chimeric faces test have been recently been shown in males with schizotypal
used to probe RH functioning in patients with traits (Mason and Claridge, 1999) and young
psychiatric disorders (David and Cutting, 1990; people who are theoretically ‘‘psychosis prone’’
Gooding et al., 2001). Such tests have the advan- (Luh and Gooding, 1999).
tage of being simple and quick to administer and Kucharska-Pietura and Klimkowski (2002b)
they require very little effortful cognitive process- compared 100 remitted schizophrenic inpatients
ing. Jaeger et al. (1987) were first to report a re- with both healthy controls (n ¼ 50) and the pa-
duction in LHB in patients with unipolar tients with RH (n ¼ 30) and LH damage (n ¼ 30)
depressive illness. David and Cutting (1990) rep- and found that the pattern of the performance on
licated the reduced LHB in depression using a the chimeric faces was similar between schizo-
happy–sad chimeric drawings test (David, 1989). phrenic patients (particularly chronic ones) and
They also showed a nonsignificant increased LHB the right-brain-damaged group that showed re-
in patients with mania and an absence of bias in duced LHB. Further, left-brain-damaged patients
schizophrenia. These findings were replicated in an did not affect LHB and was very comparable
independent sample (David, 1993). to the pattern of results obtained in the healthy
Apart from non-right handedness (Levy et al., controls (Fig. 1).
1983; David, 1989; Luh et al., 1994; Jessimer and The results show that the schizophrenic patients
Markham, 1997), the other factors which have and right-brain-damaged patients had a signifi-
been shown to modify LHB using facial expression cantly weaker LHB when judging affect from
chimeras are psychiatric disorders and certain schematic chimeric faces, compared with healthy

Fig. 1. Mean values of LHF bias in each subject group: S, first episode schizophrenics; CS, chronic schizophrenic; N, normal controls;
R right-brain-damaged patients; L, left-brain-damaged patients.
470

subjects and left-brain-damaged patients. This was Cutting, 1992), particularly in systems that direct
most marked in the chronic patients and right- spatial attention (Phillips and David, 1997). The
brain-damaged patients who showed comparable role of the LH is unclear with respect to this task,
pattern of performance, on average a right-sided performance being unaffected by LH resections
perceptual bias. This in turn suggests RH dys- (Kolb et al., 1983). However, an imbalance be-
function or, at the very least, an alteration in the tween the two hemispheres, with the right failing
balance between the hemispheres. to predominate and the left apparently ‘‘overac-
The most obvious interpretation of these results tive’’ would also account for the results. It is more
in relation to schizophrenia is that it is a disorder difficult to link this to particular features of the
of RH functioning. Indeed there is considerable disorder since there is no obvious link to symptom
evidence for this from a range of neuropsycholog- clusters or self-reported mood.
ical studies, particularly those employing emo- In summary, the majority of the results repli-
tional or facial stimuli (David and Cutting, 1990; cated the lack of the normal left-perceptual bias in
Kucharska-Pietura et al., 2002). schizophrenic patients, regardless of symptom
Furthermore, the findings are extended in that profile, which is attributable to right-brain dys-
they show that the loss of LHB is present early in function. The use of chimeric faces as a means
the course of the disorder but is more pronounced of tapping anomalous spatial attentional bias in
in chronic patients (Bryson et al., 1998; Wölwer people with or at risk of developing schizophrenia
et al., 1996; Kucharska-Pietura et al., 2002). is recommended. Future work using functional
Longitudinal studies are required to determine neuroimaging techniques or cortical electrophysi-
whether lack of perceptual bias precedes or follows ology may shed light on the nature and precise
a chronic course. Lior and Nachson (1999) found anatomical location of the dysfunction.
a weaker than normal left-hemiface bias in their
sample of schizophrenic patients and reported that
the preponderance of symptoms, negative or posi- Emotion perception in schizophrenia and one-sided
tive, influenced the affect label designated to the brain damage
stimuli. Kucharska-Pietura et al. (2002) showed
that LHB was not influenced by symptom type, or Over the last few decades there have been numer-
medication dose. Further, self-reported mood did ous studies examining the perception of facial
not appear to correlate with LHB in controls or affect in healthy population and pathology (Pally,
patients — replicating previous studies (Harris and 1998; Herbener et al., 2005; Martin et al., 2005;
Snyder, 1992; David and Cutting, 1990). Scholten et al., 2005), however the neural corre-
The effect of motor lateralization was explored lates of emotional and facial processes are still less
in more detail by sub-dividing each group — all of clear cut and need further investigation.
whom reported right-handedness — in terms of Perception of facial emotion is thought to be a
foot and eye preference. The results for the con- complex cognitive ability which relies on the integ-
trols and first-episode patients show that, within rity of a select set of more basic neurocognitive
these strict parameters, atypical motor laterality processes such as visual scanning, working mem-
affects LHB in the predicted way, that is, non-right ory, and vigilance which may be asymmetrically
sidedness is associated with reduced left- or indeed distributed across the cerebral hemispheres (Kee
right-perceptual bias. The pattern in the chronic et al., 1998). According to Bruce and Young
patients is less clear, presumably because they (1986), facial affect processing is held to be de-
fail to show the normal LHB as a group, but is pendent upon view-centered descriptions created at
consistent with earlier reports of anomalous the structural encoding stage and involves neuro-
lateralization (Piran et al., 1982; Robertson and logical pathways different from the pathways used
Taylor, 1987; Gorynia and Dudeck, 1996). in unfamiliar face matching and familiar face rec-
The most obvious interpretation of these results ognition. This model has been confirmed by Cicone
is that they reflect RH dysfunction (David and et al. (1980), Etcoff (1984), and Bowers et al. (1985).
471

A deficit in emotional perception in schizophre- involving exitotoxicity in addition to neurodevel-


nia has been reported by numerous authors opmental ones (Garver et al., 1999). Here, illness
(Feinberg et al., 1986; Borod et al., 1993; Kerr duration rather than symptom ratings per se would
and Neale, 1993; Edwards et al., 2001; Silver et al., then be predicted to be correlated with task per-
2002; Herbener et al., 2005). However, as yet formance deficits. Further research examining the
the precise implications of this are unclear. All the background of deficits in the perception of facial
results of the studies carried out in the last decade emotions is needed.
found a general cognitive deficit in schizophrenia, Neuropsychological research activity on nonver-
but in addition also found a specific emotional bal behavior mainly leads to the conclusion that
deficit, usually in the context of affect processes affect relies on specific neural pathways and, more
(Borod et al., 1993; Kerr and Neale, 1993; Salem particularly, that the RH plays a dominant role in
et al., 1996). In support of this are studies showing various emotional processes (Davidson, 1984;
that schizophrenia patients have a generalized per- Gainotti, 1984). This notion is supported by the
formance deficit encompassing all facial emotions, findings that right-brain-damaged patients are sig-
but also have nonemotional faces which might nificantly less accurate in decoding both positive
suggest right cerebral hemisphere dysfunction and negative affect compared to left-brain-damaged
(Novic et al., 1984; Feinberg et al., 1986; Archer patients and healthy controls (Bowers et al., 1985).
and Hay, 1992; Heimberg et al., 1992; Kucharska- However, involvement of the RH in emotion
Pietura and Klimkowski, 2002a). However, others processing might be only a particular instantiation
appear to have a specific difficulty with perceiving of the holistic processes for which it is assumed to
negative emotions in visual (Muzekari and Bates, be specialized (Buck, 1985).
1977; Mandal and Rai, 1987; Cramer et al., 1989) The right-brain-damaged patients and schizo-
and nonvisual modality (Mandal, 1998; Edwards phrenic subjects at their early and late stages of
et al., 2001). illness revealed significantly greater impairment in
The question of the outcome of deficits in emo- the performance of facial affect than healthy con-
tion perception still remains open. There are some trols (Kucharska-Pietura and Klimkowski, 2002a,
findings demonstrating that schizophrenic patients b). Interestingly, the extent of these deficits among
in later stages of illness were significantly impaired right-brain-damaged patients and chronic schizo-
compared to those in early stages of illness in phrenics was comparable. Furthermore, both pa-
recognizing all examined expressions (Kucharska- tient groups showed significant difficulties
Pietura and Klimkowski, 2002a, b; Silver et al., compared to healthy controls in recognition of
2002). This finding suggests a progressive impair- unfamiliar neutral faces, although this deficit was
ment in emotion identification in schizophrenia, less than the deficit of facial affect perception.
which may have resulted from treatment with Bowers et al. (1985) showed that the impairments
typical antipsychotics, institutionalization, or the of facial affect performance among RH-damaged
illness itself. However, only a prospective design patients remained even when their perceptual iden-
would allow determining whether emotion percep- tity performance was statistically partialled out.
tive deficits were truly progressive in schizophre- This might suggest that the deficit in facial affect
nia. To date, the link between the perceptual perception could not entirely stem from visuoper-
emotion deficit and psychopathology has not ceptual impairment but also from different cogni-
been confirmed in schizophrenia. Moreover, the tive processes involved in matching views of
stability of such deficit and lack of correlations unfamiliar faces (Bowers et al., 1985).
between performance on any emotion recognition We found a significant relationship between
tasks and variables have been already proven the affected hemisphere and valence for emotional
(Gaebel and Wölwer, 1992; Kucharska-Pietura decoding (Kucharska-Pietura et al., 2003). Right-
et al., 2002a, b). This pattern of results might be brain-damaged patients were significantly impaired
explained from some recent models advocating compared to healthy volunteers in perceiving neg-
progressive, essentially neurodegenerative processes ative emotions only, whereas for perception of
472

positive ones, the group differences did not reach positive ones. Moreover, Mandal and Pal-
significance (Kucharska-Pietura and Klimkowski choudhury (1985) revealed significantly greater
2002a). These results lend some support to valence impairment in recognition of fear and anger by
theory (Sackeim et al., 1982). Thus, one would chronic schizophrenics compared to normals.
predict that right brain damage leads to impaired These findings showing much worse ability for
perception negative but not positive emotion. The schizophrenia in identification of negative affect
predominance of right-sided activation in recent might also suggest the differential deficit in emo-
neuroimaging studies may also reflect the essential tional decoding (Bell et al., 1997). Such deficit may
role of the RH in the perception of negative emo- stem from inefficiency of information processing
tion per se (Phillips et al., 1998; Lane et al., 1999). following RH dysfunction.
Borod and Koff (1989) examining recognition of Another conclusion reached from the study is
facial affect and vocal prosody reported that schiz- lack of significant correlation between perform-
ophrenic patients performed poorly compared to ance on all perceptual measures and the demo-
normals but did not differ from patients with RH graphic and clinical variables. Although, there was
damage. a main effect of age and of Mini Mental State
In another study of Borod et al. (1993), schizo- Examination (MMSE) (Folstein et al., 1975) on
phrenic patients, right-brain-damaged patients, and facial task performance, McDowell et al. (1994)
controls did not differ in identification of positive evaluated age differences in facial emotional de-
emotions, but schizophrenics and RH-damaged coding in healthy controls. Their findings suggest
patients were significantly impaired relative to con- that the elderly were more impaired in processing
trols in identifying negative emotions, but they were negative affect compared to the younger group,
not different from each other. To assess the cate- while their ability to process positive affect re-
gory preferences in emotional decoding, the mean mains intact. This might support the hypothesis
maximal and minimal ratings given to each of nine that the RH declines more rapidly than LH in the
emotion categories were determined. aging process.
Overall, the highest percentage of agreement In our study positive and negative psychopa-
was observed for happiness in each subject group. thology as measured by the Positive and Negative
All subjects were significantly less accurate in per- Syndrome Scale (PANSS) (Kay et al., 1987) has no
ceiving facial emotions, such as shame, contempt, influence on the task performance in schizophrenia
and disgust. which it is in accord with previous evidence
Interestingly, both patient groups were signifi- of a lack of relationship between the perceptual
cantly impaired compared to healthy volunteers in emotion deficit and psychopathology (Novic et al.,
recognizing all examined expressions, although 1984; Lewis and Garver, 1995; Bellack et al., 1996;
valence expression comparison showed signifi- Kucharska-Pietura et al., 2002).
cantly greater impairment in perceiving negative Moreover, there was also no significant correla-
affect. These results lend some support to the con- tion of age, years of education, and the mean score
clusion that affect relies on specific neural path- of MMSE, which might suggest stable deficits in
ways and, more particularly, that the RH plays a the perception of emotions. Current mood did not
dominant role in processing of emotions strongly also appear to alter the task performance among
connected with survival, preparation for action, schizophrenics. Moreover, lack of significant cor-
and vigilance processes (Heilman, 1982). In agree- relation between emotion perception performance
ment with our data, the schizophrenic patients and neuroleptic dose in both our study and pre-
have been found before to recognize happiness vious works may militate against neuroleptic
most accurately and shame and disgust less effects being a crucial factor responsible for defi-
correctly (Dougherty et al., 1974). cit in emotion performance (Schneider et al., 1992;
Muzekari and Bates (1977) reported that Lewis and Garver, 1995; Kucharska-Pietura et al.,
chronic schizophrenic patients relative to healthy 2002; Kucharska-Pietura and Klimkowski, 2002a).
controls labeled negative emotions poorly but not This data like our results seems to suggest stable
473

properties of perceptual deficit more than its state- and situation, their difficulties in communication
dependent nature. of emotions and elated mood. Furthermore,
The question of the stability and durability of these patients have difficulties receiving and eval-
deficits in the emotional decoding still remains uating emotional signs (Gainotti, 1972; Herzyk
open. Further investigations are needed to find a and Rozenkiewicz, 1994; Kucharska-Pietura
solution for the nature of generalized poor per- and Klimkowski 2002b; Kucharska-Pietura and
formance in emotional decoding in schizophrenia Hunca-Bednarska, 2002). The impaired emotional
and unilateral brain damage. expressivity of right-brain-damaged patients is dis-
played by poor mimic expression (Kaplan et al.,
1990), disturbances in expressive prosody, and
Emotional expression in schizophrenia and brain weakened gestures (Pally, 1998). ‘‘Catastrophic’’
damage reactions, depression, outbursts of tears, anger and
aggression towards the environment, and com-
Despite the fact that emotional disturbance is re- plaints about health are characteristic of people
garded as one of the most prominent schizophrenic with left-brain-hemisphere damage (Gainotti,
symptoms, increasingly more often it has been in- 1972). Left-side brain damage does not result in
dicated that in this illness there is nothing as simple disturbances of emotional expression (Kucharska-
as emotional disappearance (Mazurkiewicz, 1958). Pietura and Hunca-Bednarska, 2002; Kucharska-
At present, searching for the mechanisms of ‘‘re- Pietura and Klimkowski, 2002b). Difficulty in rec-
moving the feelings from our observation’’ in ognizing and evaluating emotional signs in these
schizophrenia, pointed out by Bleuler, which has patients was not reported.
earlier been one of the topical issues (Bleuler, The assessment of emotional behavior in patients
1906), has become the challenging aim. with paranoid schizophrenia during the early and
There are still controversies as to the nature and late stages of schizophrenic process and in patients
magnitude of these disturbances. Experimental with unilateral brain damage was carried out with
studies, carried out mainly in the last two dec- the Observational Scale of Emotional Behavior
ades, brought detailed findings confirming diffi- (Kucharska-Pietura and Klimkowski, 2002b) in
culties of schizophrenic patients in expression of which three categories related to the laterality of
emotions, particularly positive emotions, and also brain dysfunction were distinguished: (A) mood
the failure of these patients in the identification of expression behavior related to attitude processes,
their own emotional experiences. Measurement of (B) emotional reaction type, and (C) the form of the
the activity of mimic muscles using electromyo- emotional communication. Indicatory scales, or-
graphy in schizophrenia showed a higher activity dered from 1 to 31, are included in these categories.
of the corrugator supercili muscle and lesser ac- Attitude processes are expressed in the attitude
tivity of the zygomatic muscle compared to of the studied individual to the situation of the
healthy controls (Earnst et al., 1996; Kring and examination, to the person carrying out this ex-
Neale, 1996). This pattern of results might suggest amination, by the self-estimation of the individ-
that schizophrenic patients showed better ability ual’s own health state and in the mood of those
to express negative emotions than positive ones. examined.
Studies on consequences of the damage to the The assessment of emotional reaction types in-
right- or left-brain hemisphere stressed diverse pro- cluded several behavioral aspects such as the fre-
files of emotional disturbances and different hy- quency of diminishing the degree of the disorder,
potheses have been advocated in relation to brain inadequate behaviors (e.g., joking in situations
asymmetry in affect processing (Jackson, 1879; which are commonly considered joyless), euphoric
Gainotti, 1972; Bryden et al., 1982; Kucharska- reactions, and depressed behaviors (e.g., feeling of
Pietura et al., 2003). First hypothesis dates back to resignation).
Jackson (1879) who observed inadequate attitude Part (C) of the Observational Scale of Emo-
of the patients with RH damage to their own state tional Behavior was aimed to assess the forms of
474

emotional communication. On one side, commu- anxiety, and complained about the bad state of
nication is the expression and conveying of their health. As a rule, these symptoms are treated
information; on the other, it is reception and rec- by the environment as a normal reaction to the
ognition of information. The scale content in- everyday functioning impairment due to a loss of
volves emotional expression (voluntary and motion and language abilities. The patients with
involuntary), ability to verbalize emotional expe- right-side brain damage evaluate their health state
riences (speaking about emotional experience), inadequately, overestimate their abilities, and
and ability to recognize emotionally colored com- show elated mood and often lack of insight of
municates and to verbalize the understanding. The their illness (Herzyk and Rozenkiewicz, 1994;
Observational Scale of Emotional Behavior was Pally, 1998).
enriched by experimental tasks with the purpose Using the Observational Scale of Emotional
of catching some behavior related to emotional Behavior for the assessment of the schizophrenic
communication. These tasks involved: voluntary patients’ emotionality presents a kind of novelty.
face mimicry, the ability to express emotions with However, the obtained findings remain consistent
voice intonation and gesture when asked by the with generally established standpoints. The pa-
interviewer, and also understanding of verbal tients with longer lasting illness revealed passive
jokes and colloquial linguistic phrases regarding attitude to the situation of examination and
emotions (e.g., ‘‘I’m weak in the knees’’). The ex- avoided emotional involvement significantly
perimental trials were applied to capture emo- more often. This passivity is often inseparably
tional reactions difficult to observe in natural connected with a negative dimension of the
settings. The assessment of emotional reactions schizophrenic process. Behavioral change mani-
was based on observation and introspective mate- fested itself in the loss of interest, feelings of
rial was not used. The observations were carried pointlessness, passivity, and an indifferent attitude
out in a natural social context during individual towards the environment (Kucharska-Pietura and
talks with the studied patients. Klimkowski, 2002b).
The analysis of our data revealed statistically It is worth stressing that the present study in-
significant differences between the groups of pa- vestigated those behaviors of the patients that are
tients, which were shown in their attitude to the met commonly in their everyday living situations.
examination, self-evaluation, present mood, and The study was concentrated on social, although
the types of emotional reaction. Individuals perhaps superficial, aspects of their functioning.
with left-side and right-side brain damage were The question, which should be asked at this stage
different with regards to activity, adequacy of their of consideration, regards a possible relationship
evaluation, the quality of their mood, and the between changes in attitude processes and type of
pattern of the emotional reactions. Both studied emotional reactions and unilateral brain damage.
groups were found to be opposite in the assessed While damage in one of the cerebral hemispheres
diameters (e.g., decreased mood — elated mood; was evidenced in neurological patients, the alleged
positive attitude to the situation of the examina- laterality of ‘‘damage’’ in schizophrenia remains
tion — negativism; inadequacy of the evaluation unclear.
of their own health state — adequacy of this Patients with schizophrenia seem to be more
evaluation). Of those studied, the chronically different from right-–brain-damaged patients. But
schizophrenic individuals were characterized by with regards to the adequacy of their evaluation of
the highest passivity and the highest mood in- their own health state and positive attitude to the
difference. examination, they appear to be closer to those with
The assessment of attitude processes and emo- left-side brain damage. However, passiveness grow-
tional reactions in patients with schizophrenia ing along with persistence in time of the schizo-
places them in the position between those with left- phrenic process (chronic patients) and indifferent
and right-side brain damage. The patients with mood observed in all schizophrenic patients deter-
left-side brain damage reveal decreased mood, mine their specific ‘‘intermediate’’ position.
475

‘‘Decreasing the degree of their own disorders’’ those with right-sided damage — therefore the first
was most often observed in the right-brain-dam- group was closer to the norm. Wits and common
aged patients and in the schizophrenic patients. phrase understanding regarding emotions did not
Those patients when asked about their dysfunctions distinguish left- and right-brain-damaged patients,
frequently denied their occurrence or answered that nevertheless ‘‘left-sided’’ patients showed a ten-
they did not disturb them. Whilst taking up the dency to understand wit better (statistical signifi-
task, they were frequently uncritical of their mis- cance was close). Individuals with left-side brain
takes. The feelings of resignation in the group of damage behaved more frequently similar to
schizophrenic patients occurred significantly less healthy individuals with regards to speaking about
than in individuals with left-side brain lesion. emotional experiences, humor, and verbal wit un-
These patients evaluated their present situation derstanding. Furthermore, schizophrenic patients
less frequently in a negative light. They did not feel exhibited deficits in the verbalization of emotions
any sense in treatment and in rehabilitation and and did not differ significantly in this respect from
their attitude to their future was critical. It is in- all the other studied groups. Interestingly, verbal-
teresting that in the studied population of patients ization of emotion, understood as speaking about
with brain damage, inadequate behaviors occurred emotional experiences and verbal wit understand-
more frequently than in schizophrenic patients. ing, did not worsen along with the duration of the
To sum up, a dichotomy division of emotional illness, while the capacity for understanding the
disturbances based on hemispheric damage local- common linguistic phrases did deteriorate.
ization appears to present a simplification of The findings of this study indicate that schizo-
diagnostic understanding. It does not explain as phrenic individuals show similarity to the right-
to what degree emotional changes are direct con- brain-damaged patients with regards to expression
sequences of brain mechanism disturbances and to and verbalization of emotions; however, these
what degree are the secondary reaction due to functions were significantly more deteriorated
the perceptual, cognitive, and movement deficits in schizophrenic patients. Similarity was greatest
(Herzyk and Rozenkiewicz, 1994). The attitude in nonverbal emotional expression, particularly
which processes disturbances and emotional reac- in the ability for voluntary facial expression. No
tion in the group of schizophrenic patients might particular similarity was seen between these two
not so much be the result of asymmetry of the groups in their ability to communicate emotionally
brain hemispheres but rather the result of frontal with gestures. Verbalization of emotions (speaking
lobes dysfunction which are responsible for steer- about emotional experiences) significantly differ-
ing and control of emotional behaviors (Heilman entiated between those with right- and left-sided
et al., 2000; Kohler et al., 2000). brain damage. Therefore, patients with right-
Data based on part (C) of the Observational brain-hemisphere injury were closer to those with
Scale of Emotional Behavior showed that schizo- schizophrenia than all the other groups studied.
phrenic patients differed most from the other At the same time, verbalization of emotions
groups with respect to verbalizing emotions considered as humor and verbal wit understand-
(Kucharska-Pietura and Hunca-Bednarska, 2002; ing did not differ significantly between left- and
Kucharska-Pietura and Klimkowski, 2002b). They right-brain-damaged patients (only a trend toward
spoke about emotional experiences significantly lower/higher significance was observed); therefore
less frequently and did not understand verbal it is very difficult to discuss a clear similarity
humor and wit significantly more often than oth- between those ‘‘right-sided’’ and those with schiz-
ers. They also had more difficulties in understand- ophrenia.
ing common verbal expressions regarding ‘‘Right-sided’’ patients presented good under-
emotions. This latter difficulty was ever increas- standing of common linguistic phrases and
ing along with the duration of the disease. no dysfunctions in this respect were observed.
Patients with left-side brain damage spoke Difficulties in understanding verbal information
about emotional experiences more frequently than with emotional content appeared to be specific to
476

the schizophrenic patients. With respect to the of linguistic phrases describing emotional states).
ability to talk about emotional experiences and Cognitive deficits in schizophrenic patients are
verbal wit understanding, it may be that there was probably related to more general disturbances in
a deficit of pre-illness mental life in those patients brain functioning than it happens to be in the case
or a deficit related to becoming ill itself, and then it of patients with right-side brain injury. Bozikas
remained on a relatively constant level. et al. (2004) in their study showed significant
Our findings obtained from the Observational relationships of affect perception, both facial and
Scale of Emotional Behavior regarding emotional vocal as well as in everyday scenarios with several
communication (emotional expression and under- cognitive abilities. Their findings might support
standing of emotional context) are in agreement the notion that deficits in processing emotional
with data from previous studies. Research carried stimuli in schizophrenia could be attributed to im-
out using a similar evaluation showed that patients pairment in more basic neurocognitive domains.
with right-sided brain damage have difficulties in Undoubtedly, carefully well-thought-out exper-
the perception and assessment of emotional con- iments complementary to observations carried
tent and fail to express emotions (Kaplan et al., out in a natural context are necessary in order to
1990; Kohler et al., 2000). In our studies impaired explain these scientific controversies.
mimic expression and weakened gesture were com-
monly observed (interestingly, weakened gesture Conclusions
was recorded also in individuals with left-sided
brain damage; nevertheless, this weakness was less The patients with right-side brain damage rev-
marked). In our study ‘‘right-sided’’ patients did eal elated mood, inadequate self-assessment, au-
not differ from the ‘‘left-sided’’ with regards to thorization of behaviors, and the active and often
understanding verbal wit and common linguistic negativistic attitude to the environment and depre-
phrases related to emotions. Recent experimental ciation of the degree of their disorders. The patients
studies confirm decreased expression of those with left-side brain damage showed lowered mood,
suffering from schizophrenia (Stuss et al., 2000). feelings of resignation, and active or rarely passive
No emotional dysfunction was found in patients attitude to the environment and proper self-assess-
with left-sided brain damage. Comparison of the ment. Mood indifference, mostly passive attitude to
emotional behaviors of individuals with schizo- the environment and depreciation of the degree of
phrenia to those with right-side brain damage al- their illness in the patients with schizophrenia, was
lows us to formulate a hypothesis that in something that was drawing attention. Deficits re-
schizophrenia there is an anatomical irregularity vealed in schizophrenic patients as to the emotional
or malfunction localized in the right brain hemi- communication (both emotional expression and
sphere. Rotenberg (1994) suggested that schizo- understanding of emotional situations) were more
phrenia results from functional insufficiency of the marked than in patients with unilateral brain dam-
RH in combination with subtle brain damage. In age. With regards to emotional behaviors, those
this model RH hypoactivity leads to compensatory suffering from schizophrenia were generally more
hyperactivity on the left that results in enhance- similar to those with right-side brain damage
ment of dopaminergic system. than to those with left-side brain damage. This
Individuals with schizophrenia differ from those fact can be explained by similar localization of
with right-side brain damage with regards to un- brain dysfunctions in both groups. The specificity
derstanding wit and verbal humor as well as con- of schizophrenic patients consisted in difficulties of
veying feelings through common linguistic understanding jokes and language expressions
phrases, in the area they disclosed a peculiar help- with emotional content, which should be under-
lessness. This specificity of schizophrenic patients stood not only in connection to the disturbances in
could be explained by their characteristic cognitive the emotional sphere but also to cognitive dysfunc-
deficits, particularly their tendency to have a tions and to more general brain dysfunctions
given, particular attitude (metaphoric character (Kucharska-Pietura and Hunca-Bednarska, 2002).
477

References Buck, R. (1985) Prime theory: an integrated view of motivation


and emotion. Psychol. Rev., 92: 389–413.
Adolphs, R., Jansari, A. and Tranel, D. (2001) Hemispheric Christman, S.D. and Hackworth, M.D. (1993) Equivalent per-
perception of emotional valence from facial expressions. Ne- ceptual asymmetries for free viewing of positive and negative
uropsychology, 15: 516–524. emotional expressions in chimeric faces. Neuropsychologia,
Altenmüller, E., Schürmann, K., Lim, V.K. and Parlitz, D. 31: 621–624.
(2002) Hits to the left, flops to the right: different emotions Cicone, M., Wapner, W. and Gardner, H. (1980) Sensitivity to
during listening to music are reflected in cortical lateralizat- emotional expression and situations in organic patients. Cor-
ion patterns. Neuropsychologia, 40: 2242–2256. tex, 16: 145–158.
Archer, J. and Hay, D.C. (1992) Face processing in psychiatric Cramer, P., Weegmann, M. and O’Neil, M. (1989) Schizophre-
conditions. Br. J. Clin. Psychol., 31: 45–61. nia and the perception of emotions: how accurately do schiz-
Bell, M., Bryson, G. and Lysaker, P. (1997) Positive and negative ophrenics judge the emotional states of others? Br. J.
affect recognition in schizophrenia: a comparison with substance Psychiatry, 155: 225–228.
abuse and normal control subjects. Psychiatr. Res., 73: 73–82. David, A.S. (1989) Perceptual asymmetry for happy–sad
Bellack, A.S., Blanchard, J.J. and Mueser, K.T. (1996) chimeric faces: effects of mood. Neuropsychologia, 27:
Cue availability and affect perception in schizophrenia. 1289–1300.
Schizophr. Bull., 22: 535–544. David, A.S. (1993) Spatial and selective attention in the cerebral
Bleuler, E. (1906) Affektivität, Suggestibilität, Paranoia. hemispheres in depression, mania and schizophrenia. Brain
Verlag, Halle. Cogn., 23: 166–180.
Blood, A.J., Zatorre, R.J., Bermudez, P. and Evans, A.C. David, A.S. and Cutting, J.C. (1990) Affect, affective disorder
(1999) Emotional responses to pleasant and unpleasant music and schizophrenia. A neuropsychological investigation of
correlate with activity in paralimbic brain regions. Nat. Ne- right hemisphere function. Br. J. Psychiatry, 156: 491–495.
urosci., 2: 382–387. David, A.S. and Cutting, J.C. (1992) Categorical-semantic and
Borod, J.C., Kent, J., Koff, E., Martin, C. and Alpert, M. spatial imaging judgements of non-verbal stimuli in the cere-
(1988) Facial asymmetry while posing positive and negative bral hemispheres. Cortex, 28: 39–51.
emotions: support for the right hemisphere hypothesis. Ne- Davidson, R.J. (1984) Affect, cognition, and hemispheric spe-
uropsychologia, 26: 759–764. cialization. In: Izard, C.E., Kagan, J. and Zajonc, R.B.
Borod, J.C. and Koff, E. (1989) The neuropsychology of emo- (Eds.), Emotion, Cognition, and Behavior. Cambridge Uni-
tion: Evidence from normal, neurological, and psychiatric versity Press, Cambridge, pp. 320–365.
populations. In: Perecman, T. (Ed.), Integrating Theory and Dawson, G. and Fisher, K.W. (1994) Human Behavior and
Practice in Clinical Neuropsychology. Hillsdale, New York, Developing Brain. Guilford, New York.
pp. 175–215. Dougherty, F.E., Bartlett, E.S. and Izard, C.E. (1974) Response
Borod, J., Koff, E. and Caron, H.S. (1983) Right hemispheric of schizophrenics to expressions of the fundamental emo-
specialization for the expression and appreciation of emo- tions. J. Clin. Psychol., 7: 243–246.
tion: a focus on the face. In: Perecman, T. (Ed.), Cognitive Earnst, K.S., Kring, A.M., Kadar, M.A., Salem, J.E., David,
Processing in the Right Hemisphere. Academic Press, New A.S. and Loosen, P.T. (1996) Facial expression in schizo-
York, pp. 83–110. phrenia. Biol. Psychiatry, 40: 556–558.
Borod, J.C., Martin, C.C., Alpert, M., Brozgold, A. and Edwards, J., Pattison, P.E., Jackson, H.J. and Wales, R.J.
Welkowitz, J. (1993) Perception of facial emotion in schiz- (2001) Facial affect and affective prosody recognition in
ophrenic and right brain-damaged patients. J. Nerv. Ment. first-episode schizophrenia. Schizophr. Res., 48: 235–253.
Dis., 181: 494–502. Etcoff, N.L. (1984) Selective attention to facial identity and
Bowers, D., Bauer, R.M., Coslett, H.B. and Heilman, K.M. facial emotion. Neuropsychologia, 22: 281–295.
(1985) Processing of faces by patients with unilateral hem- Feinberg, T.E., Rifkin, A., Schaffer, C. and Walker, E. (1986)
isphere lesions. Dissociation between judgements of facial Facial discrimination and emotional recognition in schizo-
affect and facial identity. Brain Cogn., 4: 258–272. phrenia and affective disorders. Arch. Gen. Psychiatry, 43:
Bozikas, V.P., Kosmidis, M.H., Anezoulaki, D., Giannakou, 276–279.
M. and Karavatos, A. (2004) Relationship of affect recog- Folstein, M.F., Folstein, S.E. and McHugh, P.R. (1975) Mini-
nition with psychopathology and cognitive performance in mental state examination: a practical method for grading the
schizophrenia. J. Int. Neuropsychol. Soc., 10: 549–558. cognitive state of patients. Psychol. Res., 12: 189–198.
Bruce, V. and Young, A. (1986) Understanding face recogni- Gaebel, W. and Wölwer, W. (1992) Facial expression and emo-
tion. Br. J. Psychol., 77: 305–327. tional face recognition in schizophrenia and depression. Eur.
Bryden, M.P., Lay, R.G. and Sugarman, J.H. (1982) A left-ear Arch. Psy. Clin. Neurosci., 242: 46–52.
advantage for identifying the emotional quality of toned se- Gainotti, G. (1972) Emotional behavior and hemispheric side of
quences. Neuropsychologia, 20: 83–87. the lesion. Cortex, 8: 41–55.
Bryson, G., Bell, M., Kaplan, E., Greig, T. and Lysaker, P. Gainotti, G. (1984) Some methodological problems in the
(1998) Affect recognition in deficit syndrome schizophrenia. study of the relationships between emotions and cerebral
Psychiatr. Res., 77: 113–120. dominance. J. Clin. Neuropsychol., 6: 11–121.
478

Garver, D.L., Nair, T.R., Christensen, J.D., Holcomb, J., Kerr, S.L. and Neale, J.M. (1993) Emotion perception in
Ramberg, J. and Kingsbury, S. (1999) Atrophic and static schizophrenia: specific deficit or further evidence of gene-
(neurodevelopmental) schizophrenic psychoses: premorbid ralized poor performance? J. Abnorm. Psychol., 102:
functioning, symptoms, and neuroleptic response. Neuropsy- 312–318.
chology, 21: 82–92. Kohler, Ch.G., Gur, R.C. and Gur, R.E. (2000) Emotional
Gooding, D.C., Luh, K.E. and Tallent, K.A. (2001) Evidence of processing in schizophrenia: a focus on affective states. In:
schizophrenia patients’ reduced perceptual biases in response Borod, J.C. (Ed.), The Neuropsychology of Emotion. Oxford
to emotion chimera. Schizophr. Bull., 27: 709–716. University Press, New York.
Gorynia, I. and Dudeck, U. (1996) Patterns of lateral prefer- Kolb, B., Milner, B. and Taylor, L. (1983) Perception of faces
ence in psychotic patients. Neuropsychologia, 34: 105–111. by patients with localized cortical excisions. Can. J. Psychol.,
Grega, D.M., Sackeim, H.A., Sanchez, E., Cohen, B.H. and 37: 8–18.
Hough, S. (1988) Perceiver bias in the processing of human Kring, A.M. and Neale, J.M. (1996) Do schizophrenic patients
faces: neuropsychological mechanisms. Cortex, 24: 91–117. show a disjunctive relationship among expressive, experien-
Harris, L.J. and Snyder, P.J. (1992) Subjective mood state and tial, and psychophysiological components of emotion?
perception of emotion in chimeric faces. Cortex, 28: 471–481. J. Abnorm. Psychol., 105: 249–257.
Heilman, K.M. (1982) Discussant comments. In: Borod, J.C. Kucharska-Pietura, K., David, A.S., Dropko, P. and Klimkow-
and Buck, R. (Eds.), Asymmetries in Facial Expression: ski, M. (2002) The perception of emotional chimeric faces
Method and Meaning. Symposium conducted at the Inter- in schizophrenia: further evidence of right hemisphere dys-
national Neuropsychological Society, Pittsburgh. function. Neuropsychiatry Neuropsychol. Behav. Neurol.,
Heilman, K.M. (1997) The neurobiology of emotional experi- 15: 72–78.
ence. J. Neuropsychiat. Clin. Neurosci., 9: 439–448. Kucharska-Pietura, K. and Hunca-Bednarska, A. (2002) Emo-
Heilman, K.M., Blonder, L.X., Bowers, D. and Crucian, S.P. tional behavior in schizophrenia and one-sided brain dam-
(2000) Neurological disorders and emotional dysfunction. In: age. Cerebral hemispheric asymmetry. Part I Psychiatr. Pol.,
Borod, J.C. (Ed.), The Neuropsychology of Emotion. Oxford 36: 421–434.
University Press, New York, pp. 367–412. Kucharska-Pietura, K. and Klimkowski, M. (2002a) Perception
Heimberg, C., Gur, R.E., Erwin, R.J., Shtasel, D.L. and Gur, of facial affect in chronic schizophrenia and right brain dam-
R.C. (1992) Facial emotion discrimination: III. Behavioral age. Acta Neurobiol. Exp., 62: 33–43.
findings in schizophrenia. Psychiatr. Res., 42: 253–265. Kucharska-Pietura, K. and Klimkowski, M. (2002b) Clinical
Herbener, E.S., Hill, S.K., Marvin, R.W. and Sweeney, J.A. Aspects of Emotions in Healthy and Disordered Brain.
(2005) Effects of antipsychotic treatment on emotion per- Wydawnictwo Medyczne, Krakow, pp. 67–174.
ception deficits in first-episode schizophrenia. Am. J. Psychi- Kucharska-Pietura, K., Phillips, M., Gernand, W. and David,
atry, 62: 1746–1748. A.S. (2003) Perception of emotions from faces and voices
Herzyk, A. and Rozenkiewicz, J. (1994) Neuropsychologiczna following unilateral brain damage. Neuropsychologia, 41:
diagnoza zaburzeń emocjonalnych. In: Klimkowski, M. and 1082–1090.
Herzyk, A. (Eds.), Neuropsychologia kliniczna. Wybrane Lane, R.D., Chua, P.M. and Dolan, R.J. (1999) Common
zagadnienia. UMCS, Lublin, pp. 31–73. effects of emotional valence, arousal and attention on neural
Hoptman, M.J. and Levy, J. (1988) Perceptual asymmetries in activation during visual processing of pictures. Neuro-
left- and right-handers for cartoon and real faces. Brain psychologia, 37: 989–997.
Cogn., 8: 178–188. Levine, S.C. and Levy, J. (1986) Perceptual asymmetry for
Jackson, J.H. (1879) On affections of speech from diseases of chimeric faces across the life span. Brain Cogn., 5: 291–306.
the brain. Brain, 2: 203–222. Levy, J., Heller, W. and Banich, M.T. (1983) Asymmetry of
Jaeger, J., Borod, J.C. and Peselow, E. (1987) Depressed pa- perception in free viewing of chimeric faces. Brain Cogn., 2:
tients have atypical biases in the perception of emotional 404–419.
chimeric faces. J. Abnorm. Psychol., 96: 321–324. Lewis, S. and Garver, D. (1995) Treatment and diagnostic sub-
Jessimer, M. and Markham, R. (1997) Alexithymia: A right type in facial affect recognition in schizophrenia. J. Psychiatr.
hemisphere dysfunction specific to recognition of certain fa- Res., 29: 5–11.
cial expressions? Brain Cogn., 34: 246–258. Lior, R. and Nachson, I. (1999) Impairments in judgement of
Kaplan, J.A., Brownell, H.H., Jacobs, J.R. and Gardner, H. chimeric faces by schizophrenic and affective patients. Int. J.
(1990) The effects of right hemisphere damage on the prag- Neurosci., 97: 185–209.
matic interpretation of conversational remarks. Brain Lang, Luh, K.E. and Gooding, D.C. (1999) Perceptual biases in psy-
38: 315–333. chosis-prone individuals. J. Abnorm. Psychol., 108: 283–289.
Kay, S.R., Opler, L.A. and Fiszbein, A. (1987) Positive and Luh, K.E., Redl, J. and Levy, J. (1994) Left- and right-handers
Negative Syndrome Scale (PANSS) Rating Manual. Social see people differently: free-vision perceptual asymmetries for
and Behavioural Sciences Documents. San Rafael, Canada. chimeric stimuli. Brain Cogn., 25: 141–160.
Kee, K.S., Kern, R.S. and Green, M.F. (1998) Perception of Luh, K.E., Rueckert, L.M. and Levy, J. (1991) Perceptual as-
emotion and neurocognitive functioning in schizophrenia: ymmetries for free viewing of several types of chimeric stim-
what’s the link? Psychiatr. Res., 81: 57–65. uli. Brain Cogn., 16: 83–103.
479

Mandal, M.K. and Palchoudhury, S. (1985) Responses to facial organization in schizophrenia. Arch. Gen. Psychiatry, 39:
expression of emotion in depression. Psychol. Rep., 56: 1006–1010.
633–654. Reuter-Lorenz, P.A. and Davidson, R. (1981) Differential
Mandal, M.K. and Rai, A. (1987) Responses to facial emotion contributions of the two cerebral hemispheres to the
and psychopathology. Psychiatry Res., 20: 317–323. perception of happy and sad faces. Neuropsychologia, 19:
Mandal, M.K., Pandey, R. and Prasad, A.B. (1998) Facial ex- 609–613.
pressions of emotions and schizophrenia: a review. Schizo- Robertson, G. and Taylor, P.J. (1987) Laterality and psychosis:
phr. Bull., 24: 399–412. neuropsychological evidence. Br. Med. Bull., 43: 634–650.
Martin, F., Baudouin, J.Y., Tiberghien, G. and Franck, N. Rotenberg, V.S. (1994) An integrative psychophysiological ap-
(2005) Processing emotional expression and facial identity in proach to brain hemisphere functions in schizophrenia. Ne-
schizophrenia. Psychiatry Res., 134: 43–53. urosci. Behav. Rev., 18: 487–495.
Mason, O. and Claridge, G. (1999) Individual differences in Sackeim, H., Greenberg, M., Weiman, A., Gur, R., Hunger-
schizotypy and reduced asymmetry using the chimeric faces buhler, J. and Geschwind, N. (1982) Functional brain asym-
task. Cogn. Neuropsychiatry, 4: 289–301. metry in the expression of positive and negative emotions:
Mazurkiewicz, J. (1958) Wst˛ep do psychofizjologii normalnej. lateralization of insult in cases of uncontrollable emotional
Tom II, PZWL, Warszawa, pp. 10–78. outbursts. Arch. Neurol., 19: 210–218.
McDowell, C.L., Harrison, D.W. and Demaree, H.A. (1994) Is Salem, J.E., Kring, A.M. and Kerr, S.L. (1996) More evidence
right hemisphere decline in the perception of emotion a func- for generalized poor performance in facial emotion percep-
tion of aging? Int. J. Neurosci., 79: 1–11. tion in schizophrenia. J. Abnorm. Psychol., 105: 480–483.
Muzekari, L.H. and Bates, M.E. (1977) Judgment of emotion Schmidt, L.A. and Trainor, L.J. (2001) Frontal brain electrical
among chronic schizophrenics. J. Clin. Psychol., 33: 662–666. activity (EEG) distinguishes valence and intensity of musical
Noesselt, T., Driver, J., Heinze, H.J. and Dolan, R. (2005) emotions. Cogn. Emotion, 5: 482–500.
Asymmetrical activation in the human brain during process- Schneider, F., Ellgring, H., Friedrich, J., Fus, I., Beyer, T.,
ing of fearful faces. Curr. Biol., 15: 424–429. Heimann, H. and Himer, W. (1992) The effects of neurolep-
Northoff, G., Heinzel, A., Bermpohl, F., Niese, R., Pfennig, A., tics on facial action in schizophrenic patients. Pharmacop
Pascual-Leone, A. and Schlaug, G. (2004) Reciprocal mod- sychiatry, 25: 233–239.
ulation and attenuation in the prefrontal cortex: an fMRI Scholten, M.R., Aleman, A., Montagne, B. and Kahn, R.S.
study on emotional–cognitive interaction. Hum. Brain (2005) Schizophrenia and processing of facial emotions: sex
Mapp., 21: 202–212. matters. Schizophr. Res., 78: 61–67.
Novic, J., Luchins, D.L. and Perline, R. (1984) Facial affect Silver, H., Shlomo, N., Turner, T. and Gur, R.C. (2002) Per-
recognition in schizophrenia: is there adifferential deficit? Br. ception of happy and sad facial expressions in chronic schiz-
J. Psychiatry, 144: 533–537. ophrenia: evidence for two evaluative systems. Schizophr.
Pally, R. (1998) Bilaterality: Hemispheric specialization and Res., 55: 171–177.
integration. Int. J. Psycho-Anal., 79: 565–577. Simon-Thomas, E.R., Role, K.O. and Knight, R.T. (2005) Be-
Phillips, M.L. and David, A.S. (1997) Viewing strategies for havioral and electrophysiological evidence of a right hemi-
simple and chimeric faces: an investigation of perceptual bias sphere bias for the influence of negative emotion on higher
in normals and schizophrenic patients using visual scan cognition. J. Cogn. Neurosci., 17: 518–529.
paths. Brain Cogn., 35: 225–238. Stuss, D.T., van Reekum, R. and Murphy, K.J. (2000) Differ-
Phillips, M.L., Young, A.W., Scott, S.K., Calder, A.J., Andrew, entiation of states and causes of apathy. In: Borod, J.C.
C. and Giampietro, V. (1998) Neural responses to facial and (Ed.), The Neuropsychology of Emotion. Oxford University
vocal expressions of fear and disgust. Proc. R. Soc. Lond. B Press, New York.
Biol. Sci., 256: 1809–1817. Wölwer, W., Streit, M., Polzer, U. and Gaebel, W. (1996) Fa-
Piran, N., Bigler, E.D. and Cohen, D. (1982) Motoric latera- cial affect recognition in the ourse of schizophrenia. Eur.
lity and eye dominance suggest unique pattern of cerebral Arch. Psychiatry Clin. Neurosci., 246: 165–170.
Anders, Ende, Junghöfer, Kissler & Wildgruber (Eds.)
Progress in Brain Research, Vol. 156
ISSN 0079-6123
Copyright r 2006 Elsevier B.V. All rights reserved

CHAPTER 27

The biochemistry of dysfunctional emotions: proton


MR spectroscopic findings in major depressive
disorder

Gabriele Ende1,, Traute Demirakca1,2 and Heike Tost1

1
NMR Research in Psychiatry, Central Institute of Mental Health, J5, 68159 Mannheim, Germany
2
Heidelberg Academy of Science, Heidelberg, Germany

Abstract: Key neural systems involved in the processing and communication of emotions are impaired in
patients with major depressive disorder (MDD). Emotional and behavioral symptoms are thought to be
caused by damage or dysfunction in specific areas of the brain that are responsible for directing attention,
motivating behavior, and learning the significance of environmental stimuli. Functional brain studies with
positron emission tomography (PET) and functional magnetic resonance imaging (fMRI) give support for
functional abnormalities in MDD that are predominantly located in areas known to play an important role
in the communication and processing of emotions. Disturbances in emotional processing as they are
observed in MDD, if any, have very subtle morphometrical brain correlates. With proton magnetic res-
onance spectroscopy (1H MRS), brain metabolites can be measured noninvasively in vivo, thus furthering
the understanding of the effects of changes in neurotransmitters within the brain. The current literature on
1
H MRS studies in MDD is small with a large diversity of MRS methods applied, brain regions studied,
and metabolite changes found. Nevertheless, there is strong evidence that changes in neurometabolite
concentrations in MDD occur within brain regions, which are involved in the processing and commu-
nication of emotions that can be monitored by 1H MRS. This review summarizes the literature about
biochemical changes quantified via 1H MRS in MDD patients in brain regions that play an important role
for the communication and processing of emotions.

Keywords: proton magnetic resonance spectroscopy (1H MRS); major depressive disorder (MDD);
emotions

Introduction the neurobiology of depression have been made,


such as significant breakthroughs in genomics,
Today, major depressive disorder (MDD) is imaging, and the identification of key neural sys-
viewed as a malfunction of particular circuits that tems involved in cognition, emotion, and behavior.
connect the limbic system with the prefrontal cor- Many anatomical studies involving imaging of the
tex, the brain stem, and hypothalamus, which amygdala, the hippocampus, and the prefrontal
control basic functions such as sleep, appetite, and cortex give evidence for subtle morphometrical
libido. Major advances in our knowledge about changes in MDD. Several brain neurotrans-
mitter systems, such as glutamate, g-amino-
Corresponding author. Tel.: +49-621-1703-2971; Fax: +49- butyric acid (GABA), serotonin, norepinephrine,
621-1703-3005; E-mail: ende@zi-mannheim.de and dopamine have been implicated in depression

DOI: 10.1016/S0079-6123(06)56027-3 481


482

and mania. These transmitter systems — as well there is a mood congruent processing bias in
as other neurochemical systems such as mem- MDD patients: positive or ambiguous events tend
brane-bound signal transduction systems and to be received as negative (Beck et al., 1979; Teas-
intracellular signaling systems that modulate gene dale and Russell, 1983; Segal et al., 1996). In
transcription and protein synthesis — play an memory tests, MDD patients perform better with
important role in the etiology of depression. negative events or stimuli (Lloyd and Lishman,
Several lines of evidence suggest that central 1975; Clark and Teasdale, 1982). Moreover, MDD
cortical inhibitory mechanisms, especially those patients do not seem to perceive neutral faces as
associated with GABA neurotransmission, may unambiguous signals of being emotionally neutral.
play a role in the pathophysiology of major This is suggested by the fact that they recognized
depression. neutral faces less accurately than emotional faces
and needed more time to process them (Leppanen
et al., 2004). Other studies suggested that clinical
Emotional and behavioral biases and their neuronal
depression may affect the processing of emotional
correlates
information at the affect discrimination level. Pa-
tients are more likely to misinterpret the emotional
An important aspect of affective disorders is
valence of facial expression (Gur et al., 1992; Per-
the disturbance of emotional processing itself. An
sad and Polivy, 1993). Moreover, the ability to
element of the symptoms connected with depres-
discriminate neutral–sad faces is impaired, they
sion is the prolonged involuntary processing of
misinterpret neutral faces as sad and happy faces
emotional information in form of elaboration
as neutral (Surguladze et al., 2005).
(Mathews et al., 1996) or rumination (Nolen-Hoe-
In situations where verbal stimulus material is
ksema, 2000) on negative topics, and higher sen-
used, medicated and unmedicated depressed
sitivity to negative events (Nandrino et al., 2004).
patients show mood congruent biases in affective
These processing deficits form the focus of cogni-
go/no-go tasks (Murphy et al., 1999; Erickson et
tive theories of depression (Willner, 1984; Haaga
al., 2005). While processing happy words, depressed
and Beck, 1995; Beck, 2005). They certainly have
patients are characterized by a prolonged reaction
an influence on the behavior of the patients and
time and an increasing error rate, whereas healthy
probably on the course of the depressive episodes.
controls make more errors for sad words. Similar
The facilitation of processing negative information
results were found within a lexical decision task
may reinforce depressed mood and contribute to
(Weisbrod et al., 1999) where patients reacted faster
the maintenance of the disorder (Bradley et al.,
to negative words than positive or neutral words.
1996; Watkins et al., 1996; Bradley et al., 1997;
Watkins et al., 2000).
This volume focuses on the communication of
Functional neuroimaging
emotions. The results of several studies associated
depressed mood with specific abnormalities in the
The results of resting state studies, mainly positron
communication of emotions, e.g., the identifica-
emission tomography (PET) and single photon
tion of facial expression; negative cognition
emission computed tomography (SPECT) can be
regarding the self; and dysfunctional appraisal of
divided into two findings, i.e., either hypo- or
social events and situations. These abnormalities
hyperactivation was found. Hypoarousal — di-
may lead to impaired interpersonal functioning
minished metabolism or reduced regional cerebral
(Surguladze et al., 2004).
blood flow — was observed in frontal brain re-
gions, e.g., the anterior cingulate gyrus (ACG), the
Behavior dorsolateral prefrontal cortex (DLPFC), the or-
bitofrontal cortex (OFC), and the medial prefron-
Several studies dealing with the influence of tal cortex (Bench et al., 1992; Dolan et al., 1994;
emotions on cognitive processing consider that Bremner et al., 1997; Vollmert et al., 2004).
483

Hyperarousal has been found in brain regions that Increased activation in patients has been seen espe-
are part of the limbic system with the am- cially in the insula (Fu et al., 2004; Gotlib et al.,
ygdala–hippocampal formation and the basal 2005), the hippocampal and parahippocampal
ganglia (Ho et al., 1996; Mayberg et al., 1999). gyrus (Fu et al., 2004), the ACG (Fu et al., 2004;
The resting state activity of the amygdala–hippo- Gotlib et al., 2005), and the inferior frontal and
campal formation was not only found to be higher inferior parietal gyrus (Gotlib et al., 2005; Surgul-
in depressed patients (Ho et al., 1996); moreover, a adze et al., 2005).
correlation with the amount of negative affect has Activation of the amygdala was also found in
been found (Abercrombie et al., 1998), and a some studies of word/language processing of
reduction following successful treatment (Mayberg different emotional valence (Hamann and Mao,
et al., 1999, 2000). 2002; Maddock et al., 2003; Herbert et al., 2005) in
Functional imaging studies, investigating self- healthy controls. Furthermore, activation differ-
induced transient sadness in healthy controls ences for emotional compared to neutral words
(Mayberg et al., 1999; Liotti et al., 2000), detected were found in putamen (Hamann and Mao, 2002),
activation of parts of the limbic system (subgenual medial prefrontal cortex, insula, superior temporal
ACC, dorsal insula) and a simultaneous reduced and inferior parietal cortex (Fossati et al., 2004),
activation of cortical regions (parietal cortex, anterior and posterior cingulate gyrus, inferior
DLPFC). The diminished activation of these frontal, and orbitofrontal regions (Hamann and
cortical regions, responsible for the regulation of Mao, 2002; Maddock et al., 2003; Herbert et al.,
sustained attention and vigilance, provides an 2005). Depressed patients show an increased acti-
explanation for the cognitive deficits of depressed vation, especially to sad words in the lateral OFC,
patients. In summary, both transient sadness and the anterior temporal cortex (Elliott et al., 2002),
depressive episodes lead to activation and deacti- and the left inferior parietal gyrus (Canli et al.,
vation of very similar parts of the limbic system 2004). Increased activation of the amygdala was
and the attentional system. not directly found, but the activation of the am-
The processing of emotional stimuli, particularly ygdala was prolonged up to 25 s after stimulus
communication relevant stimuli as faces and words, presentation (Siegle et al., 2002), and this activa-
has been well studied in healthy controls. The dam- tion correlated with self-reported rumination of
age of the amygdala leads to the inability to identify the depressed patients. Prolonged activation after
fear in facial expression (Adolphs et al., 1994). negative words was also observed in the frontal,
Amygdala activation has been found during the the inferior parietal cortex, and the posterior
processing of happy, sad, fearful, and angry faces in cingulate gyrus.
healthy controls (Breiter et al., 1996; Morris et al., It can be summarized that both the limbic and
1996; Yang et al., 2002). Processing of the emo- frontal brain regions were found to activate ab-
tional expression of faces by depressed patients was normally in MDD patients when emotional stimuli
investigated with implicit tasks where the subjects had to be processed. A summary of neuroimaging
do not directly process the emotional valence but findings and key brain regions of emotional
the gender of the faces. Studies with MDD patients processing biases in MDD is given in Table 1.
focused on the difference between sad, happy, and
neutral faces. Emotions like anger, anxiety, or dis-
gust are poorly investigated. Hyperactivation of the Brain structural alterations in MDD
amygdala was found in several studies (Sheline et
al., 2001; Fu et al., 2004; Surguladze et al., 2005). Early empirical studies found a correlation be-
This hyperactivation is reduced after treatment tween specific brain injury and the prevalence of
(Sheline et al., 2001; Fu et al., 2004). Other regions MDD. In 1985 Robinson was first to conclude that
that are known to show hyperactivity in depressed physical impairment is not the major determinant
patients during resting state or mood induction did of emotional response to brain injury, but damage
also respond abnormally to emotional faces. within specific brain regions may be playing a role
484

Table 1. Neuroimaging findings of emotional processing biases in major depressive disorder

Authors Sample Method Main results and key regions

Resting state/no stimulation


Mayberg et al. (1999) 8P PET After remission: mright DPFC, IP, dorsal
ACG, PC, pons, left PFC k subgenual ACG,
ventral, middle, posterior insula, HC,
hypothalamus
Abercrombie et al. (1998) 27 P, 24 C PET Metabolism AMY correlation with negative
affect
Ho et al. (1996) 10 P, 12 C PET P m PC, AMY, HC, OC, TC, P k OFC, ACG,
BG
Videbech et al. (2002) 42 P, 47 C PET P m HC, ACG, cerebellum, BG, rCBF in HC
correlation with HAMD

Mood induction
Pardo et al. (1993) PET Sadness: m infFG, OFC
George et al. (1995) 11 P, 11 C PET Sadness: m right ACG bilateral infFG
Mayberg et al. (1999) 8C PET Transient sadness: m subgenual ACG, insula,
CV, right premotC, k right DPFC, IP, dorsal
ACG, PC, right TC, bilateral infFG
Liotti et al. (2002) 17 P ( 10 RP), 8 C PET Transient sadness: P&C k PAR, PC, posterior
infTG, m insula, CV, motC/premotC
P k medial OFC, anterior TH; RP k
pregenual ACG, C k, right PFC m subgenual
ACG, RP more similar to acute P than to C

Stimulation with faces


Thomas et al. (2001) 5 P, 5 C fMRI Fearful faces: P children k AMY
Sheline et al. (2001) 11 P, 11 C fMRI, Masked emotional faces: pretreatment P m
ROI left AMY (sad faces mm) posttreatment: P Ø,

Fu et al. (2004) 19 P, 19 C fMRI Gender decision task, sad faces: pre treatment:
P m left AMY&HC, PHG, HTH, VSTR
(PUT, GP), insula, CAU, TH, dorsal CG, IP
Posttreatment: P k left AMY, VSTR (PUT,
GP), CAU, TH, dorsal CG, ACG, IP, right
VSTR, TH, IP
Gotlib et al. (2005) 18 P, 18 C fMRI Gender decision task, happy faces: P m left
subgenual ACG, left midFG, right supFG. C
m right infTG, left insula
Sad faces: P m l infFG & subgenual ACG, C m
right midTG, right infFG. AMY Ø
Surguladze et al. (2005) 16 P, 14 C fMRI Gender decision task; increasing happiness: C
m & P k bilateral fusiform G, right PUT
Increasing sadness: P m C k right fusiform G,
left PHG&AMY, left PUT. P: negative
correlation BDI & BOLD right fusiform G
Canli et al. (2005) 17 P Emotional faces: AMY correlation with
symptom improvement (high AMY BOLD —
low BDI at T2)

Verbal communication/verbal stimuli


Nandrino et al. (2004) 26 P, 26 C EEG First episode P: positive stimuli k P300;
recurrent P: negative stimuli m P300
Shestyuk et al. (2005) 15 P, 16 C EEG P: positive stimuli k brain activity
485

Table 1 (continued )

Authors Sample Method Main results and key regions

Elliott et al. (2002) 10 P, 11 C fMRI Emotional go/nogo task: relevant emotional


targets: P m right lateral PFC, C m infFG,
right ventral CG, right Pulvinar, midTG,
preCG, postCG. Happy targets: C m rostral
right ACG, right medPFC, right anterior TC,
midTG, bilateral medFG. irrelevant emotional
distractors: P m bilateral lateral OFC, anterior
TC , sad distractors: P m lateral right OFC,
bilateral anterior TC
Canli et al. (2004) 15 P, 15 C fMRI Lexical decision task: sad words P m left IP, C
m supTG, cerebellum, happy words; P Ø, C m
AMY, infFG, supTG
Siegle et al. (2002) 7 P, 11 C fMRI Valence identification task, negative words P
Ø bilateral AMY (25s), midFG, supFG, IP,
PCG
P negative correlation AMY & DLPFC, P
AMY correlated with self-reported
rumination
George et al. (1997) 11 P (7 MDD), 11 C PET MD P (MDD+BP) sad vs. standard stroop:
mvisual cortex, cerebellum, H Ø, sad stroop:
H vs. P Ø

Note: - ¼ sustained (prolonged) BOLD response; m ¼ significant increased BOLD response; k ¼ significant decreased BOLD response; Ø ¼ no
significant differences; ACG ¼ anterior cingulate gyrus; AMY ¼ amygdala; BG ¼ basal ganglia; C ¼ healthy controls; CAU ¼ caudate; CG ¼ cingulate
gyrus; CV ¼ cerebellar vermis; DLPFC ¼ dorsolateral prefrontal cortex; DPFC ¼ dorsal prefrontal cortex; FC ¼ frontal cortex; GP ¼ globus pallidus;
GR ¼ gyrus rectus (medial OFC); HAMD ¼ Hamilton scale of depression; HC ¼ hippocampus; HTH ¼ hypothalamus; infFG ¼ inferior frontal gyrus;
IP ¼ inferior parietal cortex; MD ¼ mood disorder; MDD ¼ major depressive disorder; MedPFC ¼ medial prefrontal cortex; MFG ¼ middle frontal
gyrus; midTG ¼ middle temporal gyrus; motC ¼ motor cortex; OC ¼ occipital cortex; OFC ¼ orbitofrontal cortex; P ¼ major depressive disorder
patients; PAR ¼ parietal cortex; PC ¼ posterior cingulate; PFC ¼ prefrontal cortex; PHG ¼ parahippocampal gyrus; preCG ¼ precentral gyrus; pre-
motC ¼ premotor cortex; PUT ¼ putamen; rCBF ¼ regional cerebal blood flow; ROI ¼ region of interest analysis; RP ¼ remitted patients; sup-
FG ¼ superior frontal gyrus; supTG ¼ superior temporal gyrus; TC ¼ temporal cortex; TH ¼ thalamus; VSTR ¼ ventral striatum.

in the type of emotional response (Robinson and Although observed frequently in the aging brain
Lipsey, 1985). Another fact that encouraged the (de Leeuw et al., 2001), the number and size of
search for persistent morphometric impairments is WML are also associated with the development of
that a substantial number of MDD patients suffer neuropsychiatric impairments, especially depres-
from sustained neuropsychological deficits sion (Steffens et al., 1999). While the precise
(O’Brien et al., 2004). Current empirical evidence pathogenesis of WML remains to be elucidated,
points toward traceable alterations of brain struc- vascular risk factors like hypertensia, diabetes, and
ture on several levels, ranging from gross macro- hypercholesterinemia clearly promote the devel-
scopic lesions to subtle volumetric alterations, opment of gliosis and demyelination due to
cellular pathology, and biochemical correlates of chronic focal ischemia (Taylor et al., 2003a). On
neurogenesis disturbance. the neural network level, the lesions are thought
to disrupt neural pathways involved in regular
affective and cognitive information processing,
White matter lesions (WML) especially within the prefrontal lobe.
According to an early study by Coffey et al.
On the macroscopic level, the patchy degeneration (1989), only 14% of aged healthy subjects but 55%
of subcortical brain parenchyma manifests in the of patients with late-life depression exhibit large
so-called WML on MRI has been identified as a and confluent WML (Coffey et al., 1989). The
frequent concomitant of mood disturbances. severity of vascular risk factors and associated
486

WML is of clinical significance, as they predict the reduction of the gray matter in both amygdalae for
development of delusional states and persistent the MDD-patients.
cognitive deficits in geriatric depression (O’Brien The vast majority of morphometric studies in
et al., 1997). A recent longitudinal study by Taylor MDD examined preselected regions of interest
and co-workers in 2003 evidenced a poorer (ROI) that had been characterized as ‘‘dysfunc-
therapeutic outcome for MDD patients with a tional’’ in previous functional neuroimaging
pronounced progression of WML (Taylor et al., experiments. As outlined above, especially the
2003b). The development and prognosis of MDD emotional and cognitive processing modules of the
is thus crucially influenced by the focal destruction medial temporal and prefrontal lobe have been
of brain parenchyma; however, the clear age-de- identified as functional key regions. About half of
pendence and the lack of a correlation with other these studies focused on the volumetric analysis of
established brain volumetric alterations suggest an the hippocampus (HC) formation (see Table 2),
independent vascular risk factor for depressive which is a reliable predictor of memory dysfunc-
states in elderly MDD patients (Janssen et al., tion in MDD (O’Brien et al., 2004; Hickie et al.,
2004). 2005).
Significant reductions in HC volumes were
repeatedly reported for both early- and late-onset
Subtle morphometric changes MDD (Bremner et al., 2000; Frodl et al., 2004;
Janssen et al., 2004; Hickie et al., 2005). The
In contrast to classical neurological diseases, the precise impact of these findings is still under
evidenced brain alterations in MDD — if any — debate, e.g., the question being whether the whole
are very subtle, i.e., the pathophysiological infor- HC formation is affected or only parts of it
mation arises primarily from group averaging, and (Neumeister et al., 2005), or whether these alter-
not from the analysis of an individual data set ations are best described as volume reductions or
(e.g., as in the case of WML). All morphometric shape deformations (Posener et al., 2003).
MRI approaches are based on the acquisition of It has been shown that stress, whether environ-
precise anatomical images of the brain using high- mental or social, actually changes the shape, size,
resolution 3D sequences (e.g., MPRAGE, FLASH and number of neurons in the hippocampus and it
3D). Global segmentation algorithms provide only is hypothesized that the manifold mechanisms of
general information about tissue volumes without antidepressive action follow a final common path-
regional specificity (i.e., total gray matter, white way, i.e., a final induction of specific alterations of
matter, and cerebrospinal fluid). Two recent neuroadaptation in specific brain regions. In such
studies employed this approach; one reported a a modern disease, model behavioral depressive
negative association of illness duration and changes relate in part to alterations in hippocam-
cerebral gray matter volumes in female MDD pal function. These are thought to be induced
(Lampe et al., 2003), while the other indicated a through activation of cyclic AMP response ele-
poorer clinical outcome for patients with enlarged ment-binding protein (CREB) and neurotrophins
cerebrospinal fluid spaces (Cardoner et al., 2003). such as brain derived neurotrophic factor (BDNF)
Voxel-based morphometry (VBM), in contrast, is (McEwen, 2000; Duman, 2002a).
a fully automated technique that allows the unbi- Other ROI-based analyses focused on the am-
ased examination of the whole brain on a voxel- ygdala and the ventral prefrontal cortex. In the
by-voxel basis. Interestingly, only one VBM study human brain, both areas are substantially involved
has been published in MDD research so far (Taki in the emotional evaluation of sensory stimuli and
et al., 2005), reporting a bilateral gray matter re- the generation of appropriate behavioral and
duction of the medial prefrontal lobe in elderly endocrinological responses. In the case of the am-
males with subthreshold depression. Our own pre- ygdala, a substantial number of studies reported
liminary VBM analysis of 10 MDD patients com- significant volume decreases in MDD (Frodl et al.,
pared to 10 matched healthy controls shows a 2004; Hastings et al., 2004; Rosso et al., 2005), a
487

Table 2. Brain morphometric findings in major depressive disorder

Authors Sample Method Main result Key region

Prefrontal cortex
Taki et al. (2005) 34 P, 109 C OVBM Aged male P with subthreshold depression: MedPFC
MedPFC GMk, precentral gyrus GMk
Coryell et al. (2005) 10 P, 10 C ROI P with psychotic features: left posterior SACG
subgenual ACGk, subgroup with ACGm
at 4-year follow-up
Lacerda et al. (2004) 31 P, 34 C ROI P: left medial and right lateral OFC GMk; OFC
gender effect: medial OFC GMk in male P
Ballmaier et al. (2004) 24 P, 19 C SEG, ROI Aged P: ACG GMk WMk CSFm, GR ACG, OFC, GR
GMk WMk CSFm, OFC GMk CSFk,
DLPFC+, precentral gyrus+
Almeida et al. (2003) 51 P, 37C ROI Right FCk in late-onset P compared to Total FC
early-onset P and C; lacking correlation
with cognitive performance
Steffens et al. (2003) 30 P, 40C ROI Aged P: left OFC volume predicts Benton OFC
Visual Retention Test performance
Taylor et al. (2003c) 41 P, 40 C ROI Aged P: OFCk predicts impairment in OFC
activities of daily living
Steingard et al. (2002) 19 P, 38 C SEG, ROI Adolescent P: whole brain volumek, FC Total FC
WMk, FC GMm
Bremner et al. (2002) 15 P, 20 C ROI Remitted P: OFCk, subgenual ACG+, OFC
dorsal ACG+
Nolan et al. (2002) 22 P, 22 C ROI Left PFCm in pediatric P with nonfamiliar PFC
MDD compared to familiar MDD and C;
familiar MDD: PFC+ compared to C
Botteron et al. (2002) 48 P, 17 C ROI Female P with early-onset: left subgenual SACG
ACGk

Medial temporal lobe


Neumeister et al. (2005) 31 P, 57 C ROI Anterior HCk, posterior HC+ HC
Taylor et al. (2005) 135 P, 83 C ROI HCk in late-onset P homozygous for the L HC
allele of the serotonin transporter gene 5-
HTTLPR
Hickie et al. (2005) 66 P, 20 C ROI HCk in early- and late-onset P, association HC
with deficient visual and verbal memory
performance
Hastings et al. (2004) 18 P, 18 C ROI Female P: AMYk compared to female C; AMY
male P: left subgenual ACGk compared to
female P; HC+; OFC+
Rosso et al. (2005) 20 P, 24 C ROI Pediatric P: AMYk; HC+ AMY
Janssen et al. (2004) 28 P, 41 C ROI Female early-onset P: right HCk, no HC
association with subcortical white matter
lesions, PHG+, OFC+
O’Brien et al. (2004) 61 P, 40 C ROI Aged P: hypercortisolemia, right HCk, HC
association with persisting mild cognitive
impairment
Frodl et al. (2004) 30 P, 30 C ROI Longitudinal study: nonremitted patients HC, AMY
at 1-year follow-up exhibited AMYk HCk
at baseline
Posener et al. (2003) 27 P, 42 C ROI HC+, but highly significant differences in HC
HC shape, especially in the subiculum
MacMillan et al. (2003) 23 P, 23 C ROI Drug-naive pediatric P: significantly HC
increased AMY/HC volume ratio
compared to C, association with anxiety
severity
488

Table 2 (continued )

Authors Sample Method Main result Key region

Vythilingam et al. (2002) 32 P, 14 C ROI Influence of environmental factors: left HC


HCk in female P with childhood trauma
compared to nontraumatized P and C.
Bremner et al. (2000) 16 P, 16 C ROI P: Left HCk, AMY+, CAU+, FC+, HC, AMY
TC+

Basal ganglia
Lacerda et al. (2003) 25 P, 48 C ROI CAU+, PUT+, GP+; significant GP GP
asymmetry decrease; GP/PUT volume:
association with clinical parameters
Naismith et al. (2002) 47 P, 20 C ROI Psychomotor slowing in aged P is predicted CAU
by CAUk and methylenetetrahydrofolate
reductase genotype

Treatment effects
Lekwauwa et al. (2005) 25 P ROI Pre-post ECT: smaller right HC volumes HC
predict better treatment outcome
Lavretsky et al. (2005) 41 P, 41 C ROI OFC GMm in P treated with OFC
antidepressants compared to drug-naive P;
OFC GMk in both P groups compared to
controls
Vythilingam et al. (2004) 38 P, 33 C ROI Pre-post SSRI in medication-free P: HC+ HC
compared to C at baseline, HC+ after
successful SSRI treatment
Sheline et al. (2003) 38 P ROI Female P: length of untreated MDD HC
episodes predicts smaller HC GM volume
(antidepressants - neuroprotective)
Hsieh et al. (2002) 60 P ROI Pre-post antidepressants: smaller total/ HC
right HC volumes predict poorer treatment
outcome
Vakili et al. (2000) 38 P, 20 C ROI Female fluoxetine responder: right HCm HC
compared to nonresponder. Whole P
group: HC+ compared to C

Notes: P ¼ major depressive disorder patients; C ¼ healthy controls; ROI ¼ region of interest analysis (manual tracing); SEG ¼ global seg-
mentation in gray matter (GM) — white matter (WM) — cerebrospinal fluid (CSF); OVBM ¼ optimized voxel-based morphometry; m ¼ significant
volumetric increase; k ¼ significant volumetric decrease; + ¼ no significant volumetric differences; MDD ¼ major depressive disorder; FC ¼ frontal
cortex; PFC ¼ prefrontal cortex; TC ¼ temporal cortex; DLPFC ¼ dorsolateral prefrontal cortex; AGC ¼ anterior cingulate gyrus; sAGC ¼ sub-
genual anterior cingulate gyrus; MedPFC ¼ medial prefrontal cortex; OFC ¼ orbitofrontal cortex; GR ¼ gyrus rectus (medial OFC); AMY ¼
amygdala; HC ¼ hippocampus; PHG ¼ parahippocampal gyrus; PUT ¼ putamen; CAU ¼ caudate; GP ¼ globus pallidus; SSRI ¼ selective serotonin
reuptake inhibitor.

finding that could be replicated in our own VBM MDD patients (Ballmaier et al., 2004; Lacerda et
morphometric study (see Fig. 1). Within the pre- al., 2004; Coryell et al., 2005). On the cognitive-
frontal cortex, most volumetric analyses targeted behavioral level, the structural integrity of the
the key regions of cortical mood regulation, the OFC proved to be of particular importance, as it
OFC, and subgenual anterior cingulate cortex predicted deficits in visuospatial memory (Steffens
(sACG). For both regions, a significant gray mat- et al., 2003) and impairments in activities of daily
ter decrease was reported in early-onset and aged living (Taylor et al., 2003c) associated with MDD.
489

Fig. 1. Results of the voxel-based morphometry analysis. Regions with significant more gray matter volume in 10 healthy subjects
compared to 10 depressed patients.

In summary, especially fine structural anomalies depression. Antidepressant medications work for
of HC, amygdala, and OFC are regarded as stable many patients, but there is no absolute certainty of
correlates of the disorder, as they were also evi- the actual relationship between neurotransmitters
denced in unmedicated patients (Bremner et al., and depression.
2002; Lacerda et al., 2004; Neumeister et al., 2005), The effects of neurotransmitters are extremely
populations at risk (Omura et al., 2005; Taki et al., difficult to study in vivo. Neurotransmitters are
2005) and pediatric samples (Steingard et al., 2002; present in very small quantities, they are only
Rosso et al., 2005). In line with established diathe- available in certain locations within the brain, and
sis-stress models, the formation of morphometric they disappear very quickly once they are used.
alterations seems to be facilitated by both genetic Because they are removed so quickly, they cannot
vulnerability factors (e.g., familiar MDD (Nolan be measured directly. What can be measured are
et al., 2002) and 5-HTTLPR serotonin transporter the so-called metabolites — the remaining sub-
genotype (Taylor et al., 2005)) and environmental stances in the brain after neurotransmitters have
stressors (e.g., childhood trauma, Vythilingam been used. By measuring these metabolites, un-
et al., 2002). On the microscopic level, the findings derstanding of the effects of changes in neuro-
are in good accordance with the known cellular transmitters in the brain can be gained.
and neurochemical alterations associated with Proton magnetic resonance spectroscopy (1H
MDD, i.e., anomalies in the density and size of MRS) allows quantitative and noninvasive access
neuronal and glial cells (Cotter et al., 2002; Hamidi to a number of metabolites in different brain
et al., 2004), neurotrophin-induced disturbances of regions in vivo. MRS is a noninvasive technique
cellular plasticity (Duman, 2002b), and neurogen- that exhibits relatively high-spatial resolution and
esis (Kempermann and Kronenberg, 2003). requires neither radioactive tracers nor ionizing
In conclusion, consistent functional and struc- radiation. Since metabolite concentrations are
tural imaging findings in MDD reported changes more than 103 times smaller than the tissue water
predominantly in the emotionally relevant net- concentration, spatial and temporal resolution is
works, including the prefrontal cortex and the proportionally lower for MRS compared to MRI
limbic system with the hippocampus and the am- applications. There are hundreds of metabolites
ygdala. produced by the human brain, but only some can
reliably be detected and quantified using 1H MRS.
The significant neurometabolites that have been
Magnetic resonance spectroscopy (MRS) and MDD measured in patients with major depression are
as follows: glutamate, glutamine, and GABA, N-
There appears to be a strong relationship between acetylaspartate (NAA), choline-containing com-
neurotransmitter levels in the brain and clinical pounds (Cho), creatine and phosphocreatine (tCr),
490

and myo-inositol (mI). In the past, most investi- and white matter (WM) and have only negligible
gators have expressed MRS results in terms of concentrations in CSF (Wang and Zimmerman,
peak ratios, resulting in the ambiguity of whether 1998; Hetherington et al., 2001) and also vary with
one metabolite is increased or another decreased. the brain region. Cortical gray matter ranges from
Although quantitative MRS would make inter- about 3 to 6 mm in thickness and the orientation
subject comparisons more rigorous, no generally of this thickness varies (Noworolski et al., 1999).
accepted standard quantitation method for in vivo With typical voxel sizes of 1–4 cm3 for 1H MRS,
data exists. This is primarily due to the difficulty of most spectroscopic voxels consist a mixture of
computing absolute metabolite concentrations different tissues and CSF, which makes the com-
from MR signal intensities, including correction parison of measured metabolite concentrations
of coil loading, B1-field inhomogeneity, and oth- between subjects difficult.
ers. But by themselves these corrections do not Therefore, regardless of the applied method for
provide a measurement of concentration (e.g., in MRS data acquisition and evaluation, the tissue
mmol/L) and thus cannot be used for direct com- composition of the voxel from which the spectrum
parisons of data taken on different instruments, or is obtained plays an important role (Schuff et al.,
at other laboratories. Therefore, most of the stud- 1997), especially for the detection of discrete ab-
ies, which report metabolite values avoiding ratios, normalities in MR-detectable metabolites in MDD
report ‘‘arbitrary units’’ (a.u.) or ‘‘institutional patients. Yet, many previous MRS studies did not
units’’ (i.u.) and make use of the term ‘‘semiquan- take this influence into account.
titative’’ measure. Within the following, an introduction to the
Several studies have shown that metabolite con- MR observable metabolite resonances will be
centrations vary in different brain regions. The given. Fig. 2 illustrates key brain regions for emo-
concentrations are different in gray matter (GM) tional processing and MDD that have been

Fig. 2. Key brain regions for emotional processing and MDD that have been investigated with 1H MR spectroscopy.
491

investigated with 1H MR spectroscopy. The results 15% of the brain energy consumption is GABA
of MRS studies in several key regions for emo- related (Shulman et al., 2004). The in vivo detec-
tional processing in patients suffering from major tion of GABA via 1H MRS requires special editing
depressive episodes will be presented and discussed methods that have been pioneered by Rothman
and a summary is given in Table 3. and colleagues from Yale (Rothman et al., 1993).

MR-detectable brain metabolites that have been N-acetylaspartate (NAA)


found altered in MDD NAA, which is solely found in the brain, is present
in both gray and white matter and was identified
Glutamate, glutamine, and GABA as a neuronal marker in early histological studies,
Glutamate is the major excitatory, and GABA the and also as a surrogate marker of neuronal and
major inhibitory neurotransmitter in the human axonal functioning and integrity. The NAA reso-
brain. Both glutamate and GABA are linked to nance is typically the tallest peak in the normal
metabolism through a neurotransmitter cycle brain spectrum. Since its decrease or disappear-
between neurons and glia. In this cycle, neuro- ance can be due to either cell death or axonal
transmitter molecules released by the neurons are injury, it is considered to be a measure of neural
taken up by transporters in surrounding glial cells. tissue viability and/or function.
In the glia they are converted to glutamine, which
is released to the neuron where it is used for the Choline-containing compounds (Cho)
resynthesis of the neurotransmitter. The MR detectable Cho signal represents the
Glutamate is present at an even higher concen- trimethyl ammonium resonance of a variety of
tration as N-acetyl aspartate in the brain, though in Cho. The MRS resonance is composed of acetyl-
practice glutamate, glutamine, and GABA signals choline (ACh), phosphocholine (PC), glycero-
are barely detectable using clinical MR scanners. phosphocholine (GPC), and free choline. Most of
The MR sensitivity for their detection is poor due the signal arises from PC and GPC, free choline is
to their signal intensity being spread over a large less than 5% and the contribution from ACh is
number of closely spaced multiplet resonances and negligible (Boulanger et al., 2000). An increased
from signal cancellation of overlapping resonances Cho signal most likely reflects an increase in mem-
due to phase differences at longer echo times (TE). brane turnover.
In order to detect these resonances at a field Phosphatidylcholine, the major choline-contain-
strength of 1.5 T, pulse sequences with short TE ing metabolite of the normal brain, is MR-invisible
under 30 m/s have to be employed. As a rule spec- in myelin, cell membranes, and other brain lipids
tra acquired at short TE have an increased signal- under normal circumstances. However, under
to-noise ratio compared to longer TE spectra and certain conditions, visible choline may be released
may allow for the quantification of a variety of from this pool (Danielsen and Ross, 1999). In
overlapping signals. At a field strength of 1.5 T, proton MR spectroscopy (nonmembrane bound),
attempts have been made to quantify the overlap- PC and GPC cannot be distinguished — both are
ping signals of glutamate, glutamine, and GABA detected within one choline peak.
at the 2.3 ppm position, labeled as Glx. The origin
of the glutamate signal (e.g., intra vs. extracellular)
in 1H MRS brain spectra cannot be subclassified. Creatine and phosphocreatine (tCr)
All mobile molecules will contribute equally to the In normal brain metabolism, phosphocreatine sup-
observed glutamate resonance. An example for a plies phosphate to adenosine diphosphate (ADP),
brain 1H MR spectrum acquired at 1.5 T with the resulting in the production of an adenosine tripho-
(relative) short TE of 30 ms is shown in Fig. 3. sphate (ATP) molecule and the release of creatine.
GABA is synthesized primarily from the gluta- Thus, total creatine (creatine plus phosphocreatine,
mate precursor in GABAergic neurons. About tCr) should be a reliable marker of brain
492

Table 3. Proton MR spectroscopic findings in major depressive disorder

Authors (by brain Sample Method Main result: abnormal Key region
region) metabolite levels

Prefrontal cortex
Frey et al. (1998) 22 P, 22 C SVS TE ¼ 55 ms, Decreased m-Ino/tCr Right frontal lobe
TR ¼ 3.5 s
Steingard et al. (2000) 17 P, 22 C SVS TE ¼ 30 ms, Increased Cho
TR ¼ 2.5 s
Farchione et al. (2002) 11 aP, 11C 2D MRSI, TE ¼ 272 ms, Increased Cho/NAA Left anterior
TR ¼ 2.3 s medial frontal lobe
Kumar et al. (2002) 20 P, 18 C SVS TE ¼ 30 ms, Increased Cho/tCr, increased Left DLPFC
TR ¼ 3 s m-Ino/tCr
Michael et al. (2003a) 12 P, 12 C SVS TE ¼ 20 ms, Decreased Glx Left DLPFC
TR ¼ 2.5 s
Grachev et al. (2003) 10 P, 10 C SVS TE ¼ 30 ms, Decreased NAA/tCr Right DLPFC
TR ¼ 1.5 s
Gruber et al. (2003) 17 P, 17 C SVS TE ¼ 20 ms, Increased Cr Left DLPFC
TR ¼ 6 s
Caetano et al. (2005) 14 aP, 22 C SVS TE ¼ 30 ms, Decreased Cho, increased m- Left DLPFC
TR ¼ 6 s Ino
Brambilla et al. (2005) 9 cP, 10 ncP, 19 C SVS TE ¼ 20 ms, Decreased NAA/Cho and left DLPFC
TR ¼ 3 s NAA/tCr in cP
Coupland et al. (2005) 13 P, 13 C SVS TE ¼ 28 ms, Decreased m-Ino/tCr Pregenual anterior
TR ¼ 3 s cingulate/MedPFC

Prefrontal cortex — therapy response


Gonul et al. (2006) 20 P, 18 C SVS Increase of NAA/tCr Left medial frontal
posttreatment cortex

Hippocampus/temporal lobe
Blasi et al. (2004) 17 P, 17 C 2D MRSI, TE ¼ 272 ms, Decreased NAA/tCr Hippocampus
TR ¼ 2.2 s
Ende et al. (2000) 17 P, 24 C 2D MRSI, TE ¼ 20 ms, Decreased Cho Hippocampus
TR ¼ 1.8 s
Kusumakar et al. 11 aP, 11 C 2D MRSI, TE ¼ 135 ms, Decreased Cho/tCR Left amygdala
(2001) TR ¼ 1.5 s
Ende et al. (in press) 8 P, 8 C 2D MRSI, TE ¼ 135 ms, Decreased Cho Hippocampus
(WIN) TR ¼ 1.5 s
Michael et al. (2003b) 13 P, 28 C SVS TE ¼ 20 ms, Decreased Glx Hippocampus/
TR ¼ 2.5 s amygdala

Basal ganglia
Charles et al. (1994) 7 P, 7 C SVS Increased Cho/tCr Thalamus,
putamen, and
white matter
Renshaw et al. (1997) 41 P, 22 C SVS Decreased Cho/tCr Caudate and
putamen
Vythilingam et al. 8 P, 12 C 2D MRSI, TE ¼ 20 ms, Decreased NAA/tCr Caudate
(2003) TR ¼ 1.5 s
Vythilingam et al. 17 P, 17 C 2D MRSI, TE ¼ 20 ms, Increased Cho/tCr Putamen
(2003) TR ¼ 1.5 s
Ende et al. (in press) 8 P, 6 C 2D MRSI, TE ¼ 135 ms, Increased Cho Putamen
TR ¼ 1.5 s
493

Table 3 (continued )

Authors (by brain Sample Method Main result: abnormal Key region
region) metabolite levels

Basal ganglia — therapy response


Charles et al. (1994) 7 P, 7 C SVS Increase of NAA/Cho, Thalamus,
decrease of Cho/tCr putamen, and
white matter
Sonawalla et al. (1999) 15 P: 8 R, 7 NR SVS Increase of Cho/tCr in R Caudate and
putamen

Anterior cingulate
Auer et al. (2000) 19 P, 18 C SVS TE ¼ 35 ms, Decreased Glx Anterior cingulate
TR ¼ 2 s
Pfleiderer et al. (2003) 17 P, 17 C SVS TE ¼ 20 ms, Decreased Glx Left anterior
TR ¼ 2.5 s cingulate
Rosenberg et al. (2005) 14 aP, 14 C SVS short TE Decreased Glx Anterior cingulate

Notes: P ¼ major depressive disorder patients; aP ¼ adolescent major depressive disorder patients; cP ¼ chronic major depressive disorder patients;
ncP ¼ nonchronic major depressive disorder patients; C ¼ healthy controls; DLPFC ¼ dorsolateral prefrontal cortex; AGC ¼ anterior cingulate gyrus;
MedPFC ¼ medial prefrontal cortex; NAA ¼ N-acetylaspartate; Cho ¼ choline-containing compounds; tCr ¼ total creatine (creatine and phosphocrea-
tine); m-Ino ¼ myo-Inositol; Glx ¼ glutamate and glutamine; SVS ¼ single voxel spectroscopy; MRSI ¼ MR spectroscopic imaging; TE ¼ echo time;
TR ¼ repetition time.

studies. However, tCr represents an important


buffer capacity in the energy metabolism of the
cell, which cannot be considered 100% stable a
priori. Recent studies show a brain activity-de-
pendent change in tCr signal intensity (Ke et al.,
2002). That tCr is not always useful, as a reference
compound has already been suggested more than a
decade ago (Ross and Michaelis, 1994).

Myo-inositol (mI)
mI is the major nutritionally active form of in-
ositol and is vital to many biological processes of
the body, participating in a diverse range of activ-
Fig. 3. Example for a brain 1H MR spectrum acquired at 1.5 T
ities. Myo-inositol is one of nine distinct isomers of
with the (relative) short TE of 30 ms. inositol. It is essential for the growth of rodents,
but not for most animals, including humans. In
metabolism. Creatine, phosphocreatine, and their humans mI is made endogenously from glucose.
main precursor, guanidinoacetate, are synthesized The dietary intake of mI can influence the levels of
in extracerebral tissues (primarily liver and kidney) circulating and bound mI in the body. The specific
and then transported to the brain. Therefore, any importance of mI is based on the fact that its lipid
metabolic defect resulting in decreased production conjugates are directly involved into the inositol-3-
of creatine/phosphocreatine (e.g., hepatic or renal phosphate (IP3) second messenger pathway. The
diseases) will lower the tCr peak. ‘‘inositol depletion hypothesis’’ explains the mech-
The tCr signal has been used for gauging in anism of action of lithium in bipolar disorders
metabolite ratios like NAA/tCr in countless (Harwood, 2005).
494

MRS results in key regions for emotional processing MRS studies have targeted the hippocampus
and MDD and the amygdala, and a decreased Cho signal is
the most frequently reported abnormality in MDD
Prefrontal cortex (Ende et al., 2000; Kusumakar et al., 2001; Ende
The prefrontal cortex is the anterior part of the et al., in press). Additionally, a decreased NAA/
frontal lobes of the brain. It comprises the dorso- tCr ratio (Blasi et al., 2004) and a decreased Glx
lateral, ventrolateral, orbitofrontal, and mesial signal (Michael et al., 2003b) have been found.
prefrontal areas, cortical key regions of cognition
and mood regulation. These brain regions have
Basal ganglia
been implicated in planning complex cognitive be-
haviors, personality expression, moderating cor-
The basal ganglia comprise a number of subcor-
rect social behavior and regulation of emotion.
tical structures that are vital for the modulation of
The prefrontal cortex may play an important role
behavioral patterns, emotional responses, and
in delayed gratification by maintaining emotions
executive cognitive functions. Part of the basal
over time and organizing human behavior toward
ganglia is the striatum consisting of the caudate
specific goals.
nucleus and the putamen.
As outlined above and summarized in Tables 1
Resting state metabolism in the basal ganglia
and 2, functional as well as morphological alter-
measured with PET has been found altered in
ations have been found in these brain regions in
MDD patients (Ho et al., 1996; Videbech et al.,
MDD. Thus it is not surprising that the majority
2002). Functional magnetic resonance imaging
of MRS studies in MDD targeted the prefrontal
(fMRI) studies using emotional stimulation with
lobe.
faces detected significantly different bold responses
Although the results are not always consistent,
in the putamen and the caudate nuclei of MDD
there is evidence that the Cho, mI, and tCr signals
patients as compared to healthy subjects (Fu et al.,
are abnormal in MDD patients. Some of the di-
2004; Surguladze et al., 2005). The most concord-
vergent results, at first sight, are possibly explained
ant MRS finding in this brain region in MDD is an
by the methodological differences, e.g., voxel sizes,
increased Cho signal or Cho/tCr, respectively.
voxel tissue composition, and quantitation methods
Nevertheless, one study also found a decrease in
used. The observation of an increased absolute tCr
the tCr/Cho ratio (Renshaw et al., 1997).
signal by Gruber et al. (2003) could explain why the
There are two MRS studies in the basal ganglia
studies by Frey et al. (1998) and Coupland et al.
reporting metabolite changes accompanying ther-
(2005) find decreased mI/tCr, whereas an increased
apy response: in the early study by Charles et al.
mI signal was found by Caetano et al. (2005).
(1994), an increase of NAA/Cho and a decrease of
Additionally, the finding of an increased tCr by
Cho/tCr in response to antidepressant therapy were
Gruber et al. (2003) gives further reason to be
observed, whereas Sonawalla et al. (1999) found an
cautious with the use of tCr as an internal reference.
increase of Cho/tCr in therapy responders.

Temporal lobe, hippocampus, and amygdala Anterior cingulate gyrus

Another target region for emotional processing is The ACG is located bilaterally in the medial wall
the temporal lobe. Although the amygdala plays a of the frontal lobes below the cingulated sulcus.
particularly important role for all kinds of emo- The ACG is vital to cognitive functions, such as
tional processing, the hippocampus and other reward anticipation, decision making, empathy,
parts of the temporal lobe have been found to be and emotion. It may be particularly important
morphometrically abnormal and abnormally acti- with regard to conscious subjective emotional
vated in MDD patients in the brain’s response to awareness in humans. Functional and morpho-
emotional stimuli (see Tables 1 and 2). metrical studies reported alterations of the ACG in
495

MDD (Tables 1 and 2). Three MRS studies have decreased tCr signal possibly mirrors energy me-
targeted the ACG and have consistently found tabolism hypofunction. The observation of the
abnormally low Glx signals in MDD compared to GABA signal with MRS has been limited to oc-
healthy subjects (Auer et al., 2000; Pfleiderer et al., cipital regions due to methodological limitations
2003; Rosenberg et al., 2005). Since the Glx signal with GABA editing (Sanacora et al., 2004). Future
comprises glutamine and glutamate, the interpre- improvements are foreseeable as the methodology
tation of the observed changes remains difficult. further develops.
Spectral editing methods for more reliable separa- MRS can also be applied to animal models. This
tion of these two resonances will improve with is a great advantage since MRS is the only non-
higher magnet field strengths (3 T and more). invasive tool monitoring metabolic changes in pa-
tients in vivo. Most of our knowledge about
etiopathogenesis of depression comes from animal
Conclusions models, and MRS may be the tool of choice to
bridge the gap between clinical and preclinical re-
MRS is a noninvasive tool that offers unique in- search in psychiatric disorders.
sights into the pathophysiology of MDD in vivo, MRS has become a complex and sophisticated
with the ability for longitudinal observations of neuroimaging technique that enables reliable and
therapy response or illness progression. reproducible quantification of an increasing num-
To date the number of studies attempting to put ber of neurometabolites. Still, it has not yet
those pieces together is still limited. Nevertheless, reached maturity. With access to higher field
it is impressive that metabolic alterations in MDD strength (3 T and more), spectral editing and the
patients appear in those brain regions relevant for observation of other nuclei than protons will be
communication and processing of emotions. facilitated. In general, MRS will gain in spatial
There is evidence that the Cho signal, repre- resolution and sensitivity with higher field
senting membrane turnover and supposable strength. Although advances in MRS research
changes in synaptic plasticity, is abnormal in are foreseeable in the near future, the quality of
MDD compared to healthy subjects. Furthermore, MRS studies will depend on the accuracy of the
there are concordant findings of a reduced Glx applied acquisition and quantitation methods, and
signal in the ACG in MDD. Further changes have the application in large and well-defined patient
been reported for mI with less concordant direc- populations ideally in longitudinal prospective
tions of concentration changes. study designs to confirm either stable or progres-
NAA as a surrogate marker of neuronal viabil- sive brain deficits.
ity and probably of synaptic plasticity has not yet Last but not least, with decreasing measurement
been reported to change in absolute concentration times at higher field strength it will be possible to
in any of the investigated brain regions in MDD. join forces to create new study protocols and tech-
A decrease of NAA/tCr has been reported in niques for mapping noninvasively brain activa-
MDD patients from the DLPFC (Grachev et al., tion, morphometry, diffusion, and metabolism
2003; Brambilla et al., 2005), the hippocampus with advanced MR methods and then to correlate
(Blasi et al., 2004), and the caudate nucleus these findings with neuropsychological test results,
(Vythilingam et al., 2003). Furthermore, an in- psychiatric ratings, and genetic polymorphisms to
crease of this ratio following antidepressant ther- further the understanding of emotional processing
apy was observed by Sonawalla et al. (1999). biases in MDD.
However, it has yet to be determined whether these
changes are solely due to, or even dominated by, a
concentration change of NAA or rather an altered Acknowledgment
tCr signal as reported by Gruber et al. (2003) and
observed to increase in the hippocampus following The authors thank the Heidelberg Academy of
electroconvulsive therapy (Ende et al., 2000). A Science for grant support, Dr. Tim Wokrina for
496

providing the brain 1H MR spectrum (Fig. 3), Sigi Bradley, B.P., Mogg, K. and Millar, N. (1996) Implicit memory
Walter for analyzing the MR spectra, and Matt- bias in clinical and non-clinical depression. Behav. Res.
Ther., 34: 865–879.
hias Ruf for providing Fig. 2.
Brambilla, P., Stanley, J.A., Nicoletti, M.A., Sassi, R.B., Mal-
linger, A.G., Frank, E., Kupfer, D.J., Keshavan, M.S. and
Soares, J.C. (2005) 1 H Magnetic resonance spectroscopy
study of dorsolateral prefrontal cortex in unipolar mood dis-
References order patients. Psychiatry Res., 138: 131–139.
Breiter, H.C., Etcoff, N.L., Whalen, P.J., Kennedy, W.A.,
Abercrombie, H.C., Schaefer, S.M., Larson, C.L., Oakes, T.R., Rauch, S.L., Buckner, R.L., Strauss, M.M., Hyman, S.E.
Lindgren, K.A., Holden, J.E., Perlman, S.B., Turski, P.A., and Rosen, B.R. (1996) Response and habituation of the
Krahn, D.D., Benca, R.M. and Davidson, R.J. (1998) Met- human amygdala during visual processing of facial expres-
abolic rate in the right amygdala predicts negative affect in sion. Neuron, 17: 875–887.
depressed patients. Neuroreport, 9: 3301–3307. Bremner, J.D., Innis, R.B., Salomon, R.M., Staib, L.H., Ng,
Adolphs, R., Tranel, D., Damasio, H. and Damasio, A. (1994) C.K., Miller, H.L., Bronen, R.A., Krystal, J.H., Duncan, J.,
Impaired recognition of emotion in facial expressions fol- Rich, D., Price, L.H., Malison, R., Dey, H., Soufer, R. and
lowing bilateral damage to the human amygdala. Nature, Charney, D.S. (1997) Positron emission tomography meas-
372: 669–672. urement of cerebral metabolic correlates of tryptophan de-
Almeida, O.P., Burton, E.J., Ferrier, N., McKeith, I.G. and pletion-induced depressive relapse. Arch. Gen. Psychiatry,
O’Brien, J.T. (2003) Depression with late onset is associated 54: 364–374.
with right frontal lobe atrophy. Psychol. Med., 33: 675–681. Bremner, J.D., Narayan, M., Anderson, E.R., Staib, L.H.,
Auer, D.P., Putz, B., Kraft, E., Lipinski, B., Schill, J. and Ho- Miller, H.L. and Charney, D.S. (2000) Hippocampal volume
lsboer, F. (2000) Reduced glutamate in the anterior cingulate reduction in major depression. Am. J. Psychiatry, 157:
cortex in depression: an in vivo proton magnetic resonance 115–118.
spectroscopy study. Biol. Psychiatry, 47: 305–313. Bremner, J.D., Vythilingam, M., Vermetten, E., Nazeer, A.,
Ballmaier, M., Toga, A.W., Blanton, R.E., Sowell, E.R., Adil, J., Khan, S., Staib, L.H. and Charney, D.S. (2002) Re-
Lavretsky, H., Peterson, J., Pham, D. and Kumar, A. duced volume of orbitofrontal cortex in major depression.
(2004) Anterior cingulate, gyrus rectus, and orbitofrontal Biol. Psychiatry, 51: 273–279.
abnormalities in elderly depressed patients: an MRI-based Caetano, S.C., Fonseca, M., Olvera, R.L., Nicoletti, M., Hatch,
parcellation of the prefrontal cortex. Am. J. Psychiatry, 161: J.P., Stanley, J.A., Hunter, K., Lafer, B., Pliszka, S.R. and
99–108. Soares, J.C. (2005) Proton spectroscopy study of the left
Beck, A.T. (2005) The current state of cognitive therapy: a 40- dorsolateral prefrontal cortex in pediatric depressed patients.
year retrospective. Arch. Gen. Psychiatry, 62: 953–959. Neurosci. Lett., 384: 321–326.
Beck, A.T., Rush, A.J. and Shaw, B.G.E. (1979) Cognitive Canli, T., Cooney, R.E., Goldin, P., Shah, M., Sivers, H.,
Therapy of Depression. Guilford Press, New York. Thomason, M.E., Whitfield-Gabrieli, S., Gabrieli, J.D. and
Bench, C.J., Friston, K.J., Brown, R.G., Scott, L.C., Frack- Gotlib, I.H. (2005) Amygdala reactivity to emotional faces
owiak, R.S. and Dolan, R.J. (1992) The anatomy of melan- predicts improvement in major depression. Neuroreport, 16:
cholia–focal abnormalities of cerebral blood flow in major 1267–1270.
depression. Psychol. Med., 22: 607–615. Canli, T., Sivers, H., Thomason, M.E., Whitfield-Gabrieli, S.,
Blasi, G., Bertolino, A., Brudaglio, F., Sciota, D., Altamura, Gabrieli, J.D. and Gotlib, I.H. (2004) Brain activation to
M., Antonucci, N., Scarabino, T., Weinberger, D.R. and emotional words in depressed vs healthy subjects. Neurore-
Nardini, M. (2004) Hippocampal neurochemical pathology port, 15: 2585–2588.
in patients at first episode of affective psychosis: a proton Cardoner, N., Pujol, J., Vallejo, J., Urretavizcaya, M., Deus, J.,
magnetic resonance spectroscopic imaging study. Psychiatry Lopez-Sala, A., Benlloch, L. and Menchon, J.M. (2003) En-
Res., 131: 95–105. largement of brain cerebrospinal fluid spaces as a predictor of
Botteron, K.N., Raichle, M.E., Drevets, W.C., Heath, A.C. and poor clinical outcome in melancholia. J. Clin. Psychiatry, 64:
Todd, R.D. (2002) Volumetric reduction in left subgenual 691–697.
prefrontal cortex in early onset depression. Biol. Psychiatry, Charles, H.C., Lazeyras, F., Krishnan, K.R., Boyko, O.B.,
51: 342–344. Payne, M. and Moore, D. (1994) Brain choline in depression:
Boulanger, Y., Labelle, M. and Khiat, A. (2000) Role of in vivo detection of potential pharmacodynamic effects of
phospholipase A(2) on the variations of the choline signal antidepressant therapy using hydrogen localized spectros-
intensity observed by 1 H magnetic resonance spectroscopy in copy. Prog. Neuropsychopharmacol. Biol. Psychiatry, 18:
brain diseases. Brain Res. Brain Res. Rev., 33: 380–389. 1121–1127.
Bradley, B.P., Mogg, K. and Lee, S.C. (1997) Attentional biases Clark, D.M. and Teasdale, J.D. (1982) Diurnal variation in
for negative information in induced and naturally occurring clinical depression and accessibility of memories of positive
dysphoria. Behav. Res. Ther., 35: 911–927. and negative experiences. J. Abnorm. Psychol., 91: 87–95.
497

Coffey, C.E., Figiel, G.S., Djang, W.T., Saunders, W.B. and Fossati, P., Harvey, P.O., Le Bastard, G., Ergis, A.M., Jouvent,
Weiner, R.D. (1989) White matter hyperintensity on mag- R. and Allilaire, J.F. (2004) Verbal memory performance of
netic resonance imaging: clinical and neuroanatomic corre- patients with a first depressive episode and patients with
lates in the depressed elderly. J. Neuropsychiatry Clin. unipolar and bipolar recurrent depression. J. Psychiatr. Res.,
Neurosci., 1: 135–144. 38: 137–144.
Coryell, W., Nopoulos, P., Drevets, W., Wilson, T. and And- Frey, R., Metzler, D., Fischer, P., Heiden, A., Scharfetter, J.,
reasen, N.C. (2005) Subgenual prefrontal cortex volumes in Moser, E. and Kasper, S. (1998) Myo-inositol in depressive and
major depressive disorder and schizophrenia: diagnostic healthy subjects determined by frontal 1H-magnetic resonance
specificity and prognostic implications. Am. J. Psychiatry, spectroscopy at 1.5 tesla. J. Psychiatr. Res., 32: 411–420.
162: 1706–1712. Frodl, T., Meisenzahl, E.M., Zetzsche, T., Hohne, T., Banac,
Cotter, D., Mackay, D., Chana, G., Beasley, C., Landau, S. S., Schorr, C., Jager, M., Leinsinger, G., Bottlender, R., Re-
and Everall, I.P. (2002) Reduced neuronal size and glial cell iser, M. and Moller, H.J. (2004) Hippocampal and amygdala
density in area 9 of the dorsolateral prefrontal cortex in changes in patients with major depressive disorder and
subjects with major depressive disorder. Cereb. Cortex, 12: healthy controls during a 1-year follow-up. J. Clin. Psychi-
386–394. atry, 65: 492–499.
Coupland, N.J., Ogilvie, C.J., Hegadoren, K.M., Seres, P., Fu, C.H., Williams, S.C., Cleare, A.J., Brammer, M.J., Walsh,
Hanstock, C.C. and Allen, P.S. (2005) Decreased prefrontal N.D., Kim, J., Andrew, C.M., Pich, E.M., Williams, P.M.,
Myo-inositol in major depressive disorder. Biol. Psychiatry, Reed, L.J., Mitterschiffthaler, M.T., Suckling, J. and Bull-
57: 1526–1534. more, E.T. (2004) Attenuation of the neural response to sad
Danielsen, E.R. and Ross, B. (1999) Magnetic Resonance faces in major depression by antidepressant treatment: a
Spectroscopy Diagnosis of Neurological Diseases. Marcel prospective, event-related functional magnetic resonance im-
Dekker, New York. aging study. Arch. Gen. Psychiatry, 61: 877–889.
de Leeuw, F.E., de Groot, J.C., Achten, E., Oudkerk, M., George, M.S., Ketter, T.A., Parekh, P.I., Horwitz, B., Hers-
Ramos, L.M., Heijboer, R., Hofman, A., Jolles, J., van Gijn, covitch, P. and Post, R.M. (1995) Brain activity during tran-
J. and Breteler, M.M. (2001) Prevalence of cerebral white sient sadness and happiness in healthy women. Am. J.
matter lesions in elderly people: a population based magnetic Psychiatry, 152: 341–351.
resonance imaging study. The Rotterdam Scan Study. J. Ne- George, M.S., Ketter, T.A., Parekh, P.I., Rosinsky, N., Ring,
urol. Neurosurg. Psychiatry, 70: 9–14. H.A., Pazzaglia, P.J., Marangell, L.B., Callahan, A.M. and
Dolan, R.J., Bench, C.J., Brown, R.G., Scott, L.C. and Post, R.M. (1997) Blunted left cingulate activation in mood
Frackowiak, R.S. (1994) Neuropsychological dysfunction in disorder subjects during a response interference task (the
depression: the relationship to regional cerebral blood flow. Stroop). J. Neuropsychiatry Clin. Neurosci., 9: 55–63.
Psychol. Med., 24: 849–857. Gonul, A.S., Kitis, O., Ozan, E., Akdeniz, F., Eker, C., Eker,
Duman, R.S. (2002a) Pathophysiology of depression: the con- O.D. and Vahip, S. (2006) The effect of antidepressant treat-
cept of synaptic plasticity. Eur. Psychiatry, 17(Suppl 3): ment on N-acetyl aspartate levels of medial frontal cortex in
306–310. drug-free depressed patients. Prog. Neuropsychopharmacol.
Duman, R.S. (2002b) Synaptic plasticity and mood disorders. Biol. Psychiatry., 30: 120–125.
Mol. Psychiatry, 7(Suppl 1): S29–S34. Gotlib, I.H., Sivers, H., Gabrieli, J.D., Whitfield-Gabrieli, S.,
Elliott, R., Rubinsztein, J.S., Sahakian, B.J. and Dolan, R.J. Goldin, P., Minor, K.L. and Canli, T. (2005) Subgenual an-
(2002) The neural basis of mood-congruent processing biases terior cingulate activation to valenced emotional stimuli in
in depression. Arch. Gen. Psychiatry, 59: 597–604. major depression. Neuroreport, 16: 1731–1734.
Ende, G., Braus, D.F., Walter, S., Weber-Fahr, W. and Henn, Grachev, I.D., Ramachandran, T.S., Thomas, P.S., Szeverenyi,
F.A. (2000) The hippocampus in patients treated with elec- N.M. and Fredrickson, B.E. (2003) Association between
troconvulsive therapy: a proton magnetic resonance spec- dorsolateral prefrontal N-acetyl aspartate and depression in
troscopic imaging study. Arch. Gen. Psychiatry, 57: 937–943. chronic back pain: an in vivo proton magnetic resonance
Ende, G., Demirakca, T., Walter, S., Wokrina, T. and Henn, spectroscopy study. J. Neural. Transm., 110: 287–312.
F.A. (in press). Subcortical and medial temporal MR-detect- Gruber, S., Frey, R., Mlynarik, V., Stadlbauer, A., Heiden, A.,
able metabolite abnormalities in unipolar major depression. Kasper, S., Kemp, G.J. and Moser, E. (2003) Quantification
Eur. Arch. Psychiatry Clin. Neurosci. of metabolic differences in the frontal brain of depressive
Erickson, K., Drevets, W.C., Clark, L., Cannon, D.M., Bain, patients and controls obtained by 1H-MRS at 3 Tesla. Invest.
E.E., Zarate Jr., C.A., Charney, D.S. and Sahakian, B.J. Radiol., 38: 403–408.
(2005) Mood-congruent bias in affective go/no-go perform- Gur, R.C., Erwin, R.J., Gur, R.E., Zwil, A.S., Heimberg, C. and
ance of unmedicated patients with major depressive disorder. Kraemer, H.C. (1992) Facial emotion discrimination: II.
Am. J. Psychiatry, 162: 2171–2173. Behavioral findings in depression. Psychiatry Res., 42: 241–251.
Farchione, T.R., Moore, G.J. and Rosenberg, D.R. (2002) Haaga, D.A. and Beck, A.T. (1995) Perspectives on depressive
Proton magnetic resonance spectroscopic imaging in pe- realism: implications for cognitive theory of depression.
diatric major depression. Biol. Psychiatry, 52: 86–92. Behav. Res. Ther., 33: 41–48.
498

Hamann, S. and Mao, H. (2002) Positive and negative Lacerda, A.L., Keshavan, M.S., Hardan, A.Y., Yorbik, O.,
emotional verbal stimuli elicit activity in the left amygdala. Brambilla, P., Sassi, R.B., Nicoletti, M., Mallinger, A.G.,
Neuroreport, 13: 15–19. Frank, E., Kupfer, D.J. and Soares, J.C. (2004) Anatomic
Hamidi, M., Drevets, W.C. and Price, J.L. (2004) Glial reduc- evaluation of the orbitofrontal cortex in major depressive
tion in amygdala in major depressive disorder is due to disorder. Biol. Psychiatry, 55: 353–358.
oligodendrocytes. Biol. Psychiatry, 55: 563–569. Lacerda, A.L., Nicoletti, M.A., Brambilla, P., Sassi, R.B.,
Harwood, A.J. (2005) Lithium and bipolar mood disorder: the Mallinger, A.G., Frank, E., Kupfer, D.J., Keshavan, M.S.
inositol-depletion hypothesis revisited. Mol. Psychiatry, 10: and Soares, J.C. (2003) Anatomical MRI study of basal gan-
117–126. glia in major depressive disorder. Psychiatry Res., 124:
Hastings, R.S., Parsey, R.V., Oquendo, M.A., Arango, V. and 129–140.
Mann, J.J. (2004) Volumetric analysis of the prefrontal Lampe, I.K., Hulshoff Pol, H.E., Janssen, J., Schnack, H.G.,
cortex, amygdala, and hippocampus in major depression. Kahn, R.S. and Heeren, T.J. (2003) Association of depres-
Neuropsychopharmacology, 29: 952–959. sion duration with reduction of global cerebral gray matter
Herbert, C., Kissler, J., Junghöfer, M., Peyk, P., Wildgruber, volume in female patients with recurrent major depressive
D., Ethofer, T. and Grodd, W. (2005). Sexy, successful, disorder. Am. J. Psychiatry, 160: 2052–2054.
dynamic: left amygdala activation during reading of pleasant Lavretsky, H., Roybal, D.J., Ballmaier, M., Toga, A.W. and
adjectives. HBM 2005 237, Toronto, CA. Kumar, A. (2005) Antidepressant exposure may protect
Hetherington, H.P., Spencer, D.D., Vaughan, J.T. and Pan, against decrement in frontal gray matter volumes in geriatric
J.W. (2001) Quantitative 31P spectroscopic imaging of human depression. J. Clin. Psychiatry, 66: 964–967.
brain at 4 Tesla: assessment of gray and white matter differ- Lekwauwa, R.E., McQuoid, D.R. and Steffens, D.C. (2005)
ences of phosphocreatine and ATP. Magn. Reson. Med., 45: Hippocampal volume as a predictor of short-term ECT
46–52. outcomes in older patients with depression. Am. J. Geriatr.
Hickie, I., Naismith, S., Ward, P.B., Turner, K., Scott, E., Psychiatry, 13: 910–913.
Mitchell, P., Wilhelm, K. and Parker, G. (2005) Reduced Leppanen, J.M., Milders, M., Bell, J.S., Terriere, E. and
hippocampal volumes and memory loss in patients with Hietanen, J.K. (2004) Depression biases the recognition of
early- and late-onset depression. Br. J. Psychiatry, 186: emotionally neutral faces. Psychiatry Res., 128: 123–133.
197–202. Liotti, M., Mayberg, H.S., Brannan, S.K., McGinnis, S.,
Ho, A.P., Gillin, J.C., Buchsbaum, M.S., Wu, J.C., Abel, L. Jerabek, P. and Fox, P.T. (2000) Differential limbic–cortical
and Bunney Jr., W.E. (1996) Brain glucose metabolism correlates of sadness and anxiety in healthy subjects: impli-
during non-rapid eye movement sleep in major depression. A cations for affective disorders. Biol. Psychiatry, 48: 30–42.
positron emission tomography study. Arch. Gen. Psychiatry, Liotti, M., Mayberg, H.S., McGinnis, S., Brannan, S.L. and
53: 645–652. Jerabek, P. (2002) Unmasking disease-specific cerebral blood
Hsieh, M.H., McQuoid, D.R., Levy, R.M., Payne, M.E., flow abnormalities: mood challenge in patients with remitted
MacFall, J.R. and Steffens, D.C. (2002) Hippocampal vol- unipolar depression. Am. J. Psychiatry, 159: 1830–1840.
ume and antidepressant response in geriatric depression. Int. Lloyd, G.G. and Lishman, W.A. (1975) Effect of depression on
J. Geriatr. Psychiatry, 17: 519–525. the speed of recall of pleasant and unpleasant experiences.
Janssen, J., Hulshoff Pol, H.E., Lampe, I.K., Schnack, H.G., de Psychol. Med., 5: 173–180.
Leeuw, F.E., Kahn, R.S. and Heeren, T.J. (2004) Hippo- MacMillan, S., Szeszko, P.R., Moore, G.J., Madden, R.,
campal changes and white matter lesions in early-onset Lorch, E., Ivey, J., Banerjee, S.P. and Rosenberg, D.R.
depression. Biol. Psychiatry, 56: 825–831. (2003) Increased amygdala: hippocampal volume ratios as-
Ke, Y., Cohen, B.M., Lowen, S., Hirashima, F., Nassar, L. and sociated with severity of anxiety in pediatric major depres-
Renshaw, P.F. (2002) Biexponential transverse relaxation sion. J. Child. Adolesc. Psychopharmacol., 13: 65–73.
(T2) of the proton MRS creatine resonance in human brain. Maddock, R.J., Garrett, A.S. and Buonocore, M.H. (2003)
Magn. Reson. Med., 47: 232–238. Posterior cingulate cortex activation by emotional words:
Kempermann, G. and Kronenberg, G. (2003) Depressed new fMRI evidence from a valence decision task. Hum. Brain
neurons–adult hippocampal neurogenesis and a cellular Mapp., 18: 30–41.
plasticity hypothesis of major depression. Biol. Psychiatry, Mathews, A., Ridgeway, V. and Williamson, D.A. (1996)
54: 499–503. Evidence for attention to threatening stimuli in depression.
Kumar, A., Thomas, A., Lavretsky, H., Yue, K., Huda, A., Behav. Res. Ther., 34: 695–705.
Curran, J., Venkatraman, T., Estanol, L., Mintz, J., Mega, Mayberg, H.S., Brannan, S.K., Tekell, J.L., Silva, J.A., Mah-
M. and Toga, A. (2002) Frontal white matter biochemical urin, R.K., McGinnis, S. and Jerabek, P.A. (2000) Regional
abnormalities in late-life major depression detected with pro- metabolic effects of fluoxetine in major depression: serial
ton magnetic resonance spectroscopy. Am. J. Psychiatry, 159: changes and relationship to clinical response. Biol. Psychia-
630–636. try, 48: 830–843.
Kusumakar, V., MacMaster, F.P., Gates, L., Sparkes, S.J. and Mayberg, H.S., Liotti, M., Brannan, S.K., McGinnis, S., Mah-
Khan, S.C. (2001) Left medial temporal cytosolic choline in urin, R.K., Jerabek, P.A., Silva, J.A., Tekell, J.L., Martin,
early onset depression. Can. J. Psychiatry, 46: 959–964. C.C., Lancaster, J.L. and Fox, P.T. (1999) Reciprocal limbic-
499

cortical function and negative mood: converging PET find- Omura, K., Todd Constable, R. and Canli, T. (2005) Amygdala
ings in depression and normal sadness. Am. J. Psychiatry, gray matter concentration is associated with extraversion and
156: 675–682. neuroticism. Neuroreport, 16: 1905–1908.
McEwen, B.S. (2000) The neurobiology of stress: from seren- Pardo, J.V., Pardo, P.J. and Raichle, M.E. (1993) Neural cor-
dipity to clinical relevance. Brain Res., 886: 172–189. relates of self-induced dysphoria. Am. J. Psychiatry, 150:
Michael, N., Erfurth, A., Ohrmann, P., Arolt, V., Heindel, W. 713–719.
and Pfleiderer, B. (2003a) Metabolic changes within the left Persad, S.M. and Polivy, J. (1993) Differences between de-
dorsolateral prefrontal cortex occurring with electroconvul- pressed and nondepressed individuals in the recognition of
sive therapy in patients with treatment resistant unipolar and response to facial emotional cues. J. Abnorm. Psychol.,
depression. Psychol. Med., 33: 1277–1284. 102: 358–368.
Michael, N., Erfurth, A., Ohrmann, P., Arolt, V., Heindel, W. Pfleiderer, B., Michael, N., Erfurth, A., Ohrmann, P.,
and Pfleiderer, B. (2003b) Neurotrophic effects of electro- Hohmann, U., Wolgast, M., Fiebich, M., Arolt, V. and
convulsive therapy: a proton magnetic resonance study of the Heindel, W. (2003) Effective electroconvulsive therapy
left amygdalar region in patients with treatment-resistant de- reverses glutamate/glutamine deficit in the left anterior
pression. Neuropsychopharmacology, 28: 720–725. cingulum of unipolar depressed patients. Psychiatry Res.,
Morris, J.S., Frith, C.D., Perrett, D.I., Rowland, D., Young, 122: 185–192.
A.W., Calder, A.J. and Dolan, R.J. (1996) A differential Posener, J.A., Wang, L., Price, J.L., Gado, M.H., Province,
neural response in the human amygdala to fearful and happy M.A., Miller, M.I., Babb, C.M. and Csernansky, J.G. (2003)
facial expressions. Nature, 383: 812–815. High-dimensional mapping of the hippocampus in depres-
Murphy, F.C., Sahakian, B.J., Rubinsztein, J.S., Michael, A., sion. Am. J. Psychiatry, 160: 83–89.
Rogers, R.D., Robbins, T.W. and Paykel, E.S. (1999) Emo- Renshaw, P.F., Lafer, B., Babb, S.M., Fava, M., Stoll, A.L.,
tional bias and inhibitory control processes in mania and Christensen, J.D., Moore, C.M., Yurgelun-Todd, D.A.,
depression. Psychol. Med., 29: 1307–1321. Bonello, C.M., Pillay, S.S., Rothschild, A.J., Nierenberg,
Naismith, S., Hickie, I., Ward, P.B., Turner, K., Scott, E., Little, A.A., Rosenbaum, J.F. and Cohen, B.M. (1997) Basal gan-
C., Mitchell, P., Wilhelm, K. and Parker, G. (2002) Caudate glia choline levels in depression and response to fluoxetine
nucleus volumes and genetic determinants of homocysteine treatment: an in vivo proton magnetic resonance spectros-
metabolism in the prediction of psychomotor speed in older copy study. Biol. Psychiatry, 41: 837–843.
persons with depression. Am. J. Psychiatry, 159: 2096–2098. Robinson, R.G. and Lipsey, J.R. (1985) Cerebral localization
Nandrino, J.L., Dodin, V., Martin, P. and Henniaux, M. (2004) of emotion based on clinical-neuropathological correlations:
Emotional information processing in first and recurrent ma- methodological issues. Psychiatry Dev., 3: 335–347.
jor depressive episodes. J. Psychiatry. Res., 38: 475–484. Rosenberg, D.R., Macmaster, F.P., Mirza, Y., Smith, J.M.,
Neumeister, A., Wood, S., Bonne, O., Nugent, A.C., Luckenb- Easter, P.C., Banerjee, S.P., Bhandari, R., Boyd, C., Lynch,
augh, D.A., Young, T., Bain, E.E., Charney, D.S. and M., Rose, M., Ivey, J., Villafuerte, R.A., Moore, G.J. and
Drevets, W.C. (2005) Reduced hippocampal volume in Renshaw, P. (2005) Reduced anterior cingulate glutamate in
unmedicated, remitted patients with major depression versus pediatric major depression: a magnetic resonance spectros-
control subjects. Biol. Psychiatry, 57: 935–937. copy study. Biol. Psychiatry, 58: 700–704.
Nolan, C.L., Moore, G.J., Madden, R., Farchione, T., Bartoi, Ross, B. and Michaelis, T. (1994) Clinical applications of mag-
M., Lorch, E., Stewart, C.M. and Rosenberg, D.R. (2002) netic resonance spectroscopy. Magn. Reson. Q, 10: 191–247.
Prefrontal cortical volume in childhood-onset major depres- Rosso, I.M., Cintron, C.M., Steingard, R.J., Renshaw, P.F.,
sion: preliminary findings. Arch. Gen. Psychiatry, 59: 173–179. Young, A.D. and Yurgelun-Todd, D.A. (2005) Amygdala
Nolen-Hoeksema, S. (2000) The role of rumination in depres- and hippocampus volumes in pediatric major depression.
sive disorders and mixed anxiety/depressive symptoms. J. Biol. Psychiatry, 57: 21–26.
Abnorm. Psychol., 109: 504–511. Rothman, D.L., Petroff, O.A.C., Behar, K.L. and Mattson,
Noworolski, S.M., Nelson, S.J., Henry, R.G., Day, M.R., R.H. (1993) Localized 1H NMR measurements of g-amino-
Wald, L.L., Star-Lack, J. and Vigneron, D.B. (1999) High butyric acid in human brain in vivo. Proc. Natl. Acad. Sci.
spatial resolution 1H-MRSI and segmented MRI of cortical USA, 90: 5662–5666.
gray matter and subcortical white matter in three regions of Sanacora, G., Gueorguieva, R., Epperson, C.N., Wu, Y.T.,
the human brain. Magn. Reson. Med., 41: 21–29. Appel, M., Rothman, D.L., Krystal, J.H. and Mason, G.F.
O’Brien, J.T., Ames, D., Schweitzer, I., Desmond, P., Coleman, (2004) Subtype-specific alterations of gamma-aminobutyric
P. and Tress, B. (1997) Clinical, magnetic resonance imaging acid and glutamate in patients with major depression. Arch.
and endocrinological differences between delusional and non- Gen. Psychiatry, 61: 705–713.
delusional depression in the elderly. Int. J. Geriatr. Psychi- Schuff, N., Amend, D., Ezekiel, F., Steinman, S.K., Tanabe, J.,
atry, 12: 211–218. Norman, D., Jagust, W., Kramer, J.H., Mastrianni, J.A.,
O’Brien, J.T., Lloyd, A., McKeith, I., Gholkar, A. and Ferrier, Fein, G. and Weiner, M.W. (1997) Changes of hippocampal
N. (2004) A longitudinal study of hippocampal volume, co- N-acetyl aspartate and volume in Alzheimer’s disease. A
rtisol levels, and cognition in older depressed subjects. Am. J. proton MR spectroscopic imaging and MRI study. Neurol-
Psychiatry, 161: 2081–2090. ogy, 49: 1513–1521.
500

Segal, Z.V., Williams, J.M., Teasdale, J.D. and Gemar, M. Taki, Y., Kinomura, S., Awata, S., Inoue, K., Sato, K., Ito, H.,
(1996) A cognitive science perspective on kindling and epi- Goto, R., Uchida, S., Tsuji, I., Arai, H., Kawashima, R. and
sode sensitization in recurrent affective disorder. Psychol. Fukuda, H. (2005) Male elderly subthreshold depression
Med., 26: 371–380. patients have smaller volume of medial part of prefrontal
Sheline, Y.I., Barch, D.M., Donnelly, J.M., Ollinger, J.M., cortex and precentral gyrus compared with age-matched
Snyder, A.Z. and Mintun, M.A. (2001) Increased amygdala normal subjects: a voxel-based morphometry. J. Affect.
response to masked emotional faces in depressed subjects Disord., 88: 313–320.
resolves with antidepressant treatment: an fMRI study. Biol. Taylor, W.D., MacFall, J.R., Provenzale, J.M., Payne, M.E.,
Psychiatry, 50: 651–658. McQuoid, D.R., Steffens, D.C. and Krishnan, K.R. (2003a)
Sheline, Y.I., Gado, M.H. and Kraemer, H.C. (2003) Untreated Serial MR imaging of volumes of hyperintense white matter
depression and hippocampal volume loss. Am. J. Psychiatry, lesions in elderly patients: correlation with vascular risk fac-
160: 1516–1518. tors. AJR Am. J. Roentgenol., 181: 571–576.
Shestyuk, A.Y., Deldin, P.J., Brand, J.E. and Deveney, C.M. Taylor, W.D., Steffens, D.C., MacFall, J.R., McQuoid, D.R.,
(2005) Reduced sustained brain activity during processing of Payne, M.E., Provenzale, J.M. and Krishnan, K.R. (2003b)
positive emotional stimuli in major depression. Biol. Psychi- White matter hyperintensity progression and late-life depres-
atry, 57: 1089–1096. sion outcomes. Arch. Gen. Psychiatry, 60: 1090–1096.
Shulman, R.G., Rothman, D.L., Behar, K.L. and Hyder, F. Taylor, W.D., Steffens, D.C., McQuoid, D.R., Payne, M.E.,
(2004) Energetic basis of brain activity: implications for Lee, S.H., Lai, T.J. and Krishnan, K.R. (2003c) Smaller
neuroimaging. Trends Neurosci., 27: 489–495. orbital frontal cortex volumes associated with functional
Siegle, G.J., Steinhauer, S.R., Thase, M.E., Stenger, V.A. and disability in depressed elders. Biol. Psychiatry, 53: 144–149.
Carter, C.S. (2002) Can’t shake that feeling: event-related Taylor, W.D., Steffens, D.C., Payne, M.E., MacFall, J.R.,
fMRI assessment of sustained amygdala activity in response Marchuk, D.A., Svenson, I.K. and Krishnan, K.R. (2005)
to emotional information in depressed individuals. Biol. Influence of serotonin transporter promoter region poly-
Psychiatry, 51: 693–707. morphisms on hippocampal volumes in late-life depression.
Sonawalla, S.B., Renshaw, P.F., Moore, C.M., Alpert, J.E., Arch. Gen. Psychiatry, 62: 537–544.
Nierenberg, A.A., Rosenbaum, J.F. and Fava, M. (1999) Teasdale, J.D. and Russell, M.L. (1983) Differential effects of
Compounds containing cytosolic choline in the basal ganglia: induced mood on the recall of positive, negative and neutral
a potential biological marker of true drug response to words. Br. J. Clin. Psychol., 22(Pt 3): 163–171.
fluoxetine. Am. J. Psychiatry, 156: 1638–1640. Thomas, K.M., Drevets, W.C., Dahl, R.E., Ryan, N.D., Birma-
Steffens, D.C., Helms, M.J., Krishnan, K.R. and Burke, G.L. her, B., Eccard, C.H., Axelson, D., Whalen, P.J. and Casey,
(1999) Cerebrovascular disease and depression symptoms in B.J. (2001) Amygdala response to fearful faces in anxious and
the cardiovascular health study. Stroke, 30: 2159–2166. depressed children. Arch. Gen. Psychiatry, 58: 1057–1063.
Steffens, D.C., McQuoid, D.R., Welsh-Bohmer, K.A. and Vakili, K., Pillay, S.S., Lafer, B., Fava, M., Renshaw, P.F.,
Krishnan, K.R. (2003) Left orbital frontal cortex volume and Bonello-Cintron, C.M. and Yurgelun-Todd, D.A. (2000)
performance on the benton visual retention test in older Hippocampal volume in primary unipolar major depression:
depressives and controls. Neuropsychopharmacology, 28: a magnetic resonance imaging study. Biol. Psychiatry, 47:
2179–2183. 1087–1090.
Steingard, R.J., Renshaw, P.F., Hennen, J., Lenox, M., Videbech, P., Ravnkilde, B., Pedersen, T.H., Hartvig, H.,
Cintron, C.B., Young, A.D., Connor, D.F., Au, T.H. and Egander, A., Clemmensen, K., Rasmussen, N.A., Andersen,
Yurgelun-Todd, D.A. (2002) Smaller frontal lobe white mat- F., Gjedde, A. and Rosenberg, R. (2002) The Danish PET/
ter volumes in depressed adolescents. Biol. Psychiatry, 52: depression project: clinical symptoms and cerebral blood
413–417. flow. A regions-of-interest analysis. Acta Psychiatr. Scand.,
Steingard, R.J., Yurgelun-Todd, D.A., Hennen, J., Moore, 106: 35–44.
J.C., Moore, C.M., Vakili, K., Young, A.D., Katic, A., Vollmert, C., Tost, H., Brassen, S., Jatzko, A. and Braus, D.F.
Beardslee, W.R. and Renshaw, P.F. (2000) Increased orbito- (2004) Depression und moderne Bildgebung. Fortschr.
frontal cortex levels of choline in depressed adolescents as Neurol. Psychiatr., 72: 435–445.
detected by in vivo proton magnetic resonance spectroscopy. Vythilingam, M., Charles, H.C., Tupler, L.A., Blitchington, T.,
Biol. Psychiatry, 48: 1053–1061. Kelly, L. and Krishnan, K.R. (2003) Focal and lateralized
Surguladze, S., Brammer, M.J., Keedwell, P., Giampietro, V., subcortical abnormalities in unipolar major depressive dis-
Young, A.W., Travis, M.J., Williams, S.C. and Phillips, M.L. order: an automated multivoxel proton magnetic resonance
(2005) A differential pattern of neural response toward sad spectroscopy study. Biol. Psychiatry, 54: 744–750.
versus happy facial expressions in major depressive disorder. Vythilingam, M., Heim, C., Newport, J., Miller, A.H., Ander-
Biol. Psychiatry, 57: 201–209. son, E., Bronen, R., Brummer, M., Staib, L., Vermetten, E.,
Surguladze, S.A., Young, A.W., Senior, C., Brebion, G., Tra- Charney, D.S., Nemeroff, C.B. and Bremner, J.D. (2002)
vis, M.J. and Phillips, M.L. (2004) Recognition accuracy and Childhood trauma associated with smaller hippocampal
response bias to happy and sad facial expressions in patients volume in women with major depression. Am. J. Psychiatry,
with major depression. Neuropsychology, 18: 212–218. 159: 2072–2080.
501

Vythilingam, M., Vermetten, E., Anderson, G.M., Watkins, P.C., Vache, K., Verney, S.P., Muller, S. and Ma-
Luckenbaugh, D., Anderson, E.R., Snow, J., Staib, thews, A. (1996) Unconscious mood-congruent memory bias
L.H., Charney, D.S. and Bremner, J.D. (2004) Hippocam- in depression. J. Abnorm. Psychol., 105: 34–41.
pal volume, memory, and cortisol status in major de- Weisbrod, M., Trage, J., Hill, H., Sattler, H.-D., Maier, S.,
pressive disorder: effects of treatment. Biol. Psychiatry, 56: Kiefer, M., Grothe, J. and Spitzer, M. (1999) Emotional
101–112. priming in depressive patients. Ger. J. Psychiatry, 2: 19–47.
Wang, Z.J. and Zimmerman, R.A. (1998) Proton MR spec- Willner, P. (1984) Cognitive functioning in depression: a review
troscopy of pediatric brain metabolic disorders. Neuroimag- of theory and research. Psychol. Med., 14: 807–823.
ing Clin. N Am., 8: 781–807. Yang, T.T., Menon, V., Eliez, S., Blasey, C., White, C.D., Reid,
Watkins, P.C., Martin, C.K. and Stern, L.D. (2000) Uncon- A.J., Gotlib, I.H. and Reiss, A.L. (2002) Amygdalar activa-
scious memory bias in depression: perceptual and conceptual tion associated with positive and negative facial expressions.
processes. J. Abnorm. Psychol., 109: 282–289. Neuroreport, 13: 1737–1741.
Subject Index

Acoustic cue hypothesis 272, 280 Attitude 307, 309


Acoustic lateralization hypothesis 251, 253, 258 Autism 405–408, 410
Addiction 46 Autobiographic recall 430
Affective Norms for English Words (ANEW) 5, 148, Autonomic response 56, 57, 61, 275, 370, 372, 374, 426,
180, 186, 187, 189, 196, 210 450–451
Affective priming see Priming Autoshaping 11
Affective space 148, 187–189, 226 Average reference 123, 124, 127–130, 153, 155, 172–174
Aggression 235, 409, 434–435, 447 Awareness 105–118, 369, 422
Amphetamine 12
Amplitude envelope 238 Backward masking see Masking
Amygdala 8–17, 20, 22–23, 107, 147, 159–161, 166–168, Basal ganglia 261, 262, 289, 446, 448–451, 494 see also
208, 222, 255, 261, 263, 264, 444, 458–463, 468, 481, Putamen, Striatum
483, 486 Basic emotion 411 see also Facial expression
Crossmodal processing 355, 356 Bed nucleus of stria terminalis (BNST) 9, 11, 13
Emotion regulation 433 Binocular rivalry 107
Gaze 365, 369 Biphasic typology of emotions 5, 32
Social judgements 367 Blindsight 112, 369
Visuospatial attention 369 Bottom-up effects 271–272
Anger 5, 205, 240, 254, 296, 304, 322, 331, 365, 413, Brain stem 21, 186, 481
446–448 Broca’s area 258, 277
Anterior cingulate cortex (ACC) see Cingulate cortex Broca’s homologue 250
Antisocial personality disorder 457, 459 Brunswik’s ‘‘lens model’’ 237
Aphasia 286
Appetitive system see Appetitive and defensive system Categorization 53–55, 57, 58, 62, 192
Appetitive and defensive system 3–24, 67–68 Chimeric face 468–470
Aprosodia 286 Cholecystokinin (CCK) 11
Area 7b 373 Choline 489, 491
Area F5 380, 383, 426 Cingulate cortex 468
Area PF 380, 383 Anterior cingulate cortex (ACC) 11, 38, 80, 99, 160,
Arousal 4–8, 10, 14, 16, 17, 19, 20, 23, 32, 34, 35, 37, 56, 209, 386, 410, 413, 427, 432, 436, 461, 447–450, 482,
57, 60, 62, 71, 85, 94, 96, 98, 134, 137, 148–150, 487, 494
154–156, 158, 163, 165–168, 186–193, 195, 196, 198, Classical conditioning see Conditioning
201, 210, 211, 212, 214, 218–221, 223–228, 236, Competition model of visual attention 40
239–242, 260, 272, 275, 287, 296, 363, 367, 370, 389, Computer vision 333–337
411, 426, 433, 449, 458, 459, 468, 482 Conditioning 5, 9–10, 13–14, 16–17, 19, 22, 81, 159, 219,
Articulator 287 222–224, 228, 372, 450, 458
Asperger Syndrome 407 Conjunction analysis 349
Attention 5, 8, 9, 10, 22, 32, 37, 40, 43, 53, 70, 80, 85, Connotation 147, 161
107, 157, 165, 188, 191, 208, 214, 217, 235, 244, Connectivity
330–331, 346, 363, 372, 405, 426, 444, 461, 462, 468, Connectivity architecture 260
483 Functional connectivity 260, 369, 460
spatial attention 67–87 Intrinsic connectivity 260, 262
motivated attention 31–48 Consciousness 107
Attentional blink 118, 160, 219, 219, 227, 228 Contrastive stress patterns 279

503
504

Corrugator muscle 6, 425, 433–434 Emotional connotation 205–215


Cortical mapping 125 Emotional contagion 404, 420, 422, 425
Corticosteroid 10 Emotional experience 373
Corticotrophin releasing hormone (CRH) 11 Emotional expression, facial 32, 473–476
Counterempathy 426 Emotional Stroop 188, 219
Creatine 489, 491, 493 Emotion-specific intonation contours 242
Criterion threshold 115–116 Empathy quotient 409, 413
Crossmodal processing 350, 355–357 Enthorinal cortex 449
Bias 346 E-Representation 407
Binding 346 Event categorization 55
Congruent/incongruent information 356 Event-related potential (ERP) 6, 17, 18, 31–48, 53, 54,
Convergence zones 346 62, 68, 72, 124, 147, 152, 185, 221, 242, 329, 346, 446,
Hebbian learning 392 458, 468
Integration 263, 264 Event-related field (ERF) 131–139
Integration rules 347 Explicit judgment 261
Interaction 261 Explicit processing 158
Response depression 346 emotional prosody 253
Supra-additivity 353 Extinction 14, 228, 446, 460
Current source density (CSD) 55, 125, 126, 129 Extrastriate cortex 67, 77, 80, 83, 87, 331, 357,
368, 468
DC potentials 271
Deep brain stimulation 447 F0 see Fundamental frequency
Defense cascade 22–23, 57–58 Face inversion effect 322
Defensive system see Appetitive and Defensive System Face processing, models of 327–329, 331–333
Depression, symptoms, cognitive theory 482 see also Face recognition 321, 323, 324, 326, 328, 329, 331,
Major depression, Unipolar depression 333–337, 470
Desensitization 94 Component-configural 323
Dichotic listening 244, 245 Composite effect 323
Differential specialization hypothesis 270 Double dissociation of expression recognition and 331
Disgust 207, 254, 255, 262, 292, 296–300, 302, 328, 336, Dynamic information 327
365, 367, 373, 380, 386–388, 390, 392, 394, 396, 404, Expertise hypothesis of 330
411–413, 427, 443–446, 451, 472, 483 Fuzzy logical model of perception (FLMP) 326
Disruptive behaviour disorder 434 Holistic 323
Dissociation paradigm 106 N170 332
Dopamine 9, 11, 12, 16, 21, 446, 447, 481 Scrambling and blurring, effect of 324
Dopaminergic system 449, 476 Facial action coding system (FACS) 326, 337
Dot-probe paradigm 68–78, 81 Facial expression 67–87, 482
Duchenne smile 326 Prototypical 334
Duration cues 278 Temporal cues 327
Duration discrimination 251 Double dissociation of recognition 331
Dynamic causal modeling 260 Facial imitation 392, 425
Facial mimicry see Facial imitation
Early posterior negativity (EPN) 33–35, 38, 40, 42, 44, Fairness 435
46–48, 125 Fear 4, 7, 9, 11, 18, 20, 57, 67, 78, 81, 84, 159, 160, 205,
Effective sensor coverage 132 208, 224, 235, 240, 255, 328, 336, 347, 364, 365, 387,
Efferent leakage 92 388, 404, 411, 421, 443, 444–445, 458–459, 472, 483
Electroencephalography (EEG) 7–10, 17, 22, 33, 44, 71, Fear bradycardia 5–24
75, 77, 78, 85, 123–128, 130–133, 136–138, 147, 153, Freezing 3–23
156, 161–163, 186, 189, 208, 221, 223, 227, 228, 239, Fear conditioning see Conditioning
242, 244, 269, 271, 347, 461, 468 Frequency
Electromyography (EMG), facial 425 Domain analyses 222, 223
Emotion regulation 423, 433 Patterns 279
505

Properties 289 Helping behaviour 434


Frequency-related information 278 High-lexical (HLI) interjections 299, 303
Frontal cortex 209–215, 433 Hippocampus 8, 12, 38, 99, 137, 365, 366, 371, 481, 486,
Dorsolateral prefrontal cortex (DLPFC) 38, 44, 54, 489, 492, 494–495
349, 352, 412, 436, 459, 482, 485, 488, 493 Huntington’s disease 446, 451
Inferior frontal cortex (IFC) 249, 250, 254, 256, Hypothalamus 9, 11, 208, 388, 411, 481
258–260, 263, 264, 287, 463
Inferior frontal gyrus (IFG) 97, 100, 258, 264, 383, Ideomotor theory 421
411, 413, 427, 463, 485 Imagery, 94, 173,177, 189, 196, 220
Medial prefrontal cortex (MPFC) 54, 87, 96, 99, 396, Implicit processing
410–411, 413, 428–430, 433–435, 444, 447–448, 483, Emotional prosody 253, 263
492 Incongruent prosody 302, 303
Orbitofrontal cortex (OFC) 13, 15, 67, 84, 99, 134, Inner speech 277
258, 259, 264, 285, 289, 290, 346, 350–354, 411, Insula 99, 251, 253, 255, 261, 279, 291, 331, 373, 386,
446–449, 457–461, 468, 482, 485, 488 451, 459, 483
Ventromedial prefrontal cortex (VMPFC) 44, 84 Crossmodal processing 346, 354
Ventrolateral prefrontal cortex (VLPFC) 81 Disgust 373, 386, 411, 413
Frontal eye field 81 Pain 386, 427, 432
Frontal lobe 208, 210, 211, 214, 222, 260, 462, 468, 475, Intention 205, 208, 209, 213, 214, 249, 250, 306, 314,
494 367, 410, 412, 413, 419, 421, 423, 426, 429, 449
Frontoparietal network of attention 44 Interaction analysis of fMRI data
Functional lateralization hypothesis 256, 258 Inspection of time series 352
Fundamental frequency (F0) 236, 270 Negative interaction 352
contours 240, 241, 242, 256 Unspecific deactivation 352
direction 280 Interjections 297, 298, 300, 302, 303
level 241, 242 International Affective Digitized Sounds (IADS) 5, 220,
range 241, 270, 272, 273 224
Fusiform face area (FFA) 263, 330, 368 International Affective Picture System (IAPS) 5, 33
Crossmodal processing 357 Interpersonal stance 307, 309
Expertise hypothesis 330 Intonation 238
Fusiform gyrus 40, 76, 160, 263, 264, 329–331, 368, 427 Inverse problem 131

GABA 11, 481, 482, 489, 491, 495 Late Positive Complex (LPC) 154–155, 157, 159,
Gabor patch 75 163–165, 170, 173–175, 177–178
Gain control mechanism 77, 78, 80, 87 Late Positive Potential (LPP) 17, 35–37, 48, 53, 57, 59,
Gambling task 460 61–62, 190–193, 197, 199
Gamma band activity 223 Lateralization 250–256, 258, 468
Gastric ulceration 10 Left hemispatial bias (LHB) 469
Glutamate 481, 489, 491, 495 Left hemisphere 130, 153–154, 163, 188–189, 208–209,
Glutamine 489, 491, 495 214, 228, 270, 277–279, 282, 467
Go/no-go task 55, 72 Lexical access 164, 167, 193, 194, 217, 225, 227–229
Lexical decision 154, 164, 165, 193–195, 218, 219, 225
Habituation 38–40, 53, 60–62 Lexicon 161, 167, 206, 209, 218, 297
Haemodynamic refractoriness 355, 358 Limbic system 11, 287, 331, 481, 483, 489
Happiness 189, 205, 240–242, 255, 262, 269, 270, Linguistic accentuation 256
296–298, 300, 302, 322, 326, 327, 347, 351, 355, 257, Linguistic prosody 242, 256, 258–259, 279, 287–288, 309
365, 367, 390, 397, 404, 412, 413, 425, 443, 445, Linked mastoid reference 127–130
449–450, 459, 472 Locus coeruleus 9, 44
Hebbian cell assemblies 16 Loudness contours 298
Learning 380, 392, 394 Low fear model 458
Facial imitation 392 Low-lexical (LLI) interjections 299, 303
Language 392 Low-resolution tomography (LORETA) 126
506

Magnetic resonance spectroscopy (MRS), proton P300 55, 74, 153, 157–158, 161–163, 165, 189, 193, 199,
489 221, 226, 458, 461
Magnetencephalography (MEG) 76, 85, 123, 124, Pain 386, 427
130–133, 136–139, 329, 383 Paracingulate cortex 429, 436
Major depression 482, 489 Parahippocampal gyrus 96, 448, 449, 468, 483, 485,
Mania 457, 469, 482 488
Masking 57, 106–107, 369 Parietal cortex 18, 34, 67, 70, 75, 80, 85, 256, 332, 349,
McGurk effect 345, 356 352, 353, 373, 430, 483, 485
Memory task 459, 462 Inferior parietal cortex (IPC) 85, 352, 353, 430, 483,
Mentalizing see Theory of Mind 485
Mesolimbic dopamine system 11, 12 Perceptual awareness 106
Microstate 77–78 Perceptual threshold 157–158
Mindreading see Theory of Mind Perspective taking 420, 422–423
Mindreading system 404–406 Phobia 7, 46, 57
Mirror neurons 381, 410, 426 Phonation 287
Circuit 381 Phonemic identification 242, 244
Congruent selectivity 384 Phosphocreatine 489, 491, 493
Overgeneralization 395 Phrase 185–202
Somatotopy 384 Phrase understanding 475
Tuning of 381, 385 Pitch 298
Mock politeness 317 Contours 239, 258, 273, 279, 280
Mood congruent processing bias 482 Direction 280, 282
Morphometry 485, 486, 488, 489 Features 239
Motivation 3–24 Patterns 256
Motivated attention see attention Range 256, 258, 272, 279
Motivational priming hypothesis 19 Variability 280, 281
Motivational-structural rules 235, 236 Positive slow wave 37
Motivational systems 3–24, 186 Posner orienting task 82
Motor imagery 94 Postencounter stage 4–24
Motor imitation see Facial imitation Posttraumatic stress disorder (PTSD) 99
Motor lateralization 470 Preencounter stage 4–24
Motor learning 12 Prefrontal cortex see Frontal cortex
Motor vocabulary 381, 385 Premotor cortex 150, 207, 214, 379–380, 382–384, 390,
M-representation 407 426, 427
Multimodal processing see Crossmodal processing Priming 19, 42–43, 75–76, 157, 159, 164, 167, 176, 193,
Myo-inositol 490, 493 195, 198–199, 201, 219 195,
Proportional global signal scaling (PGSS) 124,
N100 75, 152–157, 188, 222, 347 133–136
N170 76, 332 Propositional labels 297
N-acetylaspartate 491 Prosocial behaviour 423
Natural scenes 34, 53–63 Prosopagnosia 329
Neglect 71 Pseudo-utterances 308, 311, 314
Neurotransmitter 11, 16, 481, 489, 491 Psychic blindness 364
N-methyl-D-aspantam (NMDA) 10, 13 Psychopathy 158, 403, 408, 457–463
Norepinephrine 9, 11, 16, 44, 481 Pull effects 236, 237, 238
Novelty, of stimuli 38, 61 Pulvinar 331, 367, 369, 373, 374
Push effects 236, 237, 238
Occipital face area (OFA) 329, 339 Putamen 12, 251, 446, 463, 483, 485, 488,
Occipital cortex 18, 74, 82, 83, 85–87, 134, 330, 492–494
485
Orbitofrontal cortex see Frontal cortex Question intonation 280
Oscillatory brain activity 217 Questionnaire 431
507

Rapid categorization 217 Steady state visual evoked potential (ssVEP) 223–227
Rapid serial visual presentation (RSVP) 33, 46, 53, 59, Striatum 11–13, 15, 21, 222, 291, 411, 446–447, 449, 485,
153–157, 160, 172, 174, 188 494
Recognition potential 54, 160 Stroop effect 188, 202
Respiration 287 Stroop task 219, 462
Response bias 114–115 Subliminal 107–111, 115, 118–120, 147, 152, 154, 159,
Retrosplenial cortex 96, 98, 209, 213, 214 160, 188, 369, 425
Reward 3, 4, 11, 115, 117, 411, 412, 449, 460, 494 Subthalamic nucleus (STN) 447
Rhythm 238 Superior colliculus 369
Right hemisphere 270, 282, 304–306, 309, 313, 315, 468 crossmodal processing 346
Right hemisphere hypothesis 286, 287 response depression 353
Rythm 238 supra-additivity 353
Supramarginal gyrus (SMG) 277, 372, 373
SI 386 Suprasegmental processing 279, 298
SII 373, 386 Suprasegmental features 251, 258
Sadness 4, 205, 240–242, 254, 255, 267, 270, 275, 287, Suprasegmental sequences 260
296–300, 302, 304, 327, 365, 367, 397, 404, 413, 443, Surprise 5, 108, 118, 277, 327, 328, 336, 365, 367, 388,
444, 445, 451, 459, 483, 484 411, 443, 444, 446, 448, 449, 451
Schizophrenia 467–476 Syllabic stress 308
Schizophrenic patients, acute, chronic 472 Syllable durations 298
Script imagery 95, 97–99 Sympathy 404
Segmental information 279, 298
Segmental structure 251 Taming effect 11
Self-assessment manikin (SAM) 94, 149, 187, 260, 347, Temporal cues 278, 289, 327
359 Temporal cortex 8, 73, 251, 253, 258
Self-other distinction 391, 412 Inferior temporal cortex 17, 46, 54, 94
Self-report measures see Questionnaire Mid temporal area (MT) 331
Semantic categorization 192–193, 227 Middle temporal gyrus (MTG) 265, 350, 368, 379,
Semantic differential 147, 150, 153, 158, 166, 170, 381, 383–385, 390, 411, 485
172–174, 187–188, 202 Superior temporal gyrus (STG) 331, 410–413, 430
Semantics 150, 152, 164–165, 167, 205–206, 260, Superior temporal sulcus (STS) 244–245, 253–254,
295–296, 298 256, 260–261, 290–292, 346, 368, 379, 410
Semantic network 147, 150, 167, 217, 228 Temporal lobe 494
Semantic monitoring 209–211 Temporal lobectomy 444
Semantic Priming see Priming Temporal pole 22, 428, 430, 450
Sensor standardization 124, 130–133 Temporo-parietal junction 81, 410, 411, 428, 430
Sensory threshold 113–115 Thalamus 8, 9, 20, 99, 137, 159, 256, 258, 331, 346, 350,
Sentence focus 256, 279 372, 463 see also Pulvinar
Serotonin 9, 11, 481, 489 Thatcher illusion 322
Sex differences 408–410 Theory of mind (ToM) 395, 404, 421
Simulation theory 372, 395, 421 Thyroid-releasing hormone (TRH) 11
Skin conductance response (SCR) 6, 7, 35, 38, 55, 56, Time frequency analysis 223
370, 426, 436 Tone languages 237, 258
Spatial attention see attention Touch 385–386
Spatial sampling 125–126 Top-down effects see bottom-up effects
Speaker attitude 309–310 Transient sadness 483
Speaker confidence 311–313, 317, 318 Two-stage model of stimulus perception 47–48, 54
Speaker politeness 314, 316
Spectral energy distribution 238 Ultrarapid visual categorization 54
Startle reflex 10–11, 17–23, 46, 56, 61, 155, 157, 182, 450, Unattended processing see implicit processing
459, 463 Unintelligible emotional speech 291
Statement intonations 256, 279 Unipolar depression 469
508

Valence 4–8, 19, 20, 32, 34–37, 42–44, 55, 60, 71, 76, 94, Visual search task 70, 218
96, 97, 148, 150, 154, 18, 163, 165–168, 186–190, Voice quality 238, 298
193–196, 198–201, 210, 213, 218–220, 226, 253, 254, Vowel duration 272, 273
256, 260, 262, 270, 281, 285, 286, 288, 289, 291, 347, Voxel-based morphometry (VBM) see Morphometry
348, 411, 431, 449, 458, 459, 463, 467, 468, 471, 472,
482, 483 Water maze task 12
Valence hypothesis 254, 285, 286, 288, 467 Wernicke’s homologue 250, 253
Valence theory 472 White matter lesions 485
Vasopressin 11 Withdrawal see Appetitive and Defensive System
Ventral striatum see Striatum Word generation task 209–211
Verbal expression 296, 475 Word processing 127, 147–175, 195
Violence inhibition mechanism (VIM) model 458, 459
Virtual reality 434 Zygomatic muscle 6, 425, 431

Vous aimerez peut-être aussi