ChapterPDF Available

Emotion perception: Putting the face in context

Authors:

Abstract

Two studies tested the hypothesis that in judging people's emotions from their facial expressions, Japanese, more than Westerners, incorporate information from the social context. In Study 1, participants viewed cartoons depicting a happy, sad, angry, or neutral person surrounded by other people expressing the same emotion as the central person or a different one. The surrounding people's emotions influenced Japanese but not Westerners' perceptions of the central person. These differences reflect differences in attention, as indicated by eye-tracking data (Study 2): Japanese looked at the surrounding people more than did Westerners. Previous findings on East–West differences in contextual sensitivity generalize to social contexts, suggesting that Westerners see emotions as individual feelings, whereas Japanese see them as inseparable from the feelings of the group.
Page 1 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
The Oxford Handbook of Cognitive Psychology
Daniel Reisberg
Print publication date: Jun 2013
Print ISBN-13: 9780195376746
Published to Oxford Handbooks Online: Jun-13
Subject: Psychology, Cognitive Psychology
DOI: 10.1093/oxfordhb/9780195376746.001.0001
Emotion Perception: Putting the Face in Context
Maria Gendron, Batja Mesquita, Lisa Feldman Barrett
DOI: 10.1093/oxfordhb/9780195376746.013.0034
Abstract and Keywords
Emotion perception research is dominated by the assumption that emotions
are written on the face as particular arrangements of facial actions. This
face-focused paradigm assumes that posed, static configurations of facial
muscle movements provide sufficient cues to emotion, such that the study of
context is secondary. As a result, it has been assumed that the mechanisms
for perceiving posed, static facial actions reveal the mechanisms that
support emotion perception outside of the laboratory setting. In this chapter,
we review this face-focused paradigm in emotion perception and contrast it
with experimental findings that place the face in context. Furthermore, we
question whether a posed face is ever without context, even in a lab. We
discuss how features of an experiment can serve (however unintentionally)
as a context that influences performance in emotion perception tasks. Based
on the literature reviewed, we conclude that the study of emotion perception
is necessarily contextualized, such that context not only influences emotion
perception, but it might be intrinsic to seeing an emotion in the first place.
emotion, perception, facial expression, context
An iconic fear face looks startled—eyes wide, mouth agape, and eyebrows
raised. An iconic anger face is scowling—brows furrowed, eyes glowering,
and jaw tightened. An iconic sad face is frowning—lips pouting, brows pulled
together. These posed, exaggerated faces dominate a research paradigm
in the psychology of emotion perception. Simply exposing a perceiver to
one of these faces is assumed to produce an automatic “recognition” that
Page 2 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
reflects some statistical regularity in the faces that populate the emotion
category, and this automatic perception is taken as evidence that there is
an evolutionarily preserved capability to “decode” facial actions for their
psychological meaning. Several notable experiments (Ekman, 1972, 1973;
Ekman & Freisen, 1971; Ekman, Sorensen, & Friesen, 1969; Izard, 1971)
within this face-focused approach have captured the attention of textbook
writers and have become part of the standard curriculum in psychology.
These experiments have also captured the imagination of other academic
disciplines. The face-focused approach to emotion perception (where
emotional information is written on the face) guides cognitive neuroscience
studies of emotion perception (e.g., Breiter, Etcoff, Whalen, Kennedy, Rauch,
Buckner, et al., 1996; Phillips et al., 1997; Sprengelmeyer, Rauch, Eysel, &
Przuntek, 1998; for a discussion, see Watson, 2004). Research investigating
social and emotional deficits in people with mental disorders, brain lesions,
and neurodegenerative disease almost exclusively relies on the face-focused
approach (e.g., Kohler, Walker, Martin, Healey, & Moberg, 2010). The face-
focused approach also informs government spending on security training.
For example, over 3,000 TSA employees have received security training to
read” emotion in the face, in the hopes of keeping citizens safe (Weinberger,
2010). Even the popular media is enamored with this paradigm, as evidenced
by television shows such as “Lie to Me” and popular science writing such as
Malcolm Gladwell’s The Naked Face (2002).
The face-focused approach to emotion perception has several key features.
First, this approach views the face as transmitting evolved signals that
broadcast the internal state of a target person for perceivers to see (Keltner,
Ekman, Gonzaga, & Beer, 2003). Data to support this view come in several
varieties: (1) descriptions of facial behaviors that occur in emotion across
cultures (e.g., Scherer & Wallbott, 1994); (2) cross-cultural production of
facial behaviors during the experience of emotion (e.g., “negative” facial
behaviors in Japanese and American students, Ekman, 1973; for review,
see Matsumoto, Keltner, Shiota, Frank, & O’Sullivan, 2008, pp. 215–216);
or (3) expressions produced by congenitally blind individuals as compared
to sighted individuals (e.g., Matsumoto & Willingham, 2009). In this regard,
“emotion perception” is redefined as “emotion recognition.”
A second aspect of the face-focused approach is that emotion perception
is an innate psychological process (e.g., Schyns, Petro, & Smith, 2009).
Data used to support this point come from (1) studies on cross-cultural
recognition of emotion (for reviews, see Ekman, 1998; Izard, 1977; but see
Russell, 1994), (2) work on the “efficiency” or automaticity of perception
Page 3 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
(e.g., Schyns et al., 2009; Tracy & Robins, 2008), (3) work on the categorical
perception of emotion (e.g., Etcoff & Magee, 1992), and (4) emotion
perception in infants (e.g., Hoehl & Striano, 2008; for review, see Hoehl &
Striano, 2010).
A third aspect of this view is that context (i.e., information aside from the
face) is not necessary for successful emotion perception. For example,
seeing emotion in another person is assumed to proceed “automatically,
silently, and without the benefit of language” (Izard, 1994, p. 289). In
some views, emotion perception is even resistant to contextual influences
(e.g., Nakamura, Buck, & Kenny, 1990). Emotion perception is assumed to
proceed in a feedforward fashion, with more complex “cognitive” processing
coming online relatively late and secondary to the perceptual processing of
emotional faces (for a discussion, see Vuilleumier & Pourtois, 2007).
Because of these assumptions, the face-focused approach to studying
emotion perception typically presents
1
perceivers with faces that are
posing exaggerated facial actions that are devoid of a bodily, vocal, or
situational context. Yet in everyday life, faces move in concert with bodies
and vocalizations in a way that is typically calibrated to the situation at hand.
Even proponents have noted the oddity of the face-focused experimental
paradigm (e.g., Ekman & O’Sullivan, 1988). Unfortunately, much of the
literature has yet to take these ecological concerns to heart—researchers
routinely present posed, static faces like those from Ekman and Freisen’s
(1976) set to examine how accurately perceivers can categorize these
disembodied, posed faces.
In this chapter, we discuss recent empirical challenges to a face-focused
approach to emotion perception. We focus on research that places the face
into context, demonstrating that context dramatically shapes how faces
portraying emotion are processed. We also explore the idea that the typical
laboratory paradigm serves as its own form of context, one that is often
unintended (and unrecognized) by the experimenter. We close the chapter
by discussing the implications of context-based modulation for the future
of emotion perception research. We suggest that context does more than
modulate emotion perception; instead, it might serve as a necessary source
of information in emotion perception.
Putting the Face in Context
Early investigations of context effects in emotion perception were motivated
by relatively low “accuracy” in experiments that relied exclusively on the
Page 4 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
face (Hunt, 1941; Landis 1934). Accuracy was (and still is) defined as seeing
(1) what the target intended to portray (in posed facial actions), (2) what
the stimulus developer intended (in directed facial actions often based on
predetermined configurations that have been described in the literature,
e.g., EMFACS; Ekman & Friesen, 1982), (3) the emotion associated with
the stimulus “condition” in which a target’s facial actions occurred (for
spontaneous facial actions), or (4) the self-reported state of the target
individual (again for spontaneous facial actions). Citing a number of
experiments in which perceivers failed to agree on the emotion conveyed
by faces, Hunt and Landis concluded that the face does not provide reliable
information about emotion for the perceiver. This conclusion led researchers
to explore what information, in addition to the face, is needed for perceptions
of discrete emotions like anger, sadness, and fear. These experiments,
published mainly between the 1920s and 1930s, examined whether contexts
such as bodies, hands, vocalizations, and situational knowledge contributed
to emotion perception over and above the facial actions that they were
paired with. Although this research was criticized on methodological grounds
(Ekman, Freisen, & Ellsworth, 1972), more recent, well-designed research
has emerged to again raise the question of whether context is an important
determinant in emotion perception. As we will see in the coming pages, this
more modern research makes it clear that a face does not speak for itself, at
least where emotion
2
is concerned.
Stimulus-Based Context
One of the most obvious sources of context in studies of emotion perception
is the physical array in which a face is embedded, or what we refer to as
“stimulus-based” context. Some stimulus-based context derives from the
target person who produced the facial actions, such as bodily posture or
movement, vocalizations, and even the target’s sweat. Other forms of
context characterize the broader “event” in which a face appears. Studies
investigating stimulus-based forms of context have examined both influence
of congruent (e.g., a startled face on a fearful body) and incongruent
contexts (e.g., a pouting face following an anger situation), demonstrating
both facilitation and interference effects.
Bodies
In 1935, Kline and Johannesen demonstrated that perceivers are more
accurate (i.e., to select the emotion category intended by the researcher)
when a face portraying emotion was presented in a congruent bodily context
Page 5 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
(compared to when the face was presented alone). These data indicated
that the body helps to increase interrater agreement about the feelings
of a target over information derived from the face alone. In the last few
years, there has been a spike of interest in the idea that body postures are a
salient cue that perceivers use to understand others’ emotional states (see
de Gelder, 2009). For example, when posed faces and bodies portray the
same emotion category, emotion perception in the face is facilitated; when
body and face are incongruent, emotion perception in the face is impaired
(Meeren, van Heijnsbergen, & de Gelder, 2005). Perception of a wide-eyed
face on a body adopting an angry posture was less accurate than when
that same face was placed on a body adopting a fearful posture. Strikingly,
event-related potential (ERP) recordings reflect the incongruence between
facial and bodily portrayals of emotion as soon as 115 ms after stimulus
onset. And, remarkably, bodies also provide information about emotion for
nonhuman perceivers: A computer classifier trained separately on faces and
bodies portraying emotion demonstrated higher classification accuracy for
the two cues combined than with either one alone (Gunes & Piccardi, 2007).
There are even times when bodies dominate over faces during emotion
perception. A scowl (i.e., a portrayal of anger) is more likely to be perceived
as disgusted when it is paired with a body posture involving a soiled object
(Aviezer, Hassin, Ryan, et al., 2008, Study 1) and a face portraying disgust
can even by perceived as displaying a positive emotion (i.e., pride) when
paired with a muscled body whose arms are raised in triumph (Aviezer,
Hassin, Ryan, et al., 2008, Study 2). Not all studies have found that bodies
provide incremental information over faces during emotion perception,
however (e.g., Rozin, Taylor, Ross, Bennett, & Hejnadi, 2005).
Hand gestures, too, can influence emotion perception in a face (Hietanen &
Leppänen, 2008). Participants were more accurate to judge scowling faces as
angry when paired with congruent hand gestures (a neutral sentence signed
in Finnish Sign Language in an angry manner) and were least accurate to
judge anger when scowls were paired with happy hand gestures (that same
sentence signed in a happy manner). Perception of neutral faces was also
biased toward “anger” when paired with angry hand gestures.
Vocalizations
The voice constitutes a second type of stimulus-based context for perception
of emotion in facial actions. The voice alone is an important conduit of
affective tone, most effectively conveying arousal (Bachorowski, 1999;
Pittam, Gallois, & Callan, 1990) and valence to a lesser extent (Bachorowski
Page 6 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
& Owren, 2008; Pereira, 2000). Importantly, vocal depictions shape emotion
perception based on facial actions (e.g., de Gelder & Vroomen, 2000;
Massaro & Egan, 1996; for a review, see de Gelder et al., 2006).
3
For
example, perceivers judge ambiguous facial actions (i.e., a morph between
a smiling “happy” face and a frowning “sad” face) in line with the affective
value conveyed by vocalizations (i.e., prosody conveying sadness or
happiness) even when they are told to judge the face alone (de Gelder &
Vroomen, 2000). The effect of prosody on emotion perception in a face is
most robust when the face is ambiguous (i.e., does not portray a single
emotion in a caricatured manner; de Gelder & Vroomen, 2000), neutral
(Massaro & Egan, 1996), or visually noisy (Collignon et al., 2008), although
prosody also primes judgments of faces portraying congruent emotional
content (Pell, 2005). Additional work is needed to clarify whether the effects
of vocal prosody reflect affect congruence or a more specific discrete
emotion congruence effect. Future work could use stimuli that are controlled
for valence and arousal (e.g., fear and disgust in Phillips et al., 1998) to
evaluate whether congruence or interference effects occur when vocal
portrayals matched in affect, but not emotion, accompany facial actions.
Neuroimaging studies suggest that facial and vocal information are
integrated in a variety of brain regions involved in perception. For example,
facial and vocal portrayals of fear and disgust separately produced an
increase in activation within the superior temporal gyrus, inferior posterior
temporal gyrus, middle temporal gyrus, and medial frontal cortex (Phillips
et al., 1998). Congruent combinations of negative facial and vocal cues
produced an increase in activation in the left amygdala and right fusiform
and several other structures in the distributed face-processing network
(Dolan, Morris, & de Gelder, 2001). However, positive facial and vocal
congruency produced increases in a distinct network of regions (i.e., left
superior parietal lobule, the left medial parietal cortex, the left superior
frontal gyrus, and the right anterior cingulate cortex). In another study,
brain activity in some of these same regions—the medial prefrontal cortex
(MPFC) and left superior temporal sulcus (STS) was linked to the category
of emotion being perceived, regardless of the modality in which the cues
occured (Peelen, Atkinson, & Vuilleumier, 2010). The amygdala, on the other
hand, appears to be particularly important to the processing of combined
facial and vocal cues. When startled faces portraying fear were accompanied
by a sentence spoken in a fearful tone, perceivers judged the face as
more negative than when that same face was presented alone, and these
judgments were correlated with amygdala activation, which the authors
interpret as evidence that the amygdala computes affective meaning based
Page 7 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
on multiple cues (Ethofer et al., 2007). It is also possible, however, that
the faces were unfamiliar and ambiguous (i.e., not clearly indicative of
one discrete emotion) to perceivers (Somerville & Whalen, 2006) and the
vocal cues made their psychological meaning even more uncertain. Similar
effects have been demonstrated in nonhuman primates, such that the same
neurons in the central nucleus of the amygdala fired to the presentation of
vocalizations and facial actions representing aggressive threat, scream, or
coo (Kuraoka & Nakamura, 2007), which are thought to be homologous with
expressions in humans (Parr & Waller, 2006).
Some work has investigated the time course of vocalization context effects,
with discrepant results. Research using magnetoencephalography (MEG)
demonstrates that the integration of voices and faces portraying emotion
occurs within 250 ms after stimulus onset in the superior temporal sulcus
(Hagan et al., 2009).
4
Yet other data suggest that the time frame of prosody
effects is much slower than 250 ms. In an implicit priming task where
prosodic cues preceded faces, only vocalizations longer than 600 ms
produced facilitation in judgments of a congruent emotional faces (Pell,
2005). It is unclear what this time constraint means, but one possibility is
that despite early integration in the STS, behavioral facilitation will only
occur with stimuli of sufficiently long duration, perhaps because conceptual
knowledge based on vocalizations determines whether the emotion
judgment will be affected. Alternatively, a longer presentation duration
might only be important for instances when vocalizations and faces are not
presented simultaneously. This would suggest that distinct mechanisms
might be responsible for context effects based on simultaneous versus
sequential presentation of cues.
Chemosignals
Even a target person’s sweat can influence perception of emotion from
the face (Zhou & Chen, 2009). For example, researchers collected sweat
from target individuals while they viewed videos designed to induce fear
or happiness. A second group of individuals then sniffed this sweat as they
were completing an emotion perception task. Perceivers demonstrated
better accuracy for the perception of startled “fear” faces when sniffing the
fearful” sweat compared to those sniffing the “happy” sweat. Interestingly,
perceivers could not distinguish between the two types of sweat via self-
report, suggesting this was not an explicit conceptual priming effect. This
study did not investigate whether this facilitation might be driven by the
quantity of sweat perceivers were exposed to (sweat was not measured,
Page 8 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
only collected on a gauze pad under the arms of participants). Furthermore,
it is possible that sweat influences the perception of affect, rather than the
perception of discrete emotion categories (given that happiness and fear
differ in their basic affective quality).
Situations
In the 1910s and 1920s, Lev Kuleshov performed filmmaking “experiments”
that famously demonstrated how situational context influences the emotion
perceived in another person’s facial actions. He spliced the same neutral
expression with different scenes that were emotional in nature, leading
audiences to believe that the expression on the actor’s face was changing.
With no “signal” present in the face, perceivers still saw emotion when the
context called for it. In fact, audiences raved about the superb acting in
the film. More recently, the “Kuleshov effect” has proved to be somewhat
more complex. When neutral faces were paired with common causes of
emotion, such as missing a train, participants made judgments that reflected
either the situation or the face, in equivalent amounts (Carrera-Levillain &
Fernández-Dols, 1994). When the situation was more extreme, such as brake
failure on a mountain pass, the “neutral” facial behaviors tended to dominate
the judgment—perceivers overwhelmingly saw that face as neutral.
Other experiments investigating the influence of situational effects on
emotion perception are as old as the first studies on the Kuleshov effect. For
example, two different sets of perceivers judged spontaneous facial actions
in infants that were either isolated from the context or paired with the
eliciting situation (Sherman, 1927). Perceivers generated up to 25 different
emotion labels for a given set of spontaneous facial actions in infants when
they were presented in isolation. Agreement was higher for the perceivers
who judged the facial actions in their eliciting circumstances (e.g., without
context 15% of perceivers saw “fear,” with context 65% did) (see also Munn,
1940). These experiments have been critiqued based on their methods (see
Ekman et al., 1972), but more recent, better controlled, studies confirm that
perceivers agree more on the emotion conveyed by a face when the facial
actions are placed in a congruent context than when they are presented
alone (Carroll & Russell 1996; Knudsen & Muzekari, 1983). Apparently,
context can diminish the ambiguity (i.e., lack of interrater agreement about
the emotional meaning) of posed facial actions, even when the actions depict
highly caricatured expressions as in recent experiments.
Page 9 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
A number of findings also point to a situational modulation of judgments
when facial actions and the situation are “incongruent.” For example,
participants asked to make emotion judgments based on combinations
of incongruent situational descriptions and facial actions made 55.7% of
judgments consistent with the situation and 31.6% consistent with the
facial behaviors (Goodenough & Tinker, 1931). These data suggest that the
situation may be a stronger cue when it is inconsistent with information
from a face. This general finding was replicated using bodily and situational
information conveyed by a visual scene (Munn, 1940), film clips (Goldberg,
1951), and video of the actual eliciting circumstances (Sherman, 1927).
More recently, this finding has received further support (Fernandez-Dols,
Sierra, & Ruiz-Belda, 1993; Fernandez-Dols, Wallbot & Sanchez, 1991, Study
2; Spignesi & Shor, 1981; Wallbott, 1988a, Study 2; Wallbott, 1988b). For
example, portrayals of anger are more likely to be perceived as fearful
when paired with the description of a dangerous situation (Carroll & Russell,
1996, Study 1). Yet other studies have failed to show that contextual
information can override information in the face in driving emotion
judgments (Fernandez-Dols et al., 1991, Study 1, 3; Nakamura et al., 1990;
Wallbott, 1988a, Study 1; for a review see, Fernandez-Dols & Carroll, 1997).
5
Situational information tends to dominate perception of emotion in faces
both when situations are common, everyday (Carrera-Levillain & Fernández-
Dols, 1994) and even when situations are more ambiguous (i.e., contain less
source clarity) than the exaggerated facial actions being perceived (Carroll
& Russell, 1996, Study 3). This integration of situational context and facial
actions also appears to develop over the lifespan, with children (ages 4–10)
using situational over facial information for all emotions portrayals except
surprise (Widen & Russell, 2010). Other data suggests that older children
(third grade) are most accurate when both types of cues are present, and
younger children (preschool) demonstrate advantages for stories (situational
information) over faces portraying emotion (Reichenbach & Masters, 1983).
The use of situational context might be enhanced by the epistemic goal
to perceive emotion as opposed to affect. For example, informationally
irrelevant context is automatically encoded when perceivers were asked to
judge the emotion in posed faces (i.e., “Is this person afraid or disgusted?”)
but not when they were asked to perceive affect (i.e., “Would you approach
or avoid this person?”), suggesting that the structural configuration of faces
carries information about the affective state of the target, but processing
a face in context is necessary for perceiving emotion (Barrett & Kensinger,
2010). (This is consistent with earlier work showing that perceivers use basic
affective dimensions like valence and arousal rather than discrete categories
Page 10 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
of emotion to understand posed facial behaviors in isolation; Russell &
Bullock, 1986; Schlosberg, 1952.) Interestingly, perceivers’ memory for the
situational context was just as strong as memory for the focal face itself
during emotion perception, suggesting that faces and contexts are processed
configurally during emotion perception. (For related results, see Frühholz,
Fehr, & Herrmann, 2009; Warner & Shields, 2009.)
The impact of situational context during emotion perception extends beyond
explicit judgments of emotion to the gaze pattern used for faces. Perceivers’
visual fixations on a face reflect the larger context (including the body) in
which that face is embedded (Aviezer, Hassin, Ryan, et al., 2008) suggesting
that the context changes which facial features are visually salient.
The neural responses that realize emotion perception are also shaped by
situational context. Neural activity associated with a surprised face preceded
by a negative sentence was greater in ventral regions of the amygdala, in
comparison to when a surprise face was preceded by a positive sentence
(Kim et al., 2004). Amygdala activity was also modulated by the congruence
of a situational context and facial actions in a recent demonstration of the
Kuleshov effect (Mobbs et al., 2006). A network of regions associated with
processing the “emotionality” of a face (i.e., ACC, left STS, right STS, right
amygdala, and the bilateral temporal pole) was more engaged when a face
was presented following an “emotional” video compared to a “neutral”
video. Incongruent context revealed specific activity in the ventromedial
prefrontal cortex, whereas the combination of a fearful face and negative
context revealed activity in the right amygdala, fusiform gyrus, and bilateral
temporal pole and insula. It is believed that this network of regions works
together to instantiate an emotion percept based on both facial actions and
the context that accompanies them. Similarly, visual regions such as bilateral
inferior occipital, bilateral fusiform, right inferior frontal gyrus, as well as
left amygdala were engaged when participants were asked to match an
emotional face to a context (Sommer, Döhnel, Meinhardt, & Hajak, 2008),
suggesting that these regions play a role in integrating situational context
and facial actions when the goal is to produce a coherent emotional percept.
Amygdala activation was related to the goal to perceive emotion based on
context and facial actions, because when these same stimuli were presented
but were incidental to a color-matching task, amygdala activity did not differ
from a control condition. Taken together, these findings suggest that when
the epistemic goal is to perceive emotion, neural activity in regions involved
in the perceptual processing of faces is shaped by the context in which faces
are embedded. Since context modulates the same regions that are typically
Page 11 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
associated with the formation of an emotional percept to begin with, it can
be argued that context is integral to the emotional percept itself, rather than
a postperceptual decision process.
Other Faces
Finally, other faces serve as another strong form of stimulus-based context
in emotion perception, even when those other faces are irrelevant to the
task at hand. For example, in an early study designed to examine “social
context,” perceivers were presented with a line drawing of two faces oriented
toward one another (Cline, 1956). The faces depicted either a low-arousal
negative emotion (“glum”), a high-arousal negative emotion (“frown”), or
high-arousal positive emotion (“smile”). Participants assigned a different
meaning to the face depending on the other face it was paired with. For
example, participants perceived smiles as bolder when paired with the low-
arousal “glum” expression than when paired with the high-arousal “frown
expression. In more recent experiments, perceivers judged faces as more
intense when they followed with several faces of opposite valence (e.g., a
frowning face following a smiling face) (Thayer, 1980). Similarly, perceived
arousal level of anger and surprised facial portrayals can also be shifted
based on the facial portrayal they follow (Russell & Fehr, 1987, Studies
5 & 6). Perception of a neutral face also shifted when it was presented
concurrently with (Russell & Fehr, 1987, Studies 1 & 2) or following another
face portraying emotion (Tanaka-Matsumi, Attivissimo, Nelson, & D’Urso,
1995). These effects occurred despite instructions for participants to judge
the target face independently of the other stimuli. The size of the context
effect produced by other faces appears to be culturally relative, however
(Masuda, Ellsworth, Mesquita, Leu, & Veerdonk, 2008). Japanese perceivers’
judgments of a target individual reflected the emotions conveyed by the
other figures in a scene, but this was less true for Americans perceivers, who
focused mainly on the target person.
The processing of surrounding nontarget faces is also evident in patterns
of brain activity (Amtig, Miller, Chow, & Mitchell, 2009). Posterior regions of
visual cortex are more engaged when all faces are consistent in affective
content, whereas the amygdala response is selectively lessened for positive
faces in the context of neutral or negative faces. Enhanced functional
connectivity emerged between the amygdala and medial prefrontal cortex
on trials where the target face was inconsistent in affective value with the
nontarget faces, and this functional connectivity predicted longer reaction
times. These data suggest that connectivity between the amygdala and
Page 12 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
medial prefrontal cortex reflects a process aimed at resolving competition
between conflicting affective content (c.f. Lieberman et al., 2007). It is
unclear based on these data whether this same connectivity would result
from within valence competition, however.
Perceiver-Based Context
Whereas stimulus-based forms of context are driven by something in
the external surroundings, perceiver-based forms of context are driven
by the internal state of the perceiver. In this section, we focus on how
emotion perception in the face is influenced by the perceiver’s immediate
psychological state. Other forms of context that might be categorized as
“perceiver based,” such as those based on personality, culture, ethnicity,
gender, or psychopathology, are beyond the scope of the present review.
Cognitive Load
Emotion perception depends on the amount of executive attention that
is available to a perceiver, and so attention becomes a relevant aspect
of the perceiver-based context. Although researchers often believe that
emotion perception is automatic (e.g., emotion recognition is above-chance
under conditions of cognitive load; Tracy & Robins, 2008), there are often
methodological circumstances that should constrain explanations for
observed effects. For example, in the Tracy and Robbins study, perceivers
were asked to judge the faces according to a single emotion word, which
might have sufficiently constrained their perceptual choices (see later
section on forced choice as an experimental context). Consistent with this
interpretation, a recent study by Phillips, Channon, Tunstall, Hendenstrom,
and Lyons (2008) found that placing participants under working memory load
affected emotion perception judgments far less when the task was to choose
between two labels (as opposed to four or six). Futhermore, perceivers under
cognitive load still show performance decrements compared to when they
are not placed under load (Phillips et al., 2008; Tomasik, Ruthruff, Allen, &
Lien, 2009; Tracy & Robbins, 2008). For example, perceivers under cognitive
load were less able to distinguish perceptually between scowling and smiling
faces (Tomasik et al., 2009). This effect of load appears to be somewhat
downstream in visual processing, however, given that early encoding of
the visual input (indexed by early ERP components) remains intact when
perceivers are placed under cognitive load at encoding (Holmes, Kragh
Neilsen, Tipper, & Green, 2009).
Page 13 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Affective State
Several studies indicate that a perceiver’s affective state shapes affect
perception. Perceivers induced into an unpleasant affective state, compared
to those feeling neutral, were more likely to see negative emotion in an
ambiguous schematic (line drawing) face (Bouhuys, Bloem, & Groothuis,
1995). In a second study, perceivers in an unpleasant affective state were
more likely to see “fear” in schematic faces that contained both positive
and negative features. In contrast, perceivers in another study (Leppanen
& Hietanen, 2003) who were induced to feel pleasant perceived happiness
in low-intensity smiling faces more quickly than they perceived disgust in
low-intensity disgust faces, although they had roughly equivalent reaction
times to judge happiness and disgust in those faces following induction of
unpleasant feelings. (Because this experiment did not include a neutral
condition, it is unclear whether these were facilitation and interference
effects.) The affective aspects of emotional experiences also shape affect
perception, such that participants induced to feel the discrete state of
disgust via audiotaped messages were more accurate at judging negatively
valenced emotional faces generally (Schiffenbauer, 1974). These effects
on emotion perception appear to be specific to momentary experiences of
affect, however; retrospective ratings of emotion, as measured by the profile
of mood states questionnaire, were unrelated to the perception of emotion
(Harris & Synder, 1992).
The affective state of the perceiver not only influences the emotional content
of what is perceived, but it also influences the dynamics of perception.
Perceivers induced to feel an unpleasant affective state (i.e., sadness)
perceived frowns, but not smiles, as lasting longer than they did, whereas
the reverse finding was found for those induced to feel a pleasant (i.e.,
happy) state (Niedenthal, Halberstadt, Margolin, & Innes-Ker, 2000). Affective
state also shapes how “believable” perceivers find portrayals of emotion.
The experience of positive compared to negative affect leads perceivers to
judge all facial portrayals (i.e., positive, negative, and neutral portrayals)
as more believable (Forgas & East, 2008). Perceived intensity of emotional
faces also appears to be influenced by the affective state of the perceiver.
Participants in a pleasant affective state judged faces portraying positive,
but not negative, emotion as more intense (Forgas & East, 2008). Taken
together, these data indicate that the affective state of the perceiver is
an important factor that contributes to emotion perception. Furthermore,
changes in affective state might be a mechanism by which other context-
based effects occur because odors, sounds, sights, and scenes all might
Page 14 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
shape ongoing perception by virtue of the fact that they have the potential to
perturb a perceiver’s affective state.
Embodiment
Embodiment refers to the various ways in which the body influences
cognition. Embodiment is reflected, for example, in the idea that the brain
systems used to represent the concept of an emotion (e.g., “anger”) are also
involved in seeing that emotion in another person and experiencing that
emotion oneself (Niedenthal, Barsalou, Winkielman, Krauth-Gruber, & Ric,
2005). One implication of this work is that categorizing a face as “anger
versus “fear” might rely on the perceiver’s ability to reenact the brain
state that controls the relevant facial muscle movements associated with
that emotion (cf. Niedenthal et al., 2005). Studies testing an embodiment
hypothesis have a long history. As early as 1918, it was established
that perceivers reported using facial mimicry as a strategy for emotion
perception (Langfield, 1918b). Perceivers spontaneously mimic facial
muscle movements in static portrayals of emotion (e.g., Dimberg, 1982,
1990) even when they are backwardly masked (Dimberg, Thunberg, &
Elmehed, 2000). More recently, experimental studies have shown that
facial mimicry is associated with perceptual accuracy (Wallbott, 1991), and
with the ability to judge a smile as authentic versus faked (for review, see
Niedenthal, Mermillod, Maringer, & Hess, 2010). Participants allowed to
mimic facial behaviors saw the onset and offset of facial expressions sooner
than participants who were unable to mimic because they were holding a
pen in their mouth (Niedenthal, Brauer, Halberstadt, & Innes-Ker, 2001; see
also Stel & van Knippenberg, 2008). Patients with “locked-in” syndrome,
who have complete paralysis of voluntary (but not spontaneous) facial
movements due to a brainstem lesion, have impaired accuracy for discrete
emotion perception of negative faces (i.e., anger, fear, sadness) (Pistoia et
al., 2010). Yet individuals with Mobius syndrome, who also have paralysis of
facial muscles that results from the underdevelopment of the cranial nerves,
do not show decrements in emotion perception (Rives Bogart & Matsumoto,
2010). These data suggest that central nervous system impairments disrupt
emotion perception, whereas more peripheral causes of muscle impairments
might not. Similarly, a central representation of how a set of facial muscle
movements feels may also be necessary to emotion perception; transcranial
magnetic stimulation (TMS) applied to somatosensory cortex leads to
decrements in emotion perception (Pitcher, Garrido, Walsh, & Duchaine,
2008; see also Adolphs, Damasio, Tranel, Cooper, & Damasio, 2000; Pourtois
et al., 2004). Embodied representations may have particular utility for
Page 15 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
understanding static faces, however, given that an increase in motor cortex
activity was not observed for perception of emotion in dynamic facial
movements (Kilts, Egan, Gideon, Ely, & Hoffman, 2003).
Experimental Context
Even with all the evidence that context routinely influences emotion
perception, there are still experiments where faces are presented in isolation
—no bodies or situational context—and emotion perception still proceeds.
Some of the most well cited studies to make this point date from the early
1970s. This work demonstrated that even individuals from preliterate
cultures have higher accuracy than would be expected by chance when
perceiving emotion in static posed faces (e.g., Ekman & Friesen, 1971;
Ekman et al., 1969, 1972; Izard, 1977). Yet these experiments, arguably
some of the most visible and salient in all of psychology, contain certain
contextual factors in the experimental setup that contributed to emotion
perception in an important way.
Posed Facial Actions
Most studies in the face-focused paradigm (including our own) use faces
that pose particular configurations of muscle movements to portray
emotion instead of spontaneously elicited facial actions. Standard “facial
expression” stimulus sets were created by either directing posers to move
particular facial muscles according to a predetermined configuration (e.g.,
as prescribed by EMFACS) or by merely asking them to pose their faces
to convey a particular emotion category. It is not clear how ecologically
valid these particular configurations are with respect to actual spontaneous
muscle movements during emotion experience (e.g., do people you
know pout when they are sad or scowl when they are angry?). Facial
electromyography, a direct measure of muscle activity, indicates that
people do not always pout in sadness or scowl in anger; in fact, the
electromyography data produce little differentiation beyond positive and
negative affect (for a review, see Cacioppo, Berntson, Larsen, Poehlmann,
& Ito, 2000). Photographs of posed facial actions thus introduce statistical
regularities into emotion perception experiments that are not necessarily
present in real life, and they have the potential to inflate emotion perception
accuracy in the lab compared to what it would be in everyday life (if such
perceptions were only based on a face). In fact, Landis (1924a, 1924b, 1929)
observed that judgments of spontaneous expressions were at chance levels.
More recent work is consistent with (although not as extreme as) these early
Page 16 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
findings (e.g., Yik, Meng, & Russell, 1998), even when the spontaneous facial
actions were identified as emotional by “expert” coders (Naab & Russell,
2007). One argument, of course, is that target people are using display
rules (see Ekman et al., 1969) to modify their spontaneous facial actions,
making it harder to recognize expressions when they occur. To address this
concern, it is instructive to compare emotion perception based on covertly
acquired spontaneous facial actions (acquired in a social context that would
not encourage the use of display rules) with actions that are posed. Accuracy
rates are much lower for covert spontaneous facial actions, although they
are still above chance levels (Hall, Gunnery, & Andrzejewski, 2011; Wagner,
MacDonald, & Manstead, 1986; Zuckerman, Hall, DeFrank, & Rosenthal,
1976). Participants’ accuracy to judge emotion drops even further when the
target’s facial behaviors occurred during conversation (Zuckerman et al.,
1976; also see Motley & Camden, 1988).
In contrast to recognition advantages seen for posed facial actions,
perceivers appear more willing to attribute emotional experience based on
spontaneous expressions. Spontaneous facial actions are judged to convey
emotions more clearly and reflect true feelings more than are posed facial
actions (McLellan, Johnston, Dalrymple-Alford, & Porter, 2010). McLellan and
colleagues also demonstrated that spontaneous facial actions serve as more
effective affective primes than do posed actions. Even though posed facial
actions do well in studies where the question is about the emotion displayed
on the face, spontaneous facial actions appear to do well when the question
is about the experience of the target individual (c.f. Aviezer, Hassin, Bentin,
& Trope, 2008). The ability to explicitly distinguish between posed and
spontaneous expressions (i.e., smiles) appears to be more developed in older
adults, however (Murphy, Lehrfeld, & Isaacowitz, 2010; for review of aging
and context in emotion perception, see Isaacowitz & Stanley, 2011). Future
work should address whether judgments of “true feelings” and priming
effects based on spontaneous stimuli are also enhanced with age.
Static Faces
The face-focused approach to studying emotion perception relies on face
stimuli that are not only posed but also faces that are unmoving or static.
In these stimuli, facial muscle movements are frozen in time with no
temporal dynamics. The static stimulus is often the “apex” of the supposed
expression—the most intense and caricatured instance of facial actions.
Thus, all “sad” faces are pouting; all “angry” faces are frowning; all “happy”
faces are smiling, and so on. When combined with other method choices
Page 17 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
(e.g., forced choice among a small set of words, discussed in the next
subsection), perceivers are able to see the intended emotion with high
accuracy (between 60% and 80%; Keltner et al., 2003, p. 415), and many
scientists have interpreted these findings as evidence that information in the
face is sufficient for emotion perception (e.g., Matsumoto et al., 2008). But
it is important to ask whether the high accuracy rates are, in part, a function
of using static faces. Like posed material, static faces might create statistical
regularities that do not normally exist in the facial actions that are produced
in the real world. Dynamic facial actions contain facial movements that have
their own psychological meaning. Furthermore, a static display extends the
length of an expression’s “apex” beyond its typical duration in dynamic face,
although exposure durations that are meant to mimic brief facial muscle
movements (called “micro expressions”; Ekman & Friesen, 1969) may not do
so.
Unfortunately, it is difficult to assess the consequences of presenting static
compared to dynamic facial actions because there are few studies that
directly compare the two within the same experiment. One study, although
limited in its methods, is suggestive. Perceptions of spontaneous smiles
were compared based on whether the smiles were dynamic or static.
Perceivers had higher accuracy for perceiving happiness in static smiles
than dynamic smiles (Miles & Johnston, 2007). In contrast, when using
artificially constructed (posed) faces, perceivers are more accurate in judging
the emotions conveyed by dynamic facial actions than those depicted in
static faces. This dynamic advantage holds true for judgments based on
video footage of posed facial muscle movements (Ambadar, Schooler, &
Cohn, 2005; Cunningham & Wallraven, 2009, Study 1), for synthetically
created faces (Wehrle, Kaiser, Schmidt, & Scherer, 2000), and even for
point light displays of motion created by facial muscle movements (Bassili,
1979). Developmental data, on the other hand, do not converge with
these findings. Specifically, in preschoolers (3–5 years of age) there is no
documented advantage for the perception of emotion from dynamic versus
static expressions (Nelson & Russell, 2011), suggesting that use of dynamic
cues may require learning that occurs over the lifespan.
In any case, the temporal dynamics of facial actions carry important
information. When the speed of onset is manipulated either by changing
the number of frames in a dynamic display (Kamachi et al., 2001) or the
speed at which the individual frames are presented (Sato & Yoshikawa,
2004), some emotions are more easily perceived with slower onsets (e.g.,
sadness), whereas some (e.g., surprise) are more readily perceived with
Page 18 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
faster onsets. Importantly, this does not appear to be due to the duration
of the “apex” (null findings for manipulating duration of static displays;
Kamachi et al., 2001), but rather the dynamics of how the facial actions
onset.
Emotion Words
Most experiments using a face-focused approach to emotion perception ask
participants to match face stimuli to emotion words that are included as part
of the experimental protocol. These words provide an important but hidden
source of context that shapes emotion perception and increases accuracy
(Barrett, Lindquist, &Gendron, 2007). A number of studies, dating back to the
early years of psychology, provide evidence for the important influence of
words in creating the high accuracy rates we are used to seeing in studies
of emotion perception. Perceivers asked to judge emotion, but not provided
with a list of emotion terms, had very low agreement on the meaning of a set
of facial actions (Kanner, 1931). Providing perceivers with emotion words (vs.
asking perceivers to free label) increases accuracy anywhere from 16% (Kline
& Johannsen, 1935) to 26% (Izard, 1971), on average. Related, providing
any label, even if it is not the correct one, can produce a false consenus in
emotion perception. For example, participants presented with caricatured
illustrations of emotional faces accompanied by incongruent labels tended to
accept the labels as descriptors of the face more often if they previously had
been unable to identify the emotion depicted on the face (Langfield, 1918a).
More recent studies confirm this finding (e.g., Russell, 1993; Widen, Christy,
Hewett, & Russell, 2010; Widen & Russell, 2008). Similarly, when participants
were presented with a list of choices that contained a number of foils,
they had generally low accuracy (Buzby, 1924). One interpretation is that
even posed, static, highly caricatured portrayals of emotion are somewhat
ambiguous as to their psychological meaning, and emotion words can narrow
the range of responses. Thus, any time a study reports high accuracy rates
for emotion perception (in any cue), it is important to consider whether the
experiment explicitly included emotion words that could influence perceivers’
performance (cf. Russell, 1994).
A number of experiments have now demonstrated language effects in
emotion perception when words are not directly introduced as choices in
the experimental paradigm. Emotion words produce biases in perceptual
memory for a face, such that memory is shifted toward words that were
presented at encoding (Halberstadt & Niedenthal, 2001). Furthermore, a
completely false perceptual memory (i.e., remembering a smile) can be
Page 19 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
created based on a context that primes a specific semantic category (e.g.,
wining a sporting event) (Fernández-Dols, Carrera, Barchard, & Gacitua,
2008). Words also support categorical perception of facial expressions
(Fugate, Gouzoules, & Barrett, 2010). Perceivers learned chimpanzee
expressions (e.g., a hoot) either with an arbitrary label or without. Only those
perceivers who learned the expressions with labels showed the hallmark of
categorical perception—an advantage at discriminating morphs that crossed
the categorical boundary between two expressions.
Emotion perception can be impaired by reducing the accessibility of
emotion words, even when such words are not necessary to perform the
experimental task. When emotion word meaning is made less accessible by
semantic satiation
6
(for a review, see Black, 2003), accuracy on a perceptual
matching task drops to just above chance levels (even though emotion
words are not necessary to say whether two faces match in their emotional
content; Lindquist, Barrett, Bliss-Moreau, & Russell, 2006). Furthermore,
when participants are placed under verbal load categorical perception for
posed, caricatured faces is wiped out (such that perceptual advantages for
distinguishing between stimuli from two different emotion categories is no
longer observed) (Roberson, Damjanovic, & Pilling, 2007). Given these data,
it is difficult to conclude that the structural features of expressions drive
categorical perception all by themselves.
Words also provide an advantage to young children’s emotion perception.
Specifically, young children are more accurate at matching emotional faces
in a sorting task when the box they are sorting into is marked by a word,
compared to when it is marked with a perceptually similar face (Russell
& Widen, 2002).
7
Furthermore, emotion perception accuracy increases in
parallel with children’s vocabulary for emotion words (for a review of this
work, see Widen & Russell, 2008). For example, in Widen and Russell’s work,
children acquire the term “disgust” relatively late, at a mean age of 56
months (4.6 years) and only around this time do children distinguish between
high arousal fear and disgust portrayals.
Finally, the neural representation of an emotional face is also shaped by
emotion words. Providing perceivers with emotion words significantly
reduced amygdala response to posed faces depicting emotion (Lieberman
et al., 2007). This result is likely due to the reduced uncertainty produced
by helping to resolve competing “perceptual hypotheses” that arose from a
structural analysis of the face alone. Furthermore, when perceivers judged
a structurally neutral face as emotional, this engaged a network of regions
Page 20 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
(e.g., right superior temporal sulcus, bilateral orbitofrontal cortex, right
anterior insula) (Thieschler & Pessoa, 2007) that are typically thought of as
the distributed network for emotional face perception (Haxby, Hoffman, &
Gobbini, 2000). In addition, evidence from a neural adaptation paradigm
suggests that the representation of emotional faces, within this same neural
network, is driven by perceiver conceptualization (i.e., judgments of the
emotion category) rather than by structural features of the face alone (Fox,
Moon, Iaria, & Barton, 2009). Even when repeated stimuli are changed so
drastically that they are “perceptually” drawn from a different category,
and theoretically should release the brain from adaptation (i.e., neural
responses should go back up because the perceptual aspects of the stimuli
have changed), they fail to do so when perceivers judge that the emotion
category has not changed. Plainly, the neural representation of an emotional
face involves conceptual processing and is not determined by the stimulus
features of a face alone. The role of language in emotion perception is also
evident in a meta-analysis comparing brain activity during the perception
and experience of emotion; the analysis consistently showed activation in
language-related regions, including inferior frontal gyrus (IFG), extending
from the pars opercularis (Broca’s area, BA 44) through pars triangularis (BA
45) and pars orbitalis on the inferior frontal convexity (BA 47/12 l) (Wager et
al., 2008).
Taken together, these studies indicate that emotion words (regardless of
whether they are offered) constrain how people understand the psychological
meaning of a face, even if that face is posing an intense, stereotyped set
of muscle actions. Words provide a context that contributes to emotion
perception and increases accuracy. In some ways, this is unsurprising, given
how powerful words are in the economy of the mind (e.g., Mani & Plunkett,
2010).
Repeated Measures
A final form of context that is embedded in nearly all investigations of
emotion perception is the use of a small set of faces presented repeatedly
to the perceiver. This repetition allows (and perhaps encourages) perceivers
to form ad hoc perceptual categories by extracting statistical regularities
from these posed faces (e.g., all angry faces are scowling, all sad faces
are pouting, all happy faces are smiling, and so on).
8
Indeed, attaching a
label to an emotional face in an explicit emotion perception task may help
overcome perceptual variation and support category formation as is seen for
other types of stimuli (see Lupyan, Rakison, & McClelland, 2007). A careful
Page 21 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
look suggests that repeating faces over and over creates a strong form of
experiment-based context in the face-focused approach to understanding
emotion perception. For example, research investigating the impact of
training on emotion perception indicates that participants demonstrate
substantial accuracy increases (Guilford, 1929). These training effects are
greater for more unfamiliar stimuli, such as faces portraying emotion posed
by someone from another cultural group (Elfenbein, 2006).
Conclusions
To understand how humans perceive emotion, we must focus our research
efforts on more than just the face. People rarely pose their faces into
predetermined configurations when in an emotional state, and even when
they do, facial actions often are moving targets, embedded in other facial
actions, as well as the body movements, smells, and vocalizations that
occur during conversation and other cognitively taxing events. In addition,
other factors appear crucial to perceivers’ emotion perception, including
knowledge of the situation and their own internal state. And in the context
of laboratory experiments, it is clear that scientists do many things that
serve to contextualize emotion perception, often in ways that are unlike what
occurs in everyday life.
Nevertheless, it is reasonable to ask what the face alone can tell us about
a person’s emotional state. Unfortunately, the published literature does
not really answer this question because the experiments themselves
usually contain hidden forms of context that contribute to emotion
perception accuracy. The handful of studies that have attempted to strip
away experimental forms of context from the face, with more or less
success, almost always show notable decreases in accuracy. Most of these
experiments remove only one form of experimental context, leaving the
others intact. It is important to remember that every experiment involves
some context, even if it is unintended by the experimenter.
More broadly, the context in which emotion perception unfolds might
help explain how perceivers can routinely, and with little effort, perceive
when another person is angry or afraid, despite the fact that the objective
measures of the face (e.g., electromyographic measurements) and voice
(e.g., acoustical measures) fail to reveal diagnostic patterns for each emotion
(Barrett, 2006a). Barrett (2006b) has termed this the “Emotion Paradox”:
Instrument-based measures of the face do not reveal muscle movements
that distinguish discrete emotions (Cacioppo et al., 2000; Mauss & Rpbinson,
2009), yet perceivers can readily look at someone’s face and judge how
Page 22 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
that person is feeling. One hypothesis to solve the emotion paradox is
that perceivers initially extract affective information from a face (whether
to approach or avoid it, for example) and the context activates more
nuanced conceptual knowledge that allows for the construction of discrete
emotion percept of anger, sadness, or fear (for a similar view, see Roberson,
Damjanovic, & Kikutani, 2010). The various forms of context reviewed in
this chapter might help to shape broadly “affective” actions into discrete
emotional percepts. Consistent with this hypothesis, perceivers routinely
encode and remember the context better when they are asked to judge the
emotion in a face than to judge whether to approach or avoid it (Barrett &
Kensinger, 2010). Furthermore, the more readily a situation points to a single
emotion category, the stronger its impact on subsequent emotion perception
from the face (Fernández-Dols et al., 1991). As such, it might be that the
forms of “context” reviewed here are vital to emotion perception as a highly
constructive process.
Whatever the mechanisms by which context shapes emotion perception, the
findings reviewed here should prompt us to reconsider long-held assumptions
about emotion perception. They also suggest that we should question the
experimental paradigms that dominate research based on these long-held
assumptions. The field of emotion perception seems firmly rooted in the idea
that a static, posed face is a useful cue for us to model the perception of
discrete emotional states such as anger and sadness. The present review,
in contrast, suggests that the face may not be “figure” and the context
merely “ground.” After all, the epistemic goal in emotion perception in
everyday life is not to determine what a face is displaying. The goal, instead,
is to determine what the person is experiencing, based on whatever is
most information relevant at the time. Thus, the processing of “context”
is not a secondary consideration, but instead is routine when the goal is
to understand the emotion that another person is experiencing (Barrett &
Kensinger, 2010).
As perceivers, we are drawn to faces. Our attention rests on them. It is
therefore no accident that scientists assume that faces must be the key
source of information in emotion perception. To be sure, faces convey
important psychological information, but they very likely do not provide
sufficient information to a person’s emotional state (especially if the person
himself or herself is using the context to construct that state; Barrett,
2006a, 2009). Future research that addresses emotion perception as a
contextualized phenomenon will help build our understanding of how people
Page 23 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
see emotions in other people, and how they use those perceptions to get
along and get ahead in the world.
References
Adolphs R., Damasio H., Tranel D., Cooper G., & Damasio, A. R. (2000). A role
for somatosensory cortices in the visual recognition of emotion as revealed
by 3-D lesion mapping. Journal of Neuroscience, 20, 2683–2690.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ambadar Z., Schooler J. W., & Cohn J. F. (2005). Deciphering the enigmatic
face: The importance of facial dynamics in interpreting subtle facial
expressions. Psychological Science, 16, 403–410.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Amtig J. A., Miller J. E., Chow M., & Mitchell, D. G. V. (2009). Getting mixed
messages: The impact of conflicting social signals on the brain’s target
emotional response. Neuroimage, 47, 1950–1959.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Aviezer H., Hassin R., Bentin S., & Trope Y. (2008). Putting facial expressions
back in context. In N. Ambady & J. Skowroski (Eds.), First impressions (pp.
255–288). New York : Guilford Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Aviezer H., Hassin R. R., Ryan J., Grady C., Susskind J., Anderson A.,…Bentin
S. (2008). Angry, disgusted, or afraid? Studies on the malleability of emotion
perception. Psychological Science, 19, 724–732.
Find This Resource
Find it in your Library
Page 24 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Worldcat
Google Preview
Bachorowski J. A. (1999). Vocal expression and perception of emotion.
Current Directions in Psychological Science, 8, 53–57.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Bachorowski J. A., & Owren M. J. (2008). Vocal expressions of emotion. In M.
Lewis ,J. M. Haviland-Jones & L. F. Barrett (Eds.), The handbook of emotion
(3rd ed., pp. 196–210). New York : Guilford Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Barrett L. F. (2006a). Emotions as natural kinds? Perspectives on
Psychological Science, 1, 28–58.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Barrett L. F. (2006b). Valence as a basic building block of emotional life.
Journal of Research in Personality, 40, 35–55.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Barrett L. F. (2009). The future of psychology: Connecting mind to brain.
Perspectives in Psychological Science, 4, 326–339.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Barrett L. F., & Kensinger E. A. (2010). Context is routinely encoded during
emotion perception. Psychological Science, 21, 595–599.
Find This Resource
Page 25 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find it in your Library
Worldcat
Google Preview
Barrett L. F., Lindquist K., & Gendron M. (2007). Language as a context for
emotion perception. Trends in Cognitive Sciences, 11, 327–332.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Bassili J. N. (1979). Emotion recognition: The role of facial movement and
the relative importance of upper and lower areas of the face. Journal of
Personality and Social Psychology, 37, 2049–2058.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Black S. R. (2003). A review of semantic satiation. In S. P. Shohov (Ed.),
Advances in psychology research (Vol. 26, pp. 95–106). Huntington, NY :
Nova Science.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Bouhuys A. L., Bloem G. M., & Groothuis T. G. (1995). Induction of depressed
and elated mood by music influences the perception of facial emotional
expressions in healthy subjects. Journal of Affective Disorders, 33, 215–226.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Breiter, H. C., Etcoff, N. L., Whalen, P. J., Kennedy, W. A., Rauch, S. L.,
Buckner, R. L., …Rosen, B. R. (1996). Response and habituation of the human
amygdala during visual processing of facial expression. Neuron, 17, 875–887.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 26 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Buzby D. E. (1924). The interpretation of facial expression. American Journal
of Psychology, 35, 602–604
Find This Resource
Find it in your Library
Worldcat
Google Preview
Cacioppo J. T., Berntson G. G., Larsen J. T., Poehlmann K. M., & Ito T. A.
(2000). The psychophysiology of emotion. In M. Lewis & R. J. M. Haviland-
Jones (Eds.), The handbook of emotions (2nd ed., pp. 173–191). New York :
Guilford Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Carrera-Levillain P., & Fernandez-Dols J. M. (1994). Neutral faces in context:
Their emotional meaning and their function. Journal of Nonverbal Behavior,
18, 281–299.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Carroll J. M., & Russell, J. A. (1996). Do facial expression signal specific
emotions? Judging emotion from the face in context. Journal of Personality
and Social Psychology, 70, 205–218.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Cline M. G. (1956). The influence of social context on the perception of faces.
Journal of Personality, 25, 142–158.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Collignon O., Girard S., Gosselin F., Roy S., Saint-Amour D., Lassonde M.,
& Lepore, F. (2008). Audio-visual integration of emotion expression. Brain
Research, 1242, 126–135.
Page 27 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find This Resource
Find it in your Library
Worldcat
Google Preview
Coulson M. (2004). Attributing emotion to static body postures: Recognition
accuracy, confusions, and viewpoint dependence. Journal of Nonverbal
Behavior, 28, 117–139.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Cunningham D. W., & Wallraven C. (2009). Dynamic information for the
recognition of conversational expressions. Journal of Vision, 9, 1–17.
Find This Resource
Find it in your Library
Worldcat
Google Preview
de Gelder B. (2009). Why bodies? Twelve reasons for including bodily
expressions in affective neuroscience. Philosophical Transactions of the Royal
Society B: Biological Sciences, 364, 3475–3484.
Find This Resource
Find it in your Library
Worldcat
Google Preview
de Gelder B., & Vroomen J. (2000). The perception of emotions by ear and by
eye. Cognition and Emotion, 14, 289–311.
Find This Resource
Find it in your Library
Worldcat
Google Preview
de Gelder B., Böcker K. B., Tuomainen J., Hensen M., & Vroomen J. (1999).
The combined perception of emotion from voice and face: Early interaction
revealed by human electric brain responses. Neuroscience Letters, 260, 133–
136.
Find This Resource
Find it in your Library
Worldcat
Page 28 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Google Preview
de Gelder B., Meeren H. K. M., Righart R., Van den Stock J., van de Riet W.
A. C., & Tamietto M. (2006). Beyond the face: Exploring rapid influences of
context on face processing. Progress in Brain Research, 155, 37–48.
Find This Resource
Find it in your Library
Worldcat
Google Preview
de Gelder B., Pourtois G., & Weiskrantz L. (2002). Fear recognition in the
voice is modulated by unconsciously recognized facial expressions but not
by unconsciously recognized affective pictures. Proceedings of the National
Academy of Sciences USA, 99, 4121–4126.
Find This Resource
Find it in your Library
Worldcat
Google Preview
De Silva R. R., & Bianchi-Berthouze M. (2004). Modeling human affective
postures: An information theoretic characterization of posture features.
Journal of Visualization and Computer Animation, 15, 269–276.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Dimberg U. (1982). Facial reactions to facial expressions. Psychophysiology,
19, 643–647.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Dimberg U. (1990). Facial electromyography and emotional reactions.
Psychophysiology, 27, 481–194.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 29 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Dimberg U., Thunberg M., & Elmehed, K. (2000). Unconscious facial reactions
to emotional facial expressions. Psychological Science, 11, 86–89.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Dolan R. J., Morris J. S., & de Gelder B. (2001). Crossmodal binding of fear in
voice and face. Proceedings of the National Academy of Sciences USA, 98,
10006–10010.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ekman P. (1965). Differential communication of affect by head and body
cues. Journal of Personality and Social Psychology, 2, 726–35.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ekman P. (1972). Universal and cultural differences in facial expressions of
emotions. In J. K. Cole (Ed.), Nebraska symposium on motivation, 1971 (pp.
207–283). Lincoln : University of Nebraska Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ekman P. (1973). Cross-cultural studies of facial expression. In P. Ekman (Ed.),
Darwin and facial expression: A century of research in review (pp. 169–222).
New York : Academic Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ekman P. (1998). Introduction to the third edition. In P. Ekman (Ed), Third
edition of Charles Darwin’s the expression of the emotions in man and
animals, with introduction, afterwords, and commentaries (pp. xxi–xxxvi).
London : HarperCollins.
Page 30 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ekman P., & Friesen, W. V. (1967). Head and body cues in the judgment of
emotion: A reformulation. Perceptual and Motor Skills, 24, 711–24.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ekman P., & Friesen, W. V. (1969). Nonverbal leakage and clues to deception.
Psychiatry, 32, 88–105.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ekman P., & Freisen, W. V. (1971). Constants across cultures in the face and
emotion. Journal of Personality and Social Psychology, 17, 124–129.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ekman P., & Freisen, W. V. (1976). Pictures of facial affect. Palo Alto, CA :
Consulting Psychologist Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ekman P., Freisen W. V., & Ellsworth P. (1972). Emotion in the human face.
New York : Pergamon Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 31 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Ekman, P., & O’Sullivan, M. (1988). The role of context in interpreting facial
expression: comment on Russell and Fehr (1987). Journal of Experimental
Psychology General, 117, 86–90.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ekman P., Sorenson E. R., Friesen W. V. (1969). Pan-cultural elements in facial
displays of emotion. Science, 164, 86–88.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Elfenbein H. A. (2006). Learning in emotion judgments: Training and the
cross-cultural understanding of facial expressions. Journal of Nonverbal
Behavior, 30, 21–36.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Etcoff N. L., & Magee J. J. (1992). Categorical perception of facial expressions.
Cognition, 44, 227–240.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Ethofer T., Anders S., Erb M., Droll C., Royen L., Saur R.,…Wildgruber D.
(2006). Impact of voice on emotional judgment of faces: An event-related
fMRI study. Human Brain Mapping, 27, 707–714.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Fernández-Dols J. M., Carrera P., Barchard K. A., & Gacitua M. (2008). False
recognition of facial expressions of emotion: Causes and implications.
Emotion, 8, 530–539.
Find This Resource
Page 32 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find it in your Library
Worldcat
Google Preview
Fernandez-Dols J. M., & Carroll, J. M. (1997). Is the meaning perceived
in facial expression independent of its context? In J. A. Russell & J. M.
Fernandez-Dols (Eds.), The psychology of facial expression: Studies in
emotion and social interaction, 2nd series (pp. 3–30). New York : Cambridge
University Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Fernandez-Dols J. M., Sierra B., & Ruiz-Belda M. A. (1993). On the clarity of
expressive and contextual information in the recognition of emotions: A
methodological critique. European Journal of Social Psychology, 23, 195–202.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Fernandez-Dols J. M., Wallbott H., & Sanchez F. (1991). Emotion category
accessibility and the decoding of emotion from facial expression and context.
Journal of Nonverbal Behavior, 15, 107–123.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Forgas J. P., & East R. (2008). How real is that smile? Mood effects on
acceptingor rejecting the veracity of emotional facial expressions. Journal of
Nonverbal Behavior, 32, 157–270.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Fox C. J., Moon S., Iaria G., & Barton J. J. (2009). The correlates of subjective
perception of identity and expression in the face network: An fMRI adaptation
study. Neuroimage, 44, 569–580.
Find This Resource
Page 33 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find it in your Library
Worldcat
Google Preview
Frühholz S., Fehr T., & Herrmann M. (2009). Early and late temporo-
spatial effects of contextual interference during perception of facial affect.
International Journal of Psychophysiology, 74, 1–13.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Fugate J. M. B., Gouzoules H., & Barrett L. F. (2010). Reading Chimpanzee
faces: A test of the structural and conceptual hypotheses. Emotion, 10, 544–
554.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Gendron M., & Barrett, L. F. (2009). Reconstructing the past: A century of
ideas about emotion in psychology. Emotion Review, 1, 1–24.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Gladwell M. (2002, August 5). The naked face. The New Yorker, 38–49.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Goldberg H. D. (1951). The role of “cutting” in the perception of the motion
picture. Journal of Applied Psychology, 35, 70–71.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 34 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Goodenough F. L., & Tinker M. A. (1931). The relative potency of facial
expression and verbal description of stimulus in the judgment of emotion.
Journal of Comparative Psychology, 12, 365–370.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Guilford J. P. (1929). An experiment in learning to read facial expressions.
Abnormal and Social Psychology, 24, 191–202.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Gunes H., & Piccardi M. (2007). Bi-modal emotion recognition from
expressive face and body gestures. Journal of Network and Computer
Applications, 30, 1334–1345.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Hagan C., Woods W., Johnson S., Calder A. J., Green G. G. R., & Young A. W.
(2009). MEG demonstrates a supra-additive response to facial and vocal
emotion in right superior temporal sulcus. Proceedings of the National
Academy of Sciences USA, 106, 20010–20015.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Halberstadt J., & Niedenthal P. M. (2001). Effects of emotion concepts on
perceptual memory for emotional expressions. Journal of Personality and
Social Psychology, 81, 587–598.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 35 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Hall J. A., Gunnery S. D., & Andrzejewski S. A. (2011). Nonverbal emotion
displays, communication modality, and the judgment of personality. Journal
of Research in Personality, 45, 77–83.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Harris L. J., & Snyder P. J. (1992). Subjective mood state and the perception of
emotion in chimeric faces. Cortex, 28, 471–481.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Hawk S. T., van Kleef G. A., Fischer A. H., & van der Schalk J. (2009). Worth
a thousand words: Absolute and relative decoding of affect vocalizations.
Emotion, 9, 293–305.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Haxby J. V., Hoffman E. A., & Gobbini M. I. (2000). The distributed human
neural system for face perception. Trends in Cognitive Sciences, 4, 223–233.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Hietanen J. K., & Leppänen J. M. (2008). Judgment of other people’s facial
expressions of emotions is influenced by their concurrent affective hand
movements. Scandinavian Journal of Psychology, 49, 221–230.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Hoehl S., & Striano T. (2008). Neural processing of eye gaze and threat-
related emotional facial expressions in infancy. Child Development, 79, 1752–
1760.
Find This Resource
Page 36 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find it in your Library
Worldcat
Google Preview
Hoehl S., & Striano T. (2010). Discrete emotions in infancy: Perception
without production? Emotion Review, 2, 132–133.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Holmes A., Kragh Nielsen M., Tipper S., & Green S. (2009). An
electrophysiological investigation into the automaticity of emotional face
processing in high versus low trait anxious individuals. Cognitive, Affective
and Behavioral Neuroscience, 9, 323–334.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Hunt W. A. (1941). Recent developments in the field of emotion.
Psychological Bulletin, 38, 249–276.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Isaacowitz D. M., & Stanley J. T. (2011). Bringing an ecological perspective
to the study of aging and recognition of emotional facial expressions: Past,
current, and future methods. Journal of Nonverbal Behavior, 35, 261–278.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Izard C. E. (1971). The face of emotion. New York : Appleton-Century-Crofts.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Izard C. (1977). Human emotions. New York : Plenum.
Page 37 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find This Resource
Find it in your Library
Worldcat
Google Preview
Izard, C. E. (1994). Innate and universal facial expressions: Evidence from
developmental and cross-cultural research. Psychological Bulletin, 115, 288–
299.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Kamachi M., Bruce V., Mukaida S., Gyoba J., Yoshikawa S., & Akamatsu S.
(2001). Dynamic properties influence the perception of facial expressions.
Perception, 30, 875–887.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Kanner L. (1931). Judging emotion from facial expressions. Psychological
Monographs, 41, 93.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Keltner D., Ekman P., Gonzaga G. C., & Beer J. (2003). Facial expression of
emotion. In R. J. Davidson, K. R. Scherer, & H. H. Goldsmith (Eds.), Handbook
of affective sciences (pp. 415–131). New York : Oxford University Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Kilts C. D., Egan G., Gideon D. A., Ely T. D., & Hoffman J. M. (2003).
Dissociable neural pathways are involved in the recognition of emotion in
static and dynamic facial expressions. Neuroimage, 18, 156–168.
Find This Resource
Find it in your Library
Worldcat
Page 38 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Google Preview
Kim H., Somerville L. H., Johnstone T., Polis S., Alexander A. L, Shin L. M.,
& Whalen P. J. (2004). Contextual modulation of amygdala responsivity to
surprised faces. Journal of Cognitive Neuroscience, 16, 1730–1745.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Kline L. W., & Johannsen, D. E. (1935). Comparative role of the face and of
the face-body-hands in identifying emotions. Journal of Abnormal and Social
Psychology, 2, 415–426.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Knudsen H. R., & Muzekari L. H. (1983). The effects of verbal statements of
context on facial expressions of emotion. Journal of Nonverbal Behavior, 7,
202–212.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Kohler C. G., Walker J. B., Martin E. A., Healey K. M., & Moberg P. J. (2010).
Facial emotion perception in schizophrenia: A meta-analytic review.
Schizophrenia Bulletin, 36, 1009–1019.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Kuraoka K., & Nakamura K. (2007). Responses of single neurons in monkey
amygdala to facial and vocal emotions. Journal of Neurophysiology, 97,
1379–1387.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 39 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Landis C. (1924a). Studies of emotional reactions: I. A preliminary study of
facial expression. Journal of Experimental Psychology, 7, 325–341.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Landis C. (1924b). Studies of emotional reactions: II. General behavior and
facial expression. Journal of Comparative Psychology, 4, 477–509.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Landis C. (1929). The interpretation of facial expression in emotion. Journal of
General Psychology, 2, 59–72.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Landis C. (1934). Emotion: II. The expressions of emotion. In C.Murchison
(Ed.), A handbook of general experimental psychology (pp. 312–351).
Worcester, England : Clark University Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Langfeld H. S. (1918a). Judgments of facial expression and suggestion.
Psychological Review, 25, 488–494.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Langfield H. S. (1918b). The judgment of emotions from facial expressions.
Journal of Abnormal Social Psychology, 13, 172–184.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 40 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Leppänen J. M., & Hietanen J. K. (2003). Affect and face perception: Odors
modulate the recognition advantage of happy faces. Emotion, 3, 315–326.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Lieberman M. D., Eisenberger N. I., Crockett M. J., Tom S. M., Pfeifer J. H.,
& Way B. M. (2007). Putting feelings into words: Affect labeling disrupts
amygdala activity to affective stimuli. Psychological Science, 18, 421–428.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Lindquist, K. A., Barrett, L. F., Bliss-Moreau, E., & Russell, J. A.. (2006).
Language and the perception of emotion. Emotion, 6, 125–138.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Lupyan G., Rakison D. H., & Mcclelland J. L. (2007). Language is not just for
talking redundant labels facilitate learning of novel categories. Psychological
Science, 18, 1077–1083.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Mani N., & Plunkett K. (2010). In the infant’s mind’s ear: Evidence for implicit
naming in 18-month-olds. Psychological Science, 21, 908–913.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Massaro D. W., & Egan, P. B. (1996). Perceiving affect from the voice and the
face. Psychonomic Bulletin and Review, 3, 215–221.
Find This Resource
Find it in your Library
Worldcat
Page 41 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Google Preview
Masuda T., Ellsworth P., Mesquita B., Leu J., & Veerdonk E. (2008). Placing
the face in context: Cultural differences in perceiving emotions from facial
behavior. Journal of Personality and Social Psychology, 94, 365–381.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Matsumoto D., Keltner D., Shiota M., Frank M., & O’Sullivan M. (2008). Facial
expressions of emotions. In M.Lewis, J.Haviland & L. Feldman-Barrett (Eds.),
Handbook of emotion (pp. 211–234). New York : Guilford Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Matsumoto, D., & Willingham, B. (2009). Spontaneous facial expressions
of emotion of congenitally and noncongenitally blind individuals. Journal of
Personality and Social Psychology, 96, 1–10.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Mauss I. B., & Robinson M. D. (2009). Measures of emotion: A review.
Cognition and Emotion, 23, 209–237.
Find This Resource
Find it in your Library
Worldcat
Google Preview
McLellan T., Johnston L., Dalrymple-Alford J., & Porter R. (2010). Sensitivity
to genuine versus posed emotion specified in facial displays. Cognition and
Emotion, 24, 1277–1292.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 42 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Meeren H. K. M., van Heijnsbergen C., & de Gelder B. (2005). Rapid
perceptual integration of facial expression and emotional body language.
Proceedings of the National Academy of Sciences USA, 102, 16518–16523.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Miles L., & Johnston L. (2007). Detecting happiness: Perceiver sensitivity to
enjoyment and non-enjoyment smiles. Journal of Nonverbal Behavior, 31,
259–275.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Mobbs D., Weiskopf N., Lau H. C., Featherstone E., Dolan R. J., & Frith C.
D. (2006). The Kuleshov effect: The influence of contextual framing on
emotional attributions. Social Cognitive and Affective Neuroscience, 1, 95–
106.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Motley M. T., & Camden, C. T. (1988). Facial expression of emotion: A
comparison of posed versus spontaneous expressions in an interpersonal-
communication setting. Western Journal of Speech Communication, 52, 1–22.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Munn N. L. (1940). The effect of knowledge of the situation upon judgment of
emotion from facial expressions. Journal of Abnormal and Social Psychology,
35, 324–338.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 43 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Murphy N. A, Lehrfeld J. M., & Isaacowitz D. M. (2010). Recognition of posed
and spontaneous dynamic smiles in young and older adults. Psychology and
Aging, 25, 811–821.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Naab P. J., & Russell J. A. (2007). Judgments of emotion from spontaneous
facial expressions of New Guineans. Emotion, 7, 736–744.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Nakamura M., Buck R., & Kenny D. A. (1990). Relative contribution of
expressive behavior and contextual information to the judgment of the
emotional state of another. Journal of Personality and Social Psychology, 59,
1032–1039.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Nelson, N. L., & Russell, J. A. (2011). Preschoolers’ use of dynamic facial,
bodily, and vocal cues to emotion. Journal of Experimental Child Psychology,
110, 52–61.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Niedenthal P. M., Barsalou L. W., Winkielman P., Krauth-Gruber S., & Ric F.
(2005). Embodiment in attitudes, social perception, and emotion. Personality
and Social Psychology Review, 9, 184–211.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Niedenthal P. M., Brauer M., Halberstadt J. B., & Innes-Ker A. H. (2001). When
did her smile drop? Facial mimcry and the influences of emotional state on
Page 44 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
the detection of change in emotional expression. Cognition and Emotion, 15,
853–864.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Niedenthal P. M., Halberstadt J. B., Margolin J., & Innes-Ker A. H. (2000).
Emotional state and the detection of change in facial expression of emotion.
European Journal of Social Psychology, 30, 211–222.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Niedenthal P. M., Mermillod M., Maringer M., & Hess U. (2010). The simulation
of smiles (SIMS) model: Embodied simulation and the meaning of facial
expression. Behavioral and Brain Sciences, 33, 417–33.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Parr L. A., & Waller, B. M. (2006). Understanding chimpanzee facial
expression: Insights into the evolution of communication. Social Cognitive
and Affective Neuroscience, 1, 221–228.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Peelen, M. V., Atkinson, A. P., & Vuilleumier, P. (2010). Supramodal
representations of perceived emotions in the human brain. Journal of
Neuroscience, 30, 10127–10134.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Pell M. C. (2005). Prosody-face interactions in emotional processing as
revealed by the facial affect decision task. Journal of Nonverbal Behavior, 29,
193–215.
Page 45 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find This Resource
Find it in your Library
Worldcat
Google Preview
Pereira C. (2000). Dimensions of emotional meaning in speech. In R. Cowie,
E. Douglas-Cowie, & M. Schröder (Eds.), Proceedings of the ISCA Workshop on
Speech Emotion. Belfast, Northern Ireland : Textflow.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Phillips L. H., Channon S., Tunstall M., Hedenstrom A., & Lyons K. (2008). The
role of working memory in decoding emotions. Emotion, 8, 184–191.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Phillips M. L., Young A. W., Scott S. K., Calder A. J., Andrew C., Giampietro V.,
…Gray J. A. (1998). Neural responses to facial and vocal expressions of fear
and disgust. Proceedings of the Royal Society B, 265, 1809–1817.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Phillips M. L., Young A. W., Senior C., Brammer M., Andrews C., Calder A.
J.,…David A. S. (1997). A specific neural substrate for perceiving facial
expressions of disgust. Nature 389, 495–498.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Pistoia F., Conson M., Trojano L., Grossi D., Ponari M., Colonnese, C., …Sara
M. (2010). Impaired conscious recognition of negative facial expressions in
patients with locked-in syndrome. Journal of Neuroscience, 30, 7838–7844.
Find This Resource
Find it in your Library
Worldcat
Page 46 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Google Preview
Pitcher D., Garrido L., Walsh V., & Duchaine B. (2008). TMS disrupts the
perception and embodiment of facial expressions. Journal of Neuroscience,
28, 8929–8933.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Pittam J., Gallois C., & Callan V. (1990). The long-term spectrum and
perceived emotion. Speech Communication, 9, 177–187.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Pitterman H., & Nowicki S., Jr. (2004). A test of the ability to identify emotion
in human standing and sitting postures: The diagnostic analysis of nonverbal
accuracy-2 posture test (DANVA2-POS). Genetic, Social, and General
Psychology Monographs, 130, 146–162.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Pourtois G., de Gelder B., Vroomen J., Rossion B., & Crommelinck M. (2000).
The time-course of intermodal binding between seeing and hearing affective
information. Neuroreport, 11, 1329–1333.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Pourtois G., Debatisse D., Despland P. A., & de Gelder B. (2002). Facial
expressions modulate the time course of long latency auditory brain
potentials. Brain Research Cognitive Brain Research, 14 (1), 99–105.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 47 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Pourtois G., Sander D., Andres M., Grandjean D., Reveret L., Olivier E., &
Vuilleumier, P. (2004). Dissociable roles of the human somatosensory and
superior temporal cortices for processing social face signals. European
Journal of Neuroscience, 20, 3507–3515.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Reichenbach L., & Masters, J. C. (1983). Children’s use of expressive and
contextual cues in judgments of emotion. Child Development, 54, 993–1004.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Rives Bogart K., & Matsumoto, D. (2010). Facial mimicry is not necessary
to recognize emotion: Facial expression recognition by people with Moebius
syndrome. Social Neuroscience, 5, 241–251.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Roberson D., Damjanovic L., Kikutani M. (2010). Show and tell: The role of
language in categorizing facial expression of emotion. Emotion Review, 2,
255–260.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Roberson D., Damjanovic L., & Pilling, M. (2007). Categorical perception of
facial expressions: Evidence for a “Category Adjustment” model. Memory and
Cognition, 35, 1814–1829.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 48 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Rozin P., Taylor C., Ross L., Bennett G., & Hejmadi A. (2005). General and
specific abilities to recognize negative emotions, especially disgust, as
portrayed in the face and the body. Cognition and Emotion, 19, 397–412.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Russell J. A. (1993). Forced-choice response format in the study of facial
expression. Motivation and Emotion, 17, 41–51.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Russell J. A. (1994). Is there universal recognition of emotion from facial
expression? A review of the cross-cultural studies. Psychological Bulletin,
115, 102–141.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Russell J. A., & Bullock M. (1986). On the dimensions preschoolers use to
interpret facial expressions of emotion. Developmental Psychology, 22, 97–
102.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Russell J. A., & Fehr B. (1987). Relativity in the perception of emotion in facial
expressions. Journal of Experimental Psychology: General, 116, 223–237.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Russell J. A., & Widen, S. C. (2002). A label superiority effect in children’s
categorization of facial expressions. Social Development, 11, 30–52.
Find This Resource
Find it in your Library
Page 49 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Worldcat
Google Preview
Sato W., & Yoshikawa S. (2004). The dynamic aspects of emotional facial
expressions. Cognition and Emotion, 18, 701–710.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Sauter D. A., Eisner F., Ekman P., & Scott S. K. (2010). Cross-cultural
recognition of basic emotions through nonverbal emotional vocalizations.
Proceedings of the National Academy of Sciences USA, 107(6), 2408–2412.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Scherer K. L., & Wallbot, H. B. (1994). Evidence for universality and cultural
variation of differential emotion response patterning. Journal of Personality
and Social Psychology, 66, 310–328.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Schiffenbauer A. (1974). Effect of observer’s emotional state on judgments
of the emotional state of others. Journal of Personality and Social Psychology,
30, 31–35.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Schlosberg H. (1952). The description of facial expressions in terms of two
dimensions. Journal of Experimental Psychology, 44, 229–237.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 50 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Schouwstra S. J., & Hoogstraten J. (1995). Head position and spinal position
as determinants of perceived emotional state. Perceptual Motor Skills, 81,
673–674.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Schyns P. G., Petro L. S., & Smith M. L. (2009). Transmission of facial
expressions of emotion co- evolved with their efficient decoding in the brain:
Behavioral and brain evidence. Plos One, 4, 1–16.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Sherman M. (1927). The differentiation of emotional responses in infants: II.
The ability of observers to judge the emotional characteristics of the crying
of infants and of the voice of an adult. Journal of Comparative Psychology, 7,
335–351.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Simon-Thomas E. R., Keltner D. J., Sauter D., Sinicropi-Yao L., & Abramson
A. (2009). The voice conveys specific emotions: evidence from vocal burst
displays. Emotion, 9, 838–846.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Somerville L. H., & Whalen, P. J. (2006). Prior experience as a stimulus
category confound: An example using facial expressions of emotion. Social,
Cognitive, and Affective Neuroscience, 1, 271–274.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 51 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Sommer M., Döhnel K., Meinhardt J., & Hajak G. (2008). Decoding of affective
facial expressions in the context of emotional situations. Neuropsychologia,
46, 2615–2621.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Spignesi A., & Shor R. (1981). The judgment of emotion from facial
expressions, contexts and their combination. Journal of General Psychology,
104, 41–59.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Sprengelmeyer R., Rausch M., Eysel U. T., & Przuntek H. (1998). Neural
structures associated with recognition of facial expressions of basic
emotions. Proceedings of the Royal Society of London B: Biological Sciences,
265, 1927–1931.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Stel M., & van Knippenberg A. (2008). The role of facial mimicry in the
recognition of affect. Psychological Science, 19, 984–985.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Tanaka-Matsumi J., Attivissimo D., Nelson S., & D’Urso T. (1995). Context
effects on the judgment of basic emotions in the face. Motivation and
Emotion, 19, 139–155.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 52 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Thayer S. (1980). The effect of expression sequence and expressor identity
on judgments of the intensity of facial expression. Journal of Nonverbal
Behaivor, 5, 71–79.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Thielscher A., & Pessoa, L. (2007). Neural correlates of perceptual choice and
decision making in fear-disgust discrimination. Journal of Neuroscience, 27,
2908–2917.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Tomasik D., Ruthruff E., Allen P. A., & Lien M. C. (2009). Non-automatic
emotion perception in a dual-task situation. Psychonomic Bulletin and
Review, 16, 282–288.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Tracy J. L., & Robins R. W. (2008). The automaticity of emotion recognition.
Emotion, 7, 789–801.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Vuilleumier P., & Pourtois G. (2007). Distributed and interactive brain
mechanisms during emotion face perception: Evidence from functional
neuroimaging. Neuropsychologia, 45, 174–94.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Vroomen J., Driver J., & de Gelder B. (2001). Is cross-modal integration of
emotional expressions independent of attentional resources? Cognitive,
Affective and Behavioral Neuroscience, 1, 382–387.
Page 53 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find This Resource
Find it in your Library
Worldcat
Google Preview
Wager T. D., Barrett L. F., Bliss-Moreau E., Lindquist K., Duncan S., Kober H.,
…Mize J. (2008). The neuroimaging of emotion. In M. Lewis, J. M. Haviland-
Jones, & L.F. Barrett (Eds.), The handbook of emotion (3rd ed., p p. 249–271).
New York : Guilford Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Wagner H. L., MacDonald C. J., & Manstead A. S. R. (1986). Communication of
individual emotions by spontaneous facial expressions. Journal of Personality
and Social Psychology, 50, 737–743.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Wallbott H. G. (1988a). Faces in context: The relative importance of facial
expressions and context information in determining emotion attributions. In
K. R. Scherer (Ed.), Facets of emotion (pp. 139–160). Hillsdale, NJ : Erlbaum.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Wallbott H. G. (1988b). In and out of context: Influences of facial expression
and context information on emotion attributions. British Journal of Social
Psychology, 27, 357–369.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Wallbott H. G. (1991). Recognition of emotion from facial expression via
imitation? Some indirect evidence for an old theory. British Journal of Social
Psychology, 30, 207–219.
Find This Resource
Page 54 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find it in your Library
Worldcat
Google Preview
Warner L. R., & Shields S. A. (2009). Judgments of others’ emotional
appropriateness are multidimensional. Cognition and Emotion, 23, 876–888.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Watson J. C. (2004). From interpretation to identification: A history of facial
images in the sciences of emotion. History of the Human Sciences, 17, 29–
51.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Wehrle T., Kaiser S., Schmidt S., & Scherer K. R. (2000). Studying the
dynamics of emotional expression using synthesized facial muscle
movements. Journal of Personality and Social Psychology, 78, 105–119.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Weinberger S. (2010). Airport security: Intent to deceive? Nature, 465, 412–
415.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Widen, S. C., Christy, A. M., Hewett, K., & Russell, J. A. (2011). Do proposed
facial expressions of contempt, shame, embarrassment, and compassion
communicate the predicted emotion? Cognition & Emotion, 25, 898–906.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Page 55 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Widen S. C., & Russell, J. A. (2008). Young children’s understanding of other’s
emotions. In M. Lewis & J. M. Haviland-Jones (Eds.), Handbook of emotions
(3rd ed., pp. 348–363). New York : Guilford Press.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Widen S. C., & Russell J. A. (2010). Differentiation in preschooler’s categories
of emotion. Emotion, 10, 651–61.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Widen, S. C., & Russell, J. A. (2010). Children’s scripts for social emotions:
Causes and consequences are more central than are facial expressions.
British Journal of Developmental Psychology, 28, 565–581.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Yik M. S. M., Meng Z., & Russell J. A. (1998). Adults’ freely produced emotion
labels for babies’ spontaneous facial expressions. Cognition and Emotion, 12,
723–730.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Zhou W., & Chen, D. (2009). Chemosignals of fear modulate fear recognition
in ambiguous facial expressions. Psychological Science, 20, 177–183.
Find This Resource
Find it in your Library
Worldcat
Google Preview
Zuckerman M., Hall J. A., DeFrank R. S., & Rosenthal R. (1976). Encoding and
decoding of spontaneous and posed facial expressions. Journal of Personality
and Social Psychology, 34, 966–977.
Find This Resource
Page 56 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
Find it in your Library
Worldcat
Google Preview
Notes:
(1.) The use of the forced-choice method has recently been extended to
studies testing the sufficiency of nonverbal emotional vocalizations as cues
to discrete emotion (Sauter, Eisner, Ekman, & Scott, 2009; Simon-Thomas et
al., 2009) and has been used in studies examining whether posture/bodily
movements convey discrete emotion to perceivers (e.g., Coulson, 2004;
De Silva & Bianchi-Berthouze, 2004; Ekman, 1965; Ekman & Friesen, 1967;
Pitterman & Nowicki, 2004; Schouwstra & Hoogstraten, 1995).
(2.) “Affect” versus “emotion” perception must be distinguished. This
distinction rests on the judgment that perceivers are asked to make. When
a perceiver is asked to judge the discrete emotion content of a face, for
example, “Is this face angry?” the task is typically assumed to be an emotion
perception. If, however, the perceiver is asked to make a judgment about
how intense, positive/negative, aroused/unaroused a target person is,
or whether to approach/avoid that target person, the task is categorized
as affect perception. Nonetheless, it is often the case that effects are
interpreted as emotion perception when they might, in fact, be affect
perception. For example, presenting a facial depiction of “anger” within a
“happy” context is incongruent both in terms of discrete emotion categories
and also in terms of valence. Likewise, facilitation due to pairing a a scowling
face with an anger-inducing situational description might be due merely to
the affective value of the context instead of the match on discrete emotion
category (if the effect is defined by comparing perception of the face in
the absence of context or in a neutral context). Similarly, studies that
compare the perception of started “fear” faces to neutral faces, or “happy”
faces to “sad” faces, might be studies of affect (valence) perception rather
than studies of emotion perception per se. A similar point can be made for
comparisons of “fear” faces to “sad” faces because they differ in affective
arousal.
(3.) There are additional data to suggest that context effects extend in the
other direction: Facial behaviors serve as a context to emotion perception
based on vocal cues (e.g., de Gelder, Böcker, Tuomainen, Hensen, &
Vroomen, 1999; de Gelder, Pourtois, & Weiskrantz, 2002; de Gelder &
Vroomen, 2000; Pourtois, de Gelder, Vroomen, Rossion, & Crommelinck,
Page 57 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
2000; Pourtois, Debatisse, Despland, & de Gelder, 2002; Vroomen, Driver,
& de Gelder, 2001). Furthermore, the influence of facial behaviors on
the perception of emotion in the voice is maintained under a dual task,
suggesting this effect is not dependent on attentional resources (Vroomen et
al., 2001).
(4.) There are some data to suggest that these two sources of information
can be integrated even earlier. When facial behaviors serve as a context
for vocal cues, integration appears to occur very rapidly. For example,
a mismatch negativity at 178 ms was detected for prosody following an
incongruent face portraying emotion (de Gelder et al., 1999). When facial
and vocal stimuli were presented simultaneously and were congruent, an
enhancement of the N1 was demonstrated at 110 ms after onset (Pourtois et
al., 2000). One possibility is that the time course of integration depends on
which cue, face or voice, is the “object” of perception.
(5.) Interestingly, the Nakamura et al. (1990) paper is typically cited as some
of the best evidence that the face is a stronger cue to emotion than is the
situational context because the researchers matched each in its source
clarity (i.e., the extent to which perceivers thought the cue was associated
with a single emotion). Yet this research strikingly only looked at contexts
and situations that differed in valence (e.g., a “pleasant” face paired with an
“unpleasant” slide) so that a much more limited interpretation is warranted:
situations fail to override facial information that differs in valence (although
other work shows that bodies with props can over-ride valence; Aviezer et al.,
2008, Study 2).
(6.) In semantic satiation, on a given trial, participants repeat a word 30
times (as compared to the typical control condition where they repeat the
word three times). The massive repetition of the word produces a temporary
functional equivalent of semantic dementia (characterized by impairments in
the semantic representation of words).
(7.) Recent data suggest that the face can still be a relatively effective cue
to a discrete emotion label. When young children (ages 2–4, a younger
sample than the 2002 sample demonstrating the label superiority effect) are
asked to generate a label in response to a face or a description of a cause
or consequence for emotion, for some categories, faces were a better cue
to the emotion label (Widen & Russell, 2010). These data suggest that facial
actions may be one of the first types of content that “populate” emotion
categories, perhaps given their relatively concrete nature.
Page 58 of 58
Emotion Perception: Putting the Face in Context
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). (c) Oxford University Press, 2013. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in Oxford
Handbooks Online for personal use (for details see Privacy Policy).
Subscriber: K U Leuven; date: 13 August 2013
(8.) In real life, it is unclear whether people routinely scowl when they are
angry, or pout when they are sad, and so it is unclear whether emotion
categories are, in fact, perceptual categories (cf. Barrett, 2006a; see also
Mauss & Robinson, 2009).
... The effect of language as a contextual influence on emotion perception has also garnered a solid base of research. The paradigm commonly focuses on matching facial stimuli to relevant emotion words or categories [39][40][41]. Emotion words provide information that influences precision and accuracy of emotion judgments, causing a shift in the way that faces are perceived [42]. Pairing emotional faces with labels increases sensitivity and speed in recognising emotions, even in participants with alexithymia (who are impaired in labelling their own emotions) [43]. ...
... There is a growing body of research providing evidence for the role of contextual information in emotion perception (for a review see 39). We investigated the effect of emotion labels on the perceived arousal, valence, and dominance of facial expressions, using both static and dynamic stimuli. ...
... Our results stand in line with a constructionist theory of emotion, which posits that language information provides context that allows us to perceive emotion from facial expressions [7]. Here we extend the previous work to show that emotion labels not only shift the emotion concepts perceived in a facial expression, but they shift the underlying affective dimensions as well [39][40][41]. It is possible that these dimensional judgments shift because the labels influence the extent to which underlying dimensional concepts are activated, thus, prompting the participant to judge the face to be at a different place on the affective circumplex. ...
Article
Full-text available
Whether language information influences recognition of emotion from facial expressions remains the subject of debate. The current studies investigate how variations in emotion labels that are paired with expressions influences participants’ judgments of the emotion displayed. Static (Study 1) and dynamic (Study 2) facial expressions depicting eight emotion categories were paired with emotion labels that systematically varied in arousal (low and high). Participants rated the arousal, valence, and dominance of expressions paired with labels. Isolated faces and isolated labels were also rated. As predicted, the label presented influenced participants’ judgments of the expressions. Across both studies, higher arousal labels were associated with: 1) higher ratings of arousal for sad, angry, and scared expressions, and 2) higher ratings of dominance for angry, proud, and disgust expressions. These results indicate that emotion labels influence judgments of facial expressions.
... Following SPAFF's guidelines (Coan & Gottman, 2007), and mindful of research yielding cultural differences in the interpretation of (emotional) behaviors (Brooks et al., 2019;Gendron et al., 2013;Masuda et al., 2008), we opted in our research for same-culture coders. The advantage of this approach is that the obtained codes are consistent with the everyday cultural interpretation of behavior as likely to be found during couple conflict. ...
Article
Full-text available
In the present study, we examined cultural variation in couples’ emotions during disagreement. We coded the emotions of 58 Belgian and 80 Japanese couples using the Specific Affect Coding System. We observed more anger and domineering, but less fear/tension and other-validation in Belgian than in Japanese couples. Moreover, in Japanese couples, culturally typical emotions were associated with higher conflict resolution and relationship satisfaction. The findings suggest meaningful cultural differences in couples’ observed emotions during disagreement, as they can be understood from the prevailing relationship ideals in each culture.
... Emotional faces in real-life scenarios are dynamic, embedded elements of a rich and meaningful environment. The perception of emotional faces is highly contextualized; to comprehend their meaning, we rely on many non-verbal cues, such as body gestures and contextual information from the surrounding environment (Aviezer & Hassin, 2017;Aviezer et al., 2012;Barrett et al., 2011;Chen & Whitney, 2020;Gendron et al., 2013;Shaham & Aviezer, 2020). The pictures that researchers use of actors making posed, static expressions (e.gLundqvist et al., 1998;Tottenham et al., 2009) are excellent for establishing accurate separations between categories of emotions. ...
Article
Full-text available
Despite the obvious importance of facial expressions of emotion, most studies have found that they do not bias attention. A critical limitation, however, is that these studies generally present face distractors on all trials of the experiment. For other kinds of emotional stimuli, such as emotional scenes, infrequently presented stimuli elicit greater attentional bias than frequently presented stimuli, perhaps due to suppression or habituation. The goal of the current study then was to test whether such modulation of attentional bias by distractor frequency generalizes to facial expressions of emotion. In Experiment 1, both angry and happy faces were unable to bias attention, despite being infrequently presented. Even when the location of these face cues were more unpredictable—presented in one of two possible locations—still no attentional bias was observed (Experiment 2). Moreover, there was no bottom-up influence for angry and happy faces shown under high or low perceptual load (Experiment 3). We conclude that task-irrelevant posed facial expressions of emotion cannot bias attention even when presented infrequently.
... Contextual factors have a strong influence on responses to human faces (Aviezer et al., 2008;Deffler et al., 2015;Gendron et al., 2013;Wieser & Brosch, 2012) and it is reasonable to expect they would also influence responses to CG faces. ...
Article
Full-text available
Computer-generated (CG) beings are rapidly infiltrating the human social world. Yet evidence about how humans respond to CG faces is mixed. The present systematic review and meta-analyses aimed to synthesise empirical evidence from studies comparing people’s responses to CG and human faces, across key face processing domains of interest to psychology, neuroscience, and computer science. We tested whether effects were moderated by the perceived realism of CG relative to human faces, and whether CG and human faces showed the same identity or not. We hypothesised that people would be able to tell CG and human faces apart, and that other types of responses would favour human over CG faces. While results supported our hypotheses across several domains (perceptions of human-likeness, face memory, first impressions, emotion labelling), some responses did not differ for CG and human faces (quality of interactions, emotion ratings, facial mimicry, looking behaviour). We also found a reduced inversion effect for CG relative to human faces, though only minimal data were available for hallmark face effects (ORE, N170 and FFA responses). Overall, findings highlight potential strengths and challenges of using CG faces across a range of applications, including e-health, social companionship, video-gaming, and scientific work.
... Words in a rating scale along with scale instructions presented with the face-context pair also are external-observer cues. Emotion words can activate conceptual and sensory knowledge in memory that, in turn, influence perception of facial expressions (Barrett et al., 2011;Doyle & Lindquist, 2017Gendron et al., 2013;Lindquist, 2017;Russell, 1994). An observer, then, uses such knowledge to construct meaning from the person's facial movements as a discrete emotion. ...
Article
Full-text available
The role of face and context in emotion perception was investigated by manipulating features relevant to the stimuli and to the observer. A nested-stimulus design was used, with subjects nested under stimulus item (an encoder’s facial expression or a written emotion-eliciting scenario presented alone or in an incongruent pair) and type of task instruction (judgment of encoder’s expressed or felt emotion). Subjects, using one type of task instruction, completed a decoding task in which they viewed a facial expression, a written scenario, or a facial expression paired with an emotion-incongruent scenario. Type of task instruction was intended to alter subjects’ perception by directing attention to favor face information (judgment of expressed emotion) or context information (judgment of felt emotion). Then subjects selected the predominant emotion and indicated the intensity of various emotions they perceived the encoder to be expressing or feeling. Judgments were examined for target emotion match to face and/or context using a by-stimulus analysis. The results suggest that when an encoder’s facial expression is discordant with the emotion-eliciting event, subjects will favor facial information when judging what the encoder is expressing, whereas they will favor context information when judging what the encoder is feeling. When face or context was seen alone, type of task instruction did not influence subjects’ judgments. This research provides a more detailed understanding of the role of face and context by exploring how features associated with the observer-encoder interaction influence emotion perception.
... Interpreting emotions of people around us is central to human experience, a process traditionally considered to involve the isolated face (Ekman & Friesen, 1976). Nevertheless, when categorizing facial expressions, people utilize the context in which the expression is embedded (e.g., body postures, scenes, sounds) in a manner that can shift perceivers' face categorization (Atias et al., 2019;Aviezer et al., 2008Aviezer et al., , 2012aGendron et al., 2013;Lecker et al., 2019;Reschke et al., 2019). Sources of contextual influence may also include the viewer's language, which may play a key role in the perception of other's emotions (Barrett et al., 2007;Lindquist & Gendron, 2013). ...
Article
Full-text available
Semantic emotional labels can influence the recognition of isolated facial expressions. However, it is unknown if labels also influence the susceptibility of facial expressions to context. To examine this, participants categorized expressive faces presented with emotionally congruent or incongruent bodies, serving as context. Face-body composites were presented together, aligned in their natural form, or spatially misaligned with the head shifted horizontally beside the body—a condition known to reduce the contextual impact of the body on the face. Critically, participants responded either by choosing emotion labels or by perceptually matching the target expression with expression probes. The results show a label dominance effect: Face-body congruency effects were larger with semantic labels than with perceptual expression matching, indicating that facial expressions are more prone to contextual influence when categorized with emotion labels, an effect only found when faces and bodies were aligned. These findings suggest that the role of conceptual language in face-body context effects may be larger than previously assumed.
... It is not surprising that context similarity was associated with the success of exact replication efforts in a sample of 100 replication study attempts . But, even with similarity in location, time, or culture, emerging examples from the psychological literature suggest that moving from more traditional, less representative stimuli and contexts to more naturalistic and representative contexts can fundamentally alter patterns of findings: The underlying mediational processes implicated for the target behavior of interest can be dramatically different (Gendron, Mesquita, & Barrett, 2013;Levine, Blair, & Clare, 2014). ...
Article
Causal inference and generalizability both matter. Historically, systematic designs emphasize causal inference, while representative designs focus on generalizability. Here, we suggest a transformative synthesis – Systematic Representative Design (SRD) – concurrently enhancing both causal inference and “built-in” generalizability by leveraging today’s intelligent agent, virtual environments, and other technologies. In SRD, a “default control group” (DCG) can be created in a virtual environment by representatively sampling from real-world situations. Experimental groups can be built with systematic manipulations onto the DCG base. Applying systematic design features (e.g., random assignment to DCG versus experimental groups) in SRD affords valid causal inferences. After explicating the proposed SRD synthesis, we delineate how the approach concurrently advances generalizability and robustness, cause-effect inference and precision science, a computationally-enabled cumulative psychological science supporting both “bigger theory” and concrete implementations grappling with tough questions (e.g., what is context?) and affording rapidly-scalable interventions for real-world problems.
... This being said, the ecological validity of the posed, caricatured human facial expressions used in most psychology experiments is also questionable (Nelson & Russell, 2011;Quigley, Lindquist, & Barrett, 2013). One interpretation of these posed facial expressions is that they are more like symbols than veridical representations of what people do with their faces in daily life (Adams, Albohn, & Kveraga, 2016;Gendron, Mesquita, & Barrett, 2013;Jack, Garrod, Yu, Caldara, & Schyns, 2012). Thus, there may be greater parallels between our studies and studies using human facial stimuli than appears at first glance. ...
Article
Full-text available
Understanding emotion words is vital to understanding, regulating, and communicating one's emotions. Yet, little work examines how emotion words are acquired by children. Previous research in linguistics suggests that children use the sentence frame in which a novel word is presented to home in on the meaning of that word, in conjunction with situational cues from the environment. No research has examined how children integrate these cues to learn the meaning of emotion adjectives (e.g., "happy," "sad," "mad"). We conducted 2 studies examining the role of sentence frame and situational context in children's (ages 3-5) understanding of the meanings of novel words denoting emotions. In Study 1 (N = 135) children viewed a conversation wherein a novel "alien" word was presented in 1 of 3 sentence frames that varied in how likely the word was to denote an emotion (i.e., is daxy, feels daxy, or feels daxy about). Children selected the image that represented the meaning of the word in a picture-pointing task. Images depicted aliens experiencing an emotion, a physical state, or performing an action. In Study 2 (N = 113) situational context was added via cartoons depicting an emotional scenario. Findings suggest that children are more likely to associate emotion images with a novel word with increasing age, more informative sentence frames, and when the situational context implies that an emotion is present. This provides important insight on how educational and clinical settings can use language and situational context to aide in emotion understanding. (PsycINFO Database Record (c) 2019 APA, all rights reserved).
Article
Full-text available
This article considers the status and study of "context" in psychological science through the lens of research on emotional expressions. The article begins by updating three well-trod methodological debates on the role of context in emotional expressions to reconsider several fundamental assumptions lurking within the field's dominant methodological tradition: namely, that certain expressive movements have biologically prepared, inherent emotional meanings that issue from singular, universal processes which are independent of but interact with contextual influences. The second part of this article considers the scientific opportunities that await if we set aside this traditional understanding of "context" as a moderator of signals with inherent psychological meaning and instead consider the possibility that psychological events emerge in ecosystems of signal ensembles, such that the psychological meaning of any individual signal is entirely relational. Such a fundamental shift has radical implications not only for the science of emotion but for psychological science more generally. It offers opportunities to improve the validity and trustworthiness of psychological science beyond what can be achieved with improvements to methodological rigor alone. (PsycInfo Database Record (c) 2022 APA, all rights reserved).
Article
Full-text available
Can we read emotions in faces? Many studies suggest that we can, yet skeptics contend that these studies employ methods that unwittingly help subjects in matching faces with emotions. Some studies present subjects with posed faces, which may be more exaggerated than spontaneous ones. And some studies provide subjects with a list of emotion words to choose from, which forces them to interpret faces in specific emotion terms. I argue that the skeptics’ challenge rests on a false assumption: that once subjects leave the lab, they no longer receive help in matching faces with emotions. I contend that people receive as much help in the wild as they do in the lab. People unconsciously amplify their spontaneous expressions in the presence of others, thereby making them easier to read. And people teach children to interpret faces in the same specific emotion terms found in the experimenters’ word lists. I argue that we are good at readings emotions in faces because we can normally count on a little help from our friends.
Article
Full-text available
Synthetic images of facial expression were used to assess whether judges can correctly recognize emotions exclusively on the basis of configurations of facial muscle movements. A first study showed that static, synthetic images modeled after a series of photographs that are widely used in facial expression research yielded recognition rates and confusion patterns comparable to posed photos. In a second study, animated synthetic images were used to examine whether schematic facial expressions consisting entirely of theoretically postulated facial muscle configurations can be correctly recognized. Recognition rates for the synthetic expressions were far above chance, and the confusion patterns were comparable to those obtained with posed photos. In addition, the effect of static versus dynamic presentation of the expressions was studied. Dynamic presentation increased overall recognition accuracy and reduced confusions between unrelated emotions.
Article
Full-text available
English-speaking Canadian, Cantonese-speaking Hong Kong Chinese, and Japanese speaking Japanese adults were shown 13 still photographs of the facial expressions of Chinese babies subjected to various emotion-elicitation procedures. Some respondents were asked to give an emotion label of their choice for each photograph, others to judge its pleasant-unpleasant quality. Only facial expressions taken during the "happy" condition showed agreement by a majority across all three cultural samples on a specific basic emotion. Agreement on the pleasant-unpleasant quality of the baby's expressions was higher, but still varied with culture.
Article
A sample of 60 first-year psychology students judged the emotional state of 21 drawn figures and completed the Adjective Checklist and a mood questionnaire. The judgements were affected by the interaction between head position and spinal position of the figure. Each figure was associated with a unique pattern of emotions, and the judgements given were not influenced by the subjects’ own emotional state.
Article
Although images of faces have long been employed in the scientific study of emotion, the objectives and assumptions motivating their use have shifted according to the various fields and research programs within which they have been put to use. This article traces these shifts through three such fields-the social psychology of interwar America, cross-cultural research of the 1970s, and the contemporary neurosciences of emotion-in order to assess the recent use of facial images as a means of correlating particular emotions with particular locations in the brain.
Article
• 30 male and 30 female undergraduates ("senders") viewed 2 pleasant and 2 unpleasant videotaped scenes. Unknown to the senders, their facial expressions were videotaped when they watched the scenes (spontaneous encoding) and when they talked about their reactions to the scenes (talking encoding). Later, senders were again videotaped while posing the appropriate expressions for each of the 4 scenes (posed encoding). The videotaped facial expressions were then presented for decoding to the senders. Results are as follows: (a) Accuracy of communication varied according to mode of encoding, the scene, and the mode by scene interaction. (b) Females were significantly more accurate decoders than males. (c) There were large positive and significant correlations between the ability to communicate (to decode or to encode) via spontaneous and posed cues, and positive but lower correlations between these 2 modes of communication and the talking mode. (d) The more extreme the emotional experience of the sender the more accurate his/her encoding score, especially in the spontaneous mode. (e) There were low positive correlations between total encoding and total decoding scores and low negative correlations between encoding and decoding of the same scene. (27 ref) (PsycINFO Database Record (c) 2012 APA, all rights reserved) • 30 male and 30 female undergraduates ("senders") viewed 2 pleasant and 2 unpleasant videotaped scenes. Unknown to the senders, their facial expressions were videotaped when they watched the scenes (spontaneous encoding) and when they talked about their reactions to the scenes (talking encoding). Later, senders were again videotaped while posing the appropriate expressions for each of the 4 scenes (posed encoding). The videotaped facial expressions were then presented for decoding to the senders. Results are as follows: (a) Accuracy of communication varied according to mode of encoding, the scene, and the mode by scene interaction. (b) Females were significantly more accurate decoders than males. (c) There were large positive and significant correlations between the ability to communicate (to decode or to encode) via spontaneous and posed cues, and positive but lower correlations between these 2 modes of communication and the talking mode. (d) The more extreme the emotional experience of the sender the more accurate his/her encoding score, especially in the spontaneous mode. (e) There were low positive correlations between total encoding and total decoding scores and low negative correlations between encoding and decoding of the same scene. (27 ref) (PsycINFO Database Record (c) 2012 APA, all rights reserved)
Article
We used multiple methods to examine two questions about emotion and culture: (1) Which facial expressions are recognised cross-culturally; and (2) does the “forced-choice” method lead to spurious findings of universality? Forty participants in the US and 40 in India were shown 14 facial expressions and asked to say what had happened to cause the person to make the face. Analyses of the social situations given and of the affect words spontaneously used showed high levels of recognition for most of the expressions. A subsequent forced-choice task using the same faces confirmed these findings. Analysis of the pattern of magnitude, discreteness, and similarity of responses across cultures and expressions led to the conclusion that there is no neat distinction between cross-culturally recognisable and nonrecognisable expressions. Results are better described as a gradient of recognition.
Article
Research on perception and cognition suggests that whereas East Asians view the world holistically, attending to the entire field and relations among objects, Westerners view the world analytically, focusing on the attributes of salient objects. These propositions were examined in the change-blindness paradigm. Research in that paradigm finds American participants to be more sensitive to changes in focal objects than to changes in the periphery or context. We anticipated that this would be less true for East Asians and that they would be more sensitive to context changes than would Americans. We presented participants with still photos and with animated vignettes having changes in focal object information and contextual information. Compared to Americans, East Asians were more sensitive to contextual changes than to focal object changes. These results suggest that there can be cultural variation in what may seem to be basic perceptual processes.
Article
Previous cross-cultural research on the emotions have operationalized culture by country. This article suggests that the use of stable and meaningful dimensions of cultural variability, such as those offered by Hofstede (1980), may be useful in studies on emotion. To illustrate their potential usefulness, cultural differences in previous judgment studies of universal facial expressions were reanalyzed, using Hofstede's (1980, 1983) dimensions. The results indicated that meaningful dimensions of cultural variability can be a potentially useful theoretical and empirical construct in future cross-cultural research on the emotions.