Emotion perception
Emotion perception refers to the capacities and abilities of recognizing and identifying emotions in others, in addition to biological and physiological processes involved. Emotions are typically viewed as having three components: subjective experience, physical changes, and cognitive appraisal; emotion perception is the ability to make accurate decisions about another's subjective experience by interpreting their physical changes through sensory systems responsible for converting these observed changes into mental representations. The ability to perceive emotion is believed to be both innate and subject to environmental influence and is also a critical component in social interactions. How emotion is experienced and interpreted depends on how it is perceived. Likewise, how emotion is perceived is dependent on past experiences and interpretations. Emotion can be accurately perceived in humans. Emotions can be perceived visually, audibly, through smell and also through bodily sensations and this process is believed to be different from the perception of non-emotional material.
Modes of perception
Emotions can be perceived through visual, auditory, olfactory, taste and physiological sensory processes. Nonverbal actions can provide social partners with information about subjective and emotional states. This nonverbal information is believed to hold special importance and sensory systems and certain brain regions are suspected to specialize in decoding emotional information for rapid and efficient processing.
Visual
The
Facial perception
A great deal of research conducted on emotion perception revolves around how people perceive emotion in others' facial expressions. Whether the emotion contained in someone's face is classified categorically or along dimensions of valence and arousal, the face provides reliable cues to one's subjective emotional state. As efficient as humans are in identifying and recognizing emotion in another's face, accuracy goes down considerably for most emotions, with the exception of happiness, when facial features are inverted (i.e., mouth placed above eyes and nose), suggesting that a primary means of facial perception includes the identification of spatial features that resemble a prototypical face, such that two eyes are placed above a nose which is above a mouth; any other formation of features does not immediately constitute a face and requires extra spatial manipulation to identify such features as resembling a face.
Discrete versus dimensional views
Research on the classification of perceived emotions has centered around the debate between two fundamentally distinct viewpoints. One side of the debate posits that emotions are separate and discrete entities whereas the other side suggests that emotions can be classified as values on the dimensions of valence (positive versus negative) and arousal (calm/soothing versus exciting/agitating).
Culture
Culture plays a significant role in emotion perception, most notably in facial perception. Although the features of the face convey important information, the upper (eyes/brow) and lower (mouth/nose) regions of the face have distinct qualities that can provide both consistent and conflicting information. As values, etiquette, and quality of social interactions vary across cultures, facial perception is believed to be moderated accordingly. In western cultures, where overt emotion is ubiquitous, emotional information is primarily obtained from viewing the features of the mouth, which is the most expressive part of the face. However, in eastern cultures, where overt emotional expression is less common and therefore the mouth plays a lesser role in emotional expression, emotional information is more often obtained from viewing the upper region of the face, primarily the eyes.[4] These cultural differences suggest a strong environmental and learned component in emotion expression and emotion perception.
Context
Although facial expressions convey key emotional information, context also plays an important role in both providing additional emotional information and modulating what emotion is actually perceived in a facial expression. Contexts come in three categories: stimulus-based context, in which a face is physically presented with other sensory input that has informational value; perceiver-based context, in which processes within the brain or body of a perceiver can shape emotion perception; and cultural contexts that affect either the encoding or the understanding of facial actions.[5]
Auditory
The auditory system can provide important emotional information about the environment.
Olfactory
Aromas and scents also influence mood, for example through aromatherapy,[11] and humans can extract emotional information from scents just as they can from facial expressions and emotional music. Odors may be able to exert their effects through learning and conscious perception, such that responses typically associated with particular odors are learned through association with their matched emotional experiences. In-depth research has documented that emotion elicited by odors, both pleasant and unpleasant, affects the same physiological correlates of emotion seen with other sensory mechanisms.[12]
Somatic
Theories on emotion have focused on perception, subjective experience, and appraisal. Predominant theories of emotion and emotion perception include what type of emotion is perceived, how emotion is perceived somatically, and at what stage of an event emotion is perceived and translated into subjective, physical experience.
James–Lange theory
Following the influence of René Descartes and his ideas regarding the split between body and mind, in 1884 William James proposed the theory that it is not that the human body acts in response to a person's emotional state, as common sense might suggest, but rather, people interpret their emotions on the basis of their already present bodily state. In the words of James, "we feel sad because we cry, angry because we strike, afraid because we tremble, and neither we cry, strike, nor tremble because we are sorry, angry, or fearful, as the case may be." James believed it was particular and distinct physical patterns that map onto specific experienced emotions. Contemporaneously, psychologist Carl Lange arrived at the same conclusion about the experience of emotions. Thus, the idea that felt emotion is the result of perceiving specific patterns of bodily responses is called the James–Lange theory of emotion.[13] In support of the James–Lange theory of emotion, Silvan Tomkins proposed the facial feedback hypothesis in 1963; he suggested that facial expressions actually trigger the experience of emotions and not the other way around. This theory was tested in 1974 by James Laird in an experiment where Laird asked participants to hold a pencil either between their teeth (artificially producing a smile) or between their upper lip and their nose (artificially producing a frown) and then rate cartoons. Laird found that these cartoons were rated as being funnier by those participants holding a pencil in between their teeth. In addition, Paul Ekman recorded extensive physiological data while participants posed his basic emotional facial expressions and found that heart rate raised for sadness, fear, and anger yet did not change at al for happiness, surprise, or disgust, and skin temperature raised when participants posed anger but not other emotions. While contemporary psychologists still agree with the James–Lange theory of emotion, human subjective emotion is complex and physical reactions or antecedents do not fully explain the subjective emotional experience.
Cannon–Bard theory of emotion
Two-factor theory
Stanley Schachter and his doctoral student Jerome Singer formulated their theory of emotion based on evidence that without an actual emotion-producing stimulus, people are unable to attribute specific emotions to their bodily states. They believed that there must be a cognitive component to emotion perception beyond that of just physical changes and subjective feelings. Schachter and Singer suggested that when someone encounters such an emotion-producing stimulus, they would immediately recognize their bodily symptoms (sweating and elevated heart rate in the case of the grizzly bear) as the emotion fear. Their theory was devised as a result of a study in which participants were injected with either a stimulant (adrenaline) that causes elevated heart rate, sweaty palms and shaking, or a placebo. Participants were then either told what the effects of the drug were or were told nothing, and were then placed in a room with a person they did not know who, according to the research plan, would either play with a hula hoop and make paper airplanes (euphoric condition) or ask the participant intimate, personal questions (angry condition). They found that participants who knew what the effects of the drug were attributed their physical state to the effects of the drug; however, those who had no knowledge of the drug they received attributed their physical state to the situation with the other person in the room. These results led to the conclusion that physiological reactions contributed to emotional experience by facilitating a focused cognitive appraisal of a given physiologically arousing event and that this appraisal was what defined the subjective emotional experience. Emotions were thus a result of a two-stage process: first, physiological arousal in a response to an evoking stimulus, and second, cognitive elaboration of the context in which the stimulus occurred.[15]
Neural bases
Emotion perception is primarily a cognitive process driven by particular brain systems believed to specialize in identifying emotional information and subsequently allocating appropriate cognitive resources to prepare the body to respond. The relationship between various regions is still unclear, but a few key regions have been implicated in particular aspects of emotion perception and processing including areas suspected of being involved in the processing of faces and emotional information.
Fusiform face area
The fusiform face area, part of the fusiform gyrus is an area some believe to specialize in the identification and processing of human faces, although others suspect it is responsible for distinguishing between well known objects such as cars and animals[citation needed]. Neuroimaging studies have found activation in this area in response to participants viewing images of prototypical faces, but not scrambled or inverted faces, suggesting that this region is specialized for processing human faces but not other material. This area has been an area of increasing debate and while some psychologists may approach the fusiform face area in a simplistic manner, in that it specializes in the processing of human faces, more likely this area is implicated in the visual processing of many objects, particularly those familiar and prevalent in the environment. Impairments in the ability to recognize subtle differences in faces would greatly inhibit emotion perception and processing and have significant implications involving social interactions and appropriate biological responses to emotional information.
HPA axis
The
Amygdala
The amygdala appears to have a specific role in attention to emotional stimuli.[16] The amygdala is a small, almond-shaped region within the anterior part of the temporal lobe. Several studies of non-human primates and of patients with amygdala lesions, in addition to studies employing functional neuroimaging techniques, have demonstrated the importance of the amygdala in face and eye-gaze identification.[16] Other studies have emphasized the importance of the amygdala for the identification of emotional expressions displayed by others, in particular threat-related emotions such as fear, but also sadness and happiness. In addition, the amygdala is involved in the response to non-facial displays of emotion, including unpleasant auditory, olfactory and gustatory stimuli, and in memory for emotional information.[17] The amygdala receives information from both the thalamus and the cortex; information from the thalamus is rough in detail and the amygdala receives this very quickly, while information from the cortex is much more detailed but is received more slowly.[18] In addition, the amygdala's role in attention modulation toward emotion-specific stimuli may occur via projections from the central nucleus of the amygdala to cholinergic neurons, which lower cortical neuronal activation thresholds and potentiate cortical information processing.[19]
Disordered emotion perception
There is great individual difference in emotion perception and certain groups of people display abnormal processes. Some disorders are in part classified by maladaptive and abnormal emotion perception while others, such as mood disorders, exhibit mood-congruent emotional processing. Whether abnormal processing leads to the exacerbation of certain disorders or is the result of these disorders is yet unclear, however, difficulties or deficits in emotion perception are common among various disorders.
Research investigating face and emotion perception in
In patients with major depressive disorder, studies have demonstrated either generalized or specific impairments in the identification of emotional facial expressions, or a bias towards the identification of expressions as sad.[30] Neuro-pathological and structural neuroimaging studies in patients with major depressive disorder have indicated abnormalities within the subgenual anterior cingulate gyrus and volume reductions within the hippocampus, ventral striatal regions and amygdala.[31]
Similarly,
Research methods
Researchers employ several methods designed to examine biases toward emotional stimuli to determine the salience of particular emotional stimuli, population differences in emotion perception, and also attentional biases toward or away from emotional stimuli. Tasks commonly utilized include the modified Stroop task, the dot probe task, visual search tasks, and spatial cuing tasks. The Stroop task, or modified Stroop task, displays different types of words (e.g., threatening and neutral) in varying colors. The participant is then asked to identify the color of the word while ignoring the actual semantic content. Increased response time to indicate the color of threat words relative to neutral words suggests an attentional bias toward such threat.[44] The Stroop task, however, has some interpretational difficulties[45] in addition to the lack of allowance for the measurement of spatial attention allocation.[46] To address some of the limitations of the Stroop task, the dot probe task displays two words or pictures on a computer screen (either one at the top or left and the other on the bottom or right, respectively) and after a brief stimuli presentation, often less than 1000ms, a probe appears in the location of one of the two stimuli and participants are asked to press a button indicating the location of the probe. Different response times between target (e.g., threat) and neutral stimuli infer attentional biases to the target information with shorter response times for when the probe is in the place of the target stimuli indicating an attention bias for that type of information.[46] In another task that examines spatial attentional allocation, the visual search task asks participants to detect a target stimulus embedded in a matrix of distractors (e.g., an angry face among several neutral or other emotional faces or vice versa). Faster detection times to find emotional stimuli among neutral stimuli or slower detection times to find neutral stimuli among emotional distractors infer an attentional bias for such stimuli.[47][48] The spatial cuing task asks participants to focus on a point located between two rectangles at which point a cue is presented, either in the form of one of the rectangles lighting up or some emotional stimuli appearing within one of the rectangles and this cue either directs attention toward or away from the actual location of the target stimuli. Participants then press a button indicating the location of the target stimuli with faster response times indicating an attention bias toward such stimuli.[49][50] In the morph task participants gradually scroll a facial photograph from the neutral expression to an emotion or from one emotion to another and should indicate at what frame each emotion appears on the face.[51] A recently introduced method consists of presenting dynamic faces (videoclips) and measuring verbal reaction time (into a microphone); it is more precise than previous solutions: verbal responses to six basic emotions differ in hit rates and reaction times.[52]
See also
- Affect measures
- Affective computing
- Affective forecasting
- Affective neuroscience
- Affective science
- Affective science § Measuring emotions
- Emotion recognition
- Emotion and memory
- Emotion classification
- Emotion in animals
- Emotional expression
- Emotions and culture
- Emotions in virtual communication
- Facial expressions
- Feeling
- Sex and emotion
- Social emotion
- Sociology of emotions
References
- PMID 8512154.
- hdl:10983/22919.
- ^ Plutchik, R (2002). "Nature of emotions". American Scientist. 89: 349.
- hdl:2115/22527.
- S2CID 35713636.
- PMID 7608410.
- ^ Frick, R.W. (1985). Communicating emotion: the role of prosodic features. Psychological Bulletin, 97, 412± 429.
- S2CID 145571252.
- S2CID 15310907.
- S2CID 146141346.
- S2CID 205422999.
- S2CID 23785949.
- S2CID 27900216.
- .
- PMID 14497895.
- ^ PMID 11244481.
- S2CID 195368223.
- PMID 7872730.
- S2CID 145438024.
- S2CID 2566676.
- PMID 2768363.
- PMID 681468.
- PMID 12038546.
- S2CID 13909935.
- PMID 19329561.
- S2CID 15608777.
- PMID 8315144.
- ^ Wright, I. C., Rabe-Hesketh, S., Woodruff, P.W. R., et al (2000) Regional brain structure in schizophrenia: a meta-analysis of volumetric MRI studies. American Journal of Psychiatry, 157, 16^25.
- PMID 11466121.
- S2CID 22074921.
- PMID 12611778.
- PMID 12899360.
- PMID 17201568.
- ^ Williams, J. M. G., Watts, F. N., MacLeod, C., & Matthews, A. (1988). Cognitive psychology and emotional disorders. Chichester, England:Wiley.
- PMID 2871574.
- PMID 9604557.
- PMID 11098399.
- PMID 26578996.
- ^ Cicchetti, D., Toth, S. L., & Maughan, A. (2000). An ecological- transactional model of child maltreatment. In A. J. Sameroff, M. Lewis, & S. M. Miller (Eds.), Handbook of developmental psychopathology (2nd ed., pp. 689-722). New York: Kluwer Academic/Plenum.
- ^ PMID 15677593.
- PMID 12943012.
- PMID 12072570.
- ^ a b Pollak, S. D. (2003). Experience-dependent affective learning and risk for psychopathology in children. In J. A. King, C. F. Ferris, & I. I. Lederhendler (Eds.), Roots of mental illness in children (pp. 102-111). New York: Annals of the New York Academy of Sciences.
- . (Reprinted in 1992 in the Journal of Experimental Psychology: General, 121, 15–23).
- PMID 15355142.
- ^ PMID 3700842.
- PMID 11561921.
- S2CID 40092068.
- PMID 11757875.
- S2CID 2842391.
- ISSN 1099-0992.
- S2CID 51627393.